Optimization of Quadratic Forms: Unified Minimum/Maximum Cut Computation in Directed/Undirected Graphs

Similar documents
Optimization of Quadratic Forms: NP Hard Problems : Neural Networks

Hopfield-Amari Neural Network:

On the Dynamics of a Recurrent Hopfield Network

On the Dynamics of a Recurrent Hopfield Network

Dynamics of Structured Complex Recurrent Hopfield Networks

Bounded Infinite Sequences/Functions : Orders of Infinity

Dynamics of Ordinary and Recurrent Hopfield Networks: Novel Themes

Structured Multi Matrix Variate, Matrix Polynomial Equations: Solution Techniques

Convolutional Associative Memory: FIR Filter Model of Synapse

Chance and Zero Polynomials

Also, in recent years, Tsallis proposed another entropy measure which in the case of a discrete random variable is given by

Multi-Dimensional Neural Networks: Unified Theory

Weakly Short Memory Stochastic Processes: Signal Processing Perspectives

Linear Algebra: Characteristic Value Problem

Spectral Graph Theory Lecture 2. The Laplacian. Daniel A. Spielman September 4, x T M x. ψ i = arg min

Root systems and optimal block designs

DS-GA 1002 Lecture notes 0 Fall Linear Algebra. These notes provide a review of basic concepts in linear algebra.

Lecture 9: Low Rank Approximation

Lecture 12: Introduction to Spectral Graph Theory, Cheeger s inequality

Central Groupoids, Central Digraphs, and Zero-One Matrices A Satisfying A 2 = J

Hertz, Krogh, Palmer: Introduction to the Theory of Neural Computation. Addison-Wesley Publishing Company (1991). (v ji (1 x i ) + (1 v ji )x i )

Distributed Randomized Algorithms for the PageRank Computation Hideaki Ishii, Member, IEEE, and Roberto Tempo, Fellow, IEEE

642:550, Summer 2004, Supplement 6 The Perron-Frobenius Theorem. Summer 2004

Discriminative Direction for Kernel Classifiers

Econ Slides from Lecture 7

Some global properties of neural networks. L. Accardi and A. Aiello. Laboratorio di Cibernetica del C.N.R., Arco Felice (Na), Italy

A note on the minimal volume of almost cubic parallelepipeds

6.046 Recitation 11 Handout

Sergey Norin Department of Mathematics and Statistics McGill University Montreal, Quebec H3A 2K6, Canada. and

Every Neural Code Can Be Realized by Convex Sets

Ma/CS 6b Class 23: Eigenvalues in Regular Graphs

Ahlswede Khachatrian Theorems: Weighted, Infinite, and Hamming

Bichain graphs: geometric model and universal graphs

Math 350 Fall 2011 Notes about inner product spaces. In this notes we state and prove some important properties of inner product spaces.

Decentralized Stabilization of Heterogeneous Linear Multi-Agent Systems

Spectral Theorem for Self-adjoint Linear Operators

The Hypercube Graph and the Inhibitory Hypercube Network

On the Cardinality of Mersenne Primes

1 Review: symmetric matrices, their eigenvalues and eigenvectors

Geometric Steiner Trees

Modeling and Stability Analysis of a Communication Network System

Laplacian Integral Graphs with Maximum Degree 3

Spectral Graph Theory Lecture 3. Fundamental Graphs. Daniel A. Spielman September 5, 2018

U.C. Berkeley CS294: Spectral Methods and Expanders Handout 11 Luca Trevisan February 29, 2016

CHAPTER 2: CONVEX SETS AND CONCAVE FUNCTIONS. W. Erwin Diewert January 31, 2008.

Properties and Classification of the Wheels of the OLS Polytope.

On convergent power series

Innovative Approach to Non-Uniform Sampling

Supplementary lecture notes on linear programming. We will present an algorithm to solve linear programs of the form. maximize.

Notes on the Matrix-Tree theorem and Cayley s tree enumerator

Semidefinite and Second Order Cone Programming Seminar Fall 2001 Lecture 5

AN ELEMENTARY PROOF OF THE SPECTRAL RADIUS FORMULA FOR MATRICES

Support weight enumerators and coset weight distributions of isodual codes

Fiedler s Theorems on Nodal Domains

16.1 L.P. Duality Applied to the Minimax Theorem

Generating p-extremal graphs

Finding normalized and modularity cuts by spectral clustering. Ljubjana 2010, October

Lecture 10: A (Brief) Introduction to Group Theory (See Chapter 3.13 in Boas, 3rd Edition)

Decomposing Bent Functions

Graph coloring, perfect graphs

A Cross-Associative Neural Network for SVD of Nonsquared Data Matrix in Signal Processing

1 Computational Problems

A Polynomial-Time Algorithm for Pliable Index Coding

Approximation Algorithms

Deterministic Approximation Algorithms for the Nearest Codeword Problem

1 Some loose ends from last time

U.C. Berkeley Better-than-Worst-Case Analysis Handout 3 Luca Trevisan May 24, 2018

On Locating-Dominating Codes in Binary Hamming Spaces

Kernels of Directed Graph Laplacians. J. S. Caughman and J.J.P. Veerman

CS 173: Induction. Madhusudan Parthasarathy University of Illinois at Urbana-Champaign. February 7, 2016

Chapter 3 Transformations

Shortest paths with negative lengths

CS168: The Modern Algorithmic Toolbox Lectures #11 and #12: Spectral Graph Theory

Reconstruction and Higher Dimensional Geometry

The Structure of Trivalent Graphs with Minimal Eigenvalue Gap *

Linear Codes, Target Function Classes, and Network Computing Capacity

2. Intersection Multiplicities

Notes for Lecture Notes 2

A Combinatorial Bound on the List Size

Linear & nonlinear classifiers

P is the class of problems for which there are algorithms that solve the problem in time O(n k ) for some constant k.

Strongly chordal and chordal bipartite graphs are sandwich monotone

ALGEBRA. 1. Some elementary number theory 1.1. Primes and divisibility. We denote the collection of integers

Weakly Secure Data Exchange with Generalized Reed Solomon Codes

On improving matchings in trees, via bounded-length augmentations 1

Simultaneous Diagonalization of Positive Semi-definite Matrices

Small Label Classes in 2-Distinguishing Labelings

12. LOCAL SEARCH. gradient descent Metropolis algorithm Hopfield neural networks maximum cut Nash equilibria

Introduction to Group Theory

APPENDIX A. Background Mathematics. A.1 Linear Algebra. Vector algebra. Let x denote the n-dimensional column vector with components x 1 x 2.

On shredders and vertex connectivity augmentation

Complexity Theory VU , SS The Polynomial Hierarchy. Reinhard Pichler

Outline. Complexity Theory EXACT TSP. The Class DP. Definition. Problem EXACT TSP. Complexity of EXACT TSP. Proposition VU 181.

Total Dominator Colorings in Paths

A Remark on Alan H. Kawamoto: Nonlinear Dynamics in the Resolution of Lexical Ambiguity: A Parallel Distributed Processing Account

MIT Algebraic techniques and semidefinite optimization February 14, Lecture 3

Using Laplacian Eigenvalues and Eigenvectors in the Analysis of Frequency Assignment Problems

Lecture 26: April 22nd

Minimal basis for connected Markov chain over 3 3 K contingency tables with fixed two-dimensional marginals. Satoshi AOKI and Akimichi TAKEMURA

Exercises * on Linear Algebra

Transcription:

Optimization of Quadratic Forms: Unified Minimum/Maximum Cut Computation in Directed/Undirected Graphs by Garimella Ramamurthy Report No: IIIT/TR/2015/-1 Centre for Security, Theory and Algorithms International Institute of Information Technology Hyderabad - 500 032, INDIA November 2015

OPTIMIZATION OF QUADRATIC FORMS: UNIFIED MINIMUM / MAXIMUM CUT COMPUTATION IN DIRECTED / UNDIRECTED GRAPHS Garimella Rama Murthy, International Institute of Information Technology Hyderabad, India rammurthy@iiit.ac.in Abstract In this research paper, the problem of optimization of a quadratic form over the corners of hypercube is reviewed. Results related to the computation of global optimum stable / anti-stable state of a Hopfield Neural Network (HNN) are discussed. Using these results, an efficient algorithm for computation of minimum cut as well as maximum cut in an undirected as well as directed graph is discussed. Also, POTENTIALLY, a deterministic exact polynomial time algorithm for the NP- Hard problem of maximum cut in an undirected graph is discussed. Effectively, unified approach to minimum cut as well as maximum cut computation in arbitrary directed as well as undirected graphs is presented. I. INTRODUCTION In computer science, graphs (directed as well as undirected) naturally arise in many applied problems. For instance, in a directed graph, computing the minimum cut is an interesting research problem (a special case transportation problem). Ford-Fulkerson algorithm is the first polynomial time algorithm to compute the minimum cut in a directed graph. Several efficient polynomial time algorithms have been designed for such a problem. But, it has been realized that computing the maximum cut in an undirected graph is an NP-hard problem. So far, no polynomial time algorithm has been designed for such a problem. This research paper provides an effort in that direction. This research paper is organized as follows. In Section 2, relevant research literature is reviewed. The author in his research efforts obtained general results for the problem of optimizing a quadratic form over the corners of unit hypercube. These results and the related ideas are documented in Section 3. In Section 4, the problem of computing the largest stable state of a Hopfield neural network is solved. The results are utilized to design an algorithm for computing the minimum cut in an arbitrary undirected graph. The results are generalized for min cut computation in arbitrary directed graphs. Also, in Section 4, an exact deterministic polynomial time algorithm for the NP hard problem of maximum cut computation in undirected graphs is discussed. Main Contributions: (1) Efficient algorithm for minimum cut / maximum cut computation when the matrix of weights (synaptic weight matrix) is structured (with respect to the spectral representation i.e. eigenvectors corresponding to positive / negative eigenvalues that are corners of the unit hypercube) (2) Unified approach (efficient algorithms) for minimum cut / maximum cut computation in undirected / directed graphs (Effectively 4 algorithms for 4 problems) (3) POTENTIALLY polynomial time algorithm for maximum cut computation in undirected / directed graphs.

II. `REVIEW OF RESEARCH LITERATURE: MIN / MAX CUT COMPUTATION : HOPFIELD NEURAL NETWORK Contribution of Hopfield et.al: Hopfield neural network constitutes a discrete time nonlinear dynamical system. It is naturally associated with a weighted undirected graph G = (V,E), where V is the set of vertices and E is the set of edges. A weight value is attached to each edge and a threshold value is attached to each vertex/node /artificial neuron of the graph. The order of the network is the number of nodes / vertices in the associated graph. Thus a discrete time Hopfield neural network of order N is uniquely specified by (A) N x N Symmetric Synaptic Weight Matrix M i.e. denotes the weight attached to the edge from node i to node j (node j to node i) (B) Nx1 Threshold Vector i.e. denotes the threshold attached to node i. Each neuron is in one of the two states i.e. +1 or -1. Thus, the state space of such a non-linear dynamical system is the N-dimensional unit hypercube. For notational purposes, let () denote the state of node / neuron i at the discrete time index t. Let the state of the Hopfield neural network at time t be denoted by the Nx1 vector (). The state at node i is updated in the following manner (i.e. computation of next state of node i ) ( +1) = ()..(2.1) i.e. the next state at node i is +1 if the term in the bracket is non-negative and -1 if it is negative. Depending on the set of nodes at which the state updation given in equation (2.1) is performed, the neural network operation is classified into the following modes: Serial Mode: The set of nodes at which state updation as in (2.1) is performed is exactly one i.e. at time t only at one of the nodes / neurons the above state updation is performed. Fully Parallel Mode: At time t, the state updation as in (2.1) is performed simultaneously at all the nodes In the state space of discrete time Hopfiled neural network, there are certain distinguished states, called the STABLE STATES. Definition: A state V(t) is called a Stable State if and only if () = ( () ).(2.2) Thus, if state dynamics of the network reaches the stable state at some time t, it will remain there for ever i.e. no change in the state of network occurs regardless of the mode of operation of the network (i.e. it is a fixed point in the state dynamics of discrete time Hopfield neural network). The following Convergence Theorem summarizes the dynamics of discrete time Hopfield neural network in the serial and parallel modes of operation. It characterizes the operation of the neural network as an associative memory. Theorem 1: Let the pair Z = (M,T) specify a Hopfield neural network. Then the following hold true: [1] Hopfield : If N is operating in a serial mode and the elements of the diagonal of M are non-negative, the network will always converge to a stable state (i.e. there are no cycles in the state space).

[2] Goles: If N is operating in the fully parallel mode, the network will always converge to a stable state or to a cycle of length 2 (i.e the cycles in the state space are of length 2). The proof of above Theorem is based on associating the state dynamics of Hopfield Neural Network (HNN) with an energy function. It is reasoned that the energy function is non-decreasing when the state of the network is updated (at successive time instants). Since the energy function is bounded from above, the energy will converge to some value. The next step in the proof is to show that constant energy implies a stable state is reached in the first case and atmost a cycle of length 2 is reached in the second case. The so called energy function utilized to prove the above convergence Theorem is the following one: () = () () 2 ().(2.3) Thus, HNN, when operating in the serial mode will always get to a stable state that corresponds to a local maximum of the energy function. Hence the Theorem suggests that Hopfield Associative Memory (HAM) could be utilized as a device for performing local/global search to compute the maximum value of the energy function. Contribution of Bruck et.al The above Theorem implies that all the optimization problems which involve optimization of a quadratic form over the unit hypercube (constraint / feasible set) can be mapped to a HNN which performs a search for its optimum. One such problem is the computation of a minimum cut in an undirected graph. For the sake of brevity, we skip the definition of a cut in a graph. It can be noted that min cut specification is equivalent to specifying the subset of vertices, on one side of the cut. Minimum and Maximum Cut in Undirected Graphs: Stable / Anti-Stable States: In the following Theorem, proved in [BrB], the equivalence between the minimum cut and the computation of global optimum of energy function of HNN is summarized. Theorem 2: Consider a Hopfield Neural Network (HNN) Z= (M,T) with the thresholds at all nodes being zero i.e. T 0. The problem of finding the global optimum stable state (for which the energy is maximum) is equivalent to finding a minimum cut in the graph corresponding to N. Corollary: Consider a Hopfield neural network Z =(M,T) with the thresholds at all neurons being zero i.e. T 0. The problem of finding a state V for which the energy is global minimum is equal to finding a maximum cut in the graph corresponding to N. Proof: Follows from the argument as in [BrB]. We repeat the argument for clarity (required for understanding the algorithms for MIN/MAX computation in DIRECTED / UNDIRECTED graphs). From the Lemma 1 discussed below [RaN], there is no loss of generality in assuming that the threshold at all the nodes/vertices of the Hopfield neural network is zero. Thus, the energy function is a pure quadratic form i.e! =. Let us denote the sum of weights of edges in the graph with both end points in vertex set as " " and let # #, " # denote the corresponding sums for the other two cases. We readily have that! = 2 ( " " + # # " # ) = 2 ( " " + # # + " # ) 4 " #.

Since the first term in the above equation is constant (it is the sum of weights of all edges), it follows that the minimization of H is equivalent to maximization of " #. It is clear that " # is the weight of the cut in N with being the nodes of the Z with state equal to 1. Q. E.D. Thus, the operation of Hopfield neural network in the serial mode is equivalent to conducting a local search algorithm for finding a minimum cut in the associated graph. State updation at a node in the network is equivalent to moving it from one side of the cut to the other side in the local search algorithm. As in the case of stable states, we have the following definition. Definition: The local minimum vectors of a quadratic form on the hypercube are called anti-stable states. This concept was first introduced in [3, Rama1]. Remark 1 : In view of Theorem 2, minimum cut computation in an undirected graph reduces to determining the global optimum stable state, S of the associated Hopfield neural network N = (M,T) with T 0. The {+ 1} components of S determine the subset of vertices that are members of one side of the cut (and {- 1} components determine the vertices on the other side of the cut). Similarly, maximum cut computation requires determination of global optimum anti-stable state. Minimum and Maximum Cut in Directed Graphs: Stable / Anti-Stable States: In view of the above Theorem, a natural question that arises is to see if a Hopfield neural network can be designed to perform a local search for minimum cut in a directed graph. In that effort we utilize the standard definition of minimum cut in a directed graph. The following Theorem is in the same spirit of Theorem for directed graphs. Theorem 3: Let M be the matrix of edge weights (M is not necessarily symmetric) in a weighted directed graph denoted by G = (V,E). The network % = (&,) performs a local search for a DMC of G where ( = ( + ) 2 Proof: Refer [9] i.e. [BrS]. ) = 1 2 *( ) ) ) Thus, this Theorem shows that the computation of minimum cut in a directed graph is equivalent to determining the global optimum stable state of a suitably chosen Hopfield neural network. Note: Similarly as in Theorem 2, the computation of directed maximum cut (i.e. maximum cut in a directed graph) reduces to computation of global minimum anti-stable state (global minimum vector of quadratic energy function) in a suitably chosen Hopfield neural network. Based on Theorem 2 and Theorem 3, we design polynomial time algorithms for computation of minimum cut as well as maximum cut in directed and undirected graphs. The key idea is polynomial time algorithm for computation of global optimum stable state and anti-stable state.

Claim 1 : In summary MIN / MAX cut computation in directed / undirected graphs is equivalent to the determination of global optimum stable / anti-stable state computation of associated Hopfield neural network. The problem of optimization of a quadratric form over the hypercube arises in many research problems in computer science. The author, after mastering the results in [Hop], [BrB] contemplated on removing the conditions required in Theorems 1 and 2. The fruits of such effort are documented in the following Section. III. OPTIMIZATION OF QUDRATIC FORMS OVER HYPERCUBE The energy function associated with a HNN, considered in (2.3) is not exactly a quadratic form. The author questioned whether the threshold vector associated with a HNN can always be assumed to be zero (for instance by introducing a dummy node and suitably choosing the synaptic weights from it to the other nodes). The result of such effort is the following Lemma. Lemma 1: There is no loss of generality in assuming that the threshold vector associated with a Hopfield Neural Network (HNN) is an all-zero vector Proof: Refer the argument provided in [RaN]. Thus, it is clear that a properly chosen HNN acts as a local / global optimization device of an arbitrary quadratic form as the objective function on the constraint set being unit hypercube. Thus in the following discussion, we consider only a pure quadratic form as the energy function. Also, in part 1 of Theorem 1, we require the condition that the diagonal elements of symmetric synaptic weight matrix are all non-negative. We now show in the following Theorem that such a condition can be removed. In this section, we consider the problem of maximization of quadratic form (associated with a symmetric matrix) over the corners of binary, symmetric hypercube. Mathematically, this set is specified precisely as follows: S = { X = ( xi, x2, x3,..., xn ) : xi = ± 1 for1 i N}... (3.1) From now onwards, we call the above set simply as hypercube. This optimization problem arises in a rich class of applications. This problem is the analogue of the maximization over the hypersphere of quadratic form associated with a symmetric matrix. Rayleigh provided the solution to the optimization problem on the unit hypersphere. A necessary condition on the optimum vector lying on the unit hypercube is now provided. This Theorem is the analogue of the maximization over the hypersphere of a quadratic form associated with a symmetric matrix. The following Theorem and other associated results were first documented in [Rama1]. Theorem 4: Let B be an arbitrary N x N real matrix. From the standpoint of maximization of the T quadratic form i.e. u Bu on the hypercube, it is no loss of generality to assume that B is a symmetric T matrix with zero diagonal elements. If u maximizes the quadratic form u Bu, subject to the constraint that, = 1 f-. 1 % (i.e. u lies on the corners of hypercube), then u = sign( Cu). (3.2) 1 Where ( T C = B + B ) with all the diagonal elements set to zero. In the above equation (i.e. eqn 3.2), 2 Sign (0) is interpreted as +1 or -1 based on the requirement.

Proof: Refer [Rama1], [Rama3]. Also refer [RaN] for an elementary proof. Corollary: Let E be an arbitrary N x N real matrix. If u minimizes the quadratic form the constraint u = 1 for 1 i n, then i u = sign( Cu).(3.11) where C is the symmetric matrix with zero diagonal elements obtained from E. T u Eu, subject to Note: It is immediate to see that if, is a stable state (or anti-stable state), then -, (minus u ) is also a stable state (or anti-stable state). Remark 2: In view of Theorem 4, there is no loss of generality in assuming that the TRACE of matrix is zero for determining the stable / anti-stable states (i.e for optimizing the quadratic form). Since TRACE of a matrix is the sum of eigenvalues, the sum of positive valued eigenvalues is equal to the sum of negative valued eigenvalues. Hence, it is easy to see that a symmetric matrix with zero diagonal elements cannot be purely positive definite or purely negative definite. It can be assumed to be indefinite with the largest eigenvalue being a positive real number. Thus the location of stable states (vectors) is invariant under variation of Trace (M).. IV. GLOBAL OPTIMUM STABLE / ANTI-STABLE STATE COMPUTATION : COMPUTATION OF MINIMUM AND MAXIMUM CUT IN DIRECTED AND UNDIRECTED GRAPHS: As discussed in Section II, Bruck et.al [BrB] showed that the problem of computing the maximum stable state is equivalent to that of computation of minimum cut in the associated undirected graph. As per Bruck et.al, this is claimed to be an NP hard problem [BrB]. But theoretical computer scientists Rao Kosaraju and Sartaj Sahni informed the author that the minimum cut in an undirected graph is known to be in P (i.e. polynomial time algorithms exist). They also informed the author that MAX CUT in an undirected graph is NP complete From the corollary of Theorem 2, it follows that computing the MAXIMUM CUT is equivalent to the problem of determining the global minimum anti-stable state (i.e. determining the corner of unit hypercube where the global minimum of the quadratic form is attained). Goals: To see if an efficient polynomial time algorithm can be discovered for the problem of computing the minimum cut in an undirected graph. Also, finding a polynomial time algorithm for the NP complete problem of computing the MAXIMUM CUT in an undirected graph. Thus we are interested in knowing whether P = NP. In the following discussion, we consider the quadratic form associated with the matrix M (which also can be treated as the synaptic weight matrix of a Hopfield Neural Network). Lemma 2: Suppose the synaptic weight matrix (connection matrix) of the Hopfield neural network is a non-negative matrix (i.e. all the components are non-negative real numbers). Then, the global optimum stable state is the all ones vector i.e. [ 1 1. 1 ]. Proof: Since the energy function is a quadratic form and the variables are only allowed to assume { +1, -1 } values, the global optimum is achieved by taking the sum of all the components of the symmetric synaptic weight matrix. Hence, the vectors of all ones i.e. [ 1 1 1 1 ] is the global optimum stable state. Q.E.D. Corollary: The global optimum stable state i.e. all ones vector corresponds to the EMPTY minimum cut in the associated undirected weighted graph.

Note: If u is a stable state, it is easy to see that -u is also a stable state. Thus, when the synaptic weight matrix is non-negative, the vector of all MINUS ONES is also the global optimum stable state. This, once again corresponds to EMPTY MINIMUM CUT. Structured Synaptic Weight Matrices: Computation of global optimum Stable/anti-stable State: The following lemma relates eigenvectors of a symmetric matrix M and the stable / anti-stable states associated with the synaptic weight matrix M. Lemma 3: If a corner of unit hypercube is an eigenvector of M corresponding to positive / negative eigenvalue, then it is also a stable / anti-stable state Proof: Follows from the definition of eigenvector and stable state. Details avoided for brevity.. We now reason how the computation of global optimum stable state is related to the computation of global optimum anti-stable state. Claim 2: Suppose / 0 is a anti stable state of the matrix M, then / 0 is stable state of the matrix -M (minus M). Proof: Follows from the definition of stable and anti-stable states Q.E.D. The following remark follows from the above Lemma on invoking Rayleigh s Theorem.. We now briefly state Rayleigh s Theorem. The proof can be found in a textbook of linear algebra. Theorem 5 : Rayleigh s Theorem: The local optima of the quadratic form associated with a symmetric matrix M (i.e. / /) on the unit Euclidean hypersphere (i.e. { X : / / = 1 }) occur at the eigenvectors with the corresponding value of the quadratic form being the eigenvalue. Remark 3: Suppose we consider a vector on the hypercube (one corner of hypercube), say /2 which is also the eigenvector of matrix M corresponding to the largest eigenvalue i.e. We have /2 = 3 456 /2. Then, since 3 456 is positive, we have that /2 = (3 456 /& ) = /2. Thus, in view of Rayleigh s Theorem, such a corner of the hypercube is also the global optimum stable state. This inference follows from the fact that the points on hypercube can be projected onto the unit hypersphere. Let X be a vector lying on the unit hypercube. Project such a vector onto the unit hypersphere using the following transformation i.e. / 7 = %. It is clear that the Euclidean norm of Y is equal to one. We now briefly consider such a case. CASE A : This case where a corner of the hypercube is also the eigenvector corresponding to the maximum positive eigenvalue is very special. More generally a corner of hypercube is the eigenvector corresponding to a positive eigenvalue. It should be realized that we are only interested in the SIGN STRUCTURE of the eigenvector corresponding to the largest positive eigenvalue (which is also a corner of hypercube)it is well known that the computation of maximum eigenvector of a symmetric matrix can be carried out using a polynomial time algorithm (Elsner s / Lanczos algorithm).

Suppose we use such an algorithm to compute the global minimum anti-stable state (Refer claim above Rayleigh s Theorem). We have found a polynomial time algorithm for such an NP hard problem. Thus in such a case P = NP. In case of a 2 x 2 synaptic weight matrix, it can easily be shown that the global optimum stable state happens to be the eigenvector corresponding to the largest eigenvalue. Details are avoided for brevity. Note: Using Rayleigh s Theorem, in case of hypothesis of Lemma 3, the eigenvector of M corresponding to largest positive eigenvalue is the global optimum stable state. Also, the eigenvector corresponding to smallest negative eigenvalue of M is the global optimum anti-stable state. More generally, an eigenvector that is a corner of hypercube corresponding to a larger positive eigenvalue is a more optimal ( maximum ) stable state. Similarly, an eigenvector that is a corner of hypercube corresponding to a smaller negative eigenvalue is a more optimal ( minimum) anti-stable state. By computing the spectral representation of M, if hypothesis in Lemma 3 are satisfied, global optimum stable state as well as global optimum anti-stable state can be computed by a polynomial time algorithm. CASE B: Now, we consider the arbitrary case where the eigenvector corresponding to the largest eigenvalue is NOT a stable state. More generally none of the corners of hypercube is an eigenvector. ALGORITHM A : Efficient Algorithm for minimum / maximum cut computation in directed / undirected graphs: Global optimum stable / anti-stable state computation of a Hopfield Neural Network: (I) (II) By claim 2, the computation of global optimum (minimum) anti-stable state is equivalent to computation of global optimum stable state of a suitably chosen Hopfield neural network. By Theorem 2, computation of global optimum stable state is equivalent to computing minimum cut in the graph corresponding to Hopfield neural network. There are several efficient polynomial time algorithms for computing the minimum cut in a directed / undirected graph. Thus, in view of corollary of Theorem 2, claim 2, Theorem 3, efficient POLYNOMIAL TIME algorithms exist for computing the MAXIMUM CUT in an undirected / directed graph. Now we discuss other possible algorithms for computing the minimum / maximum cut in directed /undirected graphs based on the connection of such algorithms to Hopfield neural network. Lemma 4: If y is an arbitrary vector on unit hypersphere that is obtained by projecting a vector on unit hypercube (onto the unit hypersphere) and x 0 is the eigenvector of symmetric matrix M (of Euclidean norm one) corresponding to the maximum eigen value (on the unit hypersphere), then we have that T T T y My = µ max + 2 µ max ( y x0 ) x0 + ( y x0 ) M ( y x0) Proof: Follows from a standard argument Q.E.D.

Remark 4: Since, by Rayleigh s theorem, it is well known that the global optimum value of a quadratic form on the unit hypersphere is the maximum eigen value i.e. µ max, it is clear that for all corners of the hypercube projected onto the unit hypersphere, we must necessarily have that 2 max ( 0) T T µ y x x0 + ( y x0 ) M ( y x0) 0. The goal is to choose a y, such that the above quantity is as less negative as possible (so that the value of quadratic form is as close to µ max as possible). Unlike in Remark 4, suppose that L = ( 9 0 ) ( 9 0 ). Then a natural question is to see if L can be utilized some how for arriving at the global optimum stable state. Such a question was the starting point for the following algorithm to compute the global optimum stable state. ALGORITHM B: ALGORITHM FOR COMPUTATION OF GLOBAL OPTIMUM STABLE STATE OF A HOPFIELD NEURAL NETWORK: ALGORITHM FOR COMPUTATION OF MINIMUM CUT IN AN UNDIRECTED / DIRECTED GRAPH : Step 1: Suppose the right eigenvector corresponding to the largest eigen value of M is real (i.e. real valued components). Let such an eigenvector be 9 0 (i.e. largest eigenvector of the synaptic weight matrix of the undirected graph associated with the Hopfield neural network). Compute the SIGN STRUCTURE of such an eigenvector, L ; = ( 9 0 ). Step 2: Using L as in the initial condition (vector), run the Hopfield neural network in the serial mode of operation until a stable state is reached. Such a state is the global optimum stable state. By Theorem 2, it corresponds (leads) to the minimum cut in the associated graph. In view of the following Lemma, eigenvector corresponding to the largest eigenvalue can always be assumed to contain real valued components. Lemma 5: If A is symmetric and real valued matrix, then every eigenvector can be CHOSEN to contain real valued components: Proof: Follows from standard argument related to linear algebra of a symmetric matrix. Q.E.D. Note: In view of Theorem 3, the same algorithm can be applied for computation of minimum cut in a directed graph (with a properly chose synaptic weight matrix defined in Theorem 3). In view of Lemma 4 and above remark 4, the claim is that the global optimum stable state is reached through the above procedure. The formal proof that global optimum stable state is reached is provided in Appendix Theorem 6: The above algorithm converges to the global optimum stable state using the vector L as the initial condition (on running Hopfield neural network in serial mode). Proof: Refer the Appendix Note: In view of Lemma 2 and Rayleigh s Theorem, it can be reasoned that (when the synaptic weight matrix is non-negative), non-empty min cut can be computed by the following procedure Let < 0 be the eigenvector corresponding to the second largest eigenvalue (computed by removing the mode corresponding to largest eigenvalue). Let = ; = ( < 0 ).

To compute the non-empty minimum cut, run the Hopfield neural network in serial mode with ;2 as the initial condition. We now determine the computational complexity of the above algorithm. The above algorithm involves the following computations. It is a POTENTIALLY a polynomial time algorithm. (A) Signed Largest Eigenvector Computation: Computation of eigenvector corresponding to the largest eigenvalue of the symmetric matrix (connection matrix of Hopfield neural network) i.e 9 0. The problem is a topic in standard computational linear algebra. One can use power method, Lanczos method etc. By a suitable modification of Lanczos algorithm, a polynomial time algorithm for computing Signed largest eigenvector i.e. ; = ( 9 0 ) is designed. Details are avoided for brevity. (B) State Updation in Serial Mode with Proper Initial Condition Vector: Using L as the initial condition, we run the Hopfield neural network in serial mode until global optimum stable state is reached. It is possible to bound the number of computations for this task. Under the assumption that the number of { 1 vectors } in the domain of attraction of largest stable state is bounded by a polynomial of constant degree (as a function of number of vertices), a polynomial time algorithm is designed for this task. Maximum Cut Computation in Undirected and Directed Graphs : Global Optimum Anti-Stable State Computation of a Hopfield Neural Network: Claim: Suppose / 0 is an anti stable state of the matrix M, then / 0 is stable state of the matrix -M (minus M). Proof: Follows from the definition of stable and anti-stable states Q.E.D. Thus, by the above discussion, the NP complete problem of maximum cut computation in undirected as well as directed graphs reduces to computation of global optimum anti-stable state. Remark 5: Similar Theorem, as Theorem 6, is proved for computing the global optimum anti-stable state (global minimum of the associated quadratic form) of the Hopfield neural network. The associated algorithm is POTENTIALLY a POLYNOMIAL time algorithm for the NP-complete problem of MAXIMUM CUT computation in an undirected graph. Based on Theorem 3, an interesting algorithm for computation of maximum cut in a directed graph is discussed. Hopfield Neural Network: Associated One Step Associative Memory: POLYNOMIAL TIME Algorithm for Minimum/ Maximum Cut Computation in Undirected / Directed Graphs: Now, we investigate the possibility of arriving at a more efficient algorithm for computing the global optimum (maximum / minimum) stable / anti-stable state of a Hopfield neural network. Specifically, we propose a POLYNOMIAL TIME algorithm for computing the global optimum stable state as well as anti-stable state. As per the above discussion (Theorem 6, Remark 5), we restrict the discussion to computation of global optimum / maximum stable state(equally applies to the global minimum anti-stable state ). In view of following Lemma, it has been shown in [BrB] that a graph theoretic code is naturally associated with a Hopfield network (with the associated quadratic energy function). The local and global optima of the energy function are the code words.

Lemma 6: Given a linear block code, a neural network can be constructed in such a way that every local maximum of the energy function corresponds to a codeword and every codeword corresponds to a local maximum. Proof: Refer the paper by Bruck et.al [BrB]. The following fact enables bounding minimum distance. Note: Graph-theoretic error correcting codes are limited in the sense that the following upper bound on the minimum distance holds true;? 2, where? A hc DD,D?AEFC E?, A hc FE.?EG< -H C?C E? IC.C9 ACA -H hc EAA-FEC?.EJh.CAJCFICG<. Goal: To compute the global optimum stable state (i.e. global optimum of the energy function) using the associated graph theoretic encoder. We now propose a POTENTIALLY polynomial time algorithm to achieve the above goal. ALGORITHM C: Step 1 : Compute the real eigenvector 9 0 (i.e. real valued components) of symmetric matrix M corresponding to the largest eigenvalue. Compute the corner, L of hypercube from 9 0 in the following manner L = Sign (9 0 ). Complexity of this step: A polynomial time algorithm for this step is well known in the linear algebra literature. Step 2: As discussed in [BrB], the { +1, -1 } valued vector L partitions the vertex set into two groups with +1 valued vertices on one side of the cut and the -1 valued vertices on other side of the cut. This will also enable determining the characteristic vector of cut edges. We thus arrive at a row vector, B of dimension E which can atmost be K % L. By Theorem 6 and the results in [BrB], such a vector, B lies in 2 the coding sphere corresponding to global optimum stable state i.e the codeword corresponding to the global optimum of quadratic energy function. Complexity of this Step: The determination of characteristic vector of the cut edges (which need not be the global minimum cut. It will be a global minimum cut if L is an eigenvector of M corresponding to a positive eigenvalue) requires labeling of atmost E edges ( E can atmost be K % 2 L). Step 3: Compute the generator matrix of the graph theoretic code associated with the Hopfield network. The characteristic vector computed in Step 2 (by Theorem 6) lies in the coding sphere corresponding to the global optimum codeword (global maximum of associated quadratic energy function). Using the associated information word (and the generator matrix), the global optimum codeword is computed in one step. This corresponds to the global optimum stable state. Thus, this provides an efficient algorithm to compute the min cut / max cut in an undirected / directed graph

REFERENCES [1] [BrB] J. Bruck and M. Blaum, Neural Networks, Error Correcting Codes and Polynomials over the Binary Cube, IEEE Transactions on Information Theory, Vol.35, No.5, September 1989. [2] [Hop] J.J. Hopfield, Neural Networks and Physical Systems with Emergent Collective Computational Abilities, Proceedings of National Academy Of Sciences, USA Vol. 79, pp. 2554-2558, 1982 [3] [Rama1]G. Rama Murthy, Optimal Signal Design for Magnetic and Optical Recording Channels, Bellcore Technical Memorandum, TM-NWT-018026, April 1st, 1991 [4] [Rama2]G. Rama Murthy, Efficient Algorithms for Computation of Minimum Cut and Maximum Cut in an Undirected Graph, Manuscript in Preparation [5] [Rama3]G. Rama Murthy, Multi-dimensional Neural Networks : Unified Theory, Research monograph published by NEW AGE INTERNATIONAL PUBLISHERS, NEW DELHI, 2007. [6] [RaN] G. Rama Murthy and B. Nischal, Hopfield-Amari Neural Network : Minimization of Quadratic forms, The 6th International Conference on Soft Computing and Intelligent Systems, Kobe Convention Center (Kobe Portopia Hotel) November 20-24, 2012, Kobe, Japan. [7] [Rama4] G. Rama Murthy, Multi/Infinite Dimensional Neural Networks: Multi / Infinite Dimensional Logic Theory, International Journal of Neural Networks, Volume 15, No. 3, June 2005 [8] [Rama6] G. Rama Murthy, Optimization of Quadratic Forms: NP Hard Problems : Neural Networks, 2013 International Symposium on Computational and Business Intelligence (ISCBI 2013), August 24-26, 2013, New Delhi, India. Available on IEEE Explore [9] [BrS] J. Bruck and J. Sanz, A Study on Neural Networks, International Journal of Intelligemt System, Vol. 3, pp. 59-75, 1988. [10] G. Rama Murthy, Towards a Resolution of P = NP conjecture, Cornell University Archive [11] G. Rama Murthy, Towards a Resolution of P = NP conjecture, IIIT Hyderabad Archive. [12] S. L. Hakimi and H. Frank, Cut-set matrices and linear codes, IEEE Transactions on Information Theory, vol. IT-11, pp. 457-458, July 1965 APPENDIX Proof of Theorem 6 : In view of the results in [1] ([BrB]), the idea is to reason that the vector L is in the domain of attraction of the global optimum stable state. Equivalently using the results in [1], we want to show that the initial vector is in the coding sphere of the codeword corresponding to the global optimum stable state. Let < 0 be the global optimum stable state / vector on the unit hypercube. Let M 0 be one among the other stable states (i.e. the second largest stable state). Thus, the idea is to reason that the Hamming distance between L and < 0 i.e.? N (;,< 0 ) is smaller than the Hamming distance between L and M 0 i.e.? N (;,M 0 ) i.e.

To reason that? N (;,< 0 ) <? N (;,M 0 ). The proof is by contradiction i.e. say? N (;,M 0 ) + 1 =? N (;,< 0 ).(4.1) We know that the sign structure of the vectors L and 9 0 is exactly the same. More explicitly all the components of L and 9 0 have the same sign (positive or negative). Since the three vectors O ;,< 0,M 0 } lie on the unit hypercube, we consider various possibilities with respect to the sign structure of those vectors. Thus, we define the following sets: P..Set of components of vectors O < 0 and M 0 } that agree (both of them) in sign with those of the vector 9 0 (and hence L ). R..Set of components of vectors O < 0 and M 0 } that DONOT agree (both of them) in sign with those of the vector 9 0 (and hence L ). S..Set of components of vector < 0 where only < 0 differs in sign from those components of vector 9 0 (and hence L ). TU..Set of components of vector M 0 where only M 0 differs in sign from those components of vector 9 0 (and hence L ). Since the components of { 9 0, < 0, M 0 } assume { +1 or -1 ] values, there are eight possibilities. The following table provides a tabular summary of the sign structure of the components of { 9 0, < 0, M 0 } and the set { P,R,S,T U } the components belong to. 9 0 < 0 M 0 Set + + + A + + - D + - + C + - - B - + + B - + - C - - + D - - - A We make the following inferences: Components of { 9 0, < 0, M 0 } that belong to Set P contribute zero value to the distances? N (;,< 0 ),? N (;,M 0 ). Components of { 9 0, < 0, M 0 } that belong to Set R contribute constant value to the distances? N (;,< 0 ),? N ( ;,M 0 ). Let R = j. It could be noted that for contradiction hypothesis R as well as P can be set to zero. By the hypothesis, the cardinality of set S, i.e. S is atleast one larger than the cardinality of the set TU i.e. TU. For concreteness, we first consider the case where S = TU + 1.

To illustrate the argument, we first consider the case where only the last component of < 0 differs from that of 9 0 in sign (but not M 0 ) and all other components (of both < 0,M 0 ) either agree or disagree in sign with those of 9 0. To proceed with the proof argument, the vectors L, < 0, M 0 (lying on the unit hypercube) are projected onto the unit hypersphere through the following transformation: Let the projected vectors be Q, R i.e. V = < 0 %,W = M 0 %, where N is the dimension of the symmetric matrix M.. Thus, we want to reason (if the Hamming distance condition specified above i.e. equation (4.1) is satisfied i.e. our hypothesis) that the value of quadratic form associated with the vectors Q, R satisfies the following inequality i.e. W W > V V. Hence the idea is to thus arrive at a contradiction to the fact that < 0 is the global optimum stable state. Note: From Theorem 4, it is clear that by a variation of diagonal elements of M, a constant value will be added to both sides of the above inequality. Thus, it is sufficient (to arrive at a contradiction) to exhibit one possible choice of diagonal elements of M for which the above inequality holds. In view of Lemma 4, we have the following expressions for V V and W W : V V = 3 456 + 23 456 (V 9 0 ) 9 0 +(V 9 0 ) (V 9 0 ) W W = 3 456 + 23 456 (W 9 0 ) 9 0 +(W 9 0 ) (W 9 0 ) Equivalently, we effectively want to show that 23 456 (W V ) 9 0 + (W 9 0 ) (W 9 0 ) (V 9 0 ) (V 9 0 ) > 0. Let us label the terms in the above expression as follows: (I) = 23 456 (W V ) 9 0 (II) = (W 9 0 ) (W 9 0 ) - (V 9 0 ) (V 9 0 ). We want to show that (I) + (II) > 0. To prove such an inequality, we first partition the vectors Q, 9 0, R (lying on the unit hypersphere) into two parts: Part (A): where components of Q, R simultaneously agree or disagree in sign with those of 9 0. Part (B): Last component of Q that disagrees in sign with the last component of 9 0. But the last component of R agrees in sign with that of 9 0. Thus, the vectors 9 0, Q, R are given by 9 0 = Y 9 0 Z [ 9 \,V = ]V Z,W = ] W Z 0 V [^ W [^ Where V [, W [ are scalars. Also the components of OV Z,W Z }

are simultaneously either + or. Thus except for the last component, all other components of the [ vector R-Q are all zeroes. Further suppose that the last component of 9 0.C.9 0 is _, with _ > 0. Then it is easy to see that (W [ V [ ) = `. Summarizing (W V ) = a0 0 0 #` b. Hence, we have that (W V ) 9 0 = ` _. Thus ( I ) = d e fgh i which is strictly greater than zero. Similarly, even when the last component of 9 0 is +_ with > 0, it is easy to reason that (W V ) 9 0 is strictly greater than zero. Now we consider the other term i.e Term (II): We first partition M into a block structured matrix i.e. = ] ` ` j ^, where j = `` is a scalar and ` = ` We also partition the vectors ( V 9 0 ),(W 9 0 ) in the following manner: ( V 9 0 ) = ] k () l () ^ ; ( W 9 0 ) = ] k(`) l (`) ^, Where l (),l (`) are scalars. As per the partitioning procedure, it is clear that k () = k (`). ([) Also, let us consider the case where the last component of 9 0.C.9 0 is _, with _ > 0. In such a case l () = + _ ; l(`) = + _. Note: For the case, where the last component of 9 0 is +_ nh _ > 0, all the following equations are suitably modified. Details are avoided for brevity. In term (II), the following definition is utilized: H = (W 9 0 ) (W 9 0 ) and J = (V 9 0 ) (V 9 0 ). Thus (II) = H J. In view of partitioning of the matrix M and vectors ( V 9 0 ),(W 9 0 ); we have that Using the fact that k () = k (`), we have that H = k (`)o k (`) + 2 k (`)o ` l (`) + l (`)o `` l (`) J = k ()o k () + 2 k ()o ` l () + l ()o `` l (). H- J = 2 k ()o ` l (`) l () + j p(l (`) )` (l () )`q Let k ()o ` = ` k () = r. Thus, we have that.

H J = 2 rk_ _ L+ ja(_ )` (_ + )`b = d s + j (#d t ). Hence, we have the following expression for (I)+(II) (I)+(II) = d e fgh t d s + #d t u. = d p r ( j_ 3 456_ ) q. But since 9 0 is the eigenvector of M corresponding to the largest eigenvalue 3 456, we have that (Z) ` 9 0 j_ = 3456 _. (Z) ` 9 0 = j_ 3456 _. Hence we necessarily have that (I) + (II) = d a r ` (Z) 9 0 b = d a ` k () (Z) ` 9 0 b = d a ` (V (Z) (Z) 9 0 ) (Z) ` 9 0 b = d p ` V (Z) q. We first note that ` is constrained by the fact that < 0 is a stable state. Thus ( < 0 ) = < 0. Or equivalently K w x L = < 0 = ( V ). Thus, we necessarily have y ] zv(z) ` ` j ^ { = } < (Z) 0 +1 ~. In view of Theorem 4, this equation must hold true irrespective of choice of diagonal elements of matrix M (i.e. diagonal elements of and j). The above equation can equivalently be specified in the following manner: K V (Z) + (Z) `L = 7 0 and (` V (Z) + u ) = + 1. Let V (Z) + (Z) ` = with ( ) = 7 0. Equivalently, we have that [ V (Z) ] V (Z) 1 + % [ V(Z) ] ` = [ V (Z) ].

This equation can equivalently be specified in the following form: ` V (Z) = % [ V (Z) ] [ V (Z) ] V (Z) ƒ From Theorem (4), we have freedom in choosing the diagonal elements of M (since Trace(M) contributes a constant value to the value of quadratic form at all the corners of hypercube and the points corners of hypercube projected onto the unit hypersphere. The location of stable / anti-stable states is invariant under arbitrary choice of diagonal elements of matrix and scalar). Please refer to the note at the beginning of the proof (by contradiction). Thus by a suitable choice of diagonal elements of, we can ensure that ` V (Z) < 0. Also, by a proper choice of, the following equation is also satisfied ( ` V (Z) + u ) = + 1. Thus, we arrive at the desired contradiction to the fact that < 0 is a global optimum stable state. Thus, the vector L is in the domain of attraction of the global optimum stable state. Hence, with this choice of initial condition, when the Hopfield neural network is run in the serial mode, global optimum stable state is reached. Note: To arrive at contradiction, the case where S = TU + 1 is sufficient. For concreteness, we consider the more general case in the following. Thus, we now consider the case Where S TU + 2. We generalize the above proof for this arbitrary case (using block matrices). Even in this case, we want to show that (I) + (II) > 0, where (I) = 23 456 (W V ) 9 0 and (II) = (W 9 0 ) (W 9 0 ) - (V 9 0 ) (V 9 0 ) > 0. Let us first consider the term (I). Partition the vectors { R, Q, 9 0 } into FOUR parts (as per the sets P,R,S,TU considered in the above discussion) i.e. 9 0 = ˆ90 Z V Z W Z [ 9 0 V ; V = Ž [ W ; W = Ž [. 9 0 9 Š 0 V W V Š W Š It is clear that from the description of the sets PU,R,S,TU, the following equations follow: W Z = V Z,W [ = V [.

( W V ) = 0 0 ˆ 0 0 2 % 2 % 2 % 2 % ; 9 0 = ˆ + +(.) H (.) + + _ _ l = Ž k _ Where A ( F-DJ-CA -H ICF-. l ), H s (F-DJ-CA -H ICF-. k), A (components of vector ), _ A (components of vector _ ) are all non-negative real numbers. Let the vector of all ones be denoted by C. Hence we have that Thus the term (I) becomes (R Q ) 9 = K ` + ` _ L C. 2 3 (R Q ) 9 = d 3 456 ( + _ ) C > 0. Using the reasoning similar to the above, it is shown that (I) + (II) > 0. Q.E.D.