Copositive Programming and Combinatorial Optimization

Similar documents
Copositive Programming and Combinatorial Optimization

A solution approach for linear optimization with completely positive matrices

Modeling with semidefinite and copositive matrices

Projection methods to solve SDP

The maximal stable set problem : Copositive programming and Semidefinite Relaxations

Solving large Semidefinite Programs - Part 1 and 2

Introduction to Semidefinite Programming I: Basic properties a

Relations between Semidefinite, Copositive, Semi-infinite and Integer Programming

1 Strict local optimality in unconstrained optimization

4. Algebra and Duality

Semidefinite Programming Basics and Applications

6-1 The Positivstellensatz P. Parrilo and S. Lall, ECC

Semidefinite and Second Order Cone Programming Seminar Fall 2001 Lecture 5

An Adaptive Linear Approximation Algorithm for Copositive Programs

Interior points of the completely positive cone

University of Groningen. Copositive Programming a Survey Dür, Mirjam. Published in: EPRINTS-BOOK-TITLE

The Difference Between 5 5 Doubly Nonnegative and Completely Positive Matrices

New Lower Bounds on the Stability Number of a Graph

Analysis of Copositive Optimization Based Linear Programming Bounds on Standard Quadratic Optimization

Lift-and-Project Techniques and SDP Hierarchies

On a Polynomial Fractional Formulation for Independence Number of a Graph

Semidefinite Programming

Scaling relationship between the copositive cone and Parrilo s first level approximation

Completely Positive Reformulations for Polynomial Optimization

Convex Optimization. (EE227A: UC Berkeley) Lecture 6. Suvrit Sra. (Conic optimization) 07 Feb, 2013

CSC Linear Programming and Combinatorial Optimization Lecture 12: The Lift and Project Method

POLYNOMIAL OPTIMIZATION WITH SUMS-OF-SQUARES INTERPOLANTS

Example: feasibility. Interpretation as formal proof. Example: linear inequalities and Farkas lemma

Optimization over Polynomials with Sums of Squares and Moment Matrices

On the Sandwich Theorem and a approximation algorithm for MAX CUT

Optimization over Structured Subsets of Positive Semidefinite Matrices via Column Generation

Variable Objective Search

I.3. LMI DUALITY. Didier HENRION EECI Graduate School on Control Supélec - Spring 2010

arxiv: v1 [math.oc] 23 Nov 2012

MINI-TUTORIAL ON SEMI-ALGEBRAIC PROOF SYSTEMS. Albert Atserias Universitat Politècnica de Catalunya Barcelona

Conic approach to quantum graph parameters using linear optimization over the completely positive semidefinite cone

CSC Linear Programming and Combinatorial Optimization Lecture 10: Semidefinite Programming

Lecture Note 5: Semidefinite Programming for Stability Analysis

Lecture: Introduction to LP, SDP and SOCP

Convex Optimization. (EE227A: UC Berkeley) Lecture 28. Suvrit Sra. (Algebra + Optimization) 02 May, 2013

SDP Relaxations for MAXCUT

Considering Copositivity Locally

Linear and non-linear programming

Semidefinite Programming

Lecture 6: Conic Optimization September 8

12. Interior-point methods

1 The independent set problem

Advanced Mathematical Programming IE417. Lecture 24. Dr. Ted Ralphs

A ten page introduction to conic optimization

Convexification of Mixed-Integer Quadratically Constrained Quadratic Programs

Technische Universität Ilmenau Institut für Mathematik

Canonical Problem Forms. Ryan Tibshirani Convex Optimization

A Brief Review on Convex Optimization

- Well-characterized problems, min-max relations, approximate certificates. - LP problems in the standard form, primal and dual linear programs

Summer School: Semidefinite Optimization

Part IB Optimisation

A notion of Total Dual Integrality for Convex, Semidefinite and Extended Formulations

Lecture 3: Lagrangian duality and algorithms for the Lagrangian dual problem

e-companion ONLY AVAILABLE IN ELECTRONIC FORM

Representations of Positive Polynomials: Theory, Practice, and

The complexity of optimizing over a simplex, hypercube or sphere: a short survey

12. Interior-point methods

CSCI : Optimization and Control of Networks. Review on Convex Optimization

MIT Algebraic techniques and semidefinite optimization February 14, Lecture 3

Semidefinite Programming, Combinatorial Optimization and Real Algebraic Geometry

c 2000 Society for Industrial and Applied Mathematics

Lecture 8 Plus properties, merit functions and gap functions. September 28, 2008

Lecture: Duality.

SEMIDEFINITE PROGRAM BASICS. Contents

Written Examination

A CONIC DANTZIG-WOLFE DECOMPOSITION APPROACH FOR LARGE SCALE SEMIDEFINITE PROGRAMMING

Agenda. Applications of semidefinite programming. 1 Control and system theory. 2 Combinatorial and nonconvex optimization

Iterative LP and SOCP-based. approximations to. sum of squares programs. Georgina Hall Princeton University. Joint work with:

Hilbert s 17th Problem to Semidefinite Programming & Convex Algebraic Geometry

Lecture 5. Theorems of Alternatives and Self-Dual Embedding

Part 4: Active-set methods for linearly constrained optimization. Nick Gould (RAL)

minimize x x2 2 x 1x 2 x 1 subject to x 1 +2x 2 u 1 x 1 4x 2 u 2, 5x 1 +76x 2 1,

On approximations, complexity, and applications for copositive programming Gijben, Luuk

5. Duality. Lagrangian

Lecture: Duality of LP, SOCP and SDP

LNMB PhD Course. Networks and Semidefinite Programming 2012/2013

Optimization Methods. Lecture 18: Optimality Conditions and. Gradient Methods. for Unconstrained Optimization

Motivation. Lecture 2 Topics from Optimization and Duality. network utility maximization (NUM) problem:

CS-E4830 Kernel Methods in Machine Learning

Primal-Dual Geometry of Level Sets and their Explanatory Value of the Practical Performance of Interior-Point Methods for Conic Optimization

The moment-lp and moment-sos approaches

Complexity of Deciding Convexity in Polynomial Optimization

16.1 L.P. Duality Applied to the Minimax Theorem

IE 521 Convex Optimization

The Primal-Dual Algorithm P&S Chapter 5 Last Revised October 30, 2006

Semidefinite programs and combinatorial optimization

Additional Homework Problems

CS711008Z Algorithm Design and Analysis

Robust and Optimal Control, Spring 2015

Lecture: Algorithms for LP, SOCP and SDP

Semidefinite Programming and Integer Programming

Lecture 4: January 26

ELE539A: Optimization of Communication Systems Lecture 15: Semidefinite Programming, Detection and Estimation Applications

ECE580 Exam 1 October 4, Please do not write on the back of the exam pages. Extra paper is available from the instructor.

Convex Optimization and l 1 -minimization

Transcription:

Copositive Programming and Combinatorial Optimization Franz Rendl http://www.math.uni-klu.ac.at Alpen-Adria-Universität Klagenfurt Austria joint work with I.M. Bomze (Wien) and F. Jarre (Düsseldorf) IMA workshop on nonlinear integer programming p.1/58

Overview What is Copositive Programming? Copositivity and combinatorial optimization? Relaxations based on CP Heuristics based on CP IMA workshop on nonlinear integer programming p.2/58

Completely Positive Matrices Let A = (a 1,...,a k ) be a nonnegative n k matrix, then X = a 1 a T 1 +... + a k a T k = AAT is called completely positive. COP = {X : X completely positive} COP is closed, convex cone. From the definition we get COP = conv{aa T : a 0}. For basics, see the book: A. Berman, N. Shaked-Monderer: Completely Positive Matrices, World Scientific 2003 IMA workshop on nonlinear integer programming p.3/58

Copositive Matrices Dual cone COP of COP in S n (sym. matrices): Y COP trxy 0 X COP a T Y a 0 vectors a 0. By definition, this means Y is copositive. CP = {Y : a T Y a 0 a 0} CP is dual cone to COP! IMA workshop on nonlinear integer programming p.4/58

Copositive Matrices Dual cone COP of COP in S n (sym. matrices): Y COP trxy 0 X COP a T Y a 0 vectors a 0. By definition, this means Y is copositive. CP = {Y : a T Y a 0 a 0} CP is dual cone to COP! Bad News: X / CP is NP-complete decision problem. Semidefinite matrices PSD: Y PSD a T Y a 0 a. Well known facts: PSD = PSD (PSD cone is selfdual. ) COP PSD CP IMA workshop on nonlinear integer programming p.5/58

Semidefinite and Copositive Programs Problems of the form max C,X s.t. A(X) = b, X PSD are called Semidefinite Programs. IMA workshop on nonlinear integer programming p.6/58

Semidefinite and Copositive Programs Problems of the form max C,X s.t. A(X) = b, X PSD are called Semidefinite Programs. Problems of the form or max C,X s.t. A(X) = b, X CP max C,X s.t. A(X) = b, X COP are called Copositive Programs, because the primal or the dual involves copositive matrices. IMA workshop on nonlinear integer programming p.7/58

Overview What is Copositive Programming? Copositivity and combinatorial optimization? Relaxations based on CP Heuristics based on CP IMA workshop on nonlinear integer programming p.8/58

Why Copositive Programs? Copositive Programs can be used to solve combinatorial optimization problems. IMA workshop on nonlinear integer programming p.9/58

Why Copositive Programs? Copositive Programs can be used to solve combinatorial optimization problems. Stable Set Problem: Let A be adjacency matrix of graph, J be all ones matrix. Theorem (DeKlerk and Pasechnik (SIOPT 2002)) α(g) = max{ J,X : A + I,X = 1, X COP } = min{y : y(a + I) J CP }. This is a copositive program with only one equation (in the primal problem). This is a simple consequence of the Motzkin-Straus Theorem. IMA workshop on nonlinear integer programming p.10/58

Proof (1) 1 α(g) = min{xt (A + I)x : x } (Motzkin-Straus Theorem) = {x : i x i = 1, x 0} is standard simplex. IMA workshop on nonlinear integer programming p.11/58

Proof (1) 1 α(g) = min{xt (A + I)x : x } (Motzkin-Straus Theorem) = {x : i x i = 1, x 0} is standard simplex. We get 0 = min{x T (A + I eet α )x : x } = min{x T (α(a + I) J)x : x 0}. This shows that α(a + I) J is copositive. IMA workshop on nonlinear integer programming p.12/58

Proof (1) 1 α(g) = min{xt (A + I)x : x } (Motzkin-Straus Theorem) = {x : i x i = 1, x 0} is standard simplex. We get 0 = min{x T (A + I eet α )x : x } = min{x T (α(a + I) J)x : x 0}. This shows that α(a + I) J is copositive. Therefore inf{y : y(a + I) J CP } α. IMA workshop on nonlinear integer programming p.13/58

Proof (2) Weak duality of copositive program gives: sup{ J,X : A + I,X = 1, X COP } inf{y : y(a + I) J CP } α. IMA workshop on nonlinear integer programming p.14/58

Proof (2) Weak duality of copositive program gives: sup{ J,X : A + I,X = 1, X COP } inf{y : y(a + I) J CP } α. Now let ξ be incidence vector of a stable set of size α. The matrix 1 α ξξt is feasible for the first problem. Therefore α sup{...} inf{...} α. This shows that equality holds throughout and sup and inf are attained. The recent proof of this result by DeKlerk and Pasechnik does not make explicit use of the Motzkin Straus Theorem. IMA workshop on nonlinear integer programming p.15/58

Connections to theta function Theta function (Lovasz (1979) ): ϑ(g) = max{ J,X : x ij = 0 ij E, tr(x) = 1,X 0} α(g). Motivation: If ξ characteristic vector of stable set, then 1 ξ T ξ ξξt is feasible for above SDP. IMA workshop on nonlinear integer programming p.16/58

Connections to theta function Theta function (Lovasz (1979) ): ϑ(g) = max{ J,X : x ij = 0 ij E, tr(x) = 1,X 0} α(g). Motivation: If ξ characteristic vector of stable set, then 1 ξ T ξ ξξt is feasible for above SDP. Schrijver (1979) improvement: include X 0 In this case we can add up the constraints x ij = 0 and get ϑ (G) = max{ J,X : A,X = 0, tr(x) = 1,X 0,X 0}. (A... adjacency matrix). We have ϑ(g) ϑ (G) α(g). Replacing the cone X 0,X 0 by X COP gives α(g), see before. IMA workshop on nonlinear integer programming p.17/58

A general copositive modeling theorem Burer (2007) shows the following general result for the power of copositive programming: The optimal values of P and C are equal: opt(p) = opt(c) (P) minx T Qx + c T x Here x IR n and m n. a T i x = b i, x 0, x i {0, 1} i m. (C) min tr(qx) + c T x, s.t. a T i x = b i, ( ) a T i Xa i = b 2 1 x T i, X ii = x i i m, COP x X IMA workshop on nonlinear integer programming p.18/58

Overview What is Copositive Programming? Copositivity and combinatorial optimization? Relaxations based on CP Heuristics based on CP IMA workshop on nonlinear integer programming p.19/58

Approximating COP We have now seen the power of copositive programming. Since optimizing over CP is NP-Hard, it makes sense to get approximations of CP or COP. To get relaxations, we need supersets of COP, or inner approximations of CP (and work on the dual cone). The Parrilo hierarchy uses Sum of Squares and provides such an outer approximation of COP (dual viewpiont!). We can also consider inner approximations of COP. This can be viewed as a method to generate feasible solutions of combinatorial optimization problems ( primal heuristic!). IMA workshop on nonlinear integer programming p.20/58

Relaxations Inner approximation of CP. CP := {M : x T Mx 0 x 0} Parrilo (2000) and DeKlerk, Pasechnik (2002) use the following idea to approximate CP from inside: M CP iff P(x) := ij x 2 ix 2 jm ij 0 x. A sufficient condition for this to hold is that P(x) has a sum of squares (SOS) representation. Theorem Parrilo (2000) : P(x) has SOS iff M = P + N, where P 0 and N 0. IMA workshop on nonlinear integer programming p.21/58

Parrilo hierarchy To get tighter approximations, Parrilo proposes to consider SOS representations of P r (x) := ( i x 2 i) r P(x) for r = 0, 1,... (For r = 0 we get the previous case.) Mathematical motivation by an old result of Polya. Theorem Polya (1928): If M strictly copositive then P r (x) is SOS for some sufficiently large r. IMA workshop on nonlinear integer programming p.22/58

Parrilo hierarchy (2) Parrilo characterizes SOS for r = 0, 1: P 0 (x) is SOS iff M = P + N, where P 0 and N 0. P 1 (x) is SOS iff M 1,...,M n such that M M i 0 (M i ) ii = 0 i (M i ) jj + 2(M j ) ij = 0 i j (M i ) jk + (M j ) ik + (M k ) ij 0 i < j < k The resulting relaxations are SDP. But the r = 1 relaxation involves n matrices and n SDP constraints to certify SOS. This is computationally challenging. IMA workshop on nonlinear integer programming p.23/58

Overview What is Copositive Programming? Copositivity and combinatorial optimization? Relaxations based on CP Heuristics based on CP IMA workshop on nonlinear integer programming p.24/58

Inner approximations of COP We consider min C,X s.t. A(X) = b, X COP Remember: COP= {X : X = V V T, V 0}. Some previous work by: Bomze, DeKlerk, Nesterov, Pasechnik, others: Get stable sets by approximating COP formulation of the stable set problem using optimization of quadratic over standard simplex, or other local methods. Bundschuh, Dür (2008): linear inner and outer approximations of CP IMA workshop on nonlinear integer programming p.25/58

Incremental version A general feasible descent approach: Let X = V V T with V 0 be feasible. Consider the regularized, and convex descent step problem: min ǫ C, X + (1 ǫ) X 2, such that A( X) = 0, X + X COP. IMA workshop on nonlinear integer programming p.26/58

Incremental version A general feasible descent approach: Let X = V V T with V 0 be feasible. Consider the regularized, and convex descent step problem: min ǫ C, X + (1 ǫ) X 2, such that A( X) = 0, X + X COP. For small ǫ > 0 we approach the true optimal solution, because we follow the continuous steepest descent path, projected onto COP. Unfortunately, this problem is still not tractable. We approximate it by working in the V -space instead of the X-space. IMA workshop on nonlinear integer programming p.27/58

Incremental version: V -space X + = (V + V )(V + V ) T hence, X = V V T X = X( V ) = V V T + V V T + ( V )( V T ). Now linearize and make sure V is small. IMA workshop on nonlinear integer programming p.28/58

Incremental version: V -space X + = (V + V )(V + V ) T hence, X = V V T X = X( V ) = V V T + V V T + ( V )( V T ). Now linearize and make sure V is small. We get the following subproblem: min ǫ 2CV, V + (1 ǫ) V 2 such that 2A i, V = b i A i V,V i, V + V 0 This is convex approximation of nonconvex version in X( V ). IMA workshop on nonlinear integer programming p.29/58

Algorithm: basic version Input: (A,b,C), 0 < ǫ < 1, V 0 repeat until done - Solve subproblem min ǫ 2CV, V + (1 ǫ) V 2 s. t. 2A i V, V = b i A i V,V i, V + V 0. - Update: V V + V end repeat Possible stopping condition: V and b A(X) should be small. IMA workshop on nonlinear integer programming p.30/58

Correcting the linearization The linearization V V T + V V T introduces an error both in the cost function and in the constraints A(X) = b. We therefore include corrector iterations of the form V = V old + V corr before the actual update V V + V. IMA workshop on nonlinear integer programming p.31/58

Correcting the linearization The linearization V V T + V V T introduces an error both in the cost function and in the constraints A(X) = b. We therefore include corrector iterations of the form V = V old + V corr before the actual update V V + V. Final Subproblem: Additional input: V old (initially set to 0) min ǫ 2C(V + V old ), V + (1 ǫ) V 2 s.t. 2A i (V + V old ), V = b i A(V + V old ),V + V old, V + V old + V 0. IMA workshop on nonlinear integer programming p.32/58

Convex quadratic subproblem The convex subproblem is of the following form, after appropriate redefinition of data and variables x = vec(v + V ),... min c T x + ρx T x such that Rx = r, x 0. IMA workshop on nonlinear integer programming p.33/58

Convex quadratic subproblem The convex subproblem is of the following form, after appropriate redefinition of data and variables x = vec(v + V ),... min c T x + ρx T x such that Rx = r, x 0. If V is n k (k columns generating V ), x if of dimension nk and there are m equations. IMA workshop on nonlinear integer programming p.34/58

Convex quadratic subproblem The convex subproblem is of the following form, after appropriate redefinition of data and variables x = vec(v + V ),... min c T x + ρx T x such that Rx = r, x 0. If V is n k (k columns generating V ), x if of dimension nk and there are m equations. Since Hessian of cost function is identity matrix, this problem can be solved efficiently using interior-point methods. (convex quadratic with sign constraints and linear equations) Main effort is solving a linear system of order m, essentially independent of n and k. IMA workshop on nonlinear integer programming p.35/58

Test data sets COP problems coming from formulations of NP-hard problems are too difficult ( Stable Set, Coloring, Quadratic Assignment) to test new algorithms. Would like to have: Data (A,b,C) and (X,y,Z) such that (X,y,Z) is optimal for primal and dual (no duality gap). COP is nontrivial (optimum not given by optimizing over semidefiniteness plus nonnegativity ) generate instances of varying size both in n and m. Hard part: Z provably copositive!! IMA workshop on nonlinear integer programming p.36/58

A COP generator Basic idea: Let G be a graph, and (X G,y G,Z G ) be optimal solution of ω(g) = max{ J,X : A G + I,X = 1, X COP } = min{y : y(a G + I) J CP }. IMA workshop on nonlinear integer programming p.37/58

A COP generator Basic idea: Let G be a graph, and (X G,y G,Z G ) be optimal solution of ω(g) = max{ J,X : A G + I,X = 1, X COP } Would like to have = min{y : y(a G + I) J CP }. ω(g) < max{ J,X : A G + I,X = 1,X 0,X 0} = ϑ (G). This insures that SDP is not enough to solve the problem. Optimal complementary pair: Z G = y G (A G + I) J and X G = i λ iξ i ξ T i with ξ i characteristic vectors of max clique. IMA workshop on nonlinear integer programming p.38/58

A COP generator (2) Reverse Engineering: Select A i (i = 1,...,m) and set b = A(X G ), i.e. b i = A i,x G. Select y IR m and set C = Z G A T (y) Now (X G,y,Z G ) is optimal for data (A,b,C). IMA workshop on nonlinear integer programming p.39/58

A COP generator (2) Reverse Engineering: Select A i (i = 1,...,m) and set b = A(X G ), i.e. b i = A i,x G. Select y IR m and set C = Z G A T (y) Now (X G,y,Z G ) is optimal for data (A,b,C). To generate bigger instances, we select G = H K (strong graph product) where K is a perfect graph and H is such that ω(h) < ϑ (H). We get A G = A H I + I A K + A H A K. Facts: ω(g) = ω(h)ω(k) and ϑ (G) = ϑ (H)ϑ (K). Hence, we have optimal solution y G = ω H ω K,Z G = y G (A G + I) J,X G = X H X K. In our generator we set H = C 5 (5-cycle) IMA workshop on nonlinear integer programming p.40/58

Computational results A sample instance with n = 60, m = 100. z sdp = 9600, 82, z sdp+nonneg = 172.19,z cop = 69.75 IMA workshop on nonlinear integer programming p.41/58

Computational results A sample instance with n = 60, m = 100. z sdp = 9600, 82, z sdp+nonneg = 172.19,z cop = 69.75 it b-a(x) f(x) 1 0.002251-68.7274 5 0.000014-69.5523 10 0.000001-69.6444 15 0.000001-69.6887 20 0.000000-69.6963 The number of inner iterations was set to 5, column 1 shows the outer iteration count. IMA workshop on nonlinear integer programming p.42/58

Computational results A sample instance with n = 60, m = 100. z sdp = 9600, 82, z sdp+nonneg = 172.19,z cop = 69.75 it b-a(x) f(x) 1 0.002251-68.7274 5 0.000014-69.5523 10 0.000001-69.6444 15 0.000001-69.6887 20 0.000000-69.6963 The number of inner iterations was set to 5, column 1 shows the outer iteration count. But starting point: V 0 =.95 Vopt +.05 rand IMA workshop on nonlinear integer programming p.43/58

Computational results (2) Example (continued). recall n = 60, m = 100. z sdp = 9600, 82, z sdp+nonneg = 172.19,z cop = 69.75 start iter b-a(x) f(x) (a) 20 0.000000-69.696 (b) 20 0.000002-69.631 (c) 50 0.000008-69.402 Different starting points: (a) V =.95 * V opt +.05 * rand (b) V =.90 * V opt +.10 * rand (c) V = rand(n, 2n) IMA workshop on nonlinear integer programming p.44/58

Random Starting Point Example (continued), n = 60, m = 100. z sdp = 9600, 82, z sdp+nonneg = 172.19,z cop = 69.75 it b-a(x) f(x) 1 6.121227 1831.5750 5 0.021658 101.1745 10 0.002940-43.4477 20 0.000147-67.0989 30 0.000041-68.7546 40 0.000015-69.2360 50 0.000008-69.4025 Starting point: V 0 = rand(n, 2n) IMA workshop on nonlinear integer programming p.45/58

More results n m opt found b A(X) 50 100 314.48 314.90 4 10 5 60 120-266.99-266.48 4 10 5 70 140-158.74-157.55 3 10 5 80 160-703.75-701.68 5 10 5 100 100-659.65-655.20 8 10 5 Starting point in all cases: rand(n,2n) Inner iterations: 5 Outer iterations: 30 The code works on random instances. Now some more serious experiments. IMA workshop on nonlinear integer programming p.46/58

Some experiments with Stable Set max J,X such that tr(x) = 1, tr(a G X) = 0, X COP Only two equations but many local optima. IMA workshop on nonlinear integer programming p.47/58

Some experiments with Stable Set max J,X such that tr(x) = 1, tr(a G X) = 0, X COP Only two equations but many local optima. We consider a selection of graphs from the DIMACS collection. Computation times in the order of a few minutes. name n ω clique found keller4 171 11 9 brock200-4 200 17 14 c-fat200-1 200 12 12 c-fat200-5 200 58 58 brock400-1 400 27 24 p-hat500-1 500 9 8 IMA workshop on nonlinear integer programming p.48/58

Sufficient condition for Non-copositivity To show that M / CP, consider min{x T Mx : e T x = 1, x 0} and try to solve this through min{ M,X : e T x = 1, J,X = 1, ( 1 x T x X ) COP }. Our method is local, but once we have feasible solution with negative value, we have a certificate for M / CP. We apply this to get another heuristic for stable sets. IMA workshop on nonlinear integer programming p.49/58

Stable sets - second approach If we can show that for some integer t Q = t(a + I) J is not copositive, then α(g) t + 1. IMA workshop on nonlinear integer programming p.50/58

Stable sets - second approach If we can show that for some integer t Q = t(a + I) J is not copositive, then α(g) t + 1. Hence we consider min{x T Qx : e T x = 1,x 0} and translate this into ( ) min{ Q,X : e T 1 x T x = 1, J,X = 1, COP }. x X If we find solution with value < 0, then we have certificate that α(g) t + 1. IMA workshop on nonlinear integer programming p.51/58

Stable sets - second approach If we can show that for some integer t Q = t(a + I) J is not copositive, then α(g) t + 1. Hence we consider min{x T Qx : e T x = 1,x 0} and translate this into ( ) min{ Q,X : e T 1 x T x = 1, J,X = 1, COP }. x X If we find solution with value < 0, then we have certificate that α(g) t + 1. Note that we prove existence of stable set of size t + 1, without actually providing such a set. IMA workshop on nonlinear integer programming p.52/58

Some preliminary experiments We look at easy instances, given by interval graphs, where stability number is known. n ω # max cliques new method 100 20 181 20 500 50 451 50 1000 100 901 100 IMA workshop on nonlinear integer programming p.53/58

Some preliminary experiments We look at easy instances, given by interval graphs, where stability number is known. n ω # max cliques new method 100 20 181 20 500 50 451 50 1000 100 901 100 Some DIMACS graphs name n ω clique found keller4 171 11 11 brock200-4 200 17 15 brock400-1 400 27 23 p-hat500-1 500 9 9 IMA workshop on nonlinear integer programming p.54/58

Generating Copositive Cuts We can also use the heuristic test for non-copositivity to generate CP-cuts. Under strong duality assumptions we have: min{ C,X : A(X) = b,x COP } = max{b T y : C A T (y) = Z CP } From our convex QP-subproblem, we get multipliers y for the equations A(X) = b. If our heuristic finds that C A T (y) / CP, it provides v 0 such that v T (C A T (y))v < 0. IMA workshop on nonlinear integer programming p.55/58

Generating Copositive Cuts (2) We add to the relaxed dual C A T (y),vv T 0 min{b T y : C A T (y) = Z PSD, C A T (y),vv T 0} to get better relaxation. This allows to estimate the quality of our heuristics. Work in progress, currently no practical experiments. IMA workshop on nonlinear integer programming p.56/58

Research Topics Explore the potential of copositive cuts Explore this approach in the context of (nonconvex) quadratic optimization with binary variables (heuristics and bounds) The complexity status of checking whether M COP is open. An obvious observation: M COP X = (x ij ) such that, with Y = (x 2 ij),m = Y Y T. This is a system of order 4 polynomial equations. Can you do better? IMA workshop on nonlinear integer programming p.57/58

Last Slide We have seen the power of copositivity. Relaxations: The Parrilo hierarchy is computationally too expensive. Other way to approximate CP? Heuristics: Unfortunately, the subproblem may have local solutions, which are not local minima for the original descent step problem. The number of columns of V does not need to be larger than ( n+1) 2, but for practical purposes, this is too large. Further technical details in a forthcoming paper by I. Bomze, F. Jarre and F. R.: Quadratic factorization heuristics for copositive programming, technical report, (2008). IMA workshop on nonlinear integer programming p.58/58