Lift-and-Project Techniques and SDP Hierarchies

Similar documents
MINI-TUTORIAL ON SEMI-ALGEBRAIC PROOF SYSTEMS. Albert Atserias Universitat Politècnica de Catalunya Barcelona

Notes on the decomposition result of Karlin et al. [2] for the hierarchy of Lasserre by M. Laurent, December 13, 2012

Introduction to LP and SDP Hierarchies

Lecture : Lovász Theta Body. Introduction to hierarchies.

The maximal stable set problem : Copositive programming and Semidefinite Relaxations

Introduction to Semidefinite Programming I: Basic properties a

CSC Linear Programming and Combinatorial Optimization Lecture 12: The Lift and Project Method

On the matrix-cut rank of polyhedra

Sherali-Adams relaxations of the matching polytope

linear programming and approximate constraint satisfaction

Cuts for mixed 0-1 conic programs

Semidefinite programming techniques in combinatorial optimization

Hierarchies. 1. Lovasz-Schrijver (LS), LS+ 2. Sherali Adams 3. Lasserre 4. Mixed Hierarchy (recently used) Idea: P = conv(subset S of 0,1 n )

Semidefinite and Second Order Cone Programming Seminar Fall 2001 Lecture 5

Copositive Programming and Combinatorial Optimization

On the Rank of Cutting-Plane Proof Systems

Lift-and-Project Inequalities

CORC REPORT Approximate fixed-rank closures of covering problems

Lower bounds on the size of semidefinite relaxations. David Steurer Cornell

Hilbert s 17th Problem to Semidefinite Programming & Convex Algebraic Geometry

Copositive Programming and Combinatorial Optimization

Modeling with semidefinite and copositive matrices

6-1 The Positivstellensatz P. Parrilo and S. Lall, ECC

LOVÁSZ-SCHRIJVER SDP-OPERATOR AND A SUPERCLASS OF NEAR-PERFECT GRAPHS

Sherali-Adams Relaxations of the Matching Polytope

A NEW SEMIDEFINITE PROGRAMMING HIERARCHY FOR CYCLES IN BINARY MATROIDS AND CUTS IN GRAPHS

Cutting Planes for First Level RLT Relaxations of Mixed 0-1 Programs

c 2000 Society for Industrial and Applied Mathematics

WHEN DOES THE POSITIVE SEMIDEFINITENESS CONSTRAINT HELP IN LIFTING PROCEDURES?

Integrality Gaps of Linear and Semi-definite Programming Relaxations for Knapsack

Sherali-Adams Relaxations of Graph Isomorphism Polytopes. Peter N. Malkin* and Mohammed Omar + UC Davis

1 Strict local optimality in unconstrained optimization

A Linear Round Lower Bound for Lovasz-Schrijver SDP Relaxations of Vertex Cover

Integrality Gaps for Sherali Adams Relaxations

constraints Ax+Gy» b (thus any valid inequalityforp is of the form u(ax+gy)» ub for u 2 R m + ). In [13], Gomory introduced a family of valid inequali

Improving Integrality Gaps via Chvátal-Gomory Rounding

1 The independent set problem

Carnegie Mellon University, Pittsburgh, USA. April Abstract

Semidefinite Programming and Integer Programming

Optimal Sherali-Adams Gaps from Pairwise Independence

47-831: Advanced Integer Programming Lecturer: Amitabh Basu Lecture 2 Date: 03/18/2010

A notion of Total Dual Integrality for Convex, Semidefinite and Extended Formulations

Applications of semidefinite programming in Algebraic Combinatorics

Math Matrix Algebra

Part II Strong lift-and-project cutting planes. Vienna, January 2012

1 Solution of a Large-Scale Traveling-Salesman Problem... 7 George B. Dantzig, Delbert R. Fulkerson, and Selmer M. Johnson

3.7 Cutting plane methods

Lecture 3: Semidefinite Programming

ON THE CONNECTION OF THE SHERALI-ADAMS CLOSURE AND BORDER BASES

Approximating Sparsest Cut in Graphs of Bounded Treewidth

SDP Relaxations for MAXCUT

Cutting planes from extended LP formulations

On Linear and Semidefinite Programming Relaxations for Hypergraph Matching

Cutting Plane Methods II

Deciding Emptiness of the Gomory-Chvátal Closure is NP-Complete, Even for a Rational Polyhedron Containing No Integer Point

Relations between Semidefinite, Copositive, Semi-infinite and Integer Programming

Mustapha Ç. Pinar 1. Communicated by Jean Abadie

Tight Integrality Gaps for Lovasz-Schrijver LP Relaxations of Vertex Cover and Max Cut

Advances in Convex Optimization: Theory, Algorithms, and Applications

Integer Programming: Cutting Planes

Change in the State of the Art of (Mixed) Integer Programming

Semidefinite programs and combinatorial optimization

Semidefinite Programming Basics and Applications

Lecture 5. Max-cut, Expansion and Grothendieck s Inequality

Downloaded 05/14/13 to Redistribution subject to SIAM license or copyright; see

Cutting Plane Methods I

6.854J / J Advanced Algorithms Fall 2008

Optimization over Structured Subsets of Positive Semidefinite Matrices via Column Generation

Polyhedral Approach to Integer Linear Programming. Tepper School of Business Carnegie Mellon University, Pittsburgh

Relaxations of combinatorial problems via association schemes

A General Framework for Convex Relaxation of Polynomial Optimization Problems over Cones

Optimization Exercise Set n. 4 :

Analysis of Copositive Optimization Based Linear Programming Bounds on Standard Quadratic Optimization

Improved bounds on book crossing numbers of complete bipartite graphs via semidefinite programming

Optimization over Polynomials with Sums of Squares and Moment Matrices

Dimension reduction for semidefinite programming

arxiv: v2 [cs.ds] 11 Jun 2012

arxiv:math/ v1 [math.co] 23 May 2000

Lecture Semidefinite Programming and Graph Partitioning

approximation algorithms I

MIT Algebraic techniques and semidefinite optimization February 14, Lecture 3

CORC Technical Report TR Cuts for mixed 0-1 conic programming

8 Approximation Algorithms and Max-Cut

Semidefinite programming lifts and sparse sums-of-squares

A Note on Representations of Linear Inequalities in Non-Convex Mixed-Integer Quadratic Programs

Hypergraph Matching by Linear and Semidefinite Programming. Yves Brise, ETH Zürich, Based on 2010 paper by Chan and Lau

Section Notes 9. IP: Cutting Planes. Applied Math 121. Week of April 12, 2010

3.8 Strong valid inequalities

TRISTRAM BOGART AND REKHA R. THOMAS

Limitations of Linear and Semidefinite Programs by Grant Robert Schoenebeck

arxiv: v15 [math.oc] 19 Oct 2016

Lecture 2. Split Inequalities and Gomory Mixed Integer Cuts. Tepper School of Business Carnegie Mellon University, Pittsburgh

Geometric Ramifications of the Lovász Theta Function and Their Interplay with Duality

Optimization Exercise Set n.5 :

Breaking the Rectangle Bound Barrier against Formula Size Lower Bounds

Conic approach to quantum graph parameters using linear optimization over the completely positive semidefinite cone

Representations of Positive Polynomials: Theory, Practice, and

Convex sets, conic matrix factorizations and conic rank lower bounds

On the Chvatál-Complexity of Binary Knapsack Problems. Gergely Kovács 1 Béla Vizvári College for Modern Business Studies, Hungary

Relaxations and Randomized Methods for Nonconvex QCQPs

Transcription:

MFO seminar on Semidefinite Programming May 30, 2010

Typical combinatorial optimization problem: max c T x s.t. Ax b, x {0, 1} n P := {x R n Ax b} P I := conv(k {0, 1} n ) LP relaxation Integral polytope to be found Goal: Construct a new relaxation P such that P I P P, leading to P I after finitely many iterations. Gomory-Chvatal closure: P = {x u T Ax u T b u 0 with u T A integer}. P I is found after O(n 2 log n) iterations if P [0, 1] n. [Eisenbrand-Schulz 1999] But optimization over P is hard! [Eisenbrand 1999]

Plan of the lecture Goal: We present several techniques to construct a hierarchy of LP/SDP relaxations: P P 1... P n = P I. Lovász-Schrijver N / N + operators Sherali-Adams construction Lasserre construction LP / SDP LP SDP Great interest recently in such hierarchies: Polyhedral combinatorics: How many rounds are needed to find P I? Which valid inequalities are satisfied after t rounds? Complexity theory: What is the integrality gap after t rounds? Link to hardness of the problem? Proof systems: Use hierarchies as a model to generate inequalities and show e.g. P I =.

Lift-and-project strategy 1. Generate new constraints: Multiply the system Ax b by products of the constraints x i 0 and 1 x i 0. Polynomial system in x. 2. Linearize (and lift) by introducing new variables y I for products i I x i and setting x 2 i = x i. Linear system in (x, y). 3. Project back on the x-variable space. LP relaxation P satisfying P I P P.

Some notation Write Ax b as al T x b l (l = 1,..., m) ( ) 1 or as gl T 0 (l = 1,..., m) x ( ) bl setting g l =. a l Homogenization of P: { ( ) } 1 } P = λ λ 0, x P = {y R n+1 g T x l y 0 (l = 1,..., m) V = {1,..., n}.

The Lovász-Schrijver construction 1. Multiply Ax b by x i, 1 x i i V. ( ) ( 1 1 Quadratic system: gl T x x i, g l x 2. Linearize: Introduce the matrix variable Y = indexed by {0} V. Then, Y belongs to ) (1 x i ) 0 i ( )( ) T 1 1, x x M(P) = {Y S n+1 Y 0i = Y ii, Ye i, Y (e 0 e i ) P i}, 3. Project: N(P) = N + (P) = M + (P) = M(P) S + n+1. { x R V Y M(P) s.t. { x R V Y M + (P) s.t. ( ) } 1 = Ye x 0 ( ) } 1 = Ye x 0

Properties of the N- and N + -operators P I N + (P) N(P) P. N(P) conv(p {x x i = 0, 1}) for all i V. N n (P) = P I. Assume one can optimize in polynomial time over P. Then the same holds for N t (P) and for N+(P) t for any fixed t. Example: Consider the l 1 -ball centered at e/2: { P = x R V i I x i + } i V \I 1 x i 1 2 I V. Then: P I =, but 1 2 e Nn 1 + (P). n iterations of the N + operator are needed to find P I

Application to stable sets [Lovász-Schrijver 1991] P = FRAC(G) = {x R V + x i + x j 1 (ij E)} P I = STAB(G): stable set polytope of G = (V, E). N(FRAC(G)) = FRAC(G) intersected by the constraints: i V (C) x i C 1 2 for all odd circuits C. Y M(FRAC(G)) = y ij = 0 for edges ij E. N + (FRAC(G)) TH(G). Any clique inequality i Q x i 1 is valid for N + (P), while its N-rank is Q 2. The N + operator helps! n α(g) 2 N-rank n α(g) 1. N + -rank α(g) [equality if G = line graph of K 2p+1 ]

The Sherali-Adams construction 1. Multiply Ax b by x i (1 x j ) for all disjoint I, J V with I J = t. i I j J 2. Linearize & lift: Introduce new variables y U for all U P t (V ), setting x 2 i = x i and y i = x i. 3. Project back on x-variables space. Relaxation: SA t (P). Then: SA 1 (K) = N(P), Thus: SA t (P) N t (P). SA t (P) N(SA t 1 (P)).

Application to the matching polytope For G = (V, E), let P = {x R E + x(δ(v)) 1 v V }. Then: P I is the matching polytope ( = stable set polytope of the line graph of G). For G = K 2p+1 : N + -rank = p [Stephen-Tunçel 1999] N-rank [2p, p 2 ] [LS 1991] [Goemans-Tunçel 2001] SA-rank = 2p 1 [Mathieu-Sinclair 2009] Detailed analysis of the integrality gap: g t = max x SA t(p) e t x max x P e T x g t = = max x SA t(p) e t x. p 1 + 1 2p if t p 1 1 if t 2p 1 phase transition at 2p Θ( p)

A canonical lifting lemma x {0, 1} n y x = ( x i ) I V {0, 1} P(V ) i I = (1, x 1,.., x n, x 1 x 2,.., x n 1 x n,.., x i ) i V Z: matrix with columns y x for x {0, 1} n. Equivalently: Z is the 0/1 matrix indexed by P(V ) with Z(I, J) = 1 if I V, 0 else. Z 1 (I, J) = ( 1) J\I if I J, 0 else. If x P {0, 1} n, then Y = y x (y x ) T satisfies: Y 0 Y l = g l (x)y 0 localizing matrix Y (I, J) depends only on I J moment matrix y R P(V ) Y = M V (y) = (y I J ), Y l = M V (g l y)

Lemma: P I is equal to the projection on the x-variable space of {y R P(V ) y 0 = 1, M V (y) 0, M V (g l y) 0 l}. Sketch of proof: 1. Verify that M V (y) = Z diag(z 1 y)z T. 2. M V (y) 0 = λ := Z 1 y 0 = y = Zλ = λ x y x x {0,1} n where x λ x = y 0 = 1. 3. Use M V (g l y) 0 to show that λ x > 0 = x P (= x P I ). Each 0/1 polytope is projection of a simplex.

Case n = 2 Z = 1 2 12 1 1 1 1 1 0 1 0 1 2 0 0 1 1 Z 1 = 12 0 0 0 1 1 2 12 1 1 1 1 1 0 1 0 1 2 0 0 1 1 12 0 0 0 1 y 0 y 1 y 2 y 12 M V (y) = y 1 y 1 y 12 y 12 y 2 y 12 y 2 y 12 0 y 12 y 12 y 12 y 12 y 0 y 1 y 2 + y 12 0 y 1 y 12 0 y 2 y 12 0 y 12 0

SDP hierarchies Idea: Get SDP hierarchies by truncating M V (y) and M V (g l y): Consider M U (y) = (y I J ) I,J U, indexed by P(U) for U V, or M t (y) = (y I J ) I, J t, indexed by P t (V ) for some t n. 1. (local) Get the Sherali-Adams relaxation SA t (P) when considering M U (y) 0, M W (g l y) 0 U P t (V ), W P t 1 (V ). LP with variables y I for all I P t (V ) 2. (global) Get the Lasserre relaxation L t (P) when considering M t (y) 0, M t 1 (g l y) 0. Obviously: L t (P) SA t (P). SDP with variables y I for all I P 2t (V )

Link to the Lovász-Schrijver construction L 1 (P) P, L t (P) N + (L t 1 (P)). Thus: L t (P) N t 1 + (P). L t (P) is tighter but more expensive to compute! The SDP for L t (P) involves one matrix of order O(n t ), O(n 2t ) variables. The SDP for N t 1 + (P) involves O(nt 1 ) matrices of order n + 1, O(n t+1 ) variables. Note: One can define a (block-diagonal) hierarchy, in-between and cheaper than both L t (P) and N+ t 1 (P); roughly, unfold the recursive definition of the LS hierarchy, and consider suitably defined principal submatrices of M t (y) (which can be block-diagonalized to blocks of order n + 1). [Gvozdenovic-L-Vallentin 2009]

Application of the Lasserre construction to stable sets The localizing conditions in L t (FRAC(G)) boil down to the edge conditions: y ij = 0 (ij E) (for t 2). Natural generalization of the theta body TH(G). Get the bound las (t) (G). Convergence in α(g) steps: L t (FRAC(G)) = STAB(G) for t α(g). Open: Exist graphs G for which α(g) steps are needed? Question: What is the Lasserre rank of the matching polytope?

Application of the Lasserre construction to Max-Cut Max-Cut: max ij E w ij 1 x i x j 2 s.t. x {±1} V. Consider P = [ 1, 1] V, write xi 2 = 1, and project onto the subspace R (n 2) indexed by edges. The order 1 relaxation is the basic GW relaxation: max ij E w ij 1 X ij 2 s.t. X S + n, diag(x ) = e. The Lasserre rank of CUT(K n ) is at least n/2. [La 2003] (First time when ij E(K n) x ij n/2 becomes valid). Question: Does equality hold? (Yes for n 7).

The Lasserre relaxation of order 2 relaxation satisfies the triangle inequalities: Y = 12 13 23 1 y 12 y 13 y 23 12 y 12 1 y 23 y 13 13 y 13 y 23 1 y 12 0 23 y 23 y 13 y 12 1 = e T Ye 0 = y 12 + y 13 + y 23 1.

Some negative results about integrality gaps of hierarchies for max-cut Consider the basic LP relaxation of max-cut defined by the triangle inequalities. Its integrality gap is 1/2. [Schoenebeck-Trevisan-Tulsiani 2006] For the Lovász-Schrijver construction: The integrality gap remains 1/2 + ɛ after c ɛ n rounds of the N operator. But the integrality gap is 0.878 after one round of the N + operator. [Charikar-Makarychev-Makarychev 2009] For the Sherali-Adams construction: The integrality gap remains 1/2 + ɛ after n γɛ iterations.

Some positive results Chlamtac-Singh [2008] give (for the first time) an approximation algorithm whose approximation guarantee improves indefinitely as one uses higher order relaxations in the SDP hierarchy: For the maximum independent set problem in a 3-uniform hypergraph G. Namely: Given γ > 0, assuming G contains an independent set of cardinality γn, then one can find an independent set of cardinality n Ω(γ2) using the relaxation of order Θ(1/γ 2 ).

Extensions to optimization over polynomials Minimize p(x) over {x g j (x) 0}. Linearize p = α p αx α by α p αy α. Impose SDP conditions on the moment matrix: M t (y) = (y α+β ) 0. hierarchy of SDP relaxations with asymptotic convergence (due to some SOS representation results). Exploit equations: h j (x) = 0. We saw how to exploit x 2 i = x i. The canonical lifting lemma extends to the finite variety case: when the equations h j = 0 have finitely many roots. Finite convergence of the hierarchy when the equations h j = 0 have finitely many real roots.

Another hierarchy construction via copositive programming Reformulation: α(g) = min λ s.t. λ(i + A G ) J C n, where C n = {M S n x T Mx 0 x R n +} is the copositive cone. Idea [Parrilo 2000]: Replace C n by the subcones L (t) n ( n ) r = {M S n (x T Mx) x i i=1 has non-negative coefficients}, ( n K n (t) = {M S n i,j=1 L (t) n M ij x 2 i x 2 j )( n K (t) n C n. i=1 [Pólya] If M is strictly copositive then M t 0 L(t) n. x 2 i ) t is SOS},

LP bound: ν (t) (G) = min λ s.t. λ(i + A G ) J L (t) n, SDP bound: ϑ (t) (G) = min λ s.t. λ(i + A G ) J K (t) n. ν (t) (G) < t α(g) 1. ν (t) (G) = α(g) if t α(g) 2. ϑ (0) (G) = ϑ (G) Conjecture: [de Klerk-Pasechnik 2002] ϑ (t) (G) = α(g) for t α(g) 1. Yes: For graphs with α(g) 8 [Gvozdenovic-La 2007] The Lasserre hierarchy refines the copositive hierarchy: las (t+1) (G) ϑ (t) (G). Note: The convergence in α(g) steps was easy for the Lasserre hierarchy!