Lecture 5. Max-cut, Expansion and Grothendieck s Inequality

Similar documents
SDP Relaxations for MAXCUT

Introduction to Semidefinite Programming I: Basic properties a

Maximum cut and related problems

Grothendieck s Inequality

Lecture 4: Polynomial Optimization

ORIE 6334 Spectral Graph Theory November 22, Lecture 25

Sum-of-Squares Method, Tensor Decomposition, Dictionary Learning

U.C. Berkeley CS294: Spectral Methods and Expanders Handout 11 Luca Trevisan February 29, 2016

Partitioning Algorithms that Combine Spectral and Flow Methods

Lecture Semidefinite Programming and Graph Partitioning

Topic: Balanced Cut, Sparsest Cut, and Metric Embeddings Date: 3/21/2007

ORIE 6334 Spectral Graph Theory December 1, Lecture 27 Remix

6.854J / J Advanced Algorithms Fall 2008

Lecture 12 : Graph Laplacians and Cheeger s Inequality

approximation algorithms I

On a Cut-Matching Game for the Sparsest Cut Problem

Unique Games and Small Set Expansion

Convex relaxation. In example below, we have N = 6, and the cut we are considering

Lecture 20: Goemans-Williamson MAXCUT Approximation Algorithm. 2 Goemans-Williamson Approximation Algorithm for MAXCUT

U.C. Berkeley CS294: Beyond Worst-Case Analysis Handout 12 Luca Trevisan October 3, 2017

Lecture 13: Spectral Graph Theory

Lecture 17: Cheeger s Inequality and the Sparsest Cut Problem

Hierarchies. 1. Lovasz-Schrijver (LS), LS+ 2. Sherali Adams 3. Lasserre 4. Mixed Hierarchy (recently used) Idea: P = conv(subset S of 0,1 n )

EIGENVECTOR NORMS MATTER IN SPECTRAL GRAPH THEORY

Lecture 5. 1 Goermans-Williamson Algorithm for the maxcut problem

8 Approximation Algorithms and Max-Cut

Lecture 10. Semidefinite Programs and the Max-Cut Problem Max Cut

Convex relaxation. In example below, we have N = 6, and the cut we are considering

Lift-and-Project Techniques and SDP Hierarchies

Lecture 12: Introduction to Spectral Graph Theory, Cheeger s inequality

Lecture 8: The Goemans-Williamson MAXCUT algorithm

Lecture: Examples of LP, SOCP and SDP

MIT Algebraic techniques and semidefinite optimization February 14, Lecture 3

Dependence. Practitioner Course: Portfolio Optimization. John Dodson. September 10, Dependence. John Dodson. Outline.

Graph Partitioning Algorithms and Laplacian Eigenvalues

Lecture 2: Streaming Algorithms

KRIVINE SCHEMES ARE OPTIMAL

Lecture 3: Semidefinite Programming

CSCI 1951-G Optimization Methods in Finance Part 10: Conic Optimization

Lower bounds on the size of semidefinite relaxations. David Steurer Cornell

Integer Linear Programs

SEMIDEFINITE PROGRAM BASICS. Contents

U.C. Berkeley Better-than-Worst-Case Analysis Handout 3 Luca Trevisan May 24, 2018

Canonical SDP Relaxation for CSPs

Expander Flows, Geometric Embeddings and Graph Partitioning

Global Optimization of Polynomials

An Improved Approximation Algorithm for Requirement Cut

IE 521 Convex Optimization

Probability Background

Lecture 17 (Nov 3, 2011 ): Approximation via rounding SDP: Max-Cut

Lecture 1: April 24, 2013

STAT 501 Assignment 1 Name Spring Written Assignment: Due Monday, January 22, in class. Please write your answers on this assignment

Agenda. Applications of semidefinite programming. 1 Control and system theory. 2 Combinatorial and nonconvex optimization

Lecture #21. c T x Ax b. maximize subject to

Lec. 2: Approximation Algorithms for NP-hard Problems (Part II)

Near-Optimal Algorithms for Maximum Constraint Satisfaction Problems

16 Embeddings of the Euclidean metric

U.C. Berkeley CS270: Algorithms Lecture 21 Professor Vazirani and Professor Rao Last revised. Lecture 21

Lecture 8 : Eigenvalues and Eigenvectors

CSC Linear Programming and Combinatorial Optimization Lecture 10: Semidefinite Programming

Positive Semi-definite programing and applications for approximation

Lecture 7. 1 Normalized Adjacency and Laplacian Matrices

MATHEMATICS 154, SPRING 2009 PROBABILITY THEORY Outline #11 (Tail-Sum Theorem, Conditional distribution and expectation)

Unique Games Conjecture & Polynomial Optimization. David Steurer

A better approximation ratio for the Vertex Cover problem

Semidefinite Programming

Lecture 5: Two-point Sampling

If g is also continuous and strictly increasing on J, we may apply the strictly increasing inverse function g 1 to this inequality to get

Convex and Semidefinite Programming for Approximation

Lecture 13 March 7, 2017

Duality of LPs and Applications

Lecture 15: Expanders

Lecture 22: Hyperplane Rounding for Max-Cut SDP

arxiv:cs/ v1 [cs.cc] 7 Sep 2006

2. Matrix Algebra and Random Vectors

Lecture 6 September 13, 2016

Lecture 9: Low Rank Approximation

Linear and non-linear programming

Lecture 1 Introduction

Lectures 2 3 : Wigner s semicircle law

Lecture Notes 1: Vector spaces

Lecture: Cone programming. Approximating the Lorentz cone.

Vector spaces. DS-GA 1013 / MATH-GA 2824 Optimization-based Data Analysis.

Robust Principal Component Analysis

Random Feature Maps for Dot Product Kernels Supplementary Material

The Green-Tao theorem and a relative Szemerédi theorem

The Hilbert Space of Random Variables

Spectral Graph Theory Lecture 2. The Laplacian. Daniel A. Spielman September 4, x T M x. ψ i = arg min

Lecture 14: SVD, Power method, and Planted Graph problems (+ eigenvalues of random matrices) Lecturer: Sanjeev Arora

arxiv: v1 [cs.ds] 10 Aug 2009

EECS 598: Statistical Learning Theory, Winter 2014 Topic 11. Kernels

A Continuation Approach Using NCP Function for Solving Max-Cut Problem

Integrality gaps of semidefinite programs for Vertex Cover and relations to l 1 embeddability of negative type metrics

Lecture 14: Random Walks, Local Graph Clustering, Linear Programming

Lecture Note 1: Probability Theory and Statistics

ECE 650 Lecture 4. Intro to Estimation Theory Random Vectors. ECE 650 D. Van Alphen 1

linear programming and approximate constraint satisfaction

ORF 523 Lecture 9 Spring 2016, Princeton University Instructor: A.A. Ahmadi Scribe: G. Hall Thursday, March 10, 2016

1 Strict local optimality in unconstrained optimization

princeton univ. F 13 cos 521: Advanced Algorithm Design Lecture 17: Duality and MinMax Theorem Lecturer: Sanjeev Arora

Transcription:

CS369H: Hierarchies of Integer Programming Relaxations Spring 2016-2017 Lecture 5. Max-cut, Expansion and Grothendieck s Inequality Professor Moses Charikar Scribes: Kiran Shiragur Overview Here we derive SoS certificates for Maxcut, Expansion and Grothendieck s Inequality. 1 Recap Lets restate some of the definitions and theorems needed for this class. Definition 5.1. (degree d pseudo-distribution) is a degree d pseudo-distribution iff Ẽ 1 = 1 For any polynomial f with degree d/2 Ẽ f 2 0 where Ẽ g := x {0,1} n (x)g(x) Lemma 5.1. (polynomial representation of pseudo-distribution) For any degree d pseudo-distribution, there exists a multi-linear polynomial of degree d such that: Ẽ p = Ẽ p polynomial p of degree d Lemma 5.2. (moments) is a degree d pseudo-distribution iff Ẽ 1 = 1 Ẽ ( (1, x) d/2 )( (1, x) d/2) T 0 Lemma 5.3. (duality between SoS proofs and pseudo-distribution) For any function f : {0, 1} n R, a degree d SoS certificate for f Ẽ f 0 for every degree d pseudo-distribution One of the interesting facts about pseudo-distributions is that they behave closely to an actual distribution in the following sense. Most of the common interesting inequalities which are true for an actual distribution also hold for pseudo-distributions. For instance inequalities analogous to Cauchy-Schwarz and Holder s are also true for pseudo-distributions. 1

Lemma 5.4. (Cauchy-Schwarz) For any degree d pseudo-distribution (Ẽ P Q ) 2 (Ẽ P 2)( Ẽ Q 2) where P and Q are polynomials of degree d/2 Lemma 5.5. For any degree 2 pseudo-distribution, there exist an actual distribution ρ over R n such that: ( Ẽ (1, x) 2 ) ( = E (1, x) 2 ) x ρ Proof. Let ν = Ẽ x and Σ = Ẽ (x ν)(x ν) T. For any vector u: u, Σu = Ẽ u, x ν 2 0 (Since is a degree 2 pseudo-distribution) Σ 0 Rest of the proof is pretty standard. Now pick a standard gaussian random vector g, and define a new random variable y to be: y = ν + Σ 1/2 g E y = ν E(y ν)(y ν) T = EΣ 1/2 gg T Σ 1/2 = Σ 1/2 (Egg T )Σ 1/2 = Σ ( Egg T = I) In the rest of the lecture we derive SoS certificates for Max-cut, Expansion and Grothendieck s Inequality. 2 Max-cut In this section we give SoS interpretation for Goemans-Williamson s 0.878 approximation algorithm for Max-cut. Lets first define the polynomial corresponding to max-cut: f G (x) = (x i x j ) 2 (i,j) E(G) Question 5.1. How do we show an upper bound of c for f G (x)? max f G(x) c x {0,1} n One way to show an upper bound is by showing the existence of SoS certificate for c f G (x). In this lecture we show existence of a degree 2 SoS certificate for c max f G, to be more precise we prove the following 0.878 theorem. Theorem 5.1. For every graph G(V, E), linear functions g 1, g 2... g n : {0, 1} n R such that: max(f G ) 0.878f G (x) = g i (x) 2 2

To prove this result it suffices to prove the following equivalent theorem. Theorem 5.2. For every graph G(V, E) and any degree 2 pseudo-distribution, there exists a probability distribution over {0, 1} n such that: E f G 0.878Ẽf G Proof. Without loss of generality Ẽx = 1 2 1 (else consider the pseudo-distribution 1 ((x)+(1 x))). As 2 earlier, construct a gaussian vector ξ with mean Ẽx and co-variance Ẽxx T. Lets calculate the variances first: V ar(x i ) = Ẽ(x i 1/2) 2 = Ẽx 2 i 1/4 = 1/4 (Since is a pseudo-distribution over hypercube Ẽx 2 i = Ẽx i ) Cov(x i, x j ) = Ẽx i x j 1/4 Define ρ := 4Cov(x i, x j ) = 4Ẽx i x j 1. With these definitions in mind, lets define x as follows: x i = 0 if ξ i < 1/2 x i = 1 otherwise Lets look at term wise expectation E(x i x j) 2, and it is related to random variables 2ξ i 1 and 2ξ j 1 (converts variables into ±1) in the following way. { } E(x i x j) 2 = P sign(2ξ i 1) sign(2ξ j 1) = 2 arccos ρ (1 ρ)π 0.878 One can show better results for instances where the Max-cut is close to all the edges. Lemma 5.6. For any degree 2 pseduo-distribution over {0, 1} 2 such that: Ẽ (1 ɛ) E(G) there exists an actual distribution over {0, 1} n with E (1 2 ɛ) E(G) 3 Expansion For any d-regular graph G(V, E), expansion of a set S V is defined as: Expansion φ G of a graph G(V, E) is: E(S, V \S) φ G (S) = d n S V \S φ G = min S V φ G(S) 3

The goal here is to find a set S V which minimizes φ G (S). If x i {0, 1} denotes which side of the cut vertex i belongs and f G (x) defined as earlier then: φ G = min x {0,1} n f G (x) d n x (n x ) Note the function φ G is of the form P (x) Q(x), where P (x) = f G(x) and Q(x) = d x (n x ) are polynomials. n Question 5.2. How do we certify P (x) Q(x) c? The above question is equivalent to certifying. P (x) cq(x) 0 Lemma 5.7. P (x) 1 2 φ2 GQ(x) 0 has a degree 2 SoS certificate. The above lemma certifies the expansion of graph G is 1 2 φ2 G. Graph Expansion is a well studied problem and exhibits a rich literature. [LR88] LP based algorithm which finds a set S V such that: φ G (S) = O(log n)φ G [ARV09] SDP based algorithm (adds triangle inequalities to the standard SDP formulation) and achieves O( log n) approximation guarantee. It finds a set S V such that: φ G (S) = O( log n)φ G This new SDP (with additional constraints) is captured by the degree 4 SoS. One of the other famous results on expansion is the Cheeger s inequality. Lemma 5.8. (Cheeger s Inequality) For any d-regular graph G(V, E), there exist a set S V and S n/2 such that: φ G (S) 2λ where λ is the second smallest eigenvalue of the normalized laplacian L G = I 1 d A G, and A G is the adjacency matrix of graph G. Exercise 5.1. Prove that Cheeger s inequality degree 2 SoS certificate for f G (x) 1 2 φ2 (G) d n x (n x ). Hint: Note the quadratic form of Normalized laplacian can be represented in terms of f G (x). x, L G x = 1 d f G(x) 4

4 Grothendieck s Inequality Given a matrix A R n m, define A 1 := max x R m Ax 1 x The above definition is equivalent to: A 1 = max Ax, y x {±1} m,y {±1} n = max x {±1} m Ax 1 The quantity A 1 is important, for instance it approximates the cut-norm of a matrix within a constant factor: [ A 1 ] cut norm(a) = max a ij, A 1 S [n],t [m] 4 i S,j T Theorem 5.3. (Grothendieck s inequality) There exists a constant K G such that for every matrix A R n m and a degree 2 pseudo-distribution on {±1} m {±1} n. Ẽ (x,y) Ax, y KG A 1 In 1977, Krivine showed: K G π 2 log(1 + 2) 1.782 This result was later improved by [BMMN11]. Below we prove the Grothendieck s inequality due to Krivine. Proof. Consider Ẽ(x,y)xy T, and suppose we could produce joint gaussian vectors ξ, ς such that: Then the quantity Ẽ Ax, y is: Ẽ (x,y) xy T = K Krivine E[sign(ξ)sign(ς) T ] Ẽ Ax, y = T r( A, Ẽ xy T ) = K Krivine T r( A, E[sign(ξ)sign(ς) T ] ) = K Krivine E[ Asign(ξ), sign(ς) ] K Krivine A 1 Next we show existence of joint gaussian vectors ξ, ς with: Ẽ (x,y) xy T = K Krivine E[sign(ξ)sign(ς) T ] For gaussian vectors: E[sign(ξ)sign(ς) T ] = 2 π arcsin(e[ξςt ]) 1 The goal here is to choose gaussian vectors ξ, ς such that: sin(cẽ(x,y)[xy T ]) = E[ξς T ] 1 Given a matrix A, the operations arcsin(a) and sin(a) apply functions arcsin and sin respectively entry wise 5

and this would give us K Krivine = π. Note the following matrix is PSD: 2c [ sinh(c Ẽxx T ) ] [ sin(cẽxyt ) ] Cov = [ sin(c Ẽyx T ) ] [ sinh(cẽyyt ) ] and we can choose gaussian vectors ξ, ς with Cov as the co-variance matrix, but we need the variances of entries of ξ, and ς to be 1. Note the matrices ẼxxT and ẼyyT have all the diagonal entries equal to 1 and the value of c should satisfy the equation sinh(c) = 1, which evaluates to c = log(1 + 2) which implies π K Krivine = 2 log(1 + 2). References [ARV09] Sanjeev Arora, Satish Rao, and Umesh Vazirani. Expander flows, geometric embeddings and graph partitioning. J. ACM, 56(2):5:1 5:37, April 2009. [BMMN11] M. Braverman, K. Makarychev, Y. Makarychev, and A. Naor. The grothendieck constant is strictly smaller than krivine s bound. In 2011 IEEE 52nd Annual Symposium on Foundations of Computer Science, pages 453 462, Oct 2011. [LR88] T. Leighton and S. Rao. An approximate max-flow min-cut theorem for uniform multicommodity flow problems with applications to approximation algorithms. In [Proceedings 1988] 29th Annual Symposium on Foundations of Computer Science, pages 422 431, Oct 1988. 6