Partitioning Algorithms that Combine Spectral and Flow Methods

Similar documents
Topic: Balanced Cut, Sparsest Cut, and Metric Embeddings Date: 3/21/2007

Lecture: Flow-based Methods for Partitioning Graphs (1 of 2)

U.C. Berkeley CS294: Spectral Methods and Expanders Handout 11 Luca Trevisan February 29, 2016

ORIE 6334 Spectral Graph Theory November 22, Lecture 25

On a Cut-Matching Game for the Sparsest Cut Problem

Hierarchies. 1. Lovasz-Schrijver (LS), LS+ 2. Sherali Adams 3. Lasserre 4. Mixed Hierarchy (recently used) Idea: P = conv(subset S of 0,1 n )

Lecture 5. Max-cut, Expansion and Grothendieck s Inequality

CSC Linear Programming and Combinatorial Optimization Lecture 10: Semidefinite Programming

Finite Metric Spaces & Their Embeddings: Introduction and Basic Tools

Integer Linear Programs

Lecture 17: Cheeger s Inequality and the Sparsest Cut Problem

ORIE 6334 Spectral Graph Theory December 1, Lecture 27 Remix

16.1 Min-Cut as an LP

Lecture 15 (Oct 6): LP Duality

Partitioning Metric Spaces

Lecture: Some Practical Considerations (3 of 4)

Lecture: Expanders, in theory and in practice (2 of 2)

Lecture 13: Spectral Graph Theory

SDP Relaxations for MAXCUT

- Well-characterized problems, min-max relations, approximate certificates. - LP problems in the standard form, primal and dual linear programs

U.C. Berkeley CS294: Beyond Worst-Case Analysis Handout 8 Luca Trevisan September 19, 2017

A better approximation ratio for the Vertex Cover problem

1 Matrix notation and preliminaries from spectral graph theory

Norms of Random Matrices & Low-Rank via Sampling

Lecture 17 (Nov 3, 2011 ): Approximation via rounding SDP: Max-Cut

Convex and Semidefinite Programming for Approximation

Spherical Euclidean Distance Embedding of a Graph

1 Matrix notation and preliminaries from spectral graph theory

Fréchet embeddings of negative type metrics

Semidefinite Programming Basics and Applications

Expander Flows, Geometric Embeddings and Graph Partitioning

Semidefinite and Second Order Cone Programming Seminar Fall 2001 Lecture 5

New algorithms for Disjoint Paths and Routing Problems

Z Algorithmic Superpower Randomization October 15th, Lecture 12

Lecture 12 : Graph Laplacians and Cheeger s Inequality

Hierarchical Clustering via Spreading Metrics

Lecture: Local Spectral Methods (3 of 4) 20 An optimization perspective on local spectral methods

eigenvalue bounds and metric uniformization Punya Biswal & James R. Lee University of Washington Satish Rao U. C. Berkeley

arxiv: v1 [cs.lg] 5 Jun 2018

Lecture 3. 1 Polynomial-time algorithms for the maximum flow problem

Graph Partitioning Algorithms and Laplacian Eigenvalues

arxiv:cs/ v1 [cs.cc] 7 Sep 2006

Lecture 5 : Projections

Lecture 10. Semidefinite Programs and the Max-Cut Problem Max Cut

Basic Properties of Metric and Normed Spaces

Graph Partitioning Using Random Walks

GRAPH PARTITIONING USING SINGLE COMMODITY FLOWS [KRV 06] 1. Preliminaries

Aditya Bhaskara CS 5968/6968, Lecture 1: Introduction and Review 12 January 2016

Lecture 7. 1 Normalized Adjacency and Laplacian Matrices

The Unique Games Conjecture, Integrality Gap for Cut Problems and Embeddability of Negative Type Metrics into l 1

Lecture 14: SVD, Power method, and Planted Graph problems (+ eigenvalues of random matrices) Lecturer: Sanjeev Arora

Positive Semi-definite programing and applications for approximation

CSC Linear Programming and Combinatorial Optimization Lecture 12: The Lift and Project Method

CSCI 1951-G Optimization Methods in Finance Part 10: Conic Optimization

Topics in Theoretical Computer Science April 08, Lecture 8

Vertical Versus Horizontal Poincare Inequalities. Assaf Naor Courant Institute

1 Strict local optimality in unconstrained optimization

Introduction to Semidefinite Programming I: Basic properties a

Hardness of Embedding Metric Spaces of Equal Size

Lecture 6 January 21, 2013

SVD, Power method, and Planted Graph problems (+ eigenvalues of random matrices)

Approximation Algorithms and Hardness of Integral Concurrent Flow

A Unified Theorem on SDP Rank Reduction. yyye

Sum-of-Squares Method, Tensor Decomposition, Dictionary Learning

Expander Flows, Geometric Embeddings and Graph Partitioning

Lecture 14: Random Walks, Local Graph Clustering, Linear Programming

Convex relaxation. In example below, we have N = 6, and the cut we are considering

Chapter 11. Approximation Algorithms. Slides by Kevin Wayne Pearson-Addison Wesley. All rights reserved.

Duality. Geoff Gordon & Ryan Tibshirani Optimization /

16.1 L.P. Duality Applied to the Minimax Theorem

Algorithms for 2-Route Cut Problems

15-850: Advanced Algorithms CMU, Fall 2018 HW #4 (out October 17, 2018) Due: October 28, 2018

Project Discussions: SNL/ADMM, MDP/Randomization, Quadratic Regularization, and Online Linear Programming

Approximation Algorithms

Lecture: Modeling graphs with electrical networks

Approximating the isoperimetric number of strongly regular graphs

Topics in Approximation Algorithms Solution for Homework 3

CSC Metric Embeddings Lecture 7: Lower bounds on the embeddability in via expander graphs and some algorithmic connections to

Locally-biased analytics

1 Directional Derivatives and Differentiability

Small distortion and volume preserving embedding for Planar and Euclidian metrics Satish Rao

Introduction to LP and SDP Hierarchies

Functional Analysis Exercise Class

Easy and not so easy multifacility location problems... (In 20 minutes.)

Lec. 2: Approximation Algorithms for NP-hard Problems (Part II)

Integrality gaps of semidefinite programs for Vertex Cover and relations to l 1 embeddability of negative type metrics

CS168: The Modern Algorithmic Toolbox Lectures #11 and #12: Spectral Graph Theory

Lecture 9: Dantzig-Wolfe Decomposition

Inderjit Dhillon The University of Texas at Austin

Linear and Integer Programming - ideas

Convex relaxation. In example below, we have N = 6, and the cut we are considering

Global Optimization of Polynomials

Lecture: Cone programming. Approximating the Lorentz cone.

Non-Uniform Graph Partitioning

U.C. Berkeley Better-than-Worst-Case Analysis Handout 3 Luca Trevisan May 24, 2018

Topological properties

1 The independent set problem

Lecture 3: Huge-scale optimization problems

COMS 4721: Machine Learning for Data Science Lecture 6, 2/2/2017

Constrained Optimization

Transcription:

CS369M: Algorithms for Modern Massive Data Set Analysis Lecture 15-11/11/2009 Partitioning Algorithms that Combine Spectral and Flow Methods Lecturer: Michael Mahoney Scribes: Kshipra Bhawalkar and Deyan Simeonov *Unedited notes Last time we looked at flow based graph partitioning. Today we will show why it works. We are given a graph G = (V, E), a cost function c : E R + and k pairs of nodes (s i, t i ). Let x(e) indicate if an edge e is cut, y(i) indicate if commodity i is cut and P i, i = 1, 2,..., k be the set of paths between s i and t i. Then we want: e E min c(e)x(e) k, s.t. d(i)y() x(e) y(i) P P i i = 1, 2,..., k e P y(i) {0, 1}, x(e) {0, 1} We relax x and y to be in [0, 1] and, given that for any feasible solution (x, y) and any α > 0, (αx, αy) is a feasible solution, we get the following LP: min e E c(e)x(e), s.t. Our strategy: solve the LP round the solution. k d(i)y(i) = 1 x(e) y(i) P P i i = 1, 2,..., k e P x(e) 0, y(i) 0 Recall: x R n, x p = ( n xp i ) 1 p - p-norm. Its induced metric is x y p. Theorem (Bourgain 85). Any n-point metric space admits an α-distortion embedding in l p with α = O(log n). Proof idea: Use as coordinates the minimal distances of points to suitably chosen random sets. Comparison between l 1 and l 2 : l 2 has good dimension reduction properties, l 1 doesn t. There is a fast (polytime, O(n 3 )) algorithm to get the best l 2 embedding, for l 1 it is NP. 1

l 2 has strong connection to diffusion, low dimensional spaces and manifolds, l 1 has strong connection to cuts, partition in graphs, multicommodity flows. Connection between l 1 and Cut metrics: There exists a representation of l 1 metrics in terms of combination of cut metrics. Cut metrics - Extreme rays of the cone of l 1 metrics. minimum ratio function over cone minimum over extreme rays. Definition. Given a graph G(V, E) and S V, say δ S is the cut-metric for S, if δ s (x, y) is the indicator of x and y being on different sides of S. Claim. The set of l 1 metrics is a convex cone, i.e.: If d 1, d 2 l 1, α 1, α 2 0, then α 1 d 1 + α 2 d 2 l 1. Does not hold for l 2! Proof: Line metric is an l 1 metric but d (i) (x, y) = x i y i for x i, y i R n. If d is an l! metric then it is a line metric. We can then argue for each dimension. Claim. Let d be a finite l 1 metric then we can write d as d = S V α S δ S α S R, δ S cut metric i.e. CUT n = {d : d = α S δ S, α S 0} S V = {positive cone generated by this metric} = {all n-points subsets of R n under l 1 metric Proof: (CUT n l 1 ) d = S V α Sδ S CUT n. Introduce one dimension for each pair of vertices. For a pair of points (i, j) value in the dimension (i, j) is sum of α S for all cuts (S, S) over all cuts that put i and j in different sets. (l 1 CUT n ): Consider a dimension d and sort points along that dimension in increasing values. Let v 1, v 2,...v k be the set of distinct values along that dimension. Define k-1 cut metrics S i = {x : x d v i 1 } and let α i = v i+1 v i. So along that dimention, Do this for each dimension. x d y d = Usefulness - you can optimize over l 1 metrics: Let C R n - convex cone. f, g : R n R +. Then: k α i δ Si f(x) min x C g(x) = min f(x) x extreme ray ofc g(x) 2

Conductance and Sparsity Given a graph G(V,E) we define the conductance h G and sparsity φ G as follows, h G := min S V E(S, S) min{ S, S } E(S, φ G := min S) S V S S Claim. φ G = {min d l1metric i,j E d ijs.t. i,j d ij = 1} Idea: Relax to optimization over a larger set i.e. a metric. 1 n λ := min i,j E d ij s.t. d ij = 1 i,j d ij 0 d ij = d ji d ij + d jk d ik Clearly λ φ. Less obviously φ O(log n)λ (homework 3). Algorithm: Given G. Solve the LP to get metric d, Use bourgain embedding result to approximate by l 1 metric (with loss O(log n). Round the solution to get a cut. For each dimension covert the l 1 metric along that to a cut metric. Choose the best. Note: If have l 1 embedding with distortion factor ξ then can approximate the cut upto ξ. (Homework 3). What else can you relax to? l 2 - not convex l2 2 - convex but distortion Omega(n). (Not a metric) (this happens in spectral technique. Can get eigenvalue λ s.t. λ is close to h(g) upto quadratic factor. (Cheegar Inequality). l2 2 + triangle inequality gives a metric This works. Arora, Rao Vazirani. 3

Various relaxations of spectral cut. Actual problem: Φ G = min S V E(S, S) S S = min x 0,1 V Spectral Method: replace l 1 with l 2. i,j d λ 2 = min A ij(x i x j ) 2 x R V i,j (x i x j ) 2 Leighton, Rao i,j A ij x i x j i,j x i x j i,j min A ijd ij d metric i,j d ij Leighton,Rao approach is bad in expanders, but is good in sparse graphs. Definition. l 2 2 representation of a graph G is an assignment of a point vector to each node v i R k for each i. s.t. v i v j 2 2 + v j v k 2 2 v i v k 2 2 It is a unit l 2 2 representation if on unit sphere i.e. v i = 1, i. Thing to note: Condition that l 2 2 distance form a metric all triangles are acute. l 2 2 metric is a convex cone. d l 1 = d l 2 2. Relax to vectors on unit sphere that form a metric, min A ij x i x j i,j E s.t. i,j x i x j 2 2 = 1 x i x j 2 2 + x j x k 2 2 x i x k 2 2 This it the semi-definite program(sdp) from Arora, Rao, Vazirani. Theorem: For uniform sparsest cut d ij = 1 i, j SDP integrality gap is θ( log n). For general sparsest cut, SDP integrality gap is θ( log n log log n). Main structure theorem: Let v, v 2,...v n be points on the unit vall in R n. s.t. d ij = v i v j 2 2 is a metric and all points are wellseperated. i,j d ij/n 2 δ = Ω(1). then S, T disjoint subsets of V s.t. S, T Ω(n) min i S,j T d ij Ω(1/ log n) Flow methods: Embed scaled version of complete graphs into G. - O(log n) ARV: embed an arbitrary graph H(iterative construction) such that H is a good expander or we find a cut. Iterative construction: multiplicative update method online learning. 4

References 1. Arora, Rao, and Vazirani, CACM article, Geometry, flows, and graph-partitioning algorithms 2. Khandekar, Rao, and Vazirani, Graph partitioning using single commodity flows 3. Orecchia, Schulman, Vazirani, and Vishnoi, On Partitioning Graphs via Single Commodity Flows 5