ORIE 6334 Spectral Graph Theory November 22, Lecture 25

Similar documents
Lecture 7. 1 Normalized Adjacency and Laplacian Matrices

ORIE 6334 Spectral Graph Theory December 1, Lecture 27 Remix

Topic: Balanced Cut, Sparsest Cut, and Metric Embeddings Date: 3/21/2007

Lecture 13: Spectral Graph Theory

U.C. Berkeley CS294: Spectral Methods and Expanders Handout 11 Luca Trevisan February 29, 2016

Lecture 5. Max-cut, Expansion and Grothendieck s Inequality

Partitioning Algorithms that Combine Spectral and Flow Methods

Graph Partitioning Algorithms and Laplacian Eigenvalues

ORIE 6334 Spectral Graph Theory October 13, Lecture 15

Lecture 3. 1 Polynomial-time algorithms for the maximum flow problem

Lecture 17: Cheeger s Inequality and the Sparsest Cut Problem

ORIE 6334 Spectral Graph Theory October 25, Lecture 18

Lecture 12 : Graph Laplacians and Cheeger s Inequality

Lecture Approximate Potentials from Approximate Flow

Partitioning Metric Spaces

ORIE 6334 Spectral Graph Theory September 22, Lecture 11

SDP Relaxations for MAXCUT

16.1 Min-Cut as an LP

- Well-characterized problems, min-max relations, approximate certificates. - LP problems in the standard form, primal and dual linear programs

Lecture 12: Introduction to Spectral Graph Theory, Cheeger s inequality

Lecture: Flow-based Methods for Partitioning Graphs (1 of 2)

Lecture 14: Random Walks, Local Graph Clustering, Linear Programming

Hardness of Embedding Metric Spaces of Equal Size

1 Matrix notation and preliminaries from spectral graph theory

CS168: The Modern Algorithmic Toolbox Lectures #11 and #12: Spectral Graph Theory

ORIE 633 Network Flows October 4, Lecture 10

On a Cut-Matching Game for the Sparsest Cut Problem

A better approximation ratio for the Vertex Cover problem

Lecture 8: Spectral Graph Theory III

Lecture Semidefinite Programming and Graph Partitioning

An Improved Approximation Algorithm for Requirement Cut

Non-Uniform Graph Partitioning

Lecture Introduction. 2 Brief Recap of Lecture 10. CS-621 Theory Gems October 24, 2012

Lecture: Expanders, in theory and in practice (2 of 2)

Lecture 15 (Oct 6): LP Duality

Sum-of-Squares Method, Tensor Decomposition, Dictionary Learning

Approximating the isoperimetric number of strongly regular graphs

Solutions to Exercises

Introduction to Semidefinite Programming I: Basic properties a

Tight Size-Degree Lower Bounds for Sums-of-Squares Proofs

Semidefinite and Second Order Cone Programming Seminar Fall 2001 Lecture 5

Machine Learning for Data Science (CS4786) Lecture 11

ORIE 6334 Spectral Graph Theory September 8, Lecture 6. In order to do the first proof, we need to use the following fact.

Lecture 9: Low Rank Approximation

eigenvalue bounds and metric uniformization Punya Biswal & James R. Lee University of Washington Satish Rao U. C. Berkeley

Lecture 17 (Nov 3, 2011 ): Approximation via rounding SDP: Max-Cut

Prize-collecting Survivable Network Design in Node-weighted Graphs. Chandra Chekuri Alina Ene Ali Vakilian

Improved Bounds for Flow Shop Scheduling

Lecture 6 January 21, 2013

arxiv:cs/ v1 [cs.cc] 7 Sep 2006

Lecture 13 Spectral Graph Algorithms

Lecture: Modeling graphs with electrical networks

Lecture #21. c T x Ax b. maximize subject to

Improving Christofides Algorithm for the s-t Path TSP

Topics in Approximation Algorithms Solution for Homework 3

6.854J / J Advanced Algorithms Fall 2008

Hierarchical Clustering via Spreading Metrics

CMPUT 675: Approximation Algorithms Fall 2014

Algorithms Reading Group Notes: Provable Bounds for Learning Deep Representations

Algorithms for 2-Route Cut Problems

Lecture 14: SVD, Power method, and Planted Graph problems (+ eigenvalues of random matrices) Lecturer: Sanjeev Arora

Lecture 20: Goemans-Williamson MAXCUT Approximation Algorithm. 2 Goemans-Williamson Approximation Algorithm for MAXCUT

Graph Clustering Algorithms

ORIE 6334 Spectral Graph Theory November 8, Lecture 22

Introduction to Algorithms

Finite Metric Spaces & Their Embeddings: Introduction and Basic Tools

Notes on MapReduce Algorithms

Topic: Primal-Dual Algorithms Date: We finished our discussion of randomized rounding and began talking about LP Duality.

Lecture 15: Interactive Proofs

An Approximation Algorithm for MAX-2-SAT with Cardinality Constraint

Lecture 20: Course summary and open problems. 1 Tools, theorems and related subjects

c 2011 Society for Industrial and Applied Mathematics

25.1 Markov Chain Monte Carlo (MCMC)

6. Linear Programming

Lecture 28: Generalized Minimum Distance Decoding

Hierarchies. 1. Lovasz-Schrijver (LS), LS+ 2. Sherali Adams 3. Lasserre 4. Mixed Hierarchy (recently used) Idea: P = conv(subset S of 0,1 n )

1 Review: symmetric matrices, their eigenvalues and eigenvectors

Doubling metric spaces and embeddings. Assaf Naor

Vertical Versus Horizontal Poincare Inequalities. Assaf Naor Courant Institute

Complexity Theory of Polynomial-Time Problems

Petar Maymounkov Problem Set 1 MIT with Prof. Jonathan Kelner, Fall 07

Lecture 11 October 7, 2013

Cuts & Metrics: Multiway Cut

princeton univ. F 17 cos 521: Advanced Algorithm Design Lecture 6: Provable Approximation via Linear Programming

Lecture: Local Spectral Methods (3 of 4) 20 An optimization perspective on local spectral methods

Lecture 1: April 24, 2013

Reductions Between Expansion Problems

Approximation Algorithms and Hardness of Integral Concurrent Flow

Fréchet embeddings of negative type metrics

Approximating Sparsest Cut in Graphs of Bounded Treewidth

GRAPH PARTITIONING USING SINGLE COMMODITY FLOWS [KRV 06] 1. Preliminaries

Duality of LPs and Applications

1 Matrix notation and preliminaries from spectral graph theory

8 Approximation Algorithms and Max-Cut

U.C. Berkeley CS294: Beyond Worst-Case Analysis Handout 8 Luca Trevisan September 19, 2017

Diffusion and random walks on graphs

Lecture: Local Spectral Methods (2 of 4) 19 Computing spectral ranking with the push procedure

CSC Metric Embeddings Lecture 8: Sparsest Cut and Embedding to

Discrete (and Continuous) Optimization WI4 131

An Algorithmist s Toolkit September 24, Lecture 5

Transcription:

ORIE 64 Spectral Graph Theory November 22, 206 Lecture 25 Lecturer: David P. Williamson Scribe: Pu Yang In the remaining three lectures, we will cover a prominent result by Arora, Rao, and Vazirani for the sparsest cut problem. In this lecture, we will set the scene by giving a prior result by Leighton and Rao, and explain what the Arora-Rao-Vazirani algorithm has to do with the topic of this course. Recall that the sparsity of a cut S V in a graph G = (V, E) is defined as ρ(s) S V S, where δ(s) is the set of edges with exactly one end in S. The sparsest cut of a graph is the defined as ρ(g) min S G ρ(s). Recall also that the sparsest cut is related to two other concepts. The edge expansion of a set S V, S n/2 is defined as and the edge expansion of a graph α(g) α(s), S The conductance of a set S V is defined as φ(s) min α(s). S V, S n/2 min(vol(s), Vol(V S)), where Vol(S) i S deg(i), and φ(g) min S G φ(s). We ll cover the following two results in the next three lectures. 0 Part of this lecture is taken from Leighton and Rao 999, and another from an unpublished note of Sudan and W; an analogous result to that of Sudan and W also appears in the Arora, Rao, and Vazirani paper. 25-

Theorem (Leighton and Rao 988) There is an O(log n)-approximation algorithm for the sparsest cut problem. Theorem 2 (Arora, Rao, and Vazirani 2004) There is an O( log n)-approximation algorithm for the sparsest cut problem. Let us first look at a relaxation of the sparsest cut problem: ρ LR min x(e) s.t. d x (i, j) x(e) 0 e E, where d x (i, j) is the shortest path distance from i to j using x as edge lengths. We claim that this relaxation can be solved in polynomial time. To see that it is a relaxation, let S be the sparsest cut. Set { if e S S x(e) = V S 0 o.w. Then d x (i, j) S V S S V S =. This is true because for any path that connects a node in S and a node in V S, it must use at least one edge in δ(s ), and the number of pairs of nodes with one in S and the other in V S is S V S. The objective function for x set in this way is δ(s ) x(e) = S V S = ρ(s ) = ρ(g). So x set in this way is a feasible solution to this problem, and the optimal solution for this problem gives a lower bound on ρ(g). In Lecture 4 on low-stretch trees, we proved the following lemma: Lemma Let D be some parameter. There is a partition of G into clusters s.t. each cluster has diameter (w.r.t. number of edges) less than or equal to D. there are at most α E intercluster edges with α 4 ln n D. We claim that by a similar proof, one can get the following extension of the lemma. Lemma 4 Let D be some parameter, and x(e) 0 be lengths of edges in E. There is a partition of G into clusters s.t. 25-2

each cluster has diameter (w.r.t. lengths x) less than or equal to D. there are at most α 4 ln n x(e) intercluster edges with α. D With this lemma and a little extra work we will arrive at the algorithm of Leighton and Rao. Let s see why. First, by choosing appropriate D in Lemma 2, we have the following result. Lemma 5 Let D = /(4n 2 ). Divide the graph G into clusters that satisfies Lemma 4. Then either there is a cluster C where C 2n, or there exists S V such that ρ(s) = O(log n) x(e) O(log n)ρ(g), where x(e) is the solution for the relaxed problem. Proof: If there is no cluster with size at least 2 n, then order the clusters by nonincreasing size, and add clusters in this order to S until S n. Then it must be the case that V S n as well, so that ρ(s) = δ(s) S V S 4n2 4 ln n x(e) = O(log n) x(e). 9 n2 The inequality follows is at most the number of intercluster edges, which is at most (4 ln n/d) x(e) by Lemma. Therefore, when there is no large cluster with size at least 2 n, we have found such a cut. What if there exists a cluster C such that C 2n? Lemma 6 If there exists a cluster C such that C 2 n, with diameter no more than, we can find S such that ρ(s) 6 4n 2 x(e) 6ρ(G). Proof: Let d(i, C) = min j C d(i, j). Order the vertices as i,..., i n in nonincreasing order of d(i, C). Let S k = {i,..., i k } for k =,..., n. We claim that d(i, C) d(j, C). We will prove this claim later. The 6 intuition behind this claim is that given we have many nodes in C which are close to each other because the size of C is large and the diameter of C is small, in order for the constraint in the relaxation problem to hold the total of the difference between the distance to C of each pair of nodes must be large. With this claim, we have that min ρ(s δ(s k ) k) = min k n k n S k V S k n k= d(i k+, C) d(i k, C) δ(s k ) n k= d(i k+, C) d(i k, C) S k V S k (i,j) E d(j, C) d(i, C) = d(j, C) d(i, C) x(e) = 6 x(e) 6ρ(G). /6 25-

The first inequality is easy to check by noticing that the weight d(i k+, C) d(i k, C) for δ(s k ) and S k V S k depends only on k. The last inequality is from the claim. Why is the second equality true? We notice that a given edge (i, j) appears in δ(s k ) for the indices k such that exactly one of i and j is in S k. Thus if we rewrite n d(i k+, C) d(i k, C) δ(s k ) k= as a sum over the edges, we get that d(i, C) d(j, C). (i,j) E Similarly, a given pair of vertices i and j appears in the product S k V S k for exactly the indices k such that exactly one of i and j is in S k, so we can rewrite as n d(i k+, C) d(i k, C) S k V S k k= d(j, C) d(i, C). It remains to prove the claim. Pick some i C. For any i V, there is some j C such that d(i, C) = d(i, j). So d(i, i ) d(i, j) + d(j, i ) d(i, C) +. Then 4n 2 d(i, j) i (d(i, ) + d(i, j)) = 2n d(i, i ) 2n ( d(i, C) + ) = 2n d(i, C) + 4n 2 2. i v Therefore Then d(i, C) d(j, C) d(i, C) = i/ C d(i, C) 4n. i/ C,j C d(i, C) = C i/ C d(i, C) 2n 4n = 6. Thus we have proved the existence of an O(log n)-approximate algorithm for sparsest cut. 25-4

To improve on this result we look at a related relaxation of the sparsest cut problem proposed independently by Goemans and Linial: ρ N min x(e) s.t. d x (i, j) d x is a negative type metric, where d is a negative type metric iff d(i, j)z(i)z(j) 0, for all z such that z T e = 0. The following is a useful theorem of determining whether a metric has negative type, and will be used next time to show that we can solve the relaxation in polynomial time. Theorem 7 d has negative type iff f : V R n s.t. d(i, j) = f(i) f(j) 2. To see why this is still a relaxation, again let { if e S S x(e) = V S 0 o.w. Then for any z s.t. z T e = 0, d x (i, j)z(i)z(j) = = = S V S i S,j / S z(i)z(j) ( ) z(i) z(j) S V S i S j / S ( ) ( z(i) ) z(j) S V S i S j S Thus d x for x so defined is of negative type. We now show that this relaxation is very related to previous topics of the course. To do this, we introduce the concept of a flow packing. Let P ij be the set of all i-j paths in G. Call F = (f ij ), symmetric a flow packing into G if for all paths P k P ij, there exist f k ij 0 such that Then the following can be shown. f ij = f ji = k f k ij i,j,k:(a,b) p k P ij f k ij (a, b) E. 0. We also need to show that for the optimal cut the graphs induced by S and V S are both connected; it is possible to do so. 25-5

Theorem 8 For flow packing F = (f ij ), let L F be weighted Laplacian with weights w(i, j) = f ij. Then ρ N = n max λ 2(L F ). Flow packings F The proof of this theorem uses duality, and the details are omitted here. A natural example of a flow packing is the adjacency matrix A. Then from the theorem, ρ N n λ 2(L G ). What is the difference between the two relaxations? It turns out there are cases for which the negative-type metric relaxation is much stronger than the relaxation without the negative-type restriction. We have the following lemma. Lemma 9 There exist graphs G such that ρ(g) = Ω(log n)ρ LR but for which ρ(g) = O()ρ N. We did not have time to prove this lemma in class, but here is the proof. Proof: There exist -regular expander graphs G = (V, E) such that α(g) β for some constant β > 0. For these, ρ(g) = min S V S V S min S V β max( S, V S ) = β n. If we set x(e) = 4/(n 2 (log n 2)) for all e E, then since G is -regular, there are at most + + 2 + 2 d nodes within d hops, or + d 2 k = + (2 d ). k= So within d = log 2 n, there are at most 2 log 2 n 2 = n 2 n within this 8 2 distance. Therefore, d x (i, j) = d x (i, j) 2 j i ( n 2 2 =, 4 n 2 (log n 2) ) (log n 2) because we need to traverse at least log n 2 edges to reach n/2 vertices from any given vertex i. Thus x is feasible for the linear program. Then since there are n/2 edges in a -regular graph, we have that ρ LR n 2 so that ρ(g) = Ω(log n)ρ LR. 4 n 2 (log n 2) = 6 n(log n 2), 25-6

However, recall that β α(g) φ(g) for a -regular graph, and that by Cheeger s inequality, φ(g) 2 2λ 2 (L) = λ 2(L G ), so it follows that λ 2 (L G ) = Ω(β 2 ). In a -regular graph min( S, V S ), so that ρ(g) min S V S V S min S V max( S, V S ) 6 n. Since the adjacency matrix is a flow packing, we have that ρ N λ 2 (L G )/n. Thus ρ(g) 6 n O() λ2(l G ) n O() ρ N. 25-7