GRAPH PARTITIONING USING SINGLE COMMODITY FLOWS [KRV 06] 1. Preliminaries

Similar documents
Topic: Balanced Cut, Sparsest Cut, and Metric Embeddings Date: 3/21/2007

On a Cut-Matching Game for the Sparsest Cut Problem

Approximation Algorithms and Hardness of Integral Concurrent Flow

Recall: Matchings. Examples. K n,m, K n, Petersen graph, Q k ; graphs without perfect matching

Lecture: Flow-based Methods for Partitioning Graphs (1 of 2)

1 Adjacency matrix and eigenvalues

Hardness of Directed Routing with Congestion

Lecture 17: Cheeger s Inequality and the Sparsest Cut Problem

25.1 Markov Chain Monte Carlo (MCMC)

CS675: Convex and Combinatorial Optimization Fall 2014 Combinatorial Problems as Linear Programs. Instructor: Shaddin Dughmi

Lecture 8: Path Technology

Generalized Pigeonhole Properties of Graphs and Oriented Graphs

Lecture 12: Introduction to Spectral Graph Theory, Cheeger s inequality

On Allocating Goods to Maximize Fairness

HAMILTON CYCLES IN RANDOM REGULAR DIGRAPHS

ACO Comprehensive Exam March 17 and 18, Computability, Complexity and Algorithms

Maximum Integer Flows in Directed Planar Graphs with Multiple Sources and Sinks and Vertex Capacities

A better approximation ratio for the Vertex Cover problem

On Allocating Goods to Maximize Fairness

Testing Graph Isomorphism

Robust Network Codes for Unicast Connections: A Case Study

Chapter 7 Network Flow Problems, I

U.C. Berkeley CS294: Spectral Methods and Expanders Handout 11 Luca Trevisan February 29, 2016

An Algorithmic Proof of the Lopsided Lovász Local Lemma (simplified and condensed into lecture notes)

Partitioning Algorithms that Combine Spectral and Flow Methods

CS675: Convex and Combinatorial Optimization Fall 2016 Combinatorial Problems as Linear and Convex Programs. Instructor: Shaddin Dughmi

Lecture 9: Counting Matchings

Partitioning Metric Spaces

Flow-Cut Gaps for Integer and Fractional Multiflows

ON COST MATRICES WITH TWO AND THREE DISTINCT VALUES OF HAMILTONIAN PATHS AND CYCLES

- Well-characterized problems, min-max relations, approximate certificates. - LP problems in the standard form, primal and dual linear programs

Hardness of Embedding Metric Spaces of Equal Size

Computing Maximum Flow with Augmenting Electrical Flows

Capacity Releasing Diffusion for Speed and Locality

An Improved Approximation Algorithm for Requirement Cut

Mathematics for Decision Making: An Introduction. Lecture 13

The Capacity of a Network

CS 6820 Fall 2014 Lectures, October 3-20, 2014

An Algorithmist s Toolkit September 24, Lecture 5

Out-colourings of Digraphs

Networks and Their Spectra

Algorithms Reading Group Notes: Provable Bounds for Learning Deep Representations

6.046 Recitation 11 Handout

Testing Equality in Communication Graphs

Finite Metric Spaces & Their Embeddings: Introduction and Basic Tools

arxiv: v1 [cs.dm] 23 Jan 2019

Cutting Plane Methods II

THE MAXIMAL SUBGROUPS AND THE COMPLEXITY OF THE FLOW SEMIGROUP OF FINITE (DI)GRAPHS

Solving the MWT. Recall the ILP for the MWT. We can obtain a solution to the MWT problem by solving the following ILP:

Fast Approximation Algorithms for Graph Partitioning using Spectral and Semidefinite-Programming Techniques

Lecture Introduction. 2 Brief Recap of Lecture 10. CS-621 Theory Gems October 24, 2012

directed weighted graphs as flow networks the Ford-Fulkerson algorithm termination and running time

Petar Maymounkov Problem Set 1 MIT with Prof. Jonathan Kelner, Fall 07

Lecture 21 November 11, 2014

The Mixed Chinese Postman Problem Parameterized by Pathwidth and Treedepth

Notes on MapReduce Algorithms

= ϕ r cos θ. 0 cos ξ sin ξ and sin ξ cos ξ. sin ξ 0 cos ξ

Lecture 14: SVD, Power method, and Planted Graph problems (+ eigenvalues of random matrices) Lecturer: Sanjeev Arora

arxiv: v1 [cs.ds] 10 Aug 2009

Key words. sparsification, approximation algorithm, graph partitioning, routing, metric space

CMPSCI 611: Advanced Algorithms

Algorithms for 2-Route Cut Problems

ON THE NUMBER OF ALTERNATING PATHS IN BIPARTITE COMPLETE GRAPHS

U.C. Berkeley CS294: Beyond Worst-Case Analysis Handout 3 Luca Trevisan August 31, 2017

Leveraging Big Data: Lecture 13

Chapter 11. Min Cut Min Cut Problem Definition Some Definitions. By Sariel Har-Peled, December 10, Version: 1.

Exact Algorithms for Dominating Induced Matching Based on Graph Partition

Quantum Algorithms for Graph Traversals and Related Problems

The Strong Largeur d Arborescence

25 Minimum bandwidth: Approximation via volume respecting embeddings

4. Duality Duality 4.1 Duality of LPs and the duality theorem. min c T x x R n, c R n. s.t. ai Tx = b i i M a i R n

New Hardness Results for Undirected Edge Disjoint Paths

The Inapproximability of Maximum Single-Sink Unsplittable, Priority and Confluent Flow Problems

Lecture 18: Inapproximability of MAX-3-SAT

Combinatorial Optimization

Robust Principal Component Analysis

Assignment 1: From the Definition of Convexity to Helley Theorem

CS 435, 2018 Lecture 3, Date: 8 March 2018 Instructor: Nisheeth Vishnoi. Gradient Descent

Dominating Set. Chapter 7

A Linear Round Lower Bound for Lovasz-Schrijver SDP Relaxations of Vertex Cover

Lecture 20: LP Relaxation and Approximation Algorithms. 1 Introduction. 2 Vertex Cover problem. CSCI-B609: A Theorist s Toolkit, Fall 2016 Nov 8

Flows and Cuts. 1 Concepts. CS 787: Advanced Algorithms. Instructor: Dieter van Melkebeek

Chapter 7 Matchings and r-factors

Randomized Algorithms III Min Cut

Graph-Constrained Group Testing

Applications of the Lopsided Lovász Local Lemma Regarding Hypergraphs

Generating p-extremal graphs

Edge Disjoint Paths in Moderately Connected Graphs

Lecture 1. 1 Overview. 2 Maximum Flow. COMPSCI 532: Design and Analysis of Algorithms August 26, 2015

CSC 5170: Theory of Computational Complexity Lecture 13 The Chinese University of Hong Kong 19 April 2010

11.1 Set Cover ILP formulation of set cover Deterministic rounding

Assignment 4: Solutions

Lecture 13: Spectral Graph Theory

Price of Stability in Survivable Network Design

Random Methods for Linear Algebra

Reductions Between Expansion Problems

Faster Johnson-Lindenstrauss style reductions

Zigzag Codes: MDS Array Codes with Optimal Rebuilding

ORIE 6334 Spectral Graph Theory November 22, Lecture 25

The Lovász Local Lemma: constructive aspects, stronger variants and the hard core model

Transcription:

GRAPH PARTITIONING USING SINGLE COMMODITY FLOWS [KRV 06] notes by Petar MAYMOUNKOV Theme The algorithmic problem of finding a sparsest cut is related to the combinatorial problem of building expander graphs from simple building blocks! 1. Preliminaries We consider an undirected graph G defined on V = [n] and edge-weighted by w(e) 0. For S V we use S to be the set of edges with exactly one endpoint in S. Define w( S) = e S w(e). Definition 1.1. The expansion of G is defined as (1.1) Φ(G) = min S V w( S) min{ S, S c } Definition 1.2. A single commodity flow ω between s t V can be viewed as ω R V V, having the following properties: (i) Skew symmetry: (ii) Capacity ω xy = ω yx, for all x, y V ω xy w(x, y), for all x, y V (iii) Conservation: ω xy = 0, for all s, t x, y V y (iv) Demand: ω sy = d, which implies ω ty = d y y

2 PETAR MAYMOUNKOV Lemma 1.3 (Concentration of measure). If v R d and ζ is a random unit vector in R d, then E (v ζ) 2 = v 2 d, and [ ] Pr (v ζ) 2 x v 2 /d e x/4, for x d/16. 2. Embedding graphs by flow Lemma 2.1. If H ξ G, then Φ(H) C ξ Φ(G). Proof. By assumption H routes in C ξ G with no congestion. Consider any cut (S, S c ). If h E(H) is cut, then it contributes w H (h) to w H ( S) and at least w H (h) to w Cξ G( S). Thus w H ( S) w Cξ G( S), and Φ(H) = min S w H ( S) min{ S, S c min } S w Cξ G( S) min{ S, S c } = Φ(C ξg) = C ξ Φ(G). 3. Main result Theorem 3.1. Given an undirected graph G and an expansion parameter α > 0, there is a randomized algorithm that w.h.p. outputs (i) Either a cut (S, S c ) with Φ(S, S c ) α, (ii) or an expander H and an embeding H ξ G with congestion at most O(log 2 n/α) Algorithm 3.1 (Main algorithm conceptually). On input G and α: Build an expander graph H as the union of simple graphs M 1,..., M t. Try to embed each M i in G with congestion 1/α. If success, then Φ(G) α/t, otherwise an M i fails to embed, which uncovers a sparse cut. 4. Constructing expanding graphs incrementally For a perfect matching µ, define the 1-step random walk W µ := 1/2 (I+µ), where µ is viewed as an adjacency matrix. Let (M 1,..., M t ) be a sequence of perfect matchings, and define the natural matching walk as the t-step walk with transition matrix W = W Mt W M1.

Definition 4.1. A matching walk is mixing iff 1 xw 1 j 1/2n for all x, y [n]. Lemma 4.2 (Matching walk mixes matching union expands). If (M 1,..., M t ) mixes, then Φ( t i=1 M i ) 1/2. Proof. Consider the directed time-line graph H defined on {1,..., n} {0,..., t}, where for all i {1,..., n} and k {0,..., t 1} (i) (i, j) H (i, j + 1), and (ii) (i, j) H (i, j + 1) where i and i are matched by M j+1. Every edge in H is given capacity 1/2. The matching walk from x [n] induces a flow of value 1 on H from (x, 0) to (1, t),..., (n, t). If the walk mixes, the flow delivers at least 1/2n units to each (1, t),..., (n, t). It is easy to see (using induction) that the simultaneous unit flows from (1, 0),..., (n, 0) do not violate the edge capacities of H. Observe that the vertex-projection of H onto {1,..., n} equals t i=1 M i, and thus the simultaneous flow is realizable there as well. For any (S, S c ) cut in [n] where S n/2, this flow delivers at least S n S /2n units to S c. From here, the Min-cut Max-flow Theorem asserts Φ( t i=1 M i ) 1/2. We are going to use a potential function ψ(w ) to measure how far W is from mixing. For now ψ(w ) is abstract. 3 5. Idea I am betting that G has expansion α: (a) If I am right, i.e. α Φ(G), I can build an expander as the union of O(log 2 n) matchings so that each embeds in G with congestion 1/α. Thus, proving that G s expansion is at least O(α/ log 2 n) (b) If I am wrong and Φ(G) α O(log 2 n Φ(G)), I might be just lucky enough that the matchings I need happen to be embedable in G with congestion 1/α (i.e. I never happen to need a matching across the sparsest cut), in which case I still prove that O(α/ log 2 n) Φ(G) (c) However, if I am significantly off, i.e. O(log 2 n Φ(G)) α, then I am bound to run into a cut sparser than α in my attempt to realize some needed matching in G Here ψ(t) := ψ(w Mt W M1 ). 6. Main algorithm

4 PETAR MAYMOUNKOV Algorithm 6.1 (Main). On input G and α: 1. (Find-Bisection) Find a bisection (S, S c ) such that adding any perfect matching M t+1 between S and S c to {M 1,..., M t } reduces the potential, in expectation, by 1 Θ(1/ log n), i.e. Eψ(t + 1) (1 Θ(1/ log n))ψ(t) 2. Using a maximum-flow procedure (a) Either, produce a perfect matching M t+1 that embeds in G with congestion 1/α, (b) Or, find a cut in G of expansion at most α 6.0.1. What lies ahead We ll define a potential so if ψ O(1/n 2 ) then W is mixing, and thus the algorithm terminates in at most O(log 2 n) iterations w.h.p. thereby producing an embedding (M 1,..., M O(log 2 n) ) G with congestion O(log2 n)/α. 7. Find next best matching Note, Find-Bisection has nothing to do with G. It s an algorithm about matchings and their mixing properties. Algorithm 7.1 (Find-Bisection). On input {M 1,..., M t }, output a bisection (S, S c ) so that the addition of any matching M t+1 between S and S c brings the matching walk significantly closer to mixing: Choose ζ {±1} n randomly, so ζ 1 Compute u = W t... W 1 ζ, where W i = 1/2 (I + M i ) Form S from the first n/2 smallest entries of u 8. Single step random walks and mixing The main object of study here is a positive and doubly stochastic matrix P R n n which encodes a random walk step on [n]. Definition 8.1. The 1-step walk P is mixing iff P xy 1/2n for all x, y. Definition 8.2. Define the potential of P as (8.1) ψ(p ) := P J/n 2

Remark 8.3. The potential of P is intended to measure (upper-bound) the l 2 distance between a 1-step random walk on P starting from a uniform distribution and the uniform distribution, i.e. (8.2) This intention is justified by: Lemma 8.4. ψ(p ) nψ (P ) ψ (P ) := P 1/n 1/n 2 5 Proof. ψ (P ) = P 1/n 1/n 2 = P 1/n J1/n 2 2 ( where J = 11 ) = 1/n 2 (P J/n)1 2 1/n P J/n 2 ( ) Cauchy-Schwarz or Frobenius norm = ψ(p )/n When ψ(p ) is sufficiently small, the 1-step walk is mixing: Lemma 8.5. If ψ(p ) 1/4n 2 then P xy 1/2n for all x, y V. 9. Averaging transformation and matchings All matchings in this text are perfect. For any self-inverse permutation (or a matching, in particular) on [n] characterized by its matrix µ R n n, consider the averaging transformation W µ = 1/2 (I + µ) applied to P. Remark 9.1. Note that W µ P represents the random walk where the first step is taken according to P and the second according to W µ. Lemma 9.2. The composition W µ P has the following properties: (i) 1 x(w µ P ) = 1 y(w µ P ) for all x µ y, (ii) (1 x + 1 y ) (W µ P )1 = (1 x + 1 y ) P 1, and (iii) W µ P is positive and doubly stochastic. Lemma 9.3. ψ(p ) ψ ( W µ P ) 1 (1 x 1 y ) P 2 2 x y µ Proof. Focus on contribution of x µ y. First, 1 xw µ P = 1 yw µ P = 1 2 (1 x + 1 y ) P

6 PETAR MAYMOUNKOV Write ψ(p ) as ψ(p ) = x 1 xp 1/n 2 Use the Parallelogram identity f 2 + g 2 = 1 2 f + g 2 + 1 f g 2 2 to compute the contribution of x and y to ψ(w µ P ) ψ(p ) as 2 1/2 (1 x + 1 y ) P 1/n 2 1 xp 1/n 2 1 yp 1/n 2 = (1 x 1 y ) P 2 10. Approximating potential reduction For a random ζ {±1} n : Lemma 10.1. With high probability, for all x, y, (10.1) (1 x 1 y ) P 2 n 1 O(log n) (1 x 1 y ) P ζ Proof. Use (1 x 1 y ) P 1 and ζ 1 and concentration of measure lemma. Corollary 10.2. With high probability, (10.2) ψ(p ) ψ ( W µ P ) n 1 O(log n) (1 x 1 y ) P ζ Remark 10.3. The success probability in both statements above can be chosen to be 1 n C for any constant C (due to concentration of measure). x µ y 11. Isolating a cut of good matchings We now additionally require ζ 1. Let u = P ζ and renumber its coordinates so that u 1 u 2 u n. Lemma 11.1. For any matching µ between {1,..., n/2} and {n/2 + 1,..., n}, (11.1) E (u x u y ) 2 ψ(p ) n 1 x y µ

7 Corollary 11.2. ( E ψ(p ) ψ ( ) (11.2) W µ P ) ψ(p ) O(log n) Proof of Lemma 11.1. Let η be the median of u 1,..., u n. x u y ) x y(u 2 ((u x η) 2 + (u y η) 2) µ x y µ = (u x η) 2 x = x u 2 x 2η x u x + nη ( ) x u 2 x Step ( ) follows from ζ 1 and the column-stochasticity of P as x u x = 1 P ζ = 1 ζ = 0. Note that u x = 1 xp ζ = ( 1 x P 1 /n ) ζ, since ζ 1. Observe that w x := (1 xp 1/n) 1, using row-stochasticity. And since ζ 1 and w x 1, applying the concentration of measure lemma gives Thus, E u 2 x = 1 xp 1/n 2 n 1 E x µ y (u x u y ) 2 E x u 2 x = x = ψ(p ) n 1 1 xp 1/n 2 n 1 12. Finding a matching or a cut For simplicity, assume G is unweighted. Algorithm 12.1 (Cut-or-Flow). The input is S V with S = n/2 and maximum allowable congestion 1/α > 0: 1. Assign each edge in G capacity 1/α. Add a source node with an outgoing unit-capacity arc to each vertex in S, and add a sink node with an incoming unit-capacity arc from each vertex in S c.

8 PETAR MAYMOUNKOV 2. Find a maximum flow between the source and the sink 3. If the flow value is at least n/2, we produce a matching between S and S c, by decomposing the flow into flow paths 4. Otherwise, we find a minimum cut separating the source and the sink, and output the partition induced on V Remark 12.1. If a matching is found, then by construction it can be embedded in G with congestion at most 1/α. Lemma 12.2. If the maximum flow-value between the source and the sink is less than n/2, then the minimum cut in G has expansion less than α. Proof. If the maximum flow is < n/2 then the minimum cut (separating the source and the sink) is also < n/2. Let the number (and capacity) of edges in the cut incident to the source (and respectively sink) be n s (and n t ). The remaining cut capacity is n/2 n s n t, thus using at most α(n/2 n s n t ) edges of G The cut separates n/2 n s vertices in S from n/2 n t vertices in S c, thus Φ G (S, S c n/2 n s n t ) α min{n/2 n s, n/2 n t } α 13. Running time and speedup via sparsification Break down: O(log 2 n) iterations w.h.p., each including: Find-Bisection in Õ(n), Single commodity flow in O(m 3/2 ) using [3], Decomposition into paths in Õ(m) In total Õ(n + m 3/2 ). Can get Õ(m + n3/2 ) using Theorem 13.1 (Benczùr Karger [2]). Given a graph G with n vertices and m edges and an error parameter ɛ > 0, there is a graph Ĝ such that (i) Ĝ has O(n log n/ɛ2 ) edges and (ii) The value of every cut in Ĝ is within (1±ɛ) factor of the corresponding cut in G. Ĝ can be constructed in O(m log 2 n) time if G is unweighted and in O(m log 3 n) time if G is weighted.

9 14. Authors conclusions and open problems The union of the flows corresponds to an embedding of a complete graph Can this approach yield a (tight) O(log n) approximation for embedding a complete graph? What about improving to a O( log n) approximation algorithm by embedding an arbitrary expander? Can this analysis be related to random collapsing process underlying METIS? 15. Petar s remarks (1) Lower bound gap When an embedding is found, the KRV algorithm/analysis is insensitive to whether some subset of the matchings can be embedded simultaneously (with congestion 1/α), which would exhibit tighter lower-bound certificates (2) Upper bound gap consider the sparsest cut in a graph comprised of two copies of K n connected by an edge. Explain how the KRV algorithm can succeed in finding an embedding when α = log 2 n/n (3) It is conceivable that for a choice of α in the critical region Φ(G) α O(log 2 n Φ(G)), multiple runs of the algorithm will eventually find a sparser cut. But this needs to be proven and will most likely happen with tiny probability (unless the graph has many sparsest cuts) (4) Is it possible to build an expander using o(log n) matchings? How (Butterfly, de Bruijn)? If yes, then replicating the same argument will give an approximation guarantee (using multi-commodity flow to embed desired matchings, and multi-commodity duality to get a cut otherwise). More on this in [4] (5) The analysis in this paper is very much about expanders built from matchings, and it is not tight because ψ 1/4n 2 implies P xy 1/2n, but the converse is not true. In other words, the needed condition is J P 1 1/2n, but the enforced condition is P J/n 2 1/2n BIBLIOGRAPHY [1] Khandekar, Rao, Vazirani, Graph partitioning using single commodity flows, STOC 06 [2] Benczùr, Karger, Approximating s-t minimum cuts in Õ(n2 ) time, STOC 96

10 PETAR MAYMOUNKOV [3] Goldberg, Rao, Beyond the flow decomposition barrier, JACM 98 [4] Arora, Hazan, Kale, O( log n) approximation to sparsest cut in Õ(n 2 ) time, FOCS 04