Random Walks and Evolving Sets: Faster Convergences and Limitations

Similar documents
Graph Partitioning Algorithms and Laplacian Eigenvalues

Lecture 14: Random Walks, Local Graph Clustering, Linear Programming

Markov chain methods for small-set expansion

Lecture 13: Spectral Graph Theory

LOCAL CHEEGER INEQUALITIES AND SPARSE CUTS

Convergence of Random Walks and Conductance - Draft

Testing Cluster Structure of Graphs. Artur Czumaj

Four graph partitioning algorithms. Fan Chung University of California, San Diego

Personal PageRank and Spilling Paint

Lecture 17: Cheeger s Inequality and the Sparsest Cut Problem

approximation algorithms I

Reductions Between Expansion Problems

Cutting Graphs, Personal PageRank and Spilling Paint

Approximation & Complexity

Lecture 12: Introduction to Spectral Graph Theory, Cheeger s inequality

Algorithmic Primitives for Network Analysis: Through the Lens of the Laplacian Paradigm

The Rainbow Turán Problem for Even Cycles

Geometry of Similarity Search

Algorithms, Graph Theory, and Linear Equa9ons in Laplacians. Daniel A. Spielman Yale University

Lower bounds on the size of semidefinite relaxations. David Steurer Cornell

linear programming and approximate constraint satisfaction

Lecture 7. 1 Normalized Adjacency and Laplacian Matrices

Lecture 1: Contraction Algorithm

Sum-of-Squares Method, Tensor Decomposition, Dictionary Learning

Reed-Muller testing and approximating small set expansion & hypergraph coloring

Probability & Computing

Holistic Convergence of Random Walks

The Markov Chain Monte Carlo Method

Approximate Counting and Markov Chain Monte Carlo

Unique Games Conjecture & Polynomial Optimization. David Steurer

Simons Workshop on Approximate Counting, Markov Chains and Phase Transitions: Open Problem Session

Spectral Graph Theory and its Applications. Daniel A. Spielman Dept. of Computer Science Program in Applied Mathematics Yale Unviersity

Lecture: Local Spectral Methods (1 of 4)

Lecture 15: Expanders

Graph Clustering Algorithms

Lecture 2: September 8

CS6999 Probabilistic Methods in Integer Programming Randomized Rounding Andrew D. Smith April 2003

Algorithms, Graph Theory, and the Solu7on of Laplacian Linear Equa7ons. Daniel A. Spielman Yale University

GRAPH PARTITIONING USING SINGLE COMMODITY FLOWS [KRV 06] 1. Preliminaries

Statistical Inference on Large Contingency Tables: Convergence, Testability, Stability. COMPSTAT 2010 Paris, August 23, 2010

Lecture 14: SVD, Power method, and Planted Graph problems (+ eigenvalues of random matrices) Lecturer: Sanjeev Arora

Stanford University CS366: Graph Partitioning and Expanders Handout 13 Luca Trevisan March 4, 2013

Spectral Algorithms for Graph Mining and Analysis. Yiannis Kou:s. University of Puerto Rico - Rio Piedras NSF CAREER CCF

Label Cover Algorithms via the Log-Density Threshold

A Linear Round Lower Bound for Lovasz-Schrijver SDP Relaxations of Vertex Cover

On the Logarithmic Calculus and Sidorenko s Conjecture

Szemerédi s regularity lemma revisited. Lewis Memorial Lecture March 14, Terence Tao (UCLA)

Markov Chains, Conductance and Mixing Time

Spectral Analysis of Random Walks

Maximum cut and related problems

Graph Partitioning Using Random Walks

Finding is as easy as detecting for quantum walks

Vertex and edge expansion properties for rapid mixing

A local algorithm for finding dense bipartite-like subgraphs

Coupling. 2/3/2010 and 2/5/2010

A quantum walk based search algorithm, and its optical realisation

Lecture 5: Random Walks and Markov Chain

The coordinates of the vertex of the corresponding parabola are p, q. If a > 0, the parabola opens upward. If a < 0, the parabola opens downward.

Detecting Sharp Drops in PageRank and a Simplified Local Partitioning Algorithm

INTRODUCTION TO MCMC AND PAGERANK. Eric Vigoda Georgia Tech. Lecture for CS 6505

Conductance and Rapidly Mixing Markov Chains

arxiv: v4 [cs.ds] 23 Jun 2015

From query complexity to computational complexity (for optimization of submodular functions)

INTRODUCTION TO MCMC AND PAGERANK. Eric Vigoda Georgia Tech. Lecture for CS 6505

1 Adjacency matrix and eigenvalues

Not all counting problems are efficiently approximable. We open with a simple example.

An Analysis of Top Trading Cycles in Two-Sided Matching Markets

Randomized Algorithms

Dissertation Defense

Convergence of Random Walks

How Robust are Thresholds for Community Detection?

Lecture 4: Applications: random trees, determinantal measures and sampling

CSE 291: Fourier analysis Chapter 2: Social choice theory

(K2) For sum of the potential differences around any cycle is equal to zero. r uv resistance of {u, v}. [In other words, V ir.]

Lecture: Local Spectral Methods (2 of 4) 19 Computing spectral ranking with the push procedure

Spectral radius, symmetric and positive matrices

Discrete Mathematics. Kernels by monochromatic paths in digraphs with covering number 2

The Lovász Local Lemma: constructive aspects, stronger variants and the hard core model

Lecture 9: Counting Matchings

Complexity Theory of Polynomial-Time Problems

The Green-Tao theorem and a relative Szemerédi theorem

Approximating Maximum Constraint Satisfaction Problems

18.S096: Spectral Clustering and Cheeger s Inequality

THE STRUCTURE OF ALMOST ALL GRAPHS IN A HEREDITARY PROPERTY

COMMUNITY DETECTION IN SPARSE NETWORKS VIA GROTHENDIECK S INEQUALITY

Graph-Constrained Group Testing

Approximating Sparsest Cut in Graphs of Bounded Treewidth

Low degree almost Boolean functions are sparse juntas

On splitting digraphs

Quasi-randomness is determined by the distribution of copies of a fixed graph in equicardinal large sets

U.C. Berkeley CS294: Beyond Worst-Case Analysis Handout 8 Luca Trevisan September 19, 2017

On the Dynamic Chromatic Number of Graphs

SDP Relaxations for MAXCUT

Graph Limits: Some Open Problems

arxiv: v3 [math.co] 10 Mar 2018

A Schur Complement Cheeger Inequality

Powerful tool for sampling from complicated distributions. Many use Markov chains to model events that arise in nature.

Unpredictable Walks on the Integers

The Green-Tao theorem and a relative Szemerédi theorem

Graph Sparsification III: Ramanujan Graphs, Lifts, and Interlacing Families

Transcription:

Random Walks and Evolving Sets: Faster Convergences and Limitations Siu On Chan 1 Tsz Chiu Kwok 2 Lap Chi Lau 2 1 The Chinese University of Hong Kong 2 University of Waterloo Jan 18, 2017 1/23

Finding small clusters Given 1-regular graph with edge weights w (Edge) expansion φ(s) w(s, V \ S) S Task: compute φ δ (G) min S δn φ(s) 2/23

Algorithm 1: Random walk For finding small clusters [Spielman Teng 04, Andersen Chung Lang 06, ] Theorem ([Kwok Lau 12]) Random walk returns S with φ(s) = O( φ(s )/ε) and S = O( S 1+ε ) in poly time, when start from nice vertices in S When ε = 1/ log S φ(s) = O( φ(s ) log S ) and S = O( S ) 3/23

Analysis: Lovász Simonovits [ 90] Measure progress of lazy random walk mixing with a curve Large φ(g) implies fast mixing (lazy = self-loop of weight 1/2 at every vertex) More on this later 4/23

Evolving Set Process [Morris 02] Yields strong bounds on mixing times [Morris Peres 05] Evolving Set Process is Markov Chain {S t } on subsets of V Given S t, choose U uniformly from [0, 1] S t+1 {v V w(s t, v) U} small U S t large U S t+1 5/23

Algorithm 2: Evolving Set Process Can find small clusters [Andersen Peres 09, Oveis Gharan Trevisan 12] Theorem ([Oveis Gharan Trevisan 12, AOPT 16]) Evolving Set Process returns S with φ(s) = O( φ(s )/ε) and S = O( S 1+ε ) in poly time, when start from nice vertices in S 6/23

Algorithm 2: Evolving Set Process Can find small clusters [Andersen Peres 09, Oveis Gharan Trevisan 12] Theorem ([Oveis Gharan Trevisan 12, AOPT 16]) Evolving Set Process returns S with φ(s) = O( φ(s )/ε) and S = O( S 1+ε ) in poly time, when start from nice vertices in S Conjecture [Oveis Gharan 13, AOPT 16] With non-trivial probability, in fact S = O( S ) If true, will refute Small-Set-Expansion Hypothesis that φ δ (G) is hard to approximate [Raghavendra Steurer 10] (cousin of Unique-Games Conjecture) 6/23

Our results: Generalized Lovász Simonovits analysis Also measure random walk mixing with LS curve 1. Large combinatorial gap ϕ(g) implies fast mixing 2. Large robust vertex expansion φ V (G) + laziness imply fast mixing 7/23

Our results: Combinatorial gap Large combinatorial gap ϕ(g) implies fast mixing ϕ(g) w(s, T) min 1 S = T n/2 S 1. Similar definition as φ(g) (which only allows T = S) 2. ϕ(g) = φ(g) for lazy graphs 3. ϕ(g) small if G has near-bipartite component Corollary (Expansion of graph powers) φ δ/4 (G t ) = min{ω( tφ δ (G)), 1/20} without laziness assumption of [Kwok Lau 14] 8/23

Our results: Vertex expansion Robust vertex expansion φ V defined by [Kannan Lovász Montenegro 06] Theorem Evolving Set Process returns S with φ(s) = O(φ(S )/(εφ V (S))) and S = O( S 1+ε ) in poly time, when start from nice vertices in S Compare: φ(s) = O( φ(s )/ε) in [APOT 16] 9/23

Our results: Vertex expansion Robust vertex expansion φ V defined by [Kannan Lovász Montenegro 06] Theorem Evolving Set Process returns S with φ(s) = O(φ(S )/(εφ V (S))) and S = O( S 1+ε ) in poly time, when start from nice vertices in S Compare: φ(s) = O( φ(s )/ε) in [APOT 16] Implies constant factor approximation when φ V (G) min S n/2 φv (S) is Ω(1) Evolving Set Process analog of spectral partitioning result in [Kwok Lau Lee 16] 9/23

Our results: Hard instances Theorem For arbitrarily small δ, some graphs have a hidden small cluster S φ(s ) ε and S = δn but Evolving Set Process never returns S with φ(s) 1 ε and S δ ε n Refute the conjecture in [Oveis Gharan 13, AOPT 16] Also limitation of random walk, PageRank, etc 10/23

Details: Generalized Lovász Simonovits Analysis 11/23

Lovász Simonovits Let p be probability vector Measure random walk mixing with curve C(p, x) C(p, x) sum of first x largest elements in p for integer x C(p, x) C(Ap, x) 12/23

Lovász Simonovits Let p be probability vector Measure random walk mixing with curve C(p, x) C(p, x) sum of first x largest elements in p for integer x C(p, x) C(Ap, x) x x(1 φ(g)) x(1 + φ(g)) When lazy C(Ap, x) 1 (C(p, x(1 φ(g))) + C(p, x(1 + φ(g)))) 2 12/23

Lovász Simonovits Let p be probability vector Measure random walk mixing with curve C(p, x) C(p, x) sum of first x largest elements in p for integer x C(p, x) C(Ap, x) x x(1 φ(g)) x(1 + φ(g)) When lazy C(Ap, x) 1 (C(p, x(1 φ(g))) + C(p, x(1 + φ(g)))) 2 By induction: C(A t p, x) x n + x (1 φ(g)2 8 ) t 12/23

Generalizing Lovász Simonovits x x(1 ϕ(g)) x(1 + ϕ(g)) C(Ap, x) 1 (C(p, x(1 ϕ(g))) + C(p, x(1 + ϕ(g)))) 2 ϕ(g) in place of φ(g) laziness not required ϕ(g) = φ(g) for lazy graphs generalizing LS to non-lazy More intuitive analysis 13/23

Sketch of main ideas C(Ap, S ) 1 (C(p, S (1 ϕ(g))) + C(p, S (1 + ϕ(g)))) 2 Sort vertices by decreasing d S (i) w(i, S) (Ap)(S) = 0 i n d S (i)p(i) = (d S (i) d S (i + 1)) p(j) 0 i n 1 j i } {{ } C(p,i) 14/23

Sketch of main ideas C(Ap, S ) 1 (C(p, S (1 ϕ(g))) + C(p, S (1 + ϕ(g)))) 2 Sort vertices by decreasing d S (i) w(i, S) (Ap)(S) = 0 i n d S (i)p(i) = (d S (i) d S (i + 1)) p(j) 0 i n 1 j i } {{ } C(p,i) 14/23

Sketch of main ideas C(Ap, S ) 1 (C(p, S (1 ϕ(g))) + C(p, S (1 + ϕ(g)))) 2 Sort vertices by decreasing d S (i) w(i, S) (Ap)(S) = 0 i n d S (i)p(i) = (d S (i) d S (i + 1)) p(j) 0 i n 1 j i } {{ } C(p,i) 1 d S (i) 14/23

Sketch of main ideas C(Ap, S ) 1 (C(p, S (1 ϕ(g))) + C(p, S (1 + ϕ(g)))) 2 Sort vertices by decreasing d S (i) w(i, S) (Ap)(S) = 0 i n d S (i)p(i) = (d S (i) d S (i + 1)) p(j) 0 i n 1 j i } {{ } C(p,i) 1 d S (i) d S (i) d S (i + 1) 14/23

Sketch of main ideas C(Ap, S ) 1 (C(p, S (1 ϕ(g))) + C(p, S (1 + ϕ(g)))) 2 From earlier (Ap)(S) (d S (i) d S (i + 1))C(p, i) 0 i n 1 d S (i) w(i, S) 1 2 Upper Area 1 S (1 ϕ(g)) 2 15/23

Sketch of main ideas C(Ap, S ) 1 (C(p, S (1 ϕ(g))) + C(p, S (1 + ϕ(g)))) 2 From earlier (Ap)(S) (d S (i) d S (i + 1))C(p, i) 0 i n 1 d S (i) w(i, S) 1 2 Upper Area 1 2 1 i T d S (i) 1 S (1 ϕ(g)) 2 15/23

Details: Hard instances for Evolving Set Process 16/23

Hard instances: Noisy hypercube Noisy k-ary hypercube (k = 1/δ) k d vertices, each represented by a string of length d over [k] Transition probability from x { to y: x i When x = x 1 x 2... x d, y i = uniformly from [k] Coordinate cut S = {x x 1 = 0} satisfies φ(s) ε and S = δn prob ε prob 1 ε 17/23

Why local algorithms fail Lets start from S 0 = { 0} Evolving Set Process treats all vertices with the same Hamming weight equally The process only explores sets S t that are symmetric under coordinate permutations (in fact, Hamming balls) Small Hamming balls on noisy k-ary hypercube are expanding [Chan Mossel Neeman 14] 18/23

Hamming balls B expand B = {x [k] d x r} φ(b) = Pr x y[x, y B] Pr x [x B] 19/23

Hamming balls B expand B = {x [k] d x r} φ(b) = Pr x y[x, y B] Pr x [x B] Pr[x B] = E x [1 x1 0 + + 1 xd 0 r] x Pr[g r ] Gaussian g Pr [x, y B] = E x,y 1 xi 0 r and 1 yi 0 r x y 1 i d Pr[g r and h r ] 1 i d Gaussians g, h 19/23

Hamming balls B expand B = {x [k] d x r} φ(b) = Pr x y[x, y B] Pr x [x B] 1 i d Pr g,h[g, h r ] Pr g [g r ] Pr[x B] = E x [1 x1 0 + + 1 xd 0 r] x Pr[g r ] Gaussian g Pr [x, y B] = E x,y 1 xi 0 r and 1 yi 0 r x y Pr[g r and h r ] 1 i d Gaussians g, h 19/23

Expansion in Gaussian space B = {x [k] d x r} φ(b) = Pr x y[x, y B] Pr x [x B] Pr g,h[g, h r ] Pr g [g r ] r 0 1 r 0 20/23

Symmetry Obstacles to all local clustering algorithms Evolving Set Process, random walk, PageRank, etc All fall for the symmetry Folded noisy hypercubes are hard instances for SDP Folding necessary to rule out sparse cuts in those instances In our situation, we cannot fold Our instances are easy for Lasserre/sum-of-squares 21/23

Summary New analysis of Lovász Simonovits curve Faster convergence with large combinatorial gap ϕ(g) or robust vertex expansion φ V (G) x x(1 ϕ(g)) x(1 + ϕ(g)) Limitations of all local clustering algorithms: Coordinate cuts vs Hamming balls under symmetry 22/23

Open problems Ω( φ(s ) log S ) lower bound for Evolving Set Process and random walk? Thank you 23/23