Uniform generation of random regular graphs

Similar documents
Uniform generation of random graphs with power-law degree sequences

Random regular graphs. Nick Wormald University of Waterloo

Generating random regular graphs quickly

Generation and properties of random graphs and analysis of randomized algorithms

Artificial Intelligence

Lecture 17: D-Stable Polynomials and Lee-Yang Theorems

Sampling regular graphs and a peer-to-peer network

Not all counting problems are efficiently approximable. We open with a simple example.

Approximate Counting and Markov Chain Monte Carlo

Rapid Mixing of the Switch Markov Chain for Strongly Stable Degree Sequences and 2-Class Joint Degree Matrices

Why Water Freezes (A probabilistic view of phase transitions)

Sampling Good Motifs with Markov Chains

A Sequential Algorithm for Generating Random Graphs

Counting Linear Extensions of a Partial Order

On the chromatic number of random graphs with a fixed degree sequence

Sampling regular graphs and a peer-to-peer network

Lecture 21: Counting and Sampling Problems

Average Case Complexity: Levin s Theory

Packing chromatic number of subcubic graphs

Improved FPTAS for Multi-Spin Systems

Latent voter model on random regular graphs

Lecture 4. Random graphs II. 4.1 Automorphisms. 4.2 The number of simple graphs

Counting and sampling paths in graphs

Small subgraphs of random regular graphs

Loose Hamilton Cycles in Random k-uniform Hypergraphs

Dynamics for the critical 2D Potts/FK model: many questions and a few answers

Markov Chain Monte Carlo Data Association for Multiple-Target Tracking

CS 343: Artificial Intelligence

Convex Optimization CMU-10725

HAMILTON CYCLES IN RANDOM REGULAR DIGRAPHS

Approximate Counting

Lecture 1: Introduction: Equivalence of Counting and Sampling

CS 188: Artificial Intelligence. Bayes Nets

SPIN SYSTEMS: HARDNESS OF APPROXIMATE COUNTING VIA PHASE TRANSITIONS

k-regular subgraphs near the k-core threshold of a random graph

Coupling AMS Short Course

6.842 Randomness and Computation March 3, Lecture 8

Embedding the Erdős-Rényi Hypergraph into the Random Regular Hypergraph and Hamiltonicity

Mathematics Qualifying Examination January 2015 STAT Mathematical Statistics

University of Chicago Autumn 2003 CS Markov Chain Monte Carlo Methods

Stochastic inference in Bayesian networks, Markov chain Monte Carlo methods

Statistics 251/551 spring 2013 Homework # 4 Due: Wednesday 20 February

The expansion of random regular graphs

Markov Chain Monte Carlo Using the Ratio-of-Uniforms Transformation. Luke Tierney Department of Statistics & Actuarial Science University of Iowa

Markov Chain Monte Carlo Data Association for Multi-Target Tracking

Introduction to Machine Learning CMU-10701

10 Steps to Counting Unlabeled Planar Graphs: 20 Years Later

Local Kesten McKay law for random regular graphs

17 : Markov Chain Monte Carlo

Pattern Recognition and Machine Learning. Bishop Chapter 11: Sampling Methods

Generating and counting Hamilton cycles in random regular graphs

Lecture 19: November 10

Bayesian Networks: Construction, Inference, Learning and Causal Interpretation. Volker Tresp Summer 2014

Lecture 2: September 8

Mathematics Ph.D. Qualifying Examination Stat Probability, January 2018

Approximate enumeration of trivial words.

Variable Elimination (VE) Barak Sternberg

Lecture 6: September 22

Mixing Times and Hitting Times

Inapproximability After Uniqueness Phase Transition in Two-Spin Systems

Lecture 5: Counting independent sets up to the tree threshold

The Complexity of Approximating Small Degree Boolean #CSP

Sampling Algorithms for Probabilistic Graphical models

CPSC 540: Machine Learning

Lecture 20: Reversible Processes and Queues

18 : Advanced topics in MCMC. 1 Gibbs Sampling (Continued from the last lecture)

Lecture 28: April 26

Perfect simulation of repulsive point processes

Mixing Rates for the Gibbs Sampler over Restricted Boltzmann Machines

Rapidly Mixing Markov Chains for Sampling Contingency Tables with a Constant Number of Rows

Subgraphs of random graphs with specified degrees

Bayesian Networks: Construction, Inference, Learning and Causal Interpretation. Volker Tresp Summer 2016

Asymptotic enumeration of correlation-immune functions

Coupling. 2/3/2010 and 2/5/2010

PROBABILISTIC REASONING SYSTEMS

CSC 2541: Bayesian Methods for Machine Learning

Dynamic Noninterference Analysis Using Context Sensitive Static Analyses. Gurvan Le Guernic July 14, 2007

Inference in Bayesian networks

Minimum Power Dominating Sets of Random Cubic Graphs

Testing that distributions are close

A simple condition implying rapid mixing of single-site dynamics on spin systems

. p.1. Mathematical Models of the WWW and related networks. Alan Frieze

A Deterministic Fully Polynomial Time Approximation Scheme For Counting Integer Knapsack Solutions Made Easy

The cover time of sparse random graphs.

COL 352 Introduction to Automata and Theory of Computation Major Exam, Sem II , Max 80, Time 2 hr. Name Entry No. Group

Lecture 8: Bayesian Networks

Conductance and Rapidly Mixing Markov Chains

Markov Decision Processes

MCMC notes by Mark Holder

Notes for Lecture Notes 2

Asymptotic enumeration of strongly connected digraphs by vertices and edges

Graphical Models and Kernel Methods

Randomized Complexity Classes; RP

A LINEAR APPROXIMATE-SIZE RANDOM SAMPLER FOR LABELLED PLANAR GRAPHS

The Mixture Approach for Simulating New Families of Bivariate Distributions with Specified Correlations

Weakly interacting particle systems on graphs: from dense to sparse

Markov Chains and MCMC

Approximation algorithms for cycle packing problems

STA 4273H: Statistical Machine Learning

1 Randomized Computation

Transcription:

Jane Gao Joint work with Nick Wormald 15th January, 2016 27th June, 2016

Why? A classical TCS problem; Intimate connection with enumeration; Testing algorithms with random input; Coping with big data.

Commonly used methods Rejection algorithm Boltzmann sampler MCMC Coupling from the past Switching algorithm

Generating random d-regular graphs Tinhofer 79 non-uniform random generation. Rejection algorithm uniform sampler for small d (d = O( log n)). A. Békéssy, P. Békéssy and Komlós 72; Bender and Canfield 78; Bollobás 80. MCMC approximate sampler. Jerrum and Sinclair 90 for any d, FPTAS, but no explicit bound on the time complexity; Cooper, Dyer and Greenhill 07 mixing time bounded by d 24 n 9 log n; Greenhill 15 non-regular case, mixing time bounded by 14 M 10 log M. Switching algorithm fast, uniform sampler. McKay and Wormald 90 for d = O(n 1/3 ). Gao and Wormald 15 (NEW) for d = o(n 1/2 ).

Generating random d-regular graphs Tinhofer 79 non-uniform random generation. Rejection algorithm uniform sampler for small d (d = O( log n)). A. Békéssy, P. Békéssy and Komlós 72; Bender and Canfield 78; Bollobás 80. MCMC approximate sampler. Jerrum and Sinclair 90 for any d, FPTAS, but no explicit bound on the time complexity; Cooper, Dyer and Greenhill 07 mixing time bounded by d 24 n 9 log n; Greenhill 15 non-regular case, mixing time bounded by 14 M 10 log M. Switching algorithm fast, uniform sampler. McKay and Wormald 90 for d = O(n 1/3 ). Gao and Wormald 15 (NEW) for d = o(n 1/2 ).

Generating random d-regular graphs Tinhofer 79 non-uniform random generation. Rejection algorithm uniform sampler for small d (d = O( log n)). A. Békéssy, P. Békéssy and Komlós 72; Bender and Canfield 78; Bollobás 80. MCMC approximate sampler. Jerrum and Sinclair 90 for any d, FPTAS, but no explicit bound on the time complexity; Cooper, Dyer and Greenhill 07 mixing time bounded by d 24 n 9 log n; Greenhill 15 non-regular case, mixing time bounded by 14 M 10 log M. Switching algorithm fast, uniform sampler. McKay and Wormald 90 for d = O(n 1/3 ). Gao and Wormald 15 (NEW) for d = o(n 1/2 ).

Generating random d-regular graphs Tinhofer 79 non-uniform random generation. Rejection algorithm uniform sampler for small d (d = O( log n)). A. Békéssy, P. Békéssy and Komlós 72; Bender and Canfield 78; Bollobás 80. MCMC approximate sampler. Jerrum and Sinclair 90 for any d, FPTAS, but no explicit bound on the time complexity; Cooper, Dyer and Greenhill 07 mixing time bounded by d 24 n 9 log n; Greenhill 15 non-regular case, mixing time bounded by 14 M 10 log M. Switching algorithm fast, uniform sampler. McKay and Wormald 90 for d = O(n 1/3 ). Gao and Wormald 15 (NEW) for d = o(n 1/2 ).

Generating random d-regular graphs Tinhofer 79 non-uniform random generation. Rejection algorithm uniform sampler for small d (d = O( log n)). A. Békéssy, P. Békéssy and Komlós 72; Bender and Canfield 78; Bollobás 80. MCMC approximate sampler. Jerrum and Sinclair 90 for any d, FPTAS, but no explicit bound on the time complexity; Cooper, Dyer and Greenhill 07 mixing time bounded by d 24 n 9 log n; Greenhill 15 non-regular case, mixing time bounded by 14 M 10 log M. Switching algorithm fast, uniform sampler. McKay and Wormald 90 for d = O(n 1/3 ). Gao and Wormald 15 (NEW) for d = o(n 1/2 ).

Continue... Other methods asymptotic approximate sampler. Steger and Wormald 99 for d = n 1/28 ; Kim and Vu 06 for d n 1/3 ɛ ; Bayati, Kim and Saberi 10 for d n 1/2 ɛ ; Zhao 13 for d = o(n 1/3 ). These algorithms are fast (linear time). However, unlike MCMC, the approximation error depends on n and cannot be reduced by running the algorithm longer.

Continue... Other methods asymptotic approximate sampler. Steger and Wormald 99 for d = n 1/28 ; Kim and Vu 06 for d n 1/3 ɛ ; Bayati, Kim and Saberi 10 for d n 1/2 ɛ ; Zhao 13 for d = o(n 1/3 ). These algorithms are fast (linear time). However, unlike MCMC, the approximation error depends on n and cannot be reduced by running the algorithm longer.

Rejection algorithm n = 6, d = 3

Rejection algorithm n = 6, d = 3

Rejection algorithm n = 6, d = 3

Rejection algorithm n = 6, d = 3 reject!

Rejection algorithm n = 6, d = 3

Rejection algorithm n = 6, d = 3

Rejection algorithm n = 6, d = 3

Rejection algorithm n = 6, d = 3 accept!

Rejection algorithm The time complexity is exponential in d 2.

Switching algorithm (McKay and Wormald 90): DEG n = 6, d = 3

Switching algorithm (McKay and Wormald 90): DEG n = 6, d = 3

Switching algorithm (McKay and Wormald 90): DEG For d = O(n 1/3 ), with a positive probability (bounded away from 0), there are no double-loops, or multiple edges with multiplicity greater than 2. So we only need to worry about loops and double-edges.

Switching algorithm (McKay and Wormald 90): DEG n = 6, d = 3

Switching algorithm (McKay and Wormald 90): DEG n = 6, d = 3

Switching algorithm (McKay and Wormald 90): DEG switch away double-edges

Switching algorithm (McKay and Wormald 90): DEG Let S i denote the set of pairings containing exactly i loops (correspondingly, i double-edges, in the phase for double-edge reduction). A switching coverts a pairing P S i to P S i 1.

Switching algorithm (McKay and Wormald 90): DEG Let S i denote the set of pairings containing exactly i loops (correspondingly, i double-edges, in the phase for double-edge reduction). A switching coverts a pairing P S i to P S i 1.

Switching algorithm (McKay and Wormald 90): DEG Let S i denote the set of pairings containing exactly i loops (correspondingly, i double-edges, in the phase for double-edge reduction). A switching coverts a pairing P S i to P S i 1.

Switching algorithm (McKay and Wormald 90): DEG P P S i S i 1 S i 1 S 0

Switching algorithm (McKay and Wormald 90): DEG If P is uniformly distributed in S i, is P uniformly distributed in S i 1? No. Because The number (N(P)) of ways to perform a switching to P S i is not uniformly the same for all P S i ; The number (N (P )) of ways to reach P S i 1 via a switching is not uniformly the same for all P S i 1.

Switching algorithm (McKay and Wormald 90): DEG If P is uniformly distributed in S i, is P uniformly distributed in S i 1? No. Because The number (N(P)) of ways to perform a switching to P S i is not uniformly the same for all P S i ; The number (N (P )) of ways to reach P S i 1 via a switching is not uniformly the same for all P S i 1.

Switching algorithm (McKay and Wormald 90): DEG If P is uniformly distributed in S i, is P uniformly distributed in S i 1? No. Because The number (N(P)) of ways to perform a switching to P S i is not uniformly the same for all P S i ; The number (N (P )) of ways to reach P S i 1 via a switching is not uniformly the same for all P S i 1.

Switching algorithm (McKay and Wormald 90): DEG Use rejection to enforce uniformity of P S i 1. Once a random switching (P, P ) is chosen. perform an f-rejection to equalise the probability of a switching for all P S i ; perform a b-rejection to equalise the probability that P is reached for all P S i 1.

Switching algorithm (McKay and Wormald 90): DEG Use rejection to enforce uniformity of P S i 1. Once a random switching (P, P ) is chosen. perform an f-rejection to equalise the probability of a switching for all P S i ; perform a b-rejection to equalise the probability that P is reached for all P S i 1.

Switching algorithm (McKay and Wormald 90): DEG P P S i f-rejected with probability 1 S i 1 N(P ) max P Si {N(P )} ; b-rejected with probability 1 min P S i 1 {N (P )} N (P ) ;

Switching algorithm (McKay and Wormald 90): DEG Inductively, the final pairing P is uniformly distributed in S 0, if no rejection has occurred. P represents a uniformly random d-regular graph. The probability of a rejection is away from 1 if d = O(n 1/3 ).

Switching algorithm (McKay and Wormald 90): DEG Inductively, the final pairing P is uniformly distributed in S 0, if no rejection has occurred. P represents a uniformly random d-regular graph. The probability of a rejection is away from 1 if d = O(n 1/3 ).

Switching algorithm (McKay and Wormald 90): DEG Inductively, the final pairing P is uniformly distributed in S 0, if no rejection has occurred. P represents a uniformly random d-regular graph. The probability of a rejection is away from 1 if d = O(n 1/3 ).

For d n 1/3 For d n 1/3, DEG is inefficient as the probability of a rejection during the algorithm is very close to 1. This is because variation of N(P) is too big for P S i, causing big f-rejection probability; variation of N (P ) is too big for P S i 1, causing big b-rejection probability. REG reduces the probabilities of f and b-rejections, and thus runs efficiently for larger d.

For d n 1/3 For d n 1/3, DEG is inefficient as the probability of a rejection during the algorithm is very close to 1. This is because variation of N(P) is too big for P S i, causing big f-rejection probability; variation of N (P ) is too big for P S i 1, causing big b-rejection probability. REG reduces the probabilities of f and b-rejections, and thus runs efficiently for larger d.

For d n 1/3 For d n 1/3, DEG is inefficient as the probability of a rejection during the algorithm is very close to 1. This is because variation of N(P) is too big for P S i, causing big f-rejection probability; variation of N (P ) is too big for P S i 1, causing big b-rejection probability. REG reduces the probabilities of f and b-rejections, and thus runs efficiently for larger d.

REG: reduce f-rejection To reduce f-rejection, we allow some switchings that were forbidden before, these switchings are categorised into different classes (e.g. class B).

Class A: the typical one no edges class A

Class B: the one forbidden by DEG no edges class B

Reduce b-rejection To reduce b-rejection, we perform occasionally some other types of switchings that will boost the pairings that were underrepresented, i.e. with small value of N (P ).

Type I no edges type I

Type I no edges type I

Type II no edges type II, class B

New features Each switching is associated with a type τ T and a class α A; The Markov chain among (S i ) i 0 cycles; The algorithm needs to decide which type of switchings to be performed in each step.

New features Each switching is associated with a type τ T and a class α A; The Markov chain among (S i ) i 0 cycles; The algorithm needs to decide which type of switchings to be performed in each step.

New features Each switching is associated with a type τ T and a class α A; The Markov chain among (S i ) i 0 cycles; The algorithm needs to decide which type of switchings to be performed in each step.

New features Each switching is associated with a type τ T and a class α A; The Markov chain among (S i ) i 0 cycles; The algorithm needs to decide which type of switchings to be performed in each step.

Transitions into S j

New elements of the analysis We no longer equalise the transition probabilities out of or into pairings in a certain S i, instead we equalise the expected number of times each pairing is visited in a given S i. This equalisation is done within each switching class α A. For each P S i, the algorithm probabilistically determines which switching type τ to be performed. There can be a t-rejection if no type is chosen.

New elements of the analysis We no longer equalise the transition probabilities out of or into pairings in a certain S i, instead we equalise the expected number of times each pairing is visited in a given S i. This equalisation is done within each switching class α A. For each P S i, the algorithm probabilistically determines which switching type τ to be performed. There can be a t-rejection if no type is chosen.

New elements of the analysis We no longer equalise the transition probabilities out of or into pairings in a certain S i, instead we equalise the expected number of times each pairing is visited in a given S i. This equalisation is done within each switching class α A. For each P S i, the algorithm probabilistically determines which switching type τ to be performed. There can be a t-rejection if no type is chosen.

New elements of the analysis We no longer equalise the transition probabilities out of or into pairings in a certain S i, instead we equalise the expected number of times each pairing is visited in a given S i. This equalisation is done within each switching class α A. For each P S i, the algorithm probabilistically determines which switching type τ to be performed. There can be a t-rejection if no type is chosen.

Algorithm framework Given P S i, (i) If i = 0, output P. (ii) Choose a type: choose τ with probability ρ τ (i), and with the remaining probability, 1 τ ρ τ (i), perform a t-rejection. Then select u.a.r. one of the type τ switchings that can be performed on P. (iii) Let P be the element that the selected switching would produce if applied to P, let α be the class of the selected switching and let i = S(P ). Perform an f-rejection with probability 1 N τ (P)/ max{n τ (P )} and then perform a b-rejection with probability 1 min{n α(p )}/N α(p ); (iv) if no rejection occurred, replace P with P. (v) Repeat until i = 0.

The rest of the definition The algorithm is defined after we fix all parameters ρ τ (i), τ {I, II }. These probabilities are designed to equalise the expected number of times each pairing is visited in a certain S i. Thereby we deduce a system of equations that ρ τ (i) must satisfy. We describe an efficient scheme to find a desirable solution to the system of equations. done!

The rest of the definition The algorithm is defined after we fix all parameters ρ τ (i), τ {I, II }. These probabilities are designed to equalise the expected number of times each pairing is visited in a certain S i. Thereby we deduce a system of equations that ρ τ (i) must satisfy. We describe an efficient scheme to find a desirable solution to the system of equations. done!

The rest of the definition The algorithm is defined after we fix all parameters ρ τ (i), τ {I, II }. These probabilities are designed to equalise the expected number of times each pairing is visited in a certain S i. Thereby we deduce a system of equations that ρ τ (i) must satisfy. We describe an efficient scheme to find a desirable solution to the system of equations. done!

The rest of the definition The algorithm is defined after we fix all parameters ρ τ (i), τ {I, II }. These probabilities are designed to equalise the expected number of times each pairing is visited in a certain S i. Thereby we deduce a system of equations that ρ τ (i) must satisfy. We describe an efficient scheme to find a desirable solution to the system of equations. done!

The rest of the definition The algorithm is defined after we fix all parameters ρ τ (i), τ {I, II }. These probabilities are designed to equalise the expected number of times each pairing is visited in a certain S i. Thereby we deduce a system of equations that ρ τ (i) must satisfy. We describe an efficient scheme to find a desirable solution to the system of equations. done!

The rest of the definition The algorithm is defined after we fix all parameters ρ τ (i), τ {I, II }. These probabilities are designed to equalise the expected number of times each pairing is visited in a certain S i. Thereby we deduce a system of equations that ρ τ (i) must satisfy. We describe an efficient scheme to find a desirable solution to the system of equations. done!

Time complexity Theorem (Gao and Wormald 15) REG generates d-regular graphs uniformly at random. For d = o( n), the expected running time of REG for generating one graph is O(d 3 n).

Approximate sampler: REG* Most of the running time of REG is spent on computing the probability of a b-rejection in each step. If we ignore b-rejections, we get a linear-running-time approximate sampler REG*. Theorem (Gao and Wormald 15) For d = o( n), REG* generates a random d-regular graph whose total variation distance from the uniform distribution is o(1). The expected running time of REG* is O(dn).

Approximate sampler: REG* Most of the running time of REG is spent on computing the probability of a b-rejection in each step. If we ignore b-rejections, we get a linear-running-time approximate sampler REG*. Theorem (Gao and Wormald 15) For d = o( n), REG* generates a random d-regular graph whose total variation distance from the uniform distribution is o(1). The expected running time of REG* is O(dn).

Approximate sampler: REG* Most of the running time of REG is spent on computing the probability of a b-rejection in each step. If we ignore b-rejections, we get a linear-running-time approximate sampler REG*. Theorem (Gao and Wormald 15) For d = o( n), REG* generates a random d-regular graph whose total variation distance from the uniform distribution is o(1). The expected running time of REG* is O(dn).

Future directions Larger d. General degree sequences. Heavily-tailed degree sequences such as power-law sequences. New switchings.