Lecture 06 01/31/ Proofs for emergence of giant component

Similar documents
Erdős-Renyi random graphs basics

A simple branching process approach to the phase transition in G n,p

6.207/14.15: Networks Lecture 3: Erdös-Renyi graphs and Branching processes

Susceptible-Infective-Removed Epidemics and Erdős-Rényi random

1 Mechanistic and generative models of network structure

Random Graphs. EECS 126 (UC Berkeley) Spring 2019

6.207/14.15: Networks Lecture 4: Erdös-Renyi Graphs and Phase Transitions

Modern Discrete Probability Branching processes

Branching processes. Chapter Background Basic definitions

Chapter 3. Erdős Rényi random graphs

Homework 6: Solutions Sid Banerjee Problem 1: (The Flajolet-Martin Counter) ORIE 4520: Stochastics at Scale Fall 2015

THE SECOND LARGEST COMPONENT IN THE SUPERCRITICAL 2D HAMMING GRAPH

Network models: random graphs

Resistance Growth of Branching Random Networks

Almost giant clusters for percolation on large trees

Mixing time and diameter in random graphs

The Contour Process of Crump-Mode-Jagers Branching Processes

Random Graphs. 7.1 Introduction

The range of tree-indexed random walk

Critical percolation on networks with given degrees. Souvik Dhara

On trees with a given degree sequence

ANATOMY OF THE GIANT COMPONENT: THE STRICTLY SUPERCRITICAL REGIME

Eötvös Loránd University, Budapest. 13 May 2005

Scaling limits for random trees and graphs

Hitting Probabilities

ADFOCS 2006 Saarbrücken. Phase Transition. Joel Spencer

The phase transition in the Erdös-Rényi random graph model

Random Geometric Graphs

Random Geometric Graphs

Random graphs: Random geometric graphs

Branching within branching: a general model for host-parasite co-evolution

UPPER DEVIATIONS FOR SPLIT TIMES OF BRANCHING PROCESSES

The diameter of sparse random graphs

Lecture 2: Tipping Point and Branching Processes

Random Networks. Complex Networks CSYS/MATH 303, Spring, Prof. Peter Dodds

Random trees and branching processes

Uniform spanning trees and loop-erased random walks on Sierpinski graphs

The tail does not determine the size of the giant

w i w j = w i k w k i 1/(β 1) κ β 1 β 2 n(β 2)/(β 1) wi 2 κ 2 i 2/(β 1) κ 2 β 1 β 3 n(β 3)/(β 1) (2.2.1) (β 2) 2 β 3 n 1/(β 1) = d

4.5 The critical BGW tree

Hard-Core Model on Random Graphs

Phase Transitions in Random Discrete Structures

Size and degree anti-ramsey numbers

SOLUTIONS OF SEMILINEAR WAVE EQUATION VIA STOCHASTIC CASCADES

JIGSAW PERCOLATION ON ERDÖS-RÉNYI RANDOM GRAPHS

Component structure of the vacant set induced by a random walk on a random graph. Colin Cooper (KCL) Alan Frieze (CMU)

Probabilistic Combinatorics. Jeong Han Kim

Weakly interacting particle systems on graphs: from dense to sparse

arxiv: v1 [math.co] 7 Mar 2018

Three Disguises of 1 x = e λx

Lecture 2. We now introduce some fundamental tools in martingale theory, which are useful in controlling the fluctuation of martingales.

Lecture 4. 1 Estimating the number of connected components and minimum spanning tree

On the robustness of power-law random graphs in the finite mean, infinite variance region

Sharpness of second moment criteria for branching and tree-indexed processes

Problems on Evolutionary dynamics

arxiv: v1 [math.pr] 20 May 2018

Random Networks. Complex Networks, CSYS/MATH 303, Spring, Prof. Peter Dodds

Scaling limits of random trees and random graphs

First order logic on Galton-Watson trees

Sharp threshold functions for random intersection graphs via a coupling method.

EECS 126 Probability and Random Processes University of California, Berkeley: Fall 2014 Kannan Ramchandran November 13, 2014.

On the threshold for k-regular subgraphs of random graphs

Critical behavior in inhomogeneous random graphs

Survival Probabilities for N-ary Subtrees on a Galton-Watson Family Tree

Effect of scale on long-range random graphs and chromosomal inversions

Balanced Allocation Through Random Walk

Tianjin. June Phase Transition. Joel Spencer

Poisson Processes. Stochastic Processes. Feb UC3M

Lecture 4: Phase Transition in Random Graphs

Sublinear Algorithms Lecture 3. Sofya Raskhodnikova Penn State University

Selected Exercises on Expectations and Some Probability Inequalities

4: The Pandemic process

The diameter and mixing time of critical random graphs.

Cycles with consecutive odd lengths

Itô s excursion theory and random trees

arxiv:math.pr/ v1 17 May 2004

arxiv: v1 [math.pr] 3 Apr 2017

Decision Making and Social Networks

Graph Sparsifiers. Smaller graph that (approximately) preserves the values of some set of graph parameters. Graph Sparsification

Bootstrap Percolation on Periodic Trees

A Conceptual Proof of the Kesten-Stigum Theorem for Multi-type Branching Processes

6.207/14.15: Networks Lecture 12: Generalized Random Graphs

U.C. Berkeley CS294: Beyond Worst-Case Analysis Handout 3 Luca Trevisan August 31, 2017

Network Science: Principles and Applications

MIXING TIME OF NEAR-CRITICAL RANDOM GRAPHS

The Geometry of Random Right-angled Coxeter Groups

Adventures in random graphs: Models, structures and algorithms

The concentration of the chromatic number of random graphs

FRINGE TREES, CRUMP MODE JAGERS BRANCHING PROCESSES AND m-ary SEARCH TREES

A New Random Graph Model with Self-Optimizing Nodes: Connectivity and Diameter

arxiv: v1 [math.pr] 25 Mar 2016

Lecture 4. 1 Circuit Complexity. Notes on Complexity Theory: Fall 2005 Last updated: September, Jonathan Katz

De los ejercicios de abajo (sacados del libro de Georgii, Stochastics) se proponen los siguientes:

THE DIAMETER OF WEIGHTED RANDOM GRAPHS. By Hamed Amini and Marc Lelarge EPFL and INRIA

The Minesweeper game: Percolation and Complexity

Spectral asymptotics for stable trees and the critical random graph

Networks: Lectures 9 & 10 Random graphs

Asymptotics in bond percolation on expanders

Holme and Newman (2006) Evolving voter model. Extreme Cases. Community sizes N =3200, M =6400, γ =10. Our model: continuous time. Finite size scaling

How many randomly colored edges make a randomly colored dense graph rainbow hamiltonian or rainbow connected?

Transcription:

M375T/M396C: Topics in Complex Networks Spring 2013 Lecture 06 01/31/13 Lecturer: Ravi Srinivasan Scribe: Tianran Geng 6.1 Proofs for emergence of giant component We now sketch the main ideas underlying the formation of a giant component in the Erdős- Rényi model, referring the interested reader to the references for additional details. To begin, let λ = np denote the average degree as n and p = p(n) 0. First we consider the subcritical regime λ < 1 and the supercritical regime λ > 1, for which we can use branching processes to estimate the size of the largest component. For the critical case λ = 1, we provide an elegant scaling argument that utilizes more advanced topics in probability theory: martingales and Brownian motion. 6.1.1 Subcritical regime (λ < 1) As before, we let C i be the i th largest component of G(n, p). In addition, we will say that an event A happens almost surely (denoted a.s.) if P(A) = 1. Theorem 6.1. If λ < 1, a graph in G(n, p) has no connected component of size larger than O(log n) a.s. That is, there exists a constant α such that P ( C 1 > α log(n)) 0 as n. Proof: (Sketch) Recall the Chernoff bound on the tail distributions in the Galton-Watson branching process: P ( T > k) e kh(a), { [ h(x) = sup ]} θx log E e θξ. (6.1) θ 0 We examine the size of the largest component as follows. Let v be some node in the graph G(n, p). We will bound the size of the component to which v belongs by comparing its breadth-first spanning tree to a Galton-Watson branching process T with offspring distribution ξ Binomial(n 1, p) Poisson(λ). This is seen as follows. Let w be a neighbor of v and j be the number of neighbors of v. We see that v has n j 1 neighbors other than w, and each has a probability p of being w s neighbor. We can view these nodes as possible children of w in the spanning tree. Note that w has n j 1 possible children, less than the n 1 possible children we had from the branching process T. Therefore, repeating this procedure for every node in the breadth-first spanning tree of 06-1

v, we obtain a modified Galton-Watson process in which the number of possible children decreases with each generation and thus dominated in size by T. Now denote by C(v) be the connected component to which v belongs. According to the preceding argument, P ( C(v) > k) P ( T > k) e kh(1) where h is the rate function corresponding to an approximately Poisson(λ) offspring distribution. But what is h? Assuming ξ is exactly Poissonian, we have E [ e θξ] = k=0 e θk e λ λ k k! = (λe θ ) k e λ k=0 k! = e λ e λeθ = e λ(eθ 1). Now set g(θ) = θx log E [ e θξ] = θx λ(e θ 1). Note that g (θ) = x λe θ = 0 at ˆθ = log(x/λ). Since g (θ) = λe θ < 0 for all θ 0, we know that g(θ) reaches its maximum at ˆθ. Therefore, ( x ( x ) h(x) = sup g(θ) = g(ˆθ) = x log λ θ 0 λ) λ 1 and h(1) = log(λ) + λ 1 > 0 for any λ 1. Fix δ > 0 and let k = h(1) 1 (1 + δ) log n. Then, P ( C(v) > k) e (1+δ) log n = n (1+δ) and the probability that the size of the largest component exceeds k satisfies ( ) P ( C 1 > k) P max C(v j) > k j=1..n n P ( C(v j ) > k) = nn (1+δ) = n δ. Since δ is arbitrary, we conclude P ( C 1 > α log(n)) 0 as n for all α sufficiently large. 6.1.2 Supercritical regime (λ > 1) Theorem 6.2. If λ > 1, a graph in G(n, p) has a unique giant component containing a positive fraction of the vertices while no other component contains more than O(log n) vertices. Precisely, let p ext (µ) be the extinction probability of a Galton-Watson branching process with parameter µ. Then for any ε > 0 and some constant α, we have ( ) C 1 P n (1 p ext(λ)) ε and C 2 α log n 1 as n. j=1 06-2

Proof: (Sketch) For simplicity, we only consider the case 1 < λ < 2. We again identify the connected component C(v) of some vertex v by considering its breadth-first spanning tree. In this case, we will bound the size of C(v) from below using another Galton-Watson branching process. The procedure is less straightforward than in the subcritical case since we would like to show existence of a giant component, but the breadth-first spanning tree of v has a decreasing number of possible offspring with each generation. To remedy this, we consider a branching process with a strictly smaller average number of offspring but which still remains supercritical. Let d = λ > 1, and define ñ = (2 d)n and p = d/n. Note that as long as our breadth-first process has explored less than (d 1)n nodes, there are at least ñ unexplored nodes. In this case, each node in the spanning tree has at least ñ possible children, each with probability p > p. This implies that this tree is larger in size than a branching process with offspring distribution ξ Binomial (ñ, p). Ordering the vertices in C(v), we now apply the following procedure starting with i = 1: 1) Consider node v i. 2) The component to which v i belongs to in the limiting graph as n has infinite size with constant probability at least 1 p ext (d). In this case, we have successfully identified a giant component of size C(v i ) (d 1)n. 3) Otherwise, by the argument given in the proof of the subscritical case we have that C(v i ) α log n with probability at most n δ for some δ > 0. Now repeat the same process with the next unexplored vertex v i+1. The probability we have not found a giant component after k < n steps equals (p ext (d)) k, which is less than n γ for any given γ > 0 if k > γ log n/ log p ext (d). The number of nodes belonging to small components which have been explored after k steps is then O ( (log n) 2) n, so there remain enough unexplored nodes for this procedure to work. Lastly, we would not only like the existence of a giant component with high probability, but that C 1 (1 p ext (λ)) n. While we do not show this here, it is reasonable that this should be true by the law of large numbers (in the number of nodes). That is, C 1 /n should converge to the probability that the spanning tree started from a node does not become extinct, which is approximately 1 p ext (λ). 6.1.3 Critical regime (λ = 1) Theorem 6.3. If λ = 1, then G(n, p) has a largest component whose size is of order n 2/3 a.s. That is, there exist constants a 1, a 2 > 0 such that P ( a 1 n 2/3 C 1 a 2 n 2/3) 1 as n. To show this, we will use a result from probability theory known as the martingale central limit theorem. 06-3

Definition. A stochastic process {X k : k 0} is a martingale if E [ X k ] < and E [X k X k 1 X 0,..., X k 1 ] = 0 for every k 1. That is, a martingale is a process in which the average of any increment is zero given its past history. Lemma 6.4. (Martingale central limit theorem) Suppose {X k : k 0} is a martingale such that X k X k 1 < M and X 0 < M a.s. for some fixed M and for all k 1. Assume further that V ar [X k X k 1 X 0,..., X k 1 ] = 1 as k. Then for all t 0, X 1 + + X mt m d W t as m, where (W t ) t 0 is a standard Brownian motion. Using this we obtain a proof for the critical regime: Proof: (Sketch) Let {A k : k 0} be the random walk associated to the one-by-one exploration of the breadth-first spanning forest of the graph (where the forest is comprised of the spanning trees of the various disconnected components). Recall that A k = A k 1 1 + ξ k, k 1 A 0 = 1 where A k is the size of the queue of explored nodes at time k, and ξ k is the number of children of the k th deactivated node. Given A 0,..., A k 1, ξ k Binomial (n k 1 A k 1, p) Binomial(n k, p) for n k. To examine the size of largest component at the critical value, we consider λ = 1+cn 1/3 < c <. As c passes 0 we should observe the emergence of a giant component whose size scales like n 2/3. As the following argument shows, the 1/3 exponent of the perturbative term is the appropriate scaling in which to see this critical behavior. To begin, since p = λ/n = n 1 + cn 4/3, Letting E [A k A k 1 A 0,..., A k 1 ] = E [ξ k ] 1 (n k)p 1 = c n k 1/3 n ck n, 4/3 V ar [A k A k 1 A 0,..., A k 1 ] = V ar [ξ k ] (n k)p(1 p) 1. X k = A k A k 1 ( c n k 1/3 n ck ), n 4/3 06-4

8 5 6 7 14 15 2 3 4 10 12 13 1 9 11 Figure 6.1. Example of a breadth-first spanning forest. we see that {X k : k 0} is a martingale. Now, for any m 1 define the process X (m) t = mt k=1 [ cmt X k = A mt 1 mt (mt 1) n1/3 2n ] cmt (mt 1). 2n 4/3 for t m 1 N and by linear interpolation for all nonnegative t / m 1 N. Taking m = n 2/3 yields ( X (m) t = A n 2/3 t n 1/3 ct 1 ) 2 t2 + O(1). Therefore, by the martingale central limit theorem A n 2/3 t n 1/3 (ct 12 t2 ) = X (n2/3 ) t n 1/3 + O ( n 2/3) d W t for all t 0 as n, where (W t ) t 0 is Brownian motion. To summarize, we have that A n 2/3 t n 1/3 Z t for large n, where Z t = W t + ct 1 2 t2. The size of the i th component is the i th time the queue is empty: C i = min {k 0 : A k = 1 i} n 2/3 min { t 0 : Z t = (1 i) n 1/3}. We conclude that the size of the largest component can be estimated by n 2/3 times the length τ of the largest excursion of Z t above zero. Since E [Z t ] = ct 1 2 t2 = 0 when t = 2c, we expect the distribution of τ to be concentrated about 2c when c > 0 and about zero when c 0. This yields the desired result. 06-5

Figure 6.2. Plot of Z t = W t + ct 1 2 t2 for c > 0. 6.2 Connectivity and diameter For completeness, let us state the existence of thresholds for the connectivity and diameter of Erdős-Rényi random graphs. Additional discussion along with proofs can be found in the references. For connectedness we have: Theorem 6.5. A sharp threshold for the connectedness of G(n, p) is t(n) = n 1 log n. In other words, if p(n) < (1 ε)n 1 log n for some ε > 0, then G(n, p) almost surely contains isolated nodes (so it is disconnected) in the limit n. If p(n) > (1 + ε)n 1 log n for some ε > 0, G(n, p) is almost surely connected as n. Moving past the threshold for connectivity, the diameter of the graph can be estimated with high probability: Theorem 6.6. Assume that for n large, log n λ n 1/2. Then as n, ( ) log n log n P 4 diam (G(n, p)) + 1 1. log λ log λ References 1. Aldous, D. (2007). STAT 260, Random Graphs and Complex Networks, Spring 2007 [Lecture notes]. Retrieved from: http://www.stat.berkeley.edu/~aldous/networks/ 06-6

2. Draief, M., & Massoulié, L. (2010). Epidemics and rumours in complex networks. Cambridge University Press. 3. Durrett, R. (2006). Random graph dynamics (Vol. 20). Cambridge University Press. 4. Spielman, D. (2010). Applied Mathematics 462, Graphs and Networks, Fall 2010 [Lecture notes]. Retrieved from: http://www.cs.yale.edu/homes/spielman/462/ Last edited: March 21, 2013 06-7