Weakly interacting particle systems on graphs: from dense to sparse

Similar documents
Hard-Core Model on Random Graphs

Interacting diffusions on sparse graphs: hydrodynamics from local weak limits

Strong Solutions and a Bismut-Elworthy-Li Formula for Mean-F. Formula for Mean-Field SDE s with Irregular Drift

On a class of stochastic differential equations in a financial network model

1 Brownian Local Time

Local limits of random graphs

Convergence of Feller Processes

Statistical Physics on Sparse Random Graphs: Mathematical Perspective

Erdős-Renyi random graphs basics

In terms of measures: Exercise 1. Existence of a Gaussian process: Theorem 2. Remark 3.

Critical percolation on networks with given degrees. Souvik Dhara

Uncertainty quantification and systemic risk

Lipschitz continuity properties

Introduction to Empirical Processes and Semiparametric Inference Lecture 09: Stochastic Convergence, Continued

Functional Limit theorems for the quadratic variation of a continuous time random walk and for certain stochastic integrals

Traveling waves dans le modèle de Kuramoto quenched

Some Properties of NSFDEs

16 1 Basic Facts from Functional Analysis and Banach Lattices

Lecture 06 01/31/ Proofs for emergence of giant component

Brownian Motion. 1 Definition Brownian Motion Wiener measure... 3

Multiple points of the Brownian sheet in critical dimensions

Introduction to Empirical Processes and Semiparametric Inference Lecture 08: Stochastic Convergence

SDE Coefficients. March 4, 2008

Weak solutions of mean-field stochastic differential equations

Large Deviations Principles for McKean-Vlasov SDEs, Skeletons, Supports and the law of iterated logarithm

UNCERTAINTY FUNCTIONAL DIFFERENTIAL EQUATIONS FOR FINANCE

Approximating a single component of the solution to a linear system

Undirected Graphical Models

Lecture 5. If we interpret the index n 0 as time, then a Markov chain simply requires that the future depends only on the present and not on the past.

Lecture 4 October 18th

Simple random walk on the two-dimensional uniform spanning tree and its scaling limits

On some relations between Optimal Transport and Stochastic Geometric Mechanics

Gillespie s Algorithm and its Approximations. Des Higham Department of Mathematics and Statistics University of Strathclyde

Phase Transitions in Random Discrete Structures

A stochastic particle system for the Burgers equation.

Homogenization with stochastic differential equations

EXCHANGEABLE MARKOV PROCESSES ON GRAPHS: FELLER CASE

Studying large networks via local weak limit theory

CONVERGENCE, UNANIMITY AND DISAGREEMENT IN MAJORITY DYNAMICS ON UNIMODULAR GRAPHS AND RANDOM GRAPHS

Introduction to Empirical Processes and Semiparametric Inference Lecture 02: Overview Continued

Stochastic Integration and Stochastic Differential Equations: a gentle introduction

Geometric projection of stochastic differential equations

Probabilistic Graphical Models

1 Introduction. 2 Diffusion equation and central limit theorem. The content of these notes is also covered by chapter 3 section B of [1].

Metric Spaces and Topology

Mean-Field optimization problems and non-anticipative optimal transport. Beatrice Acciaio

A graph contains a set of nodes (vertices) connected by links (edges or arcs)

Lecture 4: Introduction to stochastic processes and stochastic calculus

Scaling limits for random trees and graphs

Stochastic contraction BACS Workshop Chamonix, January 14, 2008

THE INVERSE FUNCTION THEOREM

Bootstrap Percolation on Periodic Trees

Random and Deterministic perturbations of dynamical systems. Leonid Koralov

On the connection between symmetric N-player games and mean field games

A Concise Course on Stochastic Partial Differential Equations

Malliavin Calculus in Finance

SPHERES IN THE CURVE COMPLEX

Strong uniqueness for stochastic evolution equations with possibly unbounded measurable drift term

6.867 Machine learning, lecture 23 (Jaakkola)

Multilevel Monte Carlo for Stochastic McKean-Vlasov Equations

1 Tridiagonal matrices

Logarithmic scaling of planar random walk s local times

Mean-square Stability Analysis of an Extended Euler-Maruyama Method for a System of Stochastic Differential Equations

The Smoluchowski-Kramers Approximation: What model describes a Brownian particle?

Random Walks on Hyperbolic Groups IV

5. Density evolution. Density evolution 5-1

A simple branching process approach to the phase transition in G n,p

Poisson random measure: motivation

arxiv: v1 [math.pr] 24 Sep 2009

The multidimensional Ito Integral and the multidimensional Ito Formula. Eric Mu ller June 1, 2015 Seminar on Stochastic Geometry and its applications

PROBABILITY: LIMIT THEOREMS II, SPRING HOMEWORK PROBLEMS

The Wiener Itô Chaos Expansion

Homework # , Spring Due 14 May Convergence of the empirical CDF, uniform samples

Control and synchronization in systems coupled via a complex network

Large Deviations for Small-Noise Stochastic Differential Equations

Large Deviations for Perturbed Reflected Diffusion Processes

Large Deviations for Small-Noise Stochastic Differential Equations

4: The Pandemic process

Lecture 9. d N(0, 1). Now we fix n and think of a SRW on [0,1]. We take the k th step at time k n. and our increments are ± 1

Basic Definitions: Indexed Collections and Random Functions

Session 1: Probability and Markov chains

Probabilistic Graphical Models

Simulation Method for Solving Stochastic Differential Equations with Constant Diffusion Coefficients

Embeddings of finite metric spaces in Euclidean space: a probabilistic view

4 Countability axioms

Stochastic Volatility and Correction to the Heat Equation

The Cameron-Martin-Girsanov (CMG) Theorem

Introduction to Random Diffusions

A connection between the stochastic heat equation and fractional Brownian motion, and a simple proof of a result of Talagrand

p 1 ( Y p dp) 1/p ( X p dp) 1 1 p

PROBABILITY: LIMIT THEOREMS II, SPRING HOMEWORK PROBLEMS

A. Bovier () Branching Brownian motion: extremal process and ergodic theorems

1 Weak Convergence in R k

The range of tree-indexed random walk

High dimensional ising model selection using l 1 -regularized logistic regression

Infinitely iterated Brownian motion

On trees with a given degree sequence

2012 NCTS Workshop on Dynamical Systems

arxiv: v1 [math.pr] 2 Aug 2017

Stochastic Differential Equations.

Transcription:

Weakly interacting particle systems on graphs: from dense to sparse Ruoyu Wu University of Michigan (Based on joint works with Daniel Lacker and Kavita Ramanan) USC Math Finance Colloquium October 29, 2018 1 / 49

Overview 1 Networks of interacting stochastic processes 2 Local convergence 3 Conditional independence 4 Local dynamics on infinite regular trees 2 / 49

Networks of interacting stochastic processes Given a finite connected graph G = (V, E), write d v = degree of vertex v, and u v if (u, v) E. Each node v V has a particle whose stochastic evolution depends only on its own state and that of nearest neighbors 3 / 49

Networks of interacting stochastic processes Given a finite connected graph G = (V, E), write d v = degree of vertex v, and u v if (u, v) E. Each node v V has a particle whose stochastic evolution depends only on its own state and that of nearest neighbors For example, as a discrete-time Markov chain: ( ) X v (t + 1) = F X v (t), (X u (t)) u v, ξ v (t + 1). State space S Continuous transition function F Independent noises ξ v (t), v V, t = 0, 1,... 3 / 49

Networks of interacting stochastic processes Given a finite connected graph G = (V, E), write d v = degree of vertex v, and u v if (u, v) E. Each node v V has a particle whose stochastic evolution depends only on its own state and that of nearest neighbors For example, as a discrete-time Markov chain: ( ) X v (t + 1) = F X v (t), (X u (t)) u v, ξ v (t + 1). State space S Continuous transition function F Independent noises ξ v (t), v V, t = 0, 1,... Arise in: probabilistic cellular automaton, synchronous Markov chain, simultaneous updating Examples: Voter model, contact process, exclusion processes 3 / 49

Example: Contact process State space S = {0, 1} = {off/healthy, on/infected}. Parameters p, q [0, 1]. Transition rule: At time t, if particle v is at... state X v (t) = 1, it switches to X v (t + 1) = 0 w.p. q, state X v (t) = 0, it switches to X v (t + 1) = 1 w.p. p X u (t). d v u v 4 / 49

Networks of interacting diffusions Given a finite connected graph G = (V, E), write d v = degree of vertex v, and u v if (u, v) E. Or as a diffusion: dx v (t) = 1 b(x v (t), X u (t))dt + dw v (t), d v u v where (W v ) v V are independent Brownian motions. For concreteness, assume b is Lipschitz throughout, and initial states (X v (0)) v V are i.i.d. and square-integrable. 5 / 49

Networks of interacting stochastic processes Key questions Given a sequence of graphs G n = (V n, E n ) with V n, how can we describe the limiting behavior of... the dynamics of a typical or fixed particle X v (t), t [0, T ]? 1 the empirical distribution of particles V n v V n δ Xv (t)? 6 / 49

Networks of interacting stochastic processes Key questions Given a sequence of graphs G n = (V n, E n ) with V n, how can we describe the limiting behavior of... the dynamics of a typical or fixed particle X v (t), t [0, T ]? 1 the empirical distribution of particles V n v V n δ Xv (t)? Mean field as a special case If G n is the complete graph on n vertices, we are in the mean field (McKean-Vlasov) setting. 6 / 49

The mean field case (McKean-Vlasov 1966) Particles i = 1,..., n interact according to dx i t = 1 n n b(xt i, Xt k )dt + dw i k=1 where W 1,..., W n are independent Brownian, (X 1 0,..., X n 0 ) i.i.d. t, 7 / 49

The mean field case (McKean-Vlasov 1966) Particles i = 1,..., n interact according to dx i t = 1 n n b(xt i, Xt k )dt + dw i k=1 where W 1,..., W n are independent Brownian, (X 1 0,..., X n 0 ) i.i.d. This can be reformulated as where dx i t = B(X i t, µ n t )dt + dw i t, µ n t = 1 n B(x, m) = b(x, y) m(dy). R d Also referred to as weakly interacting diffusions. t, n δ X k t, k=1 7 / 49

The mean field case, law of large numbers Theorem (Sznitman 91, etc.) ( µ n t ) t [0,T ] converges in probability to the unique solution (µ t ) t [0,T ] of the McKean-Vlasov equation dx t = B(X t, µ t ) dt + dw t, µ t = Law(X t ). Moreover, the particles become asymptotically independent (propagation of chaos). Precisely, for fixed k, (X 1,..., X k ) µ k, as n. 8 / 49

Mean field limits for sequences of dense graphs Key questions Given a sequence of graphs G n = (V n, E n ) with V n, how can we describe the limiting behavior of... a typical or fixed particle X v (t)? the empirical distribution of particles 1 V n v V n δ Xv (t)? 9 / 49

Mean field limits for sequences of dense graphs Key questions Given a sequence of graphs G n = (V n, E n ) with V n, how can we describe the limiting behavior of... a typical or fixed particle X v (t)? the empirical distribution of particles 1 V n Theorem (Bhamidi-Budhiraja-W. 16) v V n δ Xv (t)? Suppose G n = G(n, p n ) is Erdős-Rényi, with np n. Then everything behaves like in the mean field case. See also: Delattre-Giacomin-Luçon 16, Coppini-Dietert-Giacomin 18. Theorem (Budhiraja-Mukherjee-W. 17) Suppose G n = G(n, p n ) is Erdős-Rényi, with np n. For the supermarket model (i.e. the power-of-d load balancing scheme), everything behaves like in the mean field case. 9 / 49

Beyond mean field limits Key questions Given a sequence of graphs G n = (V n, E n ) with V n, how can we describe the limiting behavior of... a typical or fixed particle X v (t)? the empirical distribution of particles 1 V n v V n δ Xv (t)? Observation: G n = G(n, p n ) is Erdős-Rényi, np n average degree, the graphs are dense. Our focus: The sparse regime, where degrees do not diverge. How does the n limit reflect the graph structure? Example: Erdős-Rényi G(n, p n ) with np n p (0, ). 10 / 49

Beyond mean field limits Key questions Given a sequence of graphs G n = (V n, E n ) with V n, how can we describe the limiting behavior of... a typical or fixed particle X v (t)? the empirical distribution of particles 1 V n v V n δ Xv (t)? Observation: G n = G(n, p n ) is Erdős-Rényi, np n average degree, the graphs are dense. Our focus: The sparse regime, where degrees do not diverge. How does the n limit reflect the graph structure? Example: Erdős-Rényi G(n, p n ) with np n p (0, ). Questions: (Q1) Does the whole system admit a scaling limit? (Q2) Is there a nice autonomous description of the limiting dynamics? Example: Detering-Fouque-Ichiba 18 treats directed cycle graph. 10 / 49

Beyond mean field limits: Key ingredients Sequence of sparse graphs G n = (V n, E n ) with V n, X v (t + 1) = F (X v (t), (X u (t)) u v, ξ v (t + 1)), t = 0, 1,..., dx v (t) = 1 b(x v (t), X u (t))dt + dw v (t), t 0, d v u v 11 / 49

Beyond mean field limits: Key ingredients Sequence of sparse graphs G n = (V n, E n ) with V n, X v (t + 1) = F (X v (t), (X u (t)) u v, ξ v (t + 1)), t = 0, 1,..., dx v (t) = 1 b(x v (t), X u (t))dt + dw v (t), t 0, d v u v (Q1) Does the whole system admit a scaling limit? (A1) Yes, using generalized notion of local (weak) convergence 11 / 49

Beyond mean field limits: Key ingredients Sequence of sparse graphs G n = (V n, E n ) with V n, X v (t + 1) = F (X v (t), (X u (t)) u v, ξ v (t + 1)), t = 0, 1,..., dx v (t) = 1 b(x v (t), X u (t))dt + dw v (t), t 0, d v u v (Q1) Does the whole system admit a scaling limit? (A1) Yes, using generalized notion of local (weak) convergence (Q2) Is there a nice autonomous description of the limiting dynamics? (A2) For regular trees or Galton-Watson trees, Yes: due to a certain space-time Markov random-field property 11 / 49

1 Networks of interacting stochastic processes 2 Local convergence 3 Conditional independence 4 Local dynamics on infinite regular trees 12 / 49

Local convergence of graphs Idea: Encode sparsity via local convergence of graphs. (a.k.a. Benjamini-Schramm convergence, see Aldous-Steele 04) 13 / 49

Local convergence of graphs Idea: Encode sparsity via local convergence of graphs. (a.k.a. Benjamini-Schramm convergence, see Aldous-Steele 04) Definition: A graph G = (V, E, ø) is assumed to be rooted, finite or countable, locally finite, and connected. 13 / 49

Local convergence of graphs Idea: Encode sparsity via local convergence of graphs. (a.k.a. Benjamini-Schramm convergence, see Aldous-Steele 04) Definition: A graph G = (V, E, ø) is assumed to be rooted, finite or countable, locally finite, and connected. Definition: Rooted graphs G n converge locally to G if: k, N s.t. B k (G) = B k (G n ) for all n N, where B k ( ) is ball of radius k at root (with respect to the graph distance), and = means isomorphism. 13 / 49

Examples of local convergence 1. Cycle graph converges to infinite line. ø ø. 14 / 49

Examples of local convergence 2. Line graph converges to infinite line. ø ø. 15 / 49

Examples of local convergence 3. Line graph rooted at end converges to semi-infinite line. ø ø 16 / 49

Examples of local convergence 4. Finite to infinite d-regular trees (A graph is d-regular if every vertex has degree d.) ø ø 17 / 49

Examples of local convergence 5. Uniformly random regular graph to infinite regular tree Fix d. Among all d-regular graphs on n vertices, select one uniformly at random. Place the root at a (uniformly) random vertex. When n, this converges (in law) to the infinite d-regular tree. (Bollobás 80) ø 18 / 49

root Examples of local convergence 6. Erdős-Rényi to Galton-Watson(Poisson) If G n = G(n, p n ) with np n p (0, ), then G n converges in law to the Galton-Watson tree with offspring distribution Poisson(p). ø ø.. 19 / 49

Examples of local convergence 7. Configuration model to unimodular Galton-Watson If G n is drawn from the configuration model on n vertices with degree distribution ρ P(N), then G n converges in law to the unimodular Galton-Watson tree UGW(ρ). Construct UGW(ρ) by letting root have ρ-many children, and each child thereafter has ρ-many children, where ρ(n) = (n + 1)ρ(n + 1) k kρ(k). Example 1: ρ = Poisson(p) = ρ = Poisson(p). Example 2: ρ = δ d = ρ = δ d 1, so UGW(δ d ) is the (deterministic) infinite d-regular tree. 20 / 49

Local convergence of marked graphs Recall: G n = (V n, E n, ø n ) converges locally to G = (V, E, ø) if k, N s.t. B k (G) = B k (G n ) for all n N. 21 / 49

Local convergence of marked graphs Recall: G n = (V n, E n, ø n ) converges locally to G = (V, E, ø) if k, N s.t. B k (G) = B k (G n ) for all n N. Definition: With G n, G as above: Given a metric space (S, d S ) and a sequence x n = (x n v ) v Gn S Gn, say that (G n, x n ) converges locally to (G, x) if k, ɛ > 0 N s.t. n N ϕ : B k (G n ) B k (G) isomorphism s.t. max v Bk (G n) d S (x n v, x ϕ(v) ) < ɛ. Lemma The set G [S] of (isomorphism classes of) (G, x) admits a Polish topology compatible with the above convergence. 21 / 49

Local convergence of marked graphs Recall: Particle system on a rooted locally finite graph G = (V, E, ø): X v (t + 1) = F (X v (t), (X u (t)) u v, ξ v (t + 1)), t = 0, 1,..., dxv G (t) = 1 b(xv G (t), Xu G (t))dt + dw v (t), t 0. d v u v Theorem If G n G locally, then (G n, X Gn ) converges in law to (G, X G ) in G [C(R + ; R d )] (or G [(R d ) ] for Markov chains). Valid for random graphs too. 22 / 49

Local convergence of marked graphs Recall: Particle system on a rooted locally finite graph G = (V, E, ø): X v (t + 1) = F (X v (t), (X u (t)) u v, ξ v (t + 1)), t = 0, 1,..., dxv G (t) = 1 b(xv G (t), Xu G (t))dt + dw v (t), t 0. d v u v Theorem If G n G locally, then (G n, X Gn ) converges in law to (G, X G ) in G [C(R + ; R d )] (or G [(R d ) ] for Markov chains). Valid for random graphs too. In particular, the root particles converge: X Gn ø n X G ø in C(R + ; R d ). 22 / 49

Local convergence of marked graphs Recall: Particle system on a rooted locally finite graph G = (V, E, ø): Theorem X v (t + 1) = F (X v (t), (X u (t)) u v, ξ v (t + 1)), t = 0, 1,..., dxv G (t) = 1 b(xv G (t), Xu G (t))dt + dw v (t), t 0. d v u v If G n G locally, then (G n, X Gn ) converges in law to (G, X G ) in G [C(R + ; R d )] (or G [(R d ) ] for Markov chains). Valid for random graphs too. Empirical measure convergence is harder. If G n G(n, p n ), np n p (0, ), then 1 Law(Xø T ), in P(C(R + ; R d )), G n δ X Gn v v G n where T GW(Poisson(p)). 23 / 49

Local convergence of marked graphs Recall: Particle system on a rooted locally finite graph G = (V, E, ø): X v (t + 1) = F (X v (t), (X u (t)) u v, ξ v (t + 1)), t = 0, 1,..., dxv G (t) = 1 b(xv G (t), Xu G (t))dt + dw v (t), t 0. d v u v (Q1): Does the whole system admit a scaling limit? Theorem (Answer) If G n G locally, then (G n, X Gn ) converges in law to (G, X G ) in G [C(R + ; R d )] (or G [(R d ) ] for Markov chains). Valid for random graphs too. 24 / 49

Local convergence of marked graphs Recall: Particle system on a rooted locally finite graph G = (V, E, ø): X v (t + 1) = F (X v (t), (X u (t)) u v, ξ v (t + 1)), t = 0, 1,..., dxv G (t) = 1 b(xv G (t), Xu G (t))dt + dw v (t), t 0. d v u v (Q1): Does the whole system admit a scaling limit? Theorem (Answer) If G n G locally, then (G n, X Gn ) converges in law to (G, X G ) in G [C(R + ; R d )] (or G [(R d ) ] for Markov chains). Valid for random graphs too. Recall (Q2) Is there a nice autonomous description of the limiting dynamics for regular trees or Galton-Watson trees? Goal: For unimodular Galton-Watson trees, find autonomous dynamics for the root neighborhood particles. 24 / 49

1 Networks of interacting stochastic processes 2 Local convergence 3 Conditional independence 4 Local dynamics on infinite regular trees 25 / 49

Markov random field Notation: For a set A of vertices in a graph G = (V, E), define Boundary: A = {u V \A : u A s.t. u v}. Definition: A family of random variables (Y v ) v V is a Markov random field (MRF) if (Y v ) v A (Y v ) v B (Y v ) v A, for all finite sets A, B V with B (A A) =. Example: A A B 26 / 49

Searching conditional independence property X v (t + 1) = F (X v (t), (X u (t)) u v, ξ v (t + 1)), t = 0, 1,..., dx v (t) = 1 b(x v (t), X u (t))dt + dw v (t), t 0. d v u v Assume the initial states (X v (0)) v V are i.i.d. Question #1: (Spatial MRF) For each time t, do the particle positions (X v (t)) v V form a Markov random field? 27 / 49

Searching conditional independence property X v (t + 1) = F (X v (t), (X u (t)) u v, ξ v (t + 1)), t = 0, 1,..., dx v (t) = 1 b(x v (t), X u (t))dt + dw v (t), t 0. d v u v Assume the initial states (X v (0)) v V are i.i.d. Question #1: (Spatial MRF) For each time t, do the particle positions (X v (t)) v V form a Markov random field? Answer #1: NO 27 / 49

Searching conditional independence property X v (t + 1) = F (X v (t), (X u (t)) u v, ξ v (t + 1)), t = 0, 1,..., dx v (t) = 1 b(x v (t), X u (t))dt + dw v (t), t 0. d v u v Assume the initial states (X v (0)) v V are i.i.d. Question #2: (Space-time MRF) For each time t, do the particle trajectories (X v [t]) v V form a Markov random field? Here x[t] = (x(s), s [0, t]). Namely for any finite A V and t > 0, is X A[t] X V \(A A)[t] X A[t]? 28 / 49

Searching conditional independence property X v (t + 1) = F (X v (t), (X u (t)) u v, ξ v (t + 1)), t = 0, 1,..., dx v (t) = 1 b(x v (t), X u (t))dt + dw v (t), t 0. d v u v Assume the initial states (X v (0)) v V are i.i.d. Question #2: (Space-time MRF) For each time t, do the particle trajectories (X v [t]) v V form a Markov random field? Here x[t] = (x(s), s [0, t]). Namely for any finite A V and t > 0, is X A[t] X V \(A A)[t] X A[t]? Answer #2: NO 28 / 49

Searching conditional independence property Counter-example: Consider three real-valued interacting Markov chains X (k) = (X i (k) : i = 1, 2, 3) on a line: X (k + 1) = BX (k) + ξ(k), X (0) = ξ(0), 1 1 0 where ξ i (k) is i.i.d. N (0, 1) and B = 1 1 1. 0 1 1 Then one can get 3 2 1 1 (X (1), X 2 (0)) N 0, 2 4 2 1 1 2 3 1 1 1 1 1 [ 5/3 1/3 Var(X 1 (1), X 3 (1) X 2 (1), X 2 (0)) = 1/3 5/3 So X 1 (1) is not independent of X 3 (1) given (X 2 (1), X 2 (0)). ]. 29 / 49

Second-order Markov random field Notation: For a set A of vertices in a graph G = (V, E), define Double-boundary: 2 A = A (A A). Definition: A family of random variables (Y v ) v V is a 2nd-order Markov random field if (Y v ) v A (Y v ) v B (Y v ) v 2 A, for all finite sets A, B V with B (A 2 A) =. Example: A 2 A B 30 / 49

Searching conditional independence property X v (t + 1) = F (X v (t), (X u (t)) u v, ξ v (t + 1)), t = 0, 1,..., dx v (t) = 1 b(x v (t), X u (t))dt + dw v (t), t 0. d v u v Assume the initial states (X v (0)) v V are i.i.d. Question #3: (Spatial second-order MRF) For each time t, do the particle positions (X v (t)) v V form a second-order Markov random field? Namely for any finite A V and t > 0, is X A(t) X V \(A 2 A)(t) X 2 A(t)? 31 / 49

Searching conditional independence property X v (t + 1) = F (X v (t), (X u (t)) u v, ξ v (t + 1)), t = 0, 1,..., dx v (t) = 1 b(x v (t), X u (t))dt + dw v (t), t 0. d v u v Assume the initial states (X v (0)) v V are i.i.d. Question #3: (Spatial second-order MRF) For each time t, do the particle positions (X v (t)) v V form a second-order Markov random field? Namely for any finite A V and t > 0, is Answer #3: NO X A(t) X V \(A 2 A)(t) X 2 A(t)? 31 / 49

Searching conditional independence property Counter-example: Consider four real-valued interacting Markov chains X (k) = (X i (k) : i = 1, 2, 3, 4) on a line: X (k + 1) = BX (k) + ξ(k), where ξ i (k) is i.i.d. N (0, 1) and B = Then one can get X (2) N 0, X (0) = ξ(0), 1 1 0 0 1 1 1 0 0 1 1 1 0 0 1 1 12 14 10 4 14 22 18 10 10 18 22 14 4 10 14 12 Var(X 1 (2), X 4 (2) X 2 (2), X 3 (2)) =., [ 14/5 6/5 6/5 14/5 So X 1 (2) is not independent of X 4 (2) given (X 2 (2), X 3 (2)). ]. 32 / 49

Searching conditional independence property X v (t + 1) = F (X v (t), (X u (t)) u v, ξ v (t + 1)), t = 0, 1,..., dx v (t) = 1 b(x v (t), X u (t))dt + dw v (t), t 0. d v u v Assume the initial states (X v (0)) v V are i.i.d. Question #4: (Space-time second order MRF) For each time t, do the particle trajectories (X v [t]) v V form a second-order Markov random field? Here x[t] = (x(s), s [0, t]). 33 / 49

Searching conditional independence property X v (t + 1) = F (X v (t), (X u (t)) u v, ξ v (t + 1)), t = 0, 1,..., dx v (t) = 1 b(x v (t), X u (t))dt + dw v (t), t 0. d v u v Assume the initial states (X v (0)) v V are i.i.d. Question #4: (Space-time second order MRF) For each time t, do the particle trajectories (X v [t]) v V form a second-order Markov random field? Here x[t] = (x(s), s [0, t]). Answer #4: YES Theorem: (Lacker, Ramanan, W. 18) X A[t] X V \(A 2 A)[t] X 2 A[t] for all finite A V and t > 0. In fact, suffices for (X v (0)) v V to form a second-order MRF. 33 / 49

Searching conditional independence property Intuition: A 2 A B 34 / 49

Searching conditional independence property Intuition: A 2 A B Proof idea: Markov chain: Use induction and properties of conditional independence. 34 / 49

Searching conditional independence property Intuition: A 2 A B Proof idea: Markov chain: Use induction and properties of conditional independence. Diffusion: Use Girsanov to identify the density of (X v [t]) v V w.r.t. Wiener measure, and study how it factorizes. Use Hammersley-Clifford theorem. 34 / 49

1 Networks of interacting stochastic processes 2 Local convergence 3 Conditional independence 4 Local dynamics on infinite regular trees 35 / 49

Infinite line graph: Markov chain... 3 2 1 0 1 2 3... Particle system on infinite line graph, i Z: ( ) X i (t + 1) = F X i (t), X i 1 (t), X i+1 (t), ξ i (t + 1) Assume: F is symmetric in neighbors: F (x, y, z, ξ) = F (x, z, y, ξ). Goal: Find an autonomous stochastic process (Y 1, Y 0, Y 1 ) which agrees in law with (X 1, X 0, X 1 ). 36 / 49

Local dynamics for line graph: Markov chain... 3 2 1 0 1 2 3... Start with (Y 1, Y 0, Y 1 )(0) = (X 1, X 0, X 1 )(0). 37 / 49

Local dynamics for line graph: Markov chain... 3 2 1 0 1 2 3... Start with (Y 1, Y 0, Y 1 )(0) = (X 1, X 0, X 1 )(0). Given (Y 1, Y 0, Y 1 )[t] at time t, let ( ) γ t ( y 0 [t], y 1 [t]) = Law Y 1 (t) Y 0 [t] = y 0 [t], Y 1 [t] = y 1 [t]. 37 / 49

Local dynamics for line graph: Markov chain... 3 2 1 0 1 2 3... Start with (Y 1, Y 0, Y 1 )(0) = (X 1, X 0, X 1 )(0). Given (Y 1, Y 0, Y 1 )[t] at time t, let ( ) γ t ( y 0 [t], y 1 [t]) = Law Y 1 (t) Y 0 [t] = y 0 [t], Y 1 [t] = y 1 [t]. Independently sample ghost particles Y 2 (t) and Y 2 (t) so that ( ) Law Y 2 (t) Y 1 [t], Y 0 [t], Y 1 [t] = γ t ( Y 1 [t], Y 0 [t]) Here the conditional independence is used. 37 / 49

Local dynamics for line graph: Markov chain... 3 2 1 0 1 2 3... Start with (Y 1, Y 0, Y 1 )(0) = (X 1, X 0, X 1 )(0). Given (Y 1, Y 0, Y 1 )[t] at time t, let ( ) γ t ( y 0 [t], y 1 [t]) = Law Y 1 (t) Y 0 [t] = y 0 [t], Y 1 [t] = y 1 [t]. Independently sample ghost particles Y 2 (t) and Y 2 (t) so that ( ) Law Y 2 (t) Y 1 [t], Y 0 [t], Y 1 [t] = γ t ( Y 1 [t], Y 0 [t]) Here the conditional independence is used. Sample new noises (ξ 1, ξ 0, ξ 1 )(t + 1) independently, and update: ( ) Y i (t + 1) = F Y i (t), Y i 1 (t), Y i+1 (t), ξ i (t + 1), i = 1, 0, 1 37 / 49

Infinite d-regular trees: Markov chain Autonomous dynamics for root particle and its neighbors, X ρ (t), (X v (t)) v ρ, involving conditional law of d 1 children given root and one other child u: u ρ Law((X v ) v ρ, v u X ρ, X u ) d = 3 38 / 49

Infinite d-regular trees: Markov chain Autonomous dynamics for root particle and its neighbors, u X ρ (t), (X v (t)) v ρ, involving conditional law of d 1 children given root and one other child u: ρ Law((X v ) v ρ, v u X ρ, X u ) d = 4 39 / 49

Infinite d-regular trees: Markov chain Autonomous dynamics for root particle and its neighbors, X ρ (t), (X v (t)) v ρ, involving conditional law of d 1 children given root and one other child u: Law((X v ) v ρ, v u X ρ, X u ) u ρ d = 5 40 / 49

Unimodular Galton-Watson trees: Markov chain Autonomous dynamics for root & first generation involving conditional law of 1st-generation given root and one child. ø root Condition on tree structure as well!.. 41 / 49

Example: Linear Gaussian dynamics State space R, noises ξ v (t) are independent standard Gaussian. X v (t + 1) = ax v (t) + b u v X u (t) + c + ξ v (t + 1) X v (0) = ξ v (0), a, b, c R conditional laws are all Gaussian Proposition Suppose the graph G is an infinite d-regular tree, d > 2. Simulating local dynamics for one particle up to time t is O(t 2 d 2 ). Compare: Simulation using infinite tree is O((d 1) t+1 ). 42 / 49

Example: Contact process Each particle is either 1 or 0. Parameters p, q [0, 1]. Transition rule: At time t, if particle v... is at state X v (t) = 1, it switches to X v (t + 1) = 0 w.p. q, is at state X v (t) = 0, it switches to X v (t + 1) = 1 w.p. p X u (t), d v where d v = degree of vertex v. How well do local approximation and mean field approximation do? u v 43 / 49

Example: Contact process Figure: Infinite 2-regular tree (line), p = 2/3, q = 0.1 Credit: Ankan Ganguly & Mitchell Wortsman, Brown University 44 / 49

Local dynamics for diffusions An additional ingredient: A projection/mimicking lemma (Brunick-Shreve 13; Gyongy 86) On some (Ω, F, F, P), let W be F-Brownian. Let (b(t)) be F-adapted and square-integrable, with Define (by optional projection) Then there is a weak solution Y of such that Y d = X. dx (t) = b(t)dt + dw (t). B(t, x) = E[b(t) X [t] = x[t]]. dy (t) = B(t, Y )dt + d W (t) 45 / 49

Local dynamics for line graph: diffusions... 3 2 1 0 1 2 3... Particle system on infinite line graph, i Z: dxt i = 1 ( 2 b(x i t, Xt i 1 ) + b(xt i, Xt i+1 ) ) dt + dwt i 46 / 49

Local dynamics for line graph: diffusions... 3 2 1 0 1 2 3... Particle system on infinite line graph, i Z: dxt i = 1 ( 2 b(x i t, Xt i 1 ) + b(xt i, Xt i+1 ) ) dt + dwt i Local dynamics: dy 1 t = 1 2 dy 0 t = 1 2 dy 1 t = 1 2 ( b(y 1 t, Yt 0 ) + γ t (Y 1, Y 0 ), b(yt 1, ) ) dt + dwt 1 ( b(y 0 t, Yt 1 ) + b(yt 0, Yt 1 ) ) dt + dwt 0 ( γt (Y 1, Y 0 ), b(yt 1, ) + b(yt 1, Yt 0 ) ) dt + dwt 1 t = y t 1 ) γ t (y 0, y 1 ) = Law ( Y 1 t Y 0 t = y 0 t, Y 1 Thm: Uniqueness in law & (Y 1, Y 0, Y 1 ) d = (X 1, X 0, X 1 ). 46 / 49

Local dynamics for line graph: diffusions... 3 2 1 0 1 2 3... Particle system on infinite line graph, i Z: dxt i = 1 ( 2 b(x i t, Xt i 1 ) + b(xt i, Xt i+1 ) ) dt + dwt i Local dynamics: dy 1 t = 1 2 dy 0 t = 1 2 dy 1 t = 1 2 ( b(y 1 t, Yt 0 ) + γ t (Y 1, Y 0 ), b(yt 1, ) ) dt + dwt 1 ( b(y 0 t, Yt 1 ) + b(yt 0, Yt 1 ) ) dt + dwt 0 ( γt (Y 1, Y 0 ), b(yt 1, ) + b(yt 1, Yt 0 ) ) dt + dwt 1 t = y t 1 ) γ t (y 0, y 1 ) = Law ( Y 1 t Y 0 t = y 0 t, Y 1 Thm: Uniqueness in law & (Y 1, Y 0, Y 1 ) d = (X 1, X 0, X 1 ). Analogous local dynamics hold for infinite d-regular trees and unimodular Galton-Watson trees 46 / 49

Summary Theorem 1 If finite graph sequence converges locally to infinite graph, then particle systems converge locally as well. Theorem 2 Root neighborhood particles in a unimodular Galton-Watson tree admit well-posed local dynamics. Corollary: If finite graph sequence converges locally to a unimodular Galton-Watson tree, then root neighborhood particles converge to unique solution of local dynamics. 47 / 49

Thank you! 48 / 49

References [1] S. Bhamidi, A. Budhiraja, and R. Wu. Weakly interacting particle systems on inhomogeneous random graphs. To appear in Stochastic Processes and their Applications, arxiv:1612.00801. [2] A. Budhiraja, D. Mukherjee and R. Wu. Supermarket model on graphs. Accepted by The Annals of Applied Probability, arxiv:1712.07607. [3] S. Delattre, G. Giacomin, and E. Luçon. A note on dynamical models on random graphs and Fokker Planck equations. Journal of Statistical Physics, 165(4):785 798, 2016. [4] D. Lacker, K. Ramanan, and R. Wu. Local dynamics for large sparse networks of interacting Markov chains / diffusions. Near completion, 2018. [5] A-S. Sznitman. Topics in propagation of chaos. Ecole d Eté de Probabilités de Saint-Flour XIX 1989. Lecture Notes in Mathematics, vol 1464, Springer, Berlin, Heidelberg, 1991. 49 / 49