process on the hierarchical group

Similar documents
Intertwining of Markov processes

INTERTWINING OF BIRTH-AND-DEATH PROCESSES

Lecture 5. If we interpret the index n 0 as time, then a Markov chain simply requires that the future depends only on the present and not on the past.

Mean-field dual of cooperative reproduction

On hitting times and fastest strong stationary times for skip-free and more general chains

The passage time distribution for a birth-and-death chain: Strong stationary duality gives a first stochastic proof

Irregular Birth-Death process: stationarity and quasi-stationarity

1 Continuous-time chains, finite state space

Positive and null recurrent-branching Process

Necessary and sufficient conditions for strong R-positivity

Branching Processes II: Convergence of critical branching to Feller s CSB

On Markov intertwinings

SMSTC (2007/08) Probability.

Weighted Sums of Orthogonal Polynomials Related to Birth-Death Processes with Killing

Lecture 21. David Aldous. 16 October David Aldous Lecture 21

Linear-fractional branching processes with countably many types

The range of tree-indexed random walk

The Moran Process as a Markov Chain on Leaf-labeled Trees

Erik A. van Doorn Department of Applied Mathematics University of Twente Enschede, The Netherlands

Introduction to self-similar growth-fragmentations

Other properties of M M 1

Theory of Stochastic Processes 3. Generating functions and their applications

A mathematical framework for Exact Milestoning

A MODEL FOR THE LONG-TERM OPTIMAL CAPACITY LEVEL OF AN INVESTMENT PROJECT

Modern Discrete Probability Branching processes

Chapter 7. Markov chain background. 7.1 Finite state space

Matrix analytic methods. Lecture 1: Structured Markov chains and their stationary distribution

UPPER DEVIATIONS FOR SPLIT TIMES OF BRANCHING PROCESSES


Stochastic domination in space-time for the supercritical contact process

Stochastic modelling of epidemic spread

P(X 0 = j 0,... X nk = j k )

TOTAL VARIATION CUTOFF IN BIRTH-AND-DEATH CHAINS

A review of Continuous Time MC STA 624, Spring 2015

Markov Chains CK eqns Classes Hitting times Rec./trans. Strong Markov Stat. distr. Reversibility * Markov Chains

Let (Ω, F) be a measureable space. A filtration in discrete time is a sequence of. F s F t

Path Decomposition of Markov Processes. Götz Kersting. University of Frankfurt/Main

Martingale Problems. Abhay G. Bhatt Theoretical Statistics and Mathematics Unit Indian Statistical Institute, Delhi

HEAGAN & CO., OPP. f>, L. & W. DEPOT, DOYER, N. J, OUR MOTTO! ould Iwv ia immediate vltlui. VEEY BEST NEW Creamery Butter 22c ib,

9.2 Branching random walk and branching Brownian motions

88 CONTINUOUS MARKOV CHAINS

C.7. Numerical series. Pag. 147 Proof of the converging criteria for series. Theorem 5.29 (Comparison test) Let a k and b k be positive-term series

Some Semi-Markov Processes

The Λ-Fleming-Viot process and a connection with Wright-Fisher diffusion. Bob Griffiths University of Oxford

arxiv: v1 [math.pr] 29 Jul 2014

arxiv: v2 [math.pr] 4 Sep 2017

Detailed Proofs of Lemmas, Theorems, and Corollaries

The SIS and SIR stochastic epidemic models revisited

The coupling method - Simons Counting Complexity Bootcamp, 2016

On times to quasi-stationarity for birth and death processes

Brownian Motion. 1 Definition Brownian Motion Wiener measure... 3

Convergence exponentielle uniforme vers la distribution quasi-stationnaire en dynamique des populations

Mixing Times and Hitting Times

Lecture 17 Brownian motion as a Markov process

Ernesto Mordecki 1. Lecture III. PASI - Guanajuato - June 2010

arxiv: v2 [math.pr] 25 May 2017

18.175: Lecture 30 Markov chains

ON COMPOUND POISSON POPULATION MODELS

Macroscopic quasi-stationary distribution and microscopic particle systems

Lecture 12. F o s, (1.1) F t := s>t

MATH 564/STAT 555 Applied Stochastic Processes Homework 2, September 18, 2015 Due September 30, 2015

Dynamics of the evolving Bolthausen-Sznitman coalescent. by Jason Schweinsberg University of California at San Diego.

On times to quasi-stationarity for birth and death processes

April 20th, Advanced Topics in Machine Learning California Institute of Technology. Markov Chain Monte Carlo for Machine Learning

The Minesweeper game: Percolation and Complexity

6.231 DYNAMIC PROGRAMMING LECTURE 7 LECTURE OUTLINE

Latent voter model on random regular graphs

{σ x >t}p x. (σ x >t)=e at.

Recursive tree processes and the mean-field limit of stochastic flows

Information Theory and Statistics Lecture 3: Stationary ergodic processes

Random Times and Their Properties

Kinetically constrained particle systems on a lattice

PROBABILITY: LIMIT THEOREMS II, SPRING HOMEWORK PROBLEMS

Math Homework 5 Solutions

Birth-death chain models (countable state)

arxiv: v1 [math.pr] 1 Jan 2013

Lectures on Stochastic Stability. Sergey FOSS. Heriot-Watt University. Lecture 4. Coupling and Harris Processes

Ergodic Theorems. Samy Tindel. Purdue University. Probability Theory 2 - MA 539. Taken from Probability: Theory and examples by R.

, some directions, namely, the directions of the 1. and

EXTINCTION TIMES FOR A GENERAL BIRTH, DEATH AND CATASTROPHE PROCESS

14 Branching processes

Walsh Diffusions. Andrey Sarantsev. March 27, University of California, Santa Barbara. Andrey Sarantsev University of Washington, Seattle 1 / 1

Lecture 7. µ(x)f(x). When µ is a probability measure, we say µ is a stationary distribution.

Asymptotic properties of the maximum likelihood estimator for a ballistic random walk in a random environment

Applied Stochastic Processes

The Contour Process of Crump-Mode-Jagers Branching Processes

Look-down model and Λ-Wright-Fisher SDE

Doléans measures. Appendix C. C.1 Introduction

Minicourse on: Markov Chain Monte Carlo: Simulation Techniques in Statistics

RANDOM WALKS AND PERCOLATION IN A HIERARCHICAL LATTICE

arxiv: v1 [math.pr] 13 Nov 2018

Regularity of the density for the stochastic heat equation

Theory and applications of random Poincaré maps

An inverse source problem in optical molecular imaging

Characterization of cutoff for reversible Markov chains

Analyticity and Derivatives of Entropy Rate for HMP s

Erdős-Renyi random graphs basics

Modern Discrete Probability Spectral Techniques

EXTINCTION TIMES FOR A GENERAL BIRTH, DEATH AND CATASTROPHE PROCESS

STABILIZATION OF AN OVERLOADED QUEUEING NETWORK USING MEASUREMENT-BASED ADMISSION CONTROL

Transcription:

Intertwining of Markov processes and the contact process on the hierarchical group April 27, 2010

Outline Intertwining of Markov processes

Outline Intertwining of Markov processes First passage times of birth and death processes

Outline Intertwining of Markov processes First passage times of birth and death processes The contact process on the hierarchical group

A bit of filtering theory Let (X, Y ) = (X n, Y n ) n 0 be a Markov chain with finite state space of the form S T, which jumps from a point (x, y) to a point (x, y ) with transition probability P(x, y; x, y ). Set P(y x 0,..., x n ) := P[Y n = y (X 0,..., X n ) = (x 0,..., x n )]. The so-called filtering equations read: P(y x 0,..., x n ) y = S P(y x 0,..., x n 1 )P(x n 1, y ; x n, y) y,y S P(y x 0,..., x n 1 )P(x n 1, y ; x n, y ).

Markov functionals Proposition Let K be a probability kernel from S to T. Assume that there exists a transition probability Q on S such that Q(x, x )K(x, y ) = y K(x, y)p(x, y; x, y ) Then implies that P[Y 0 = y X 0 ] = K(X 0, y) a.s. P[Y n = y (X 0,..., X n )] = K(X n, y) a.s. (n 0) and X, on its own, is a Markov chain with transition kernel Q.

Continuous time [Rogers & Pitman 81] Let K be a probability kernel from S to T and define K : R S T R S by Kf (x) := y S K(x, y)f (x, y). Let G and Ĝ be Markov generators on S resp. S T. Assume that GK = KĜ. Then the Markov process (X, Y ) with generator Ĝ started in an initial law such that satisfies P[Y 0 = y X 0 ] = K(X 0, y) a.s. P[Y t = y (X s ) 0 s t ] = K(X t, y) a.s. (t 0). and X, on its own, is a Markov process with generator G.

1 0.5 0-0.5 Intertwining of Markov processes -1 0 0.2 0.4 0.6 0.8 1 b 8 6 4 2 0-2 -4-6 -8 0 0.2 0.4 0.6 0.8 1 Example: Wright-Fisher diffusion Let (X, Y ) be the Markov process in [0, 1] {0, 1} with generator Ĝf (x, 0) = 1 2 2x(1 x) f (x, 0) + b x 2 0 (x) x f (x, 0) + ( f (x, 1) f (x, 0) ), Ĝf (x, 1) = 1 2 2x(1 x) f (x, 1) + b x 2 1 (x) x f (x, 1), where b 0 (x) = 2( 1 2 x), b 1(x) = 8x(1 x)(x 1 2 ). 1 4x(1 x) b 0 1 Note The process Y is autonomous and jumps 0 1 with rate one independent of the state of X, but the evolution of X depends on the state of Y.

Example: Wright-Fisher diffusion 1 0.8 0.6 0.4 0.2 As long as Y t = 0, X cannot reach the boundary. When Y t = 1, X cannot cross the middle. 0 0 0.1 0.2 0.3 0.4 0.5

1 0.8 0.6 0.4 0.2 0 0.2 0.4 0.6 0.8 1 Example: Wright-Fisher diffusion Define a kernel K from [0, 1] to {0, 1} by K(x, 0) := 4x(1 x), K(x, 1) := 1 4x(1 x) 0 Then (X 0, Y 0 ) = ( 1 2, 0) implies P[Y t = y (X s ) 0 s t ] = K(X t, y) a.s. (t 0). and X, on its own, is a Markov process with generator Gf (x) = 1 2 2x(1 x) f (x). x 2

Example: Wright-Fisher diffusion 1 0.8 0.6 0.4 0.2 0 0 0.1 0.2 0.3 0.4 0.5

Example: Wright-Fisher diffusion Generalization It is possible to couple a Wright-Fisher diffusion X started in X 0 = 1 2 to a process Y with state space {0, 1,..., }, started in Y 0 = 0 such that Y jumps k k + 1 with rate (k + 1)(2k + 1), independent of the state of X, and: P[Y t = k (X s ) 0 s t ] = (1 X 2 t )X 2k t, and inf { t 0 : X t {0, 1} } = inf{t 0 : Y t = }.

Adding structure to Markov processes Let X be a Markov processes with state space S and generator G, let K be a probability kernel from S to T and let (G x) x S be a collection of generators of T -valued Markov processes. Assume that GK = K G where Kf (x) := y K(x, y)f (x, y) and Gf (x, y) := G xf (y). Then X can be coupled to a process Y such that (X, Y ) is Markov, Y evolves according to the generator G x while X is in the state x, and P[Y t = y (X s ) 0 s t ] = K(X t, y) a.s. (t 0). We call Y an added-on process on X.

Intertwining of semigroups Proposition Let X and Y be Markov processes with state spaces S and T, semigroups (P t ) t 0 and (P t) t 0, and generators G and G. Let K be a probability kernel from S to T and define Kf (x) := y K(x, y)f (y). Then GK = KG implies the intertwining relation P t K = KP t (t 0) and the processes X and Y can be coupled such that P[Y t = y (X s ) 0 s t ] = K(X t, y) a.s. (t 0). We call Y an averaged Markov process on X.

First passage times of birth and death processes [Karlin & McGregor 59] Let Z be a Markov process with state space {0, 1, 2,...}, started in Z 0 = 0, that jumps k 1 k with rate b k > 0 and k k 1 with rate d k > 0 (k 1). Then τ N := inf{t 0 : Z t = N} is distributed as a sum of independent exponentially distributed random variables whose parameters λ 1 < < λ N are the negatives of the eigenvalues of the generator of the process stopped in N.

A probabilistic proof [Diaconis & Miclos 09] Let Y be a Markov process with state space {0, 1,..., N} that jumps k 1 k with rate λ N k+1. Then Y can be coupled to Z such that P[Z t τn = z (Y s ) 0 s t ] = K(Y t, z) a.s., where K is a probability kernel on {0,..., N} satisfying K(N, N) = 1 and K(y, {0,..., y}) = 1 (0 y < N). Note: here Z t τn is an averaged Markov process on Y.

A probabilistic proof Proof By induction. Assume that Z is a process on {0,..., N} such that b 1,..., b N > 0, d 1,..., d M 1 > 0, d M = = d N = 0 for some 1 M N. Then Diaconis and Miclos construct a kernel K such that K(y, {0,..., y}) = 1 (0 y < M), K(y, y) = 1 (M y N) and a proces Z with b 1,..., b N > 0, d 1,..., d M 2 > 0, d M 1 = = d N = 0 such that P[Z t = z (Z s) 0 s t ] = K(Z t, z) a.s.

A probabilistic proof Z b 1 b 2 b N 0 1 2 N d 1 d 2 λ 1 K Z b 1 b 2 Z d 1 b 1 d 2 λ 2 λ 1 K d 1 λ 3 λ 2 λ 1 K Z

A question Question Diaconis and Miclos construct Z t τn as an averaged Markov process on a process Y t with zero death rates and with birth rates λ N,..., λ 1, in such a way that Z t τn Y t. Let V t be the process with zero death rates and with birth rates λ 1,..., λ N (in this order!). Is it possible to contruct V t as an averaged Markov process on Z t τn (in this order!) in such a way that V t Z t τn?

Strong stationary times Def A strong stationary time of a Markov process X is a randomized stopping time τ such that X τ is independent of τ and X τ is distributed according to its stationary law. Claim Let Z be a Markov process with state space {0, 1,..., N}, started in Z 0 = 0, that jumps k 1 k with rate b k > 0 and k k 1 with rate d k > 0 (1 k N). Then there exists a strong stationary time for Z and the fastest such time is distributed as a sum of independent exponentially distributed random variables whose parameters λ 1 < < λ N are the negatives of the eigenvalues of the generator of Z.

Strong stationary times Z b 1 b 2 b N 0 1 2 N d 1 d 2 λ 1 K Z b 1 b 2 Z d 1 b 1 d 2 λ 2 λ 1 K d 1 λ 3 λ 2 λ 1 K Z

The hierarchical group By definition, the hierarchical group with freedom N is the set Ω N := { i = (i 0, i 1,...) : i k {0,..., N 1}, i k 0 for finitely many k }, equipped with componentwise addition modulo N. Think of sites i Ω N as the leaves of an infinite tree. Then i 0, i 1, i 2,... are the labels of the branches on the unique path from i to the root 0 1 2 0 1 2 0 1 2 0 1 2 0 of the tree. 1 0 2 1 0 2 1 0 2 1 0 2 1 0 2 1 0 2 1 0 2 1 0 2 1 0 2 site (2,0,1,0,...)

The hierarchical distance Set i := inf{k 0 : i m = 0 m k} (i Ω N ). Then i j is the hierarchical distance between two elements i, j Ω N. In the tree picture, i j measures how high we must go up the tree to find the last common ancestor of i and j. i j two sites at hierarchical distance 3

Fix a recovery rate δ 0 and infection rates α k 0 such that k=1 α k <. The contact process on Ω N with these rates is the {0, 1} Ω N -valued Markov process (X t ) t 0 with the following description: If X t (i) = 0 (resp. X t (i) = 1), then we say that the site i Ω N is healthy (resp. infected) at time t 0. An infected site i infects a healthy site j at hierarchical distance k := i j with rate α k N k, and infected sites become healthy with rate δ 0.

X t 0 0 1 0 1 1 0 0 0 0 0 1 0 1 1 0 1 0 0 0 0 0 0 0 0 1 0 α 2 N 2 α 3 N 3 Infection rates on the hierarchical group.

The critical recovery rate We say that a contact process (X t ) t 0 on Ω N with given recovery and infection rates survives if there is a positive probability that the process started with only one infected site never recovers completely, i.e., there are infected sites at any t 0. For given infection rates, we let δ c := sup { δ 0 : the contact process with infection rates (α k ) k 1 and recovery rate δ survives } denote the critical recovery rate. A simple monotone coupling argument shows that X survives for δ < δ c and dies out for δ > δ c. It is not hard to show that δ c <. The question whether δ c > 0 is more subtle.

(Non)triviality of the critical recovery rate [Athreya & S. 10] Assume that α k = e θk (a) If N < θ, then δ c = 0. (b) If 1 < θ < N, then δ c > 0. (k 1). Then: More generally, we show that δ c = 0 if lim inf k N k log(β k ) =, where β k := while δ c > 0 if (N ) k log(α k ) >, k=m for some m 1 and N < N. α n (k 1), n=k

Proof of survival We use added-on Markov processes to inductively derive bounds on the finite-time survival probability of finite systems. Let Ω n 2 := { i = (i 0,..., i n 1 ) : i k {0, 1} } and let S n := {0, 1} Ωn 2. We define a kernel from Sn to S n 1 by independently replacing blocks consisting of two spins by a single spin according to the stochastic rules: and 00 0, 11 1, { 0 with probability ξ, 01 or 10 1 with probability 1 ξ, where ξ (0, 1 2 ] is a constant, to be determined later.

Renormalization kernel K 0 0 1 0 0 1 1 1 0 1 0 1 The probability of this transition is 1 (1 ξ) ξ 1.

An added-on process Let X be a contact process on Ω n 2 with infection rates α 1,..., α n and recover rate δ. Then X can be coupled to a process Y such that P[Y t = y (X s ) 0 s t ] = K(X t, y) a.s. (t 0), where K is the kernel defined before, and ξ := γ γ 2 1 2 with γ := 1 ( 4 3 + α 1 2δ Moreover, the process Y can be coupled to a finite contact process Y on Ω n 1 2 with recovery rate δ := 2ξδ and infection rates α 1,..., α n 1 given by α k := 1 2 α k+1, in such a way that Y t Y t for all t 0. ).

Renormalization We may view the map (δ, α 1,..., α n ) (δ, α 1,..., α n 1 ) as an (approximate) renormalization transformation. By iterating this map n times, we get a sequence of recovery rates δ, δ, δ,..., the last of which gives a upper bound on the spectral gap of the finite contact process X on Ω n 2. Under suitable assumptions on the α k s, we can show that this spectral gap tends to zero as n, and in fact, we can derive explicit lower bounds on the probability that finite systems survive till some fixed time t.

A question Question Can we find an exact renormalization map (δ, α 1,..., α n ) (δ, α 1,..., α n 1 ) for hierarchical contact processes, i.e., for each hierarchical contact process X on Ω n 2, can we find an averaged Markov process Y that is itself a hierarchical contact process on Ω2 n 1?