G(t) := i. G(t) = 1 + e λut (1) u=2

Similar documents
Mixing Times and Hitting Times

The Moran Process as a Markov Chain on Leaf-labeled Trees

The cutoff phenomenon for random walk on random directed graphs

Statistical Inference on Large Contingency Tables: Convergence, Testability, Stability. COMPSTAT 2010 Paris, August 23, 2010

18.175: Lecture 30 Markov chains

Interfaces between Probability and Geometry

2. The averaging process

Oberwolfach workshop on Combinatorics and Probability

3. The Voter Model. David Aldous. June 20, 2012

APPROXIMATE WEAK AMENABILITY OF ABSTRACT SEGAL ALGEBRAS

UNIFORM EMBEDDINGS OF BOUNDED GEOMETRY SPACES INTO REFLEXIVE BANACH SPACE

Infinitely iterated Brownian motion

Finite Markov Information-Exchange processes

Markov Chains. X(t) is a Markov Process if, for arbitrary times t 1 < t 2 <... < t k < t k+1. If X(t) is discrete-valued. If X(t) is continuous-valued

CUTOFF FOR THE STAR TRANSPOSITION RANDOM WALK

Conductance and Rapidly Mixing Markov Chains

Ergodic Theorems. Samy Tindel. Purdue University. Probability Theory 2 - MA 539. Taken from Probability: Theory and examples by R.

The secondary Novikov-Shubin invariants of groups and quasi-isometry

Characterization of cutoff for reversible Markov chains

Stochastic processes. MAS275 Probability Modelling. Introduction and Markov chains. Continuous time. Markov property

XVI International Congress on Mathematical Physics

BOOK REVIEW PERSI DIACONIS

Eventual Positivity of Operator Semigroups

Lecture 4: Graph Limits and Graphons

The coupling method - Simons Counting Complexity Bootcamp, 2016

Limit Theorems for Exchangeable Random Variables via Martingales

Szemerédi s regularity lemma revisited. Lewis Memorial Lecture March 14, Terence Tao (UCLA)

Exchangeability. Peter Orbanz. Columbia University

Lecture 21. David Aldous. 16 October David Aldous Lecture 21

Exchangeable random arrays

Characterization of cutoff for reversible Markov chains

Lecture 2: Exchangeable networks and the Aldous-Hoover representation theorem

Logarithmic scaling of planar random walk s local times

Random walk on groups

Bayesian Nonparametrics

Graphs and matrices with maximal energy

Continuity of convolution and SIN groups

INTRODUCTION TO MARKOV CHAIN MONTE CARLO

INTRODUCTION TO MARKOV CHAINS AND MARKOV CHAIN MIXING

On Markov intertwinings

ARCC WORKSHOP: SHARP THRESHOLDS FOR MIXING TIMES SCRIBES: SHARAD GOEL, ASAF NACHMIAS, PETER RALPH AND DEVAVRAT SHAH

Local limits of random graphs

21 Gaussian spaces and processes

LIMITS FOR QUEUES AS THE WAITING ROOM GROWS. Bell Communications Research AT&T Bell Laboratories Red Bank, NJ Murray Hill, NJ 07974

Resistance Growth of Branching Random Networks

TOTAL VARIATION CUTOFF IN BIRTH-AND-DEATH CHAINS

Injective semigroup-algebras

SHADOWING AND INVERSE SHADOWING IN SET-VALUED DYNAMICAL SYSTEMS. HYPERBOLIC CASE. Sergei Yu. Pilyugin Janosch Rieger. 1.

arxiv: v3 [math.pr] 18 Aug 2017

The Distribution of Mixing Times in Markov Chains

Homework # , Spring Due 14 May Convergence of the empirical CDF, uniform samples

Weak and strong moments of l r -norms of log-concave vectors

CONVERGENCE OF KAC S RANDOM WALK

ON THE SIMILARITY OF CENTERED OPERATORS TO CONTRACTIONS. Srdjan Petrović

A Concise Course on Stochastic Partial Differential Equations

arxiv:math/ v4 [math.oa] 28 Dec 2005

4: The Pandemic process

arxiv:math/ v1 [math.fa] 26 Oct 1993

The passage time distribution for a birth-and-death chain: Strong stationary duality gives a first stochastic proof


curvature, mixing, and entropic interpolation Simons Feb-2016 and CSE 599s Lecture 13

A note on adiabatic theorem for Markov chains

In terms of measures: Exercise 1. Existence of a Gaussian process: Theorem 2. Remark 3.

Stochastic process. X, a series of random variables indexed by t

On a perturbation method for determining group of invariance of hierarchical models

Szemerédi s Lemma for the Analyst

An ergodic theorem for partially exchangeable random partitions

Stepping-Stone Model with Circular Brownian Migration

NIL, NILPOTENT AND PI-ALGEBRAS

EXCHANGEABLE MARKOV PROCESSES ON GRAPHS: FELLER CASE

REVERSALS ON SFT S. 1. Introduction and preliminaries

Mixing time for a random walk on a ring

TROPICAL BRILL-NOETHER THEORY

SPACES ENDOWED WITH A GRAPH AND APPLICATIONS. Mina Dinarvand. 1. Introduction

Minicourse on: Markov Chain Monte Carlo: Simulation Techniques in Statistics

A lecture on the averaging process

ON COMPLEMENTED SUBSPACES OF SUMS AND PRODUCTS OF BANACH SPACES

OPTIMAL SOLUTIONS TO STOCHASTIC DIFFERENTIAL INCLUSIONS

Lecture 5. If we interpret the index n 0 as time, then a Markov chain simply requires that the future depends only on the present and not on the past.

Linear Algebra Short Course Lecture 2

ON THE REGULARITY OF SAMPLE PATHS OF SUB-ELLIPTIC DIFFUSIONS ON MANIFOLDS

Lect4: Exact Sampling Techniques and MCMC Convergence Analysis

ENTIRE FUNCTIONS AND COMPLETENESS PROBLEMS. Lecture 3

An example of a convex body without symmetric projections.

CUT-OFF FOR QUANTUM RANDOM WALKS. 1. Convergence of quantum random walks

Acta Mathematica Academiae Paedagogicae Nyíregyháziensis 24 (2008), ISSN

Some highlights from the theory of multivariate symmetries

Entropy and Ergodic Theory Lecture 15: A first look at concentration

Extremal Solutions of Differential Inclusions via Baire Category: a Dual Approach

Note that in the example in Lecture 1, the state Home is recurrent (and even absorbing), but all other states are transient. f ii (n) f ii = n=1 < +

SIMILAR MARKOV CHAINS

for all f satisfying E[ f(x) ] <.

08a. Operators on Hilbert spaces. 1. Boundedness, continuity, operator norms

Lecture 4. David Aldous. 2 September David Aldous Lecture 4

A NOTE ON THE POINCARÉ AND CHEEGER INEQUALITIES FOR SIMPLE RANDOM WALK ON A CONNECTED GRAPH. John Pike

Stochastic relations of random variables and processes

Finite Markov Information-Exchange processes. 1. Overview

MA651 Topology. Lecture 9. Compactness 2.

CUTOFF FOR THE ISING MODEL ON THE LATTICE

Results from MathSciNet: Mathematical Reviews on the Web c Copyright American Mathematical Society 2000

Transcription:

Note: a conjectured compactification of some finite reversible MCs There are two established theories which concern different aspects of the behavior of finite state Markov chains as the size of the state space increases. One is weak convergence. Here prototype results concern convergence of birth-and-death processes to one-dimensional diffusions, and the d-dimensional analogs [9]. More generally, for any continuous time and space Markov process which seems mathematically natural, one could seek necessary and sufficient conditions for a sequence of discrete chains to converge to that limit. The second theory concerns mixing times [7, 1]. At one level this concerns definitions of various mixing times and estimation of their orders of magnitude in general examples. At another level, for more specific examples there has been extensive study of the cutoff phenomenon for distance to stationarity, mostly in the context of variation distance [4] but also for L 2 distance [3] which will be more relevant to this note. This note describes a conjecture which says, roughly, that these are the only two possibilities, at least for reversible chains. In a little more detail, the conjecture is Given a sequence of n-state reversible chains which does not have the L 2 cutoff property, there is a subsequence in which, after relabeling states, the transition densities converge to those of some limit general-state-space reversible Markov process. We emphasize that the n-state chains are arbitrary in the sense that we do not assume any connection between the chains as n varies. The conjectured behavior is a compactness assertion, in the spirit of recent work on dense graph limits [8, 5]. The rest of this note states the conjecture more precisely and outlines how one might try to start a proof. For simplicity we work with uniform stationary distributions, but we anticipate that the general reversible case will be similar. 1

0.1 Setup Consider an n-state irreducible continuous-time Markov chain with symmetric transition rate matrix in other words, reversible with uniform stationary distribution. Write p(i, j; t) for the transition probabilities P(X(t) = j X(0) = i). Consider the function G(t) := i p(i, i; t). The basic convergence theorem implies G(t) 1 as t, and the spectral representation gives the more detailed result G(t) = 1 + n e λut (1) where 0 = λ 1 < λ 2... λ n are the eigenvalues associated with the transition rate matrix. Note the probabilistic interpretation G(t) = P(X(t) = X(0)) 1 n u=2 where X(0) has uniform dist. One can regard the time τ at which G(τ) = 2 as one of many possible notions of mixing time. Rescaling time by this τ, we can standardize according to the convention G(1) = 2. The notion of L 2 cutoff studied in detail in [3] is, in our context, the property that a sequence of time-standarized chains has G (n) (t), t < 1; G (n) (t) 1, t > 1. (2) Now imagine a continuous-space analog. That is, a probability measure π on a space S and an S-valued Markov process X (t) such that for t > 0 there exist transition densities p (x, ; t) = density of P(X (t) X (0) = x) w.r.t. π which are symmetric: p (x, y; t) = p (y, x; t). The analog of G(t) is G (t) := p (x, x; t) π(dx). S 2

Assume G (t) < for t > 0. Then we expect (see remark below) the analog of (1) G (t) = 1 + e λ u t (3) where 0 = λ 1 < λ 2 λ 3... are the eigenvalues associated with the appropriate generator. And we can standardize to make G (2) = 1. Now consider a sequence of chains with n, where n is the number of states, but without assuming any relation between the chains as n varies, except for the assumption u=2 sup G n (t) < 0 < t < 1. (4) n As a standard analytic fact, because each G (n) is of form (1) there is a subsequence in which G n ( ) G ( ) for some limit function of form (3). This starts to hint at what is going on; the conjectured limit continuousspace process will be one with this function G ( ). Remark. Perhaps the simplest example is continuous-time random walk on the n-cycle, which has eigenvalues λ (n) u = 1 cos(2π(u 1)/n), u = 1,..., n and mixing time cn 2. After time-standardizing, the n limit is (for some constant c) G (t) = 1 + u 1 exp( 2π 2 c(u 1) 2 t) which is (as it should be) the density p(0, 0; t) for time-standardized Brownian motion on the unit circle. This illustrates why we anticipate limits G ) having the form (3) rather than a general spectral measure e λt Ψ(dλ). 0 3

0.2 An analogy Here is an analogy for what we shall do in the next section. Question: Can we characterize a metric space with probability measure up to measure-preserving isometry? That is, can we tell whether two such spaces (S 1, d 1, µ 1 ) and (S 1, d 1, µ 1 ) have a MPI? Answer: Yes. Given (S, d, µ), take i.i.d.(µ) random elements (ξ i, 1 i < ) of S, form the array X i,j = d(ξ i, ξ j ); i, j 1 and let Ψ be the distribution of this infinite random array. It is obvious that, for two isometric metric spaces with probability measure, we get the same Ψ. The converse that Ψ determines the space up to MPI is not obvious but is true; it was first stated explicitly by Vershik (2004) but in fact is a simple albeit technical consequence of the structure theory for partially exchangeable arrays developed by Hoover-Aldous-Kallenberg (1978-1985) and treated in detail in [6]. See [2] for my own recent account of its uses in the broad field of representing continuous limits of discrete random structures. The conjectures we state in this note are a novel instance of this methodology. 0.3 Isomorphic processes Consider a Markov process on a measurable space S with stationary distribution π, which we will view naively as a family of symmetric densities p (, ; t) satisfying the Chapman-Kolmogorov relations, with a UTC 1 on the t 0 behavior. Analogous to the previous section, define an infinite partially exchangeable random array whose entries are functions of t by take i.i.d.(π) random elements (ξ i, 1 i < ) of S set X ij = p (ξ i, ξ j ; t), i, j 1. (5) As in the previous section, this array has some distribution Ψ. There is a natural notion of isomorphism between two stationary Markov processes X 1 and X 2 on different spaces S 1 and S 2 a bijection φ that preserves joint distributions (φ(x 1 0), φ(x 1 t )) = d (X 2 0, X 2 t ) 1 unspecified technical condition 4

and hence transition densities. And as before it is obvious that, for two isomorphic processes, we get the same Ψ. Conjecture 1. If two symmetric Markov processes have the same Ψ then they are isomorphic. See discussion later. 0.4 Convergence of processes We can do exactly the same array construction for chains on finite sets S n : take i.i.d. uniform random elements (ξ i, 1 i < ) of S n set X n ij = p n (ξ i, ξ j ; t), i, j 1. Time-standardize, and consider the property discussed earlier: G n (t) G (t) as n (6) for some limit function with 1 < G (t) < for 0 < t <. This is just saying that EX n 11 G ( ). Now an easy argument gives EX 12 EX 11, and so we can take a subsequence in which (X n ij, i, j 1) d (X ij, i, j 1) as n (7) (in the usual sense of convergence of finite sub-arrays) for some limit random function-valued array. Conjecture 2. For any array (X ij, i, j 1) that arises as a limit (7) from finite chains, there exists a general-space chain with some transition densities p such that the representation (5) holds. 0.5 Summary We can now fill in some details of our original conjecture, as follows. Given a sequence of n-state reversible chains which does not have the L 2 cutoff property, we can pass to a subsequence satisfying (4) and then to a further subsequence satisfying (7); Conjecture 2 (if true) will identify the limit with some general-space process, which Conjecture 1 (if true) says is unique up to isomorphism. 5

We presented Conjecture 2 in convergence format, but there is a more intrinsic way to look at what s going on, analogous to the basic representation theorem for partially exchangeable arrays [6, 2]. We want to abstract a list of properties that an array (X ij) arising as (5) must obviously have: (i) partial exchangeability (ii) Chapman-Kolmogorov (iii) The UTC from section 0.3 (iv)...... The more definitive underlying Conjecture is that an array with a certain list of such properties has a representation in form (5). Such a result, with properties preserved under convergence in distribution, would of course imply Conjecture 2. 6

References [1] D. Aldous and J. Fill. Reversible markov chains and random walks on graphs. Available at http://www.stat.berkeley.edu/ aldous/rwg/book.html, 2002. [2] David J. Aldous. More uses of exchangeability: representations of complex random structures. In Probability and mathematical genetics, volume 378 of London Math. Soc. Lecture Note Ser., pages 35 63. Cambridge Univ. Press, Cambridge, 2010. [3] Guan-Yu Chen and Laurent Saloff-Coste. The L 2 -cutoff for reversible Markov processes. J. Funct. Anal., 258(7):2246 2315, 2010. [4] Persi Diaconis. The cutoff phenomenon in finite Markov chains. Proc. Nat. Acad. Sci. U.S.A., 93(4):1659 1664, 1996. [5] Persi Diaconis and Svante Janson. Graph limits and exchangeable random graphs. Rend. Mat. Appl. (7), 28(1):33 61, 2008. [6] Olav Kallenberg. Probabilistic symmetries and invariance principles. Probability and its Applications (New York). Springer, New York, 2005. [7] David A. Levin, Yuval Peres, and Elizabeth L. Wilmer. Markov chains and mixing times. American Mathematical Society, Providence, RI, 2009. With a chapter by James G. Propp and David B. Wilson. [8] László Lovász and Balázs Szegedy. Limits of dense graph sequences. J. Combin. Theory Ser. B, 96(6):933 957, 2006. [9] Daniel W. Stroock and S. R. Srinivasa Varadhan. Multidimensional diffusion processes, volume 233 of Grundlehren der Mathematischen Wissenschaften [Fundamental Principles of Mathematical Sciences]. Springer-Verlag, Berlin, 1979. 7