Lecture 7. We can regard (p(i, j)) as defining a (maybe infinite) matrix P. Then a basic fact is

Similar documents
Lecture 7. µ(x)f(x). When µ is a probability measure, we say µ is a stationary distribution.

Lecture 5. If we interpret the index n 0 as time, then a Markov chain simply requires that the future depends only on the present and not on the past.

P i [B k ] = lim. n=1 p(n) ii <. n=1. V i :=

Chapter 7. Markov chain background. 7.1 Finite state space

Positive and null recurrent-branching Process

Statistics 150: Spring 2007

Math 456: Mathematical Modeling. Tuesday, March 6th, 2018

Applied Stochastic Processes

Winter 2019 Math 106 Topics in Applied Mathematics. Lecture 9: Markov Chain Monte Carlo

Ergodic Theorems. Samy Tindel. Purdue University. Probability Theory 2 - MA 539. Taken from Probability: Theory and examples by R.

MARKOV CHAINS AND HIDDEN MARKOV MODELS

25.1 Ergodicity and Metric Transitivity

Lecture 10. Theorem 1.1 [Ergodicity and extremality] A probability measure µ on (Ω, F) is ergodic for T if and only if it is an extremal point in M.

Stochastic Processes (Week 6)

Convergence Rate of Markov Chains

Lecture 9 Classification of States

Markov Chains. Chapter Existence and notation. B 2 B(S) and every n 0,

Note that in the example in Lecture 1, the state Home is recurrent (and even absorbing), but all other states are transient. f ii (n) f ii = n=1 < +

Markov Chains CK eqns Classes Hitting times Rec./trans. Strong Markov Stat. distr. Reversibility * Markov Chains

MS&E 321 Spring Stochastic Systems June 1, 2013 Prof. Peter W. Glynn Page 1 of 10

11.4. RECURRENCE AND TRANSIENCE 93

Markov chains. Randomness and Computation. Markov chains. Markov processes

SMSTC (2007/08) Probability.

The Markov Chain Monte Carlo Method

12 Markov chains The Markov property

Markov processes Course note 2. Martingale problems, recurrence properties of discrete time chains.

Let (Ω, F) be a measureable space. A filtration in discrete time is a sequence of. F s F t

MATH 56A: STOCHASTIC PROCESSES CHAPTER 1

1 Random Walks and Electrical Networks

Lecture 11: Introduction to Markov Chains. Copyright G. Caire (Sample Lectures) 321

Recap. Probability, stochastic processes, Markov chains. ELEC-C7210 Modeling and analysis of communication networks

MATH 56A: STOCHASTIC PROCESSES CHAPTER 2

STOCHASTIC PROCESSES Basic notions

Markov processes and queueing networks

MS&E 321 Spring Stochastic Systems June 1, 2013 Prof. Peter W. Glynn Page 1 of 10. x n+1 = f(x n ),

STA 624 Practice Exam 2 Applied Stochastic Processes Spring, 2008

INTRODUCTION TO MARKOV CHAIN MONTE CARLO

Necessary and sufficient conditions for strong R-positivity

Chapter 2. Markov Chains. Introduction

Summary of Results on Markov Chains. Abstract

Discrete time Markov chains. Discrete Time Markov Chains, Limiting. Limiting Distribution and Classification. Regular Transition Probability Matrices

Lecture 21. David Aldous. 16 October David Aldous Lecture 21

Lectures on Stochastic Stability. Sergey FOSS. Heriot-Watt University. Lecture 4. Coupling and Harris Processes

Stochastic Processes

CONVERGENCE THEOREM FOR FINITE MARKOV CHAINS. Contents

Modern Discrete Probability Spectral Techniques

8. Statistical Equilibrium and Classification of States: Discrete Time Markov Chains

Classification of Countable State Markov Chains

MATH 564/STAT 555 Applied Stochastic Processes Homework 2, September 18, 2015 Due September 30, 2015

Zdzis law Brzeźniak and Tomasz Zastawniak

Markov Processes Hamid R. Rabiee

Markov Chains and Stochastic Sampling

A review of Continuous Time MC STA 624, Spring 2015

88 CONTINUOUS MARKOV CHAINS

Examination paper for TMA4180 Optimization I

Bayesian Methods with Monte Carlo Markov Chains II

LECTURE 3. Last time:

Probability & Computing

2 Discrete-Time Markov Chains

ADVANCED CALCULUS - MTH433 LECTURE 4 - FINITE AND INFINITE SETS

Value and Policy Iteration

Markov Chains. X(t) is a Markov Process if, for arbitrary times t 1 < t 2 <... < t k < t k+1. If X(t) is discrete-valued. If X(t) is continuous-valued

Lecture 5: Random Walks and Markov Chain

On random walks. i=1 U i, where x Z is the starting point of

Solutions to Homework Assignment 2

7. Shortest Path Problems and Deterministic Finite State Systems

µ n 1 (v )z n P (v, )

MATH5011 Real Analysis I. Exercise 1 Suggested Solution

SPECTRAL THEOREM FOR COMPACT SELF-ADJOINT OPERATORS

Homework set 3 - Solutions

Irreducibility. Irreducible. every state can be reached from every other state For any i,j, exist an m 0, such that. Absorbing state: p jj =1

Lecture 6. 2 Recurrence/transience, harmonic functions and martingales

MARKOV CHAINS AND MIXING TIMES

( f ^ M _ M 0 )dµ (5.1)

Monte Carlo importance sampling and Markov chain

Markov Chains for Everybody

STAT STOCHASTIC PROCESSES. Contents

Some Definition and Example of Markov Chain

Advanced Computer Networks Lecture 2. Markov Processes

STATS 3U03. Sang Woo Park. March 29, Textbook: Inroduction to stochastic processes. Requirement: 5 assignments, 2 tests, and 1 final

Random Times and Their Properties

Lecture 2: September 8

MARKOV PROCESSES. Valerio Di Valerio

Probability Theory. Richard F. Bass

Stochastic process. X, a series of random variables indexed by t

An Introduction to Entropy and Subshifts of. Finite Type

Chapter 2: Markov Chains and Queues in Discrete Time

JOSEPH B. KADANE AND JIASHUN JIN, IN HONOR OF ARNOLD ZELLNER

Stochastic Simulation

RANDOM WALKS. Course: Spring 2016 Lecture notes updated: May 2, Contents

Thus f is continuous at x 0. Matthew Straughn Math 402 Homework 6

Lecture 20 : Markov Chains

Stat-491-Fall2014-Assignment-III

Markov Chain Monte Carlo (MCMC)

Markov Chains (Part 3)

Markov Chain Monte Carlo

MSc MT15. Further Statistical Methods: MCMC. Lecture 5-6: Markov chains; Metropolis Hastings MCMC. Notes and Practicals available at

6 Markov Chain Monte Carlo (MCMC)

MARKOV CHAINS: STATIONARY DISTRIBUTIONS AND FUNCTIONS ON STATE SPACES. Contents

Statistics 253/317 Introduction to Probability Models. Winter Midterm Exam Friday, Feb 8, 2013

Transcription:

MARKOV CHAINS What I will talk about in class is pretty close to Durrett Chapter 5 sections 1-5. We stick to the countable state case, except where otherwise mentioned. Lecture 7. We can regard (p(i, j)) as defining a (maybe infinite) matrix P. Then a basic fact is P (X n = j X 0 = i) =P n (i, j) (12) where P n denotes matrix multiplication. There are two conceptually different ways to get this. First regard X 0 as having some arbitrary initial distribution µ 0. Write µ n for the distribution of X n : µ n (j) =P (X n = j). By considering the n + 1 st step we get the forward equation (in vectormatrix notation) µ n+1 = µ n P which implies µ n = µ 0 P n. Putting µ 0 ( ) =δ i ( ) gives (12). On the other hand we could fix a bounded function h : S R and define h n (i) =E(h(X n ) X 0 = i). By considering the first step we get the backward equation which implies Putting h( ) =δ j ( ) gives (12). h n+1 = Ph n h n = P n h. Hitting times. Fix A S and consider the first hitting time τ A = min{n 0:X n A}. Now consider h A (i) =P i (τ A < ). 12

The next result says h A is the minimal solution of equations (i)-(iii) below. (i)h(i) 0 for all i S. (ii) h(i) = 1 for i A. (iii) h(i) = j p(i, j)h(j) for i A. Proposition 18 (a) h A satisfies (i)-(iii). (b) If h is a solution of (i)-(iii) then h A h. Proof. (a) is easy. For (b) consider the transition matrix P A for the chain stopped on A p A (i, j) = p(i, j),i A = δ i (j),i A Then (ii) and (iii) imply (iv) h = P A h (jargon: h is a harmonic function for P A.) Fix h satisfying (i)-(iii). Since h 1 A, h = P n A h Pn A 1 A. This says that h(i) P i (Xn A A) where XA denotes a Markov chain with transition matrix P A. Now given an (i, P)-chain (X n ) it is clear that Xn A = X n τ A defines an (i, P A )-chain. So h(i) P i (Xn A A) =P i (τ A n). Let n. 13

Lecture 10. In this class we prove the basic results about invariant measures and stationary distributions. The results are those of Durrett Theorems 4.3-4.7, though the organization of the proofs is a little different. Most of the work is in the first result (c.f. Durrett Theorem 4.3: note we don t assume recurrence). Fix a reference state b and define the b-block occupation measure µ(b, x) = P b (X n = x, T b >n). n=0 Proposition 19 Consider the equations µ(y) = x µ(x)p(x, y),y b; µ(b) = 1, 0 µ(x) for all x. (13) Then µ(b, ) is the minimal solution of (13). Moreover P b (T b < ) = x µ(b, x)p(x, b). Theorem 20 Suppose the chain is irreducible and recurrent. Then there exists an invariant measure µ, unique up to scalar multiples. Either (i) µ(s) = and E x T x = for all x; or (ii) µ(s) < and E x T x < for all x. In case (i) the chain is null-recurrent. In case (ii) the chain is positive-recurrent. Theorem 21 Suppose the chain is irreducible. Then it is positive-recurrent iff a stationary distribution π exists. If so, then π(x) = 1/E x T x for all x. 14

Lecture 11. Durrett 5.5. Here is a counterpart to the basic convergence result, Proposition 22 Suppose the chain is irreducible but not positive-recurrent. Then P µ (X n = y) 0 for all y and all initial distributions µ. Proof. First, we can reduce to the aperiodic case. Next, the result is obvious in the transient case, because P µ (X n = y) =E µ N(y) 1+E y N(y) = 1/(1 ρ yy ) <. n 1 So we may suppose the chain is null-recurrent. Consider independent copies (X n,y n ) as a chain on S S. This product chain is irreducible. If the product chain is transient then as above P µ µ (X n = y, Y n = y) <. n 1 But the summands are (P µ (X n = y)) 2, and these must converge to 0. So suppose the product chain is recurrent. We argue by contradiction. If the result were false, then there exists an initial distribution µ, a state b and a subsequence of times j n such that P µ (X(j n )=b) α b > 0. By the diagonal argument, we can then pick a subsequence (k n ) such that for every state y P µ (X(k n )=y) α y 0. (14) Now the coupling argument in the proof of Durrett 5.5 depended only on the fact that the product chain was recurrent. So this argument shows that convergence in (14) holds for all initial distributions. The rest of the argument is analysis. Fatou s lemma shows y α y 1, and then (αp)(y) = w (lim p kn (x, w))p(w, y) by (14) lim p kn+1 (x, y) by Fatou s lemma = lim p(x, z)p kn (z, y) z = lim z = α y. p(x, z)α y by (14) and dominated convergence 15

Now if there really was a strict inequality for some y then α y > (αp)(y) = α x p(x, y) = y y x y x which is impossible. So we have shown α = αp: this is a finite invariant measure, implying positive-recurrence. Lecture 12. The ergodic theorem (Durrett 5.1) is a consequence of the SLLN and the following deterministic fact, worth isolating. Lemma 23 Let (t i ) be increasing with n 1 t n t. Let n(t) = max{i : t i t}. Then n(t)/t 1/ t. Now let (r i ) be such that r i 0 and n 1 n i=1 r i r, and let r(t) be such that α x Then r(t)/t r/ t. n(t) i=1 r i r(t) n(t)+1 i=1 r i. A number of useful identities may be derived from the following result. For simplicity, suppose our chain is irreducible and finite, aperiodic (really we only need positive-recurrent). Proposition 24 Consider the chain started at state x. Let 0 < S < be a stopping time such that X S = x. Let y be an arbitrary state. Then E x (number of visits to y before time S) =π(y)e x S. In the phrase number of... before time t, our convention is to include time 0 but exclude time t. The following series of lemmas arise from particular choices of y and S. Some involve the fundamental matrix (the sum is finite by exercise 5.8). Z(x, y) = (p t xy π(y)) t=0 16

Lemma 25 For y x, E x (number of visits to x before time T y )=π(x)(e x T y + E y T x ). Lemma 26 Lemma 27 π(x)e π τ x = Z xx. E π (number of visits to y before time τ x ) Lemma 28 π(y)e x τ y = Z yy Z xy. = π(y) π(x) Z xx Z xy. 17