µ n 1 (v )z n P (v, )

Similar documents
Note that in the example in Lecture 1, the state Home is recurrent (and even absorbing), but all other states are transient. f ii (n) f ii = n=1 < +

P i [B k ] = lim. n=1 p(n) ii <. n=1. V i :=

Examples of Countable State Markov Chains Thursday, October 16, :12 PM

2 Discrete-Time Markov Chains

2. Transience and Recurrence

Lecture 7. µ(x)f(x). When µ is a probability measure, we say µ is a stationary distribution.

Markov Chains. X(t) is a Markov Process if, for arbitrary times t 1 < t 2 <... < t k < t k+1. If X(t) is discrete-valued. If X(t) is continuous-valued

1 Continuous-time chains, finite state space

Birth-death chain models (countable state)

Stochastic Processes

MATH 564/STAT 555 Applied Stochastic Processes Homework 2, September 18, 2015 Due September 30, 2015

The Transition Probability Function P ij (t)


MATH 56A: STOCHASTIC PROCESSES CHAPTER 2

arxiv: v2 [math.pr] 4 Sep 2017

8. Statistical Equilibrium and Classification of States: Discrete Time Markov Chains

Countable state discrete time Markov Chains

Markov Chains and Stochastic Sampling

CDA6530: Performance Models of Computers and Networks. Chapter 3: Review of Practical Stochastic Processes

Markov chains. 1 Discrete time Markov chains. c A. J. Ganesh, University of Bristol, 2015

Lectures on Stochastic Stability. Sergey FOSS. Heriot-Watt University. Lecture 4. Coupling and Harris Processes

LECTURE 3. Last time:

Homework 3 posted, due Tuesday, November 29.

MATH 56A SPRING 2008 STOCHASTIC PROCESSES 65

Markov Chains CK eqns Classes Hitting times Rec./trans. Strong Markov Stat. distr. Reversibility * Markov Chains

Markov Chains Handout for Stat 110

1 Random walks: an introduction

SMSTC (2007/08) Probability.

Plan Martingales cont d. 0. Questions for Exam 2. More Examples 3. Overview of Results. Reading: study Next Time: first exam

Lecture 5. If we interpret the index n 0 as time, then a Markov chain simply requires that the future depends only on the present and not on the past.

Chapter 5. Continuous-Time Markov Chains. Prof. Shun-Ren Yang Department of Computer Science, National Tsing Hua University, Taiwan

Lecture 11: Introduction to Markov Chains. Copyright G. Caire (Sample Lectures) 321

Reinforcement Learning

CDA5530: Performance Models of Computers and Networks. Chapter 3: Review of Practical

Lecture 9 Classification of States

Markov Chain Model for ALOHA protocol

Reading: Karlin and Taylor Ch. 5 Resnick Ch. 3. A renewal process is a generalization of the Poisson point process.

Lecture 7. We can regard (p(i, j)) as defining a (maybe infinite) matrix P. Then a basic fact is

Markov chains. Randomness and Computation. Markov chains. Markov processes

12 Markov chains The Markov property

Transience: Whereas a finite closed communication class must be recurrent, an infinite closed communication class can be transient:

Lecture 6. 2 Recurrence/transience, harmonic functions and martingales

Solutions to Homework Discrete Stochastic Processes MIT, Spring 2011

Statistics 992 Continuous-time Markov Chains Spring 2004

1 Stat 605. Homework I. Due Feb. 1, 2011

Math Homework 5 Solutions

CS145: Probability & Computing Lecture 18: Discrete Markov Chains, Equilibrium Distributions

Statistics 253/317 Introduction to Probability Models. Winter Midterm Exam Friday, Feb 8, 2013

18.175: Lecture 30 Markov chains

EXAMINATIONS OF THE ROYAL STATISTICAL SOCIETY

Stochastic Simulation

4.7.1 Computing a stationary distribution

Markov Chains, Random Walks on Graphs, and the Laplacian

A primer on basic probability and Markov chains

Optional Stopping Theorem Let X be a martingale and T be a stopping time such

STAT STOCHASTIC PROCESSES. Contents

Homework set 3 - Solutions

Example: physical systems. If the state space. Example: speech recognition. Context can be. Example: epidemics. Suppose each infected

Logistics. All the course-related information regarding

Positive and null recurrent-branching Process

Discrete time Markov chains. Discrete Time Markov Chains, Limiting. Limiting Distribution and Classification. Regular Transition Probability Matrices

RENEWAL THEORY STEVEN P. LALLEY UNIVERSITY OF CHICAGO. X i

Lecture 2. We now introduce some fundamental tools in martingale theory, which are useful in controlling the fluctuation of martingales.

Lecture 20 : Markov Chains

Probability & Computing

Summary: A Random Walks View of Spectral Segmentation, by Marina Meila (University of Washington) and Jianbo Shi (Carnegie Mellon University)

Applied Stochastic Processes

MAA704, Perron-Frobenius theory and Markov chains.

Random Times and Their Properties

Lectures on Markov Chains

Continuous time Markov chains

Ergodic Theorems. Samy Tindel. Purdue University. Probability Theory 2 - MA 539. Taken from Probability: Theory and examples by R.

STOCHASTIC PROCESSES Basic notions

MASSACHUSETTS INSTITUTE OF TECHNOLOGY Department of Electrical Engineering and Computer Science

Classification of Countable State Markov Chains

Markov Chains, Stochastic Processes, and Matrix Decompositions

Homework 4 due on Thursday, December 15 at 5 PM (hard deadline).

Lecture 5: Random Walks and Markov Chain

MS&E 321 Spring Stochastic Systems June 1, 2013 Prof. Peter W. Glynn Page 1 of 10

INTRODUCTION TO MCMC AND PAGERANK. Eric Vigoda Georgia Tech. Lecture for CS 6505

6 Markov Chain Monte Carlo (MCMC)

MATH 56A: STOCHASTIC PROCESSES CHAPTER 7

Week 5: Markov chains Random access in communication networks Solutions

Markov Processes on Discrete State Spaces

Homework 2 will be posted by tomorrow morning, due Friday, October 16 at 5 PM.

An Introduction to Entropy and Subshifts of. Finite Type

Necessary and sufficient conditions for strong R-positivity

MARKOV PROCESSES. Valerio Di Valerio

INTRODUCTION TO MARKOV CHAINS AND MARKOV CHAIN MIXING

Chapter 2 SOME ANALYTICAL TOOLS USED IN THE THESIS

11.4. RECURRENCE AND TRANSIENCE 93

Math 324 Summer 2012 Elementary Number Theory Notes on Mathematical Induction

Markov Chains for Everybody

Stochastic Processes (Week 6)

0.1 Markov Chains Generalities 0.1. MARKOV CHAINS 1

Stochastic processes. MAS275 Probability Modelling. Introduction and Markov chains. Continuous time. Markov property

Stochastic Process (ENPC) Monday, 22nd of January 2018 (2h30)

At the boundary states, we take the same rules except we forbid leaving the state space, so,.

Powerful tool for sampling from complicated distributions. Many use Markov chains to model events that arise in nature.

ECE-517: Reinforcement Learning in Artificial Intelligence. Lecture 4: Discrete-Time Markov Chains

Transcription:

Plan More Examples (Countable-state case). Questions 1. Extended Examples 2. Ideas and Results Next Time: General-state Markov Chains Homework 4 typo Unless otherwise noted, let X be an irreducible, aperiodic Markov chain on countable state-space S with initial distribution µ and transition probabilities P. Example 1. Walk on a Finite, Directional Graph As before, let V = {v 1,..., v m } be a collection of vertices. Let E(v, v ) be 1 if there is an edge in the graph between v and v and otherwise. Unlike the last case, we do not assume E(v, v ) = E(v, v). Let µ be an initial distribution on V. Let P be a transition probability matrix on V that satisfies P (v, v ) = if and only if E(v, v ) =. Define Y = (Y n ) n as before. One approach to understanding the chain is to do the conditioning trick that we ve used before and apply generating functions. Let µ n be the distribution of Y n. µ n (v) = µ(v)1 (n=) + v µ n 1 (v )P (v, v)1 (n>). (1) Define G(z) = n µ nz n be the vector-valued generating function. Then, from the above recursion So, µ n z n = µ + µ n 1 (v )z n P (v, ) n v n (2) G(z) = µ + zg(z) P. (3) G(z)(I zp ) = µ (4) G(z) = µ(i zp ) 1 = z n µp n, n (5) with the last assuming the inverse matrix exists. We can use this to understand both the short and long-term behavior of the chain. If Y is irreducible, then O(s, S) = implies that the chain is recurrent because the sum O(s, S) = s O(s, s ) is finite, implying at least one (and thus all) terms must be infinite. Because any invariant measure on the finite state space will consequently be finite and hence normalizable, Y is positive recurrent as well. The search for a stationary distribution π corresponds to a search for eigenvectors of P with eigenvalue 1, vectors π such that π(i P ) =. 1 28 Feb 26

Example 2. Finite, Doubly Stochastic Chain Suppose U = (U n ) n is an irreducible Markov chain on a finite state-space S such that u P (u, u ) = 1. Because u P (u, u ) = 1, this implies that both the rows and columns sum to 1. Such a matrix is called doubly stochastic. What is the stationary distribution of the chain? Note that for any constant c, u cp (u, u ) = c. Hence, this is an invariant measure. The unique invariant distribution π is then just a uniform distribution on S. Example 3. Renewal Process and Forward Recurrence Time Chain Consider a process Z = (Z n ) n given by Z n = Z + n Ξ k (6) for iid random variables Ξ i and arbitrary Z. This is called a renewal process. As an example, consider a critical part in a system that operates for some time and then fails. When it fails it is replaced. Think of Ξ as the lifetime of the replacement parts and Z as the lifetime of the original part. Then, Z n is the time of the nth replacement a renewal of the system. Given a renewal process, we can define N t = sup{n : Z n t}. Then, N t is a counting process that counts the renewals up to time t. Suppose now that Z and Ξ 1 take values in Z + and have pmfs µ p Z and p Ξ p Ξ1. Z n is thus a countable-state Markov chain, but not a terribly interesting one as it marches inexorably toward. But we can define two related processes that will turn out to be very interesting in general. Define V + and V to be, respectively, the forward and backward recurrence time chains, as follows for n : k=1 V n + = inf{z m n: Z m > n} (7) Vn = inf{n Z m: Z m n}. (8) Then V n + represents the time until the next renewal, and V n represents the time since the last renewal. These are sometimes also called the residual lifetime and age processes. Are these Markov chains? What are the state spaces? We can check the Markov Property explicitly. But the regeneration of the system at each renewal gives us a simple way of seeing it. When V n + > 1, for instance, the next time is determined. When V n + = 1, a renewal ensues and the next time is an independent waiting time. The state space of V + is Z + ; the state space of V is Z. 2 28 Feb 26

What are the transition probabilities? If V + n = k > 1, then V + n+1 = k 1 by construction. If V + n = 1, then a renewal occurs at time n + 1, so the time until the following renewal has distribution ξ. Hence, P (k, k 1) = 1 for k > 1, (9) P (1, k) = ξ(k) for k Z +. (1) For V, we can reason similarly. Let S Ξ be the survival function of Ξ 1. Then, P (k, k + 1) = P{Ξ > k + 1 Ξ > k} = S Ξ(k + 1) S Ξ (k) (11) P (k, ) = P{Ξ = k + 1 Ξ > k} = p Ξ(k + 1). (12) S Ξ (k) Is V + irreducible? Is it recurrent? If there exists an M Z + such that S Ξ (M) = and p Ξ (M) >, then all states j > M are transient, and all states {1,..., M} communicate and are recurrent since there is only a finite number. (To see the latter, note that we can find a positive probability path between each pair of states in this set.) If no such M exists, then V + is irreducible. Note that for all states n > 1, P!1 n 1 (n, 1) = 1. Hence, So the chain is recurrent in this case as well. R(1, 1) = p Ξ (n)p!1 n 1 (n, 1) = 1. (13) What is the long-run behavior of the chain? Let ρ(j) = P!1 n (1, j). Because P n!1 (1, j) = p Ξ(j + n 1) for n 1, we can write ρ(j) = p Ξ (j + n 1) = p Ξ (n) = S Ξ (j 1). (14) n j Notice that ρ(j)p (j, k) = ρ(1)p Ξ (k) + ρ(k + 1) (15) j 1 = p Ξ (k) + S Ξ (k) (16) = ρ(k). (17) This invariant measure is positive (on its support) and is finite if and only if ρ(n) = S Ξ (n 1) = np Ξ (n) = EΞ 1 <. (18) In this case, π(k) = ρ(k)/eξ 1 is a stationary distribution. 3 28 Feb 26

Idea 4. The above argument leads to an interesting idea for the countable case. Suppose that X is recurrent, Pick a state s S such that it is easy to compute P!s n (s, s) for any s S. Define ρ(s) = P!s n (s, s). (19) Note that ρ(s ) = 1 because the chain is recurrent (i.e., R(s, s ) = 1). Then, mimicking the above argument, we get This is finite only if ρ(s)p (s, s ) = ρ(s )P (s, s ) + s S s s P n!s (s, s)p (s, s ) (2) = P (s, s ) + P!s n (s, s ) (21) n 2 = P!s (s, s ) + P!s n (s, s ) (22) n 2 = ρ(s ). (23) ρ(s) = s S = s S P n!s (s, s) (24) P n!s (s, s) (25) = P!s n (s, S) (26) = P s {T s n} (27) = E s T s (28) <, (29) which is just the positive recurrence condition. Hence, π(s) = P!s n (s, s) P!s n (s, S) = 1. (3) E s T s We already knew the last equality, but this gives a new way of finding a stationary distribution, using equation (19). 4 28 Feb 26

Example 5. Upper Hessenberg Transition Probabilities Suppose S = Z and suppose the non-zero transition probabilities are of the form P (, k) = a k for k, (31) P (j, k) = a k j+1 for j 1, k j 1, (32) where (a n ) n is a non-negative sequence satisfying n a n = 1 and n na n <. We want to determine the stability of this chain. Consider the function (s) on the state space, given by (s) = E(X n+1 X n X n = s). (33) This tells us how much the chain drifts on average in one step when starting at s. Note that because we are dealing with a time-homogeneous chain, (s) = E s X 1 s = s P (s, s )s s. (34) Let s compute this function. () = k = k (j) = k P (, k)k (35) ka k < (36) P (j, k)k j (37) = = = k=j 1 k=j 1 n= a k j+1 k j (38) a k j+1 (k j + 1) + k=j 1 a k j+1 (j 1) j (39) na n 1 <, (4) for j 1. Case (i): n na n > 1. In this case, (j) > for every j, hence on average in any state, we tend toward higher states. This positive drift seems to suggest (though does not prove) transience. Case (ii): n na n < 1. In this case, () > and (j) < ɛ n na n 1 for all j 1. Hence, on average, whenever the chain is away from zero, it tends to move back toward zero. This suggests recurrence. What we can we make of this? 5 28 Feb 26

Definition 6. Drift Let V be a non-negative function on the state space (that is, V : S R ). Define the drift operator X by X V = P V V. (41) That is, ( X V )(s) = s S P (s, s )V (s ) V (s) = E(V (X n+1 ) V (X n ) X n = s) = E s (V (X 1 )) V (s). (42) Note that, in general, X V takes values in [, ]. Theorem 7. Foster s Drift Criterion Suppose there exists a non-negative function V : S R, an ɛ >, and a finite set S S such that X V (s) < for s S (43) X V (s) ɛ for s S. (44) Then, X is positive recurrent. Proof For s S, X V (s) < implies that P V < on S. Define for n. Notice that for m, u [n] (s) = s P n (s, s )V (s ), (45) u [m+1] (s) = s P m+1 (s, s )V (s ) (46) = P m (s, t)p (t, s )V (s ) (47) s t S = t S P m (s, t) s P (t, s )V (s ) (48) = P m (s, t)(p V )(t) t S (49) = P m (s, t)(p V )(t) + P m (s, t)(p V )(t) t S t S (5) P m (s, t)(p V )(t) + P m (s, t)(v (t) ɛ) t S t S (51) P m (s, t)((p V )(t) + ɛ) + P m (s, t)(v (t) ɛ) t S t S (52) = P m (s, t)((p V )(t) + ɛ) + u [m] (s) ɛ. t S (53) This gives us an upper bound for u [m+1] u [m]. Summing these together by telescoping gives u [n+1] (s) u [] (s) + n P m (s, t)(p V (t) + ɛ) (n + 1)ɛ. (54) t m= 6 28 Feb 26

Rearranging and dividing by n + 1 gives u [] (s) n + 1 + t ( 1 n + 1 ) n P m (s, t) (P V (t) + ɛ) ɛ. (55) m= Taking limits on both sides as n and using the fact that S is finite and P V (t) < on S, gives us 1 n lim P m (s, t) >, (56) n n + 1 m= for sum t S. But it follows that (Césaro summation), that n P n (s, t) cannot converge. Hence, X is recurrent. Positive recurrence follows as well, though somewhat more delicately. Example 5 cont d Upper Hessenberg Transition Probabilities We know from the above that n na n < 1 implies that this chain is positive recurrence. What about when n na n 1? Can we deduce transience from this? It turns out that Foster s criterion cannot simply be reversed. We can get the following. Theorem 8. Suppose there exists a bounded, non-negative function V on S and r such that {s S: V (s) > r} and {s S: V (s) r} are both nonempty and X V (s) > if V (s) > r. (57) Then, X is transient. The converse is also true. The proof of this relies on a rather cute result: Lemma Let C S. Let h (s) = H(s, C) be the hitting probability of S from S. (Recall, the hitting time S C is zero if the chain starts in C and is otherwise equal to T C.) Then, if h: S R is a solution to X h(s) if s C c (58) X h(s) 1 if s C (59) Then, h h. Now, to the theorem, suppose that V M. We must have, by the conditions, that M > r. (Why?) Define { 1 if V (s) r h V (s) = M V (s) M r if V (s) > r. Then, we can show that h V solves (58) and (59) with C = {s: V (s) r}. So h h V. But then h (s) h V (s) < 1 if s C, which shows that R(s, s ) < 1 for s C c and s C. Transience follows. (6) Example 5 cont d Upper Hessenberg Transition Probabilities Can we show that n a n > 1 implies transience, using the above? 7 28 Feb 26

Example 9. Storage Model Consider a storage system (dam, warehouse, insurance policy) that receives inputs at random times but otherwise drains at a regular rate. Let T = and let the rest of the T i s be iid Z -valued with cdf G. These are inter-arrival times for the inputs to our storage system. Let the S n s be iid Z with cdf H. These are the amounts input at the time Z n = T + + T n. Assume that the S n s and T n s are independent of each other as well. Suppose also that the storage system drains or outputs at rate r between inputs. Define a process (V n ) n by V n+1 = (V n + S n rt n+1 ) +. (61) Here, V n represents the contents of the storage system just before the nth input (that is, at time Z n ). Is V a Markov chain? What is the structure of the transition probabilities? What can we say about the long-run behavior of the chain? What is special about the state {}? How might we generalize this model to make it more realistic? 8 28 Feb 26