STAT 380 Continuous Time Markov Chains

Similar documents
Queuing Theory. Richard Lockhart. Simon Fraser University. STAT 870 Summer 2011

Statistics 150: Spring 2007

The Transition Probability Function P ij (t)

Continuous-Time Markov Chain

Chapter 5. Continuous-Time Markov Chains. Prof. Shun-Ren Yang Department of Computer Science, National Tsing Hua University, Taiwan

Markov Chains. X(t) is a Markov Process if, for arbitrary times t 1 < t 2 <... < t k < t k+1. If X(t) is discrete-valued. If X(t) is continuous-valued

Birth and Death Processes. Birth and Death Processes. Linear Growth with Immigration. Limiting Behaviour for Birth and Death Processes

TMA4265 Stochastic processes ST2101 Stochastic simulation and modelling

EXAM IN COURSE TMA4265 STOCHASTIC PROCESSES Wednesday 7. August, 2013 Time: 9:00 13:00

Continuous Time Markov Chains

Quantitative Model Checking (QMC) - SS12

Markov processes and queueing networks

STA 624 Practice Exam 2 Applied Stochastic Processes Spring, 2008

Continuous time Markov chains

SMSTC (2007/08) Probability.

Stochastic Processes

A review of Continuous Time MC STA 624, Spring 2015

Markov chains. 1 Discrete time Markov chains. c A. J. Ganesh, University of Bristol, 2015

Introduction to Queuing Networks Solutions to Problem Sheet 3

Continuous time Markov chains

Math 597/697: Solution 5

1 IEOR 4701: Continuous-Time Markov Chains

Stochastic process. X, a series of random variables indexed by t

Examination paper for TMA4265 Stochastic Processes

Lecture Notes 7 Random Processes. Markov Processes Markov Chains. Random Processes

CDA6530: Performance Models of Computers and Networks. Chapter 3: Review of Practical Stochastic Processes

STAT 380 Markov Chains

Solutions to Homework Discrete Stochastic Processes MIT, Spring 2011

CDA5530: Performance Models of Computers and Networks. Chapter 3: Review of Practical

Data analysis and stochastic modeling

Time Reversibility and Burke s Theorem

Part I Stochastic variables and Markov chains

Lecture 4a: Continuous-Time Markov Chain Models

Lecture 7: Simulation of Markov Processes. Pasi Lassila Department of Communications and Networking

Performance Evaluation of Queuing Systems

Lecture 20: Reversible Processes and Queues

STAT STOCHASTIC PROCESSES. Contents

IEOR 6711, HMWK 5, Professor Sigman

TCOM 501: Networking Theory & Fundamentals. Lecture 6 February 19, 2003 Prof. Yannis A. Korilis

Irreducibility. Irreducible. every state can be reached from every other state For any i,j, exist an m 0, such that. Absorbing state: p jj =1

MARKOV PROCESSES. Valerio Di Valerio

Queuing Networks: Burke s Theorem, Kleinrock s Approximation, and Jackson s Theorem. Wade Trappe

Recap. Probability, stochastic processes, Markov chains. ELEC-C7210 Modeling and analysis of communication networks

Markov Chains in Continuous Time

Markov Processes Cont d. Kolmogorov Differential Equations

QUEUING MODELS AND MARKOV PROCESSES

Queueing Theory. VK Room: M Last updated: October 17, 2013.

STOCHASTIC PROCESSES Basic notions

(implicitly assuming time-homogeneity from here on)

Problem Set 8

LECTURE #6 BIRTH-DEATH PROCESS

Intro Refresher Reversibility Open networks Closed networks Multiclass networks Other networks. Queuing Networks. Florence Perronnin

Figure 10.1: Recording when the event E occurs

Readings: Finish Section 5.2

1 Types of stochastic models

Structured Markov Chains

Problems. HW problem 5.7 Math 504. Spring CSUF by Nasser Abbasi

Continuous Time Processes

Modelling Complex Queuing Situations with Markov Processes

Networking = Plumbing. Queueing Analysis: I. Last Lecture. Lecture Outline. Jeremiah Deng. 29 July 2013

Queueing Theory II. Summary. ! M/M/1 Output process. ! Networks of Queue! Method of Stages. ! General Distributions

Introduction to Markov Chains, Queuing Theory, and Network Performance

Queuing Theory. Using the Math. Management Science

Example: physical systems. If the state space. Example: speech recognition. Context can be. Example: epidemics. Suppose each infected

QUEUING SYSTEM. Yetunde Folajimi, PhD

MAT SYS 5120 (Winter 2012) Assignment 5 (not to be submitted) There are 4 questions.

Probability, Random Processes and Inference

GI/M/1 and GI/M/m queuing systems

Markov Chains. October 5, Stoch. Systems Analysis Markov chains 1

Markov Chain Model for ALOHA protocol

BIRTH DEATH PROCESSES AND QUEUEING SYSTEMS

= P{X 0. = i} (1) If the MC has stationary transition probabilities then, = i} = P{X n+1

6 Continuous-Time Birth and Death Chains

Name of the Student:

MATH37012 Week 10. Dr Jonathan Bagley. Semester

reversed chain is ergodic and has the same equilibrium probabilities (check that π j =

CS 798: Homework Assignment 3 (Queueing Theory)

Exact Simulation of the Stationary Distribution of M/G/c Queues

Queueing Networks and Insensitivity

Availability. M(t) = 1 - e -mt

IEOR 6711: Stochastic Models I, Fall 2003, Professor Whitt. Solutions to Final Exam: Thursday, December 18.

Queueing Theory I Summary! Little s Law! Queueing System Notation! Stationary Analysis of Elementary Queueing Systems " M/M/1 " M/M/m " M/M/1/K "

Poisson Processes. Particles arriving over time at a particle detector. Several ways to describe most common model.

ECE-517: Reinforcement Learning in Artificial Intelligence. Lecture 4: Discrete-Time Markov Chains

Contents Preface The Exponential Distribution and the Poisson Process Introduction to Renewal Theory

T. Liggett Mathematics 171 Final Exam June 8, 2011

UNIVERSITY OF YORK. MSc Examinations 2004 MATHEMATICS Networks. Time Allowed: 3 hours.

AARMS Homework Exercises

Link Models for Circuit Switching

Part II: continuous time Markov chain (CTMC)

THE QUEEN S UNIVERSITY OF BELFAST

Class 11 Non-Parametric Models of a Service System; GI/GI/1, GI/GI/n: Exact & Approximate Analysis.

Markov Chains. Sarah Filippi Department of Statistics TA: Luke Kelly

Probability Distributions

(b) What is the variance of the time until the second customer arrives, starting empty, assuming that we measure time in minutes?

Inventory Ordering Control for a Retrial Service Facility System Semi- MDP

stochnotes Page 1

Markov Chains CK eqns Classes Hitting times Rec./trans. Strong Markov Stat. distr. Reversibility * Markov Chains

Markov Reliability and Availability Analysis. Markov Processes

Promotion and Application of the M/M/1/N Queuing System

Transcription:

STAT 380 Continuous Time Markov Chains Richard Lockhart Simon Fraser University Spring 2018 Richard Lockhart (Simon Fraser University)STAT 380 Continuous Time Markov Chains Spring 2018 1 / 35

Continuous Time Markov Chains Population of single celled organisms in a stable environment. Fix short time interval, length h. Each organism has some probability of dividing to produce two organisms and some other probability of dying. Possible assumptions: Different organisms behave independently. Probability of division is λh plus o(h). Probability of death is µh plus o(h). Probability that an organism divides twice (or divides once and dies) in the interval of length h is o(h). Richard Lockhart (Simon Fraser University)STAT 380 Continuous Time Markov Chains Spring 2018 2 / 35

Tacit Assumptions Constants of proportionality do not depend on time: stable environment. Constants do not depend on organism: organisms are all similar and live in similar environments. Y(t): total population at time t. H t : history of process up to time t. Condition on event Y(t) = n. Probability of two or more divisions (more than one division by a single organism or two or more organisms dividing) is o(h). Probability of both a division and a death or of two or more deaths is o(h). Richard Lockhart (Simon Fraser University)STAT 380 Continuous Time Markov Chains Spring 2018 3 / 35

Tacit Assumptions II So prob of exactly 1 division by any one of n organisms is nλh+o(h). Similarly probability of 1 death is nµh+o(h). We deduce: P(Y(t +h) = n+1 Y(t) = n,h t ) = nλh+o(h) P(Y(t +h) = n 1 Y(t) = n,h t ) = nµh+o(h) P(Y(t +h) = n Y(t) = n,h t ) = 1 n(λ+µ)h+o(h) P(Y(t +h) {n 1,n,n +1} Y(t) = n,h t ) = o(h) These equations lead to: P(Y(t +s) = j Y(s) = i,h s ) = P(Y(t +s) = j Y(s) = i) = P(Y(t) = j Y(0) = i) This is the Markov Property. Richard Lockhart (Simon Fraser University)STAT 380 Continuous Time Markov Chains Spring 2018 4 / 35

Tacit Assumptions III Definition: A process {Y(t);t 0} taking values in S, a finite or countable state space is a Markov Chain if P(Y(t +s) = j Y(s) = i,h s ) = P(Y(t +s) = j Y(s) = i) P ij (s,s +t) Definition: A Markov chain Y has stationary transitions if P ij (s,s +t) = P ij (0,t) P ij (t) From now on: our chains have stationary transitions. Richard Lockhart (Simon Fraser University)STAT 380 Continuous Time Markov Chains Spring 2018 5 / 35

Summary of Markov Process Results Chapman-Kolmogorov equations: P ik (t +s) = j P ij (t)p jk (s) Exponential holding times: starting from state i time, T i, until process leaves i has exponential distribution, rate denoted v i. Sequence of states visited, Y 0,Y 1,Y 2,... is Markov chain transition matrix has P ii = 0. Y sometimes called skeleton. Communicating classes defined for skeleton chain. Usually assume chain has 1 communicating class. Periodicity irrelevant because of continuity of exponential distribution. Richard Lockhart (Simon Fraser University)STAT 380 Continuous Time Markov Chains Spring 2018 6 / 35

Summary of Markov Process Results II Instantaneous transition rates from i to j: q ij = v i P ij Kolmogorov backward equations: P ij(t) = q ik P kj (t) v i P ij (t) k i Kolmogorov forward equations: P ij(t) = q kj P ik (t) v j P ij (t) k j For strongly recurrent chains with a single communicating class: P ij (t) π j Stationary initial probabilities π i satisfy: v j π j = q kj π k k j Richard Lockhart (Simon Fraser University)STAT 380 Continuous Time Markov Chains Spring 2018 7 / 35

Summary of Markov Process Results III Transition probabilities given by where R has entries R ij = P(t) = e Rt { q ij v i Process is a Birth and Death process if i j i = j P ij = 0 if i j > 1 In this case we write λ i for the instantaneous birth rate: P(X(t +h) = i +1 X t = i) = λ i h+o(h) and µ i for the instantaneous death rate: P(X(t +h) = i 1 X t = i) = µ i h+o(h) Richard Lockhart (Simon Fraser University)STAT 380 Continuous Time Markov Chains Spring 2018 8 / 35

Summary of Markov Process Results IV We have 0 i j > 1 q ij = λ i j = i +1 µ i j = i 1 If all µ i = 0 then process is a pure birth process. If all λ i = 0 a pure death process. Birth and Death process have stationary distribution π n = λ ( 0 λ n 1 µ 1 µ n 1+ ) λ 0 λ n 1 n=1 µ 1 µ n Richard Lockhart (Simon Fraser University)STAT 380 Continuous Time Markov Chains Spring 2018 9 / 35

Summary of Markov Process Results V Necessary condition for existence of π is n=1 λ 0 λ n 1 µ 1 µ n < Linear birth and death processes have In this case (note immigration) provided λ < µ. λ n = (n+1)λ µ n = nµ π n = (λ/µ) n (1 (λ/µ)) Richard Lockhart (Simon Fraser University)STAT 380 Continuous Time Markov Chains Spring 2018 10 / 35

Details Suppose X a Markov Chain with stationary transitions. Then P(X(t +s) = k X(0) = i) = j P(X(t +s) = k,x(t) = j X(0) = i) = j = j = j P(X(t +s) = k X(t) = j,x(0) = i) P(X(t) = j X(0) = i) P(X(t +s) = k X(t) = j) P(X(t) = j X(0) = i) P(X(s) = k X(0) = j) P(X(t) = j X(0) = i) This shows P(t +s) = P(t)P(s) which is the Chapman-Kolmogorov equation. Richard Lockhart (Simon Fraser University)STAT 380 Continuous Time Markov Chains Spring 2018 11 / 35

Holding Times Now consider the chain starting from i and let T i be the first t for which X(t) i. Then T i is a stopping time. T i called a holding time. Technically, for each t, Then, by the Markov property, {T i t} H t. P(T i > t +s T i > s,x(0) = i) = P(T i > t +s X(u) = i;0 u s) = P(T i > t X(0) = i) Given X(0) = i, T i has memoryless property so T i has exponential distribution. Let v i be the rate parameter. Richard Lockhart (Simon Fraser University)STAT 380 Continuous Time Markov Chains Spring 2018 12 / 35

Embedded Chain: Skeleton Let S 1 < S 2 < be the stopping times at which transitions occur. Then X n = X(S n ). The sequence X n is a Markov chain by the Markov property. That P ii = 0 reflects the fact that P(X(S n+1 ) = X(S n )) = 0 by design. As before we say i j if P ij (t) > 0 for some t. It is fairly clear that i j for the X(t) if and only if i j for the embedded chain X n. We say i j (i and j communicate) if i j and j i. Richard Lockhart (Simon Fraser University)STAT 380 Continuous Time Markov Chains Spring 2018 13 / 35

Towards coupled differential equations Now consider P(X(t +h) = j X(t) = i,h t ) Suppose the chain has made n transitions so far so that S n < t < S n+1. Then the event X(t +h) = j is, except for possibilities of probability o(h) the event that The probability of this is t < S n+1 t +h and X n+1 = j (v i h+o(h))p ij = v i P ij h+o(h) Richard Lockhart (Simon Fraser University)STAT 380 Continuous Time Markov Chains Spring 2018 14 / 35

Kolmogorov s Equations The Chapman-Kolmogorov equations are P(t +h) = P(t)P(h) Subtract P(t) from both sides, divide by h and let h 0. Remember that P(0) is the identity. We find which gives P(t +h) P(t) h = P(t)(P(h) P(0)) h P (t) = P(t)P (0) The Chapman-Kolmogorov equations can also be written P(t +h) = P(h)P(t) Now subtracting P(t) from both sides, dividing by h and letting h 0 gives P (t) = P (0)P(t) Richard Lockhart (Simon Fraser University)STAT 380 Continuous Time Markov Chains Spring 2018 15 / 35

Forward and Backward Equations Look at these equations in component form: P (t) = P (0)P(t) becomes P ij(t) = P ik (0)P kj(t) k For i k our calculations of instantaneous transition rates gives For i = k we have P ik (0) = v ip ik P(X(h) = i X(0) = i) = e v ih +o(h) (X(h) = i either means T i > h which has probability e v ih or there have been two or more transitions in [0, h], a possibility of probability o(h).) Thus P ii (0) = v i Richard Lockhart (Simon Fraser University)STAT 380 Continuous Time Markov Chains Spring 2018 16 / 35

Forward and Backward Equations II Let R be the matrix with entries { q ij v i P ij R ij = v i That is R = P (0). i j i = j R is the infinitesimal generator of the chain. Thus P (t) = P (0)P(t) becomes P ij (t) = k = k i R ik P kj (t) q ik P kj (t) v i P ij (t) Called Kolmogorov s backward equations. Richard Lockhart (Simon Fraser University)STAT 380 Continuous Time Markov Chains Spring 2018 17 / 35

Forwards equations On the other hand P (t) = P(t)P (0) becomes P ij (t) = k P ik (t)r kj = k j q kj P ik (t) v j P ij (t) These are Kolmogorov s forward equations. Remark: When the state space is infinite the forward equations may not be justified. In deriving them we interchanged a limit with an infinite sum; Interchange always justified for backward equations but not for forward. Richard Lockhart (Simon Fraser University)STAT 380 Continuous Time Markov Chains Spring 2018 18 / 35

Example Example: S = {0,1}. Then P = [ 0 1 1 0 and the chain is otherwise specified by v 0 and v 1. The matrix R is ] [ ] v0 v R = 0 v 1 v 1 Richard Lockhart (Simon Fraser University)STAT 380 Continuous Time Markov Chains Spring 2018 19 / 35

Example: Kolmogorov equations The backward equations become P 00(t) = v 0 P 10 (t) v 0 P 00 (t) P 01(t) = v 0 P 11 (t) v 0 P 01 (t) P 10(t) = v 1 P 00 (t) v 1 P 10 (t) P 11 (t) = v 1P 01 (t) v 1 P 11 (t) while the forward equations are P 00(t) = v 1 P 01 (t) v 0 P 00 (t) P 01(t) = v 0 P 00 (t) v 1 P 01 (t) P 10(t) = v 1 P 11 (t) v 0 P 10 (t) P 11 (t) = v 0P 10 (t) v 1 P 11 (t) Richard Lockhart (Simon Fraser University)STAT 380 Continuous Time Markov Chains Spring 2018 20 / 35

Solving Kolmogorov s equations Add v 1 first plus v 0 third backward equations to get v 1 P 00 (t)+v 0P 10 (t) = 0 so v 1 P 00 (t)+v 0 P 10 (t) = c Put t = 0 to get c = v 1. This gives P 10 (t) = v 1 v 0 {1 P 00 (t)} Richard Lockhart (Simon Fraser University)STAT 380 Continuous Time Markov Chains Spring 2018 21 / 35

Solving Kolmogorov s equations Plug this back in to the first equation and get Multiply by e (v 1+v 0 )t and get which can be integrated to get P 00 (t) = v 1 (v 1 +v 0 )P 00 (t) { } e (v 1+v 0 )t P 00 (t) = v1 e (v 1+v 0 )t P 00 (t) = v 1 v 0 +v 1 + v 0 v 0 +v 1 e (v 1+v 0 )t Richard Lockhart (Simon Fraser University)STAT 380 Continuous Time Markov Chains Spring 2018 22 / 35

Matrix exponentials Alternative calculation: can be written as where and [ ] v0 v R = 0 v 1 v 1 MΛM 1 1 v 0 M = 1 v 1 M 1 = v 1 v 0 +v 1 v 0 v 0 +v 1 1 v 0 +v 1 1 v 0 +v 1 0 0 Λ = 0 (v 0 +v 1 ) Richard Lockhart (Simon Fraser University)STAT 380 Continuous Time Markov Chains Spring 2018 23 / 35

Matrix exponentials Then Now e Rt = 0 = R n t n /n! 0 ( MΛM 1 ) n t n 0 = M ( 0 Λ ntn n! n! ) M 1 [ 1 0 Λ ntn n! = 0 e (v 0+v 1 )t ] Richard Lockhart (Simon Fraser University)STAT 380 Continuous Time Markov Chains Spring 2018 24 / 35

Matrix exponentials So P(t) = e Rt [ 1 0 = M 0 e (v 0+v 1 )t = P e (v 0+v 1 )t v 0 +v 1 R ] M 1 where P = v 1 v 0 +v 1 v 0 v 0 +v 1 v 1 v 0 +v 1 v 0 v 0 +v 1 Richard Lockhart (Simon Fraser University)STAT 380 Continuous Time Markov Chains Spring 2018 25 / 35

Matrix exponentials Notice: rows of P are a stationary initial distribution. If rows are π then so P = [ 1 1 ] π 1π πp = (π1)π = π Moreover πr = 0 Fact: π 0 = v 1 /(v 0 +v 1 ) is long run fraction of time in state 0. Fact: 1 T f(x(t))dt T 0 j Ergodic Theorem in continuous time. π j f(j) Richard Lockhart (Simon Fraser University)STAT 380 Continuous Time Markov Chains Spring 2018 26 / 35

Birth and Death Processes Consider a population of X(t) individuals. Suppose in next time interval (t,t +h) probability of population increase of 1 (called a birth) is λ i h+o(h) Suppose probability of decrease of 1 (death) is µ i h+o(h). Jargon: X is a birth and death process. Special cases: All µi = 0; called a pure birth process. All λi = 0 (0 is absorbing): pure death process. λn = nλ and µ n = nµ is a linear birth and death process. λn λ, µ n 0: Poisson Process. λn = nλ+θ and µ n = nµ is a linear birth and death process with immigration. Richard Lockhart (Simon Fraser University)STAT 380 Continuous Time Markov Chains Spring 2018 27 / 35

Example: Queuing Theory Ingredients of Queuing Problem: Queue input process. Number of servers Queue discipline: first come first serve? last in first out? pre-emptive priorities? Service time distribution. Richard Lockhart (Simon Fraser University)STAT 380 Continuous Time Markov Chains Spring 2018 28 / 35

Example Customers arrive at a facility at times of a Poisson Process Call this input process N, with rate λ. Denote Poisson inputs with M (for Markov) in queueing literature. Single server case: Service distribution: exponential service times, rate µ. Queue discipline: first come first serve. Richard Lockhart (Simon Fraser University)STAT 380 Continuous Time Markov Chains Spring 2018 29 / 35

Analysis via Continuous Time Markov Chain Let X(t) = number of customers in line at time t. Then X is a Markov process called M/M/1 queue. Departure rate from a state: v i = λ+µ Skeleton chain transition probabilities: µ µ+λ j = i 1 0 P ij = λ µ+λ j = i +1 0 otherwise Richard Lockhart (Simon Fraser University)STAT 380 Continuous Time Markov Chains Spring 2018 30 / 35

Example: M/M/ queue Customers arrive according to PP rate λ. Each customer begins service immediately. X(t) is number being served at time t. X is a birth and death process with v n = λ+nµ and iµ iµ+λ j = i 1 0 P ij = λ iµ+λ j = i +1 0 otherwise Richard Lockhart (Simon Fraser University)STAT 380 Continuous Time Markov Chains Spring 2018 31 / 35

Stationary Initial Distributions Recall: stationary initial distribution π is a probability vector solving Rewrite this as πr = 0 v j π j = i j q ij π i Interpretation: LHS is rate at which process leaves state j. Process is in state j a fraction π j of time, then makes transition at rate v j. RHS is total rate of arrival in state j. For each state i j π i is fraction of time spent in state i and then q ij the instantaneous rate of transition from i to j. Richard Lockhart (Simon Fraser University)STAT 380 Continuous Time Markov Chains Spring 2018 32 / 35

Interpretation of equation So equation says: Rate of departure from j balances rate of arrival to j. This is called balance. Application to birth and death processes: Equation is (λ j +µ j )π j = λ j 1 π j 1 +µ j+1 π j+1 for j 1 and λ 0 π 0 = µ 1 π 1 Richard Lockhart (Simon Fraser University)STAT 380 Continuous Time Markov Chains Spring 2018 33 / 35

Stationary Distribution for Birth and Death Equations permit recursion: π 1 = λ 0 µ 1 π 0 π 2 = λ 1 +µ 1 µ 2 π 1 λ 0 µ 2 π 0 = λ 0λ 1 µ 1 µ 2 π 0 Extend by induction to Apply π k = 1 to get π 0 (1+ π k = λ 0 λ k 1 µ 1 µ k π 0 n=1 λ 0 λ n 1 µ 1 µ n ) = 1 This gives the formula announced: λ 0 λ n 1 π n = ( µ 1 µ n 1+ ) λ 0 λ n 1 n=1 µ 1 µ n Richard Lockhart (Simon Fraser University)STAT 380 Continuous Time Markov Chains Spring 2018 34 / 35

Interpretation Richard should add warnings If n=1 λ 0 λ n 1 µ 1 µ n < then we have defined a probability vector which solves πr = 0 Since we see that P = RP {πp(t)} = 0 so that πp(t) is constant. Put t = 0 to discover that the constant is π. Richard Lockhart (Simon Fraser University)STAT 380 Continuous Time Markov Chains Spring 2018 35 / 35