Probability Distributions

Size: px
Start display at page:

Download "Probability Distributions"

Transcription

1 Lecture 1: Background in Probability Theory Probability Distributions The probability mass function (pmf) or probability density functions (pdf), mean, µ, variance, σ 2, and moment generating function (mgf) M(t) for some well-known discrete and continuous probability distributions Discrete Distributions Discrete Uniform: f(x) = { 1, n x = 1, 2,, n, 0, otherwise µ = 1 + n, σ 2 = n , M(t) = e(n+1)t e t n(e t 1), t 0 Geometric: f(x) = { p(1 p) x, x = 0, 1, 2,, 0, otherwise, where 0 < p < 1 µ = 1 p p, σ2 = 1 p p 2, M(t) = Binomial b(n, p): f(x) = ( ) n 0 < p < 1 The notation = x p 1 (1 p)e t ( ) n p x (1 p) n x, x = 0, 1, 2,, n, 0, otherwise, n! x!(n x)! where n is a positive integer and µ = np, σ 2 = np(1 p), M(t) = (1 p + pe t ) n where n is a positive inte- ( ) x + n 1 p Negative Binomial: f(x) = n 1 n (1 p) x, x = 0, 1, 2,, 0, otherwise, ger and 0 < p < 1 µ = n(1 p), σ 2 n(1 p) = p p 2, M(t) = p n [1 (1 p)e t ] n λ x e λ Poisson, P o(λ): f(x) =, x = 0, 1, 2,, x! 0, otherwise, µ = λ, σ 2 = λ, M(t) = e λ(et 1) where λ is a positive constant Continuous Distributions Uniform U(a, b): f(x) = 1 b a, a x b, 0, otherwise, where a < b are constants µ = a + b 2, σ2 = (b a)2, M(t) = ebt e at 12 t(b a), t 0 1

2 Gamma: f(x) = 0 1 Γ(α)β α xα 1 e x/β, x 0, 0, x < 0, 1 β α xα 1 e x/β dx For a positive integer n, Γ(n) = (n 1)! µ = αβ, σ 2 = αβ 2, M(t) = 1 (1 βt) α, t < 1/β where α and β are positive constants and Γ(α) = Exponential: f(x) = { λe λx, x 0, 0, x < 0, where λ is a positive constant µ = 1 λ, σ2 = 1 λ 2, M(t) = λ λ t, t < λ Normal, N(µ, σ 2 ): f(x) = 1 ( ) σ 2π exp (x µ)2, < x <, where µ and σ are constants 2σ 2 E(X) = µ, V ar(x) = σ 2, M(t) = e µt+σ2 t 2 /2 The probability generating function (pgf) is P X (t) = E(t X ), moment generating function (mgf) is M X (t) = E(e tx ) and cumulant generating function (cgf) is K X (t) = ln(m X (t)) The mean µ X = E(X) and variance σ 2 X = E(X2 ) E 2 (X) computed from the pgf, P X (t), mgf, M X (t) and cgf, K X (t): µ X = P X(1) = M X(0) = K X(0) and P X (1) + P X (1) [P X (1)]2 σx 2 = M X (0) [M X (0)]2 K X (0) Continuous-time Markov Chain Example, Simple Birth Process: The per capita rate of birth is λ = Population size Time Figure 1: Three sample paths (stochastic realizations) of a simple birth process, X(0) = 1; n(t) = e t is the dashed curve See Table 1 2

3 Table 1: For two stochastic realizations, the times at which a birth occurs for a simple birth process Realization 1 Realization 2 Size X(t) Event Time t Size X(t) Event Time t

4 Lecture 2: Discrete-Time Markov Chains (DTMCs) Transition probabilities: p ji (n) = Prob{X n+1 = j X n = i} If p ji (n) does not depend on n, then the process is said to be time homogeneous The transition matrix of a DTMC {X n } n=0 with state space {1, 2, } and one-step transition probabilities, {p ij} i,j=1, is denoted as P = (p ij ), where p 11 p 12 p 13 p 21 p 22 p 23 P = p 31 p 32 p 33 The column sums equal one, a stochastic matrix, j=1 p ji = 1 The n-step transition matrix P n = (p (n) The probabilities p i (n) = Prob{X n = i}, p i (n + 1) = j=1 p ijp j (n), i = 1, 2 p(n + 1) = P p(n) Random walk model with absorbing barriers See the directed graph in Figure 2 and the corresponding (N + 1) (N + 1) transition matrix: 1 q q p p P = q p p 1 The Markov chain, graphed in Figure 2, has three communication classes: {0}, {1, 2,, N 1}, and {N} The Markov chain is reducible States 0 and N are absorbing; the remaining states are transient ij ) N Figure 2: Probability of moving to right is p and to the left is q, p + q = 1 Boundaries 0 and N are absorbing, p 00 = 1 = p NN (random walk with absorbing barriers or gambler s ruin problem) A DTMC is irreducible if its digraph is strongly connected Otherwise it is called reducible An irreducible DTMC can be positive recurrent or null recurrent or transient It may also be classified as periodic or aperiodic Recurrence is defined for each state i in the chain Recurrence means for each state i if the process leaves state i it will return to state i at some future time If not, the state is transient A state i is positive recurrent if the mean recurrence time (µ ii mean return time) is finite A state i with an infinite mean recurrence time is called null recurrent be a recurrent, ir- Theorem 1 (Basic Limit Theorem for aperiodic Markov chains) Let {X n } n=0 reducible, and aperiodic DTMC with transition matrix P = (p ij ) Then lim n p(n) ij = 1, µ ii where µ ii is the mean recurrence time for state i and i and j are any states of the chain [If µ ii =, then lim n p (n) ij = 0] 4

5 Theorem 2 (Basic Limit Theorem for periodic Markov chains) Let {X n } n=0 be a recurrent, irreducible, and d-periodic DTMC, d > 1, with transition matrix P = (p ij ) Then lim n p(nd) ii = d µ ii and p (m) ii = 0 if m is not a multiple of d, where µ ii is the mean recurrence time for state i [If µ ii =, then lim n p (nd) ii = 0] Random Walk with Absorbing Boundaries or Gambler s Ruin Let s return to the random walk model with absorbing boundaries This DTMC does not satisfy either of the Basic Limit Theorems because it is a reducible Markov chain with three communicating classes, two absorbing states {0} and {N} and the transient class {1, 2,, N 1} Interesting questions about problems with absorbing states such as an extinction state in population dynamics are (1) what is the probability of absorption? and (2) what is the expected time until absorption? We will investigate this question with the random walk model and with genetics of inbreeding for a finite DTMC Later, we will investigate the probability of extinction in a population model via a branching process Consider the random walk model with absorbing boundaries at 0 and N In the gambler s ruin problem, state 0 is ruin and state N is winning the maximum amount of money In either case, after hitting 0 or N, you stop playing the game Let a k be the probability of ruin (absorption at k = 0) beginning with a capital of k: a k = pa k+1 + qa k 1 (1) for 1 k N 1 Equation (1) is a second-order difference equation in a k The difference equation can be written as pa k+1 a k + qa k 1 = 0, k = 1,, N 1 (2) This method of deriving equation (1) is referred to as a first-step analysis (Taylor and Karlin, 1998) This difference equation (2) can be solved with the following boundary conditions Alternately, it will be shown that a 0 = 1 and a N = 0 (a 1,, a N 1 ) = (p 01, p 02,, p 0,N 1 )(I T ) 1 In general, the difference equation for probability of absorption has the form: a k = N p ik a i To find the probability of absorption into state N, the boundary conditions are a 0 = 0 and a N = 1 k=0 Expected Time until Absorption: Let τ k be the expected time until absorption at either 0 or N in the gambler s ruin problem The difference equation for τ k has the following form: for k = 1, 2,, N 1 or τ k = p(1 + τ k+1 ) + q(1 + τ k 1 ), pτ k+1 τ k + qτ k 1 = 1, (3) The boundary conditions are τ 0 = 0 = τ N In the gambler s ruin problem, this equation can be solved explicitly But in general, N N τ k = p ik (1 + τ i ) or τ k p ik τ i = 1 i=0 5 i=0

6 It will be shown that (τ 1, τ 2,, τ N 1 ) = 1(I T ) 1, 1 = (1, 1,, 1) Table 2: Gambler s ruin problem with a beginning capital of k = 50 and a total capital of N = 100 Prob a 50 b 50 τ 50 A 50 (1) B 50 (1) q = q = q = q = Capital Games Figure 3: Three sample paths for the gambler s ruin problem when N = 100, k = 50, and q = 055 Example of Genetics of Inbreeding: Two alleles A and a There are six possible breeding pairs which denote the six states of the DTMC, 1: AA AA, 2 aa aa, 3 Aa Aa, 4 Aa aa, 5 AA aa, 6 AA Aa Inbreeding of the first two types results in offspring of the same genotypes and inbreeding in the next generation will be of the same type; they are absorbing states The remaining states, 3,4,5,6 are transient The transition matrix has the following form: 1 f 0 1/ / /16 1/4 0 0 ( ) P = 0 0 1/4 1/4 1 1/4 I A 0 0 1/4 1/2 0 0 = O T 0 0 1/ / /2 Probability of absorption into states 1 or 2 from states 3, 4, 5, 6 is computed from the fundamental matrix, (I T ) 1, ( lim P n I (A + AT + AT = 2 ) ( ) + ) I A(I T ) 1 = n O O O O Thus, A(I T ) 1 is the probability of absorption into states 1 or 2 (fixation) from states 3, 4, 5 or 6 Let 1 = (1, 1,, 1), then 1(I T ) 1 is the mean time until absorption from states 3, 4, 5, or 6 6

7 Lecture 3: Discrete-Time Branching Processes Figure 4: Sample path of a branching process {X n } n=0 In the first generation, four individuals are born, X 1 = 4 The four individuals in generation one give birth to three, zero, four, and one individuals, respectively, making a total of eight individuals in generation two, X 2 = 8 Generating functions rather than transition matrices are useful in analysis of branching processes Offspring pgf: f(t) = p k t k k=0 Recall f(1) = 1 and m = f (1) is the mean number of offspring The branching process is called subcritical if m < 1, critical if m = 1, and supercritical if m > 1 Theorem 3 (Branching Process Theorem) Let X 0 = 1 Assume f(0) = p 0 > 0 and p 0 + p 1 < 1 (i) If m 1, then lim n Prob{X n = 0} = 1 (ii) If m > 1, then lim n Prob{X n = 0} = q, where q = f(q) is the unique fixed point in the interval (0, 1) To verify this theorem, we show that given X 0 = 1, then P Xn (t) = f n 1 (f(t)) = f n (t), the pgf of X n is the n-fold composition of the offspring pgf Then p 0 (n) = P Xn (0) = f n (0) and p 0 (n) = f(p 0 (n 1)) The sequence {p 0 (n)} is an increasing function, bounded above by one and therefore, has a limit q This limit is a fixed point of f, f(q) = q If X 0 = N and m > 1, then lim n p 0 (n) = lim n Prob{X n = 0} = q N The conditional expectation E(X n+1 X n ) = mx n, E(X n+1 ) = me(x n ) If the growth rate varies with n, with a mean m n, then E(X n+1 ) = m n E(X n ) Multitype Branching Process, n different types: (X 1,, X n ) The offspring random variable of type i is Y i Offspring pgfs: f i (t 1,, n) = Expectation matrix M = (m ij ), where s n s 1 P i (s 1,, s n )t s1 1 tsn n P i (s 1,, s n ) = Prob{Y 1 = s 1,, Y n = s n } m ji = f t j t1 =1,,t n=1 Assume M is irreducible Denote the spectral radius of M as ρ(m), the maximum modulus of the eigenvalues of MThe multitype branching process is called subcritical if ρ(m) < 1, critical if ρ(m) = 1 and supercritical if ρ(m) > 1 See Chapter on Branching Processes 7

8 Lecture 4: Continuous-Time Markov Chains (CTMCs), Introduction Discrete random variable X(t), t [0, ) Probabilities p i (t) = Prob{X(t) = i} Transition probability: p ji (t, s) = Prob{X(t) = j X(s) = i}, s < t We will assume time-homogenous transition probabilities p ji (t, s) = p ji (t s) The transition matrix is a stochastic matrix: where p 00 (t) p 01 (t) P (t) = (p ij (t)) = p 10 (t) p 11 (t), p ji ( t) = δ ji + q ji t + o( t) is an infinitesimal transition probability The infinitesimal generator matrix: q 00 q 01 Q = (q ij ) = q 10 q 11 P ( t) I, Q = lim t 0 t The column sums of Q equal zero dp (t) Forward Kolmogorov differential equations: = QP (t) dt dp (t) Backward Kolmogorov differential equations: = P (t)q dt In physics and chemistry, the forward Kolmogorov differential equations are often referred to as the Master equation The embedded DTMC is used to define irreducible, recurrent, and transient states or chains for the associated CTMC Let Y n denote the random variable for the state of a CTMC {X(t) : t [0, )} at the time of the nth jump, Y n = X(W n ) (See Figure 5) The set of discrete random variables {Y n } 0 is the embedded Markov chain T0 T1 T2 T3 0 W1 W2 W3 W4 Figure 5: Sample path of a CTMC, illustrating waiting times {W i } and interevent times, {T i } A CTMC is irreducible, recurrent or transient if the corresponding embedded Markov chain has these properties Some differences in the dynamics of a CTMC as opposed to a DTMC are the possibility of a finite-time blow up in a CTMC (explosive process) and the fact that CTMC are not periodic See Figure 6 The embedded MC cannot be used to classify chains as positive recurrent or null recurrent This latter classification depends on the mean recurrence time µ ii 8

9 0 W1 W2 W3 W4 W Figure 6: One sample path of a continuous time Markov chain that is explosive Theorem 4 (Basic Limit Theorem for CTMCs) If the CTMC {X(t) : t [0, )} is nonexplosive and irreducible, then for all i and j, lim p ij(t) = 1, (4) t q ii µ ii where µ ii is the mean recurrence time, 0 < µ ii In particular, a finite, irreducible CTMC is nonexplosive and the limit (4) exists and is positive If the DTMC is nonexplosive and positive recurrent, it has a limiting positive stationary distribution π satisfying Qπ = 0, where π i = 1 q ii µ ii Poisson process with X(0) = 0, p i+1,i ( t) = λ t + o( t) and p i (t) = e λt (λt) i /i! has generator matrix λ 0 0 λ λ 0 Q = 0 λ λ The associated embedded Markov chain has a transition matrix T = The Poisson process is transient A finite CTMC with two states {1, 2} The generator matrix ( ) a b Q =, a, b > 0 a b The CTMC is irreducible and positive recurrent The limiting stationary distribution can be found from the forward Kolmogorov differential equations by solving for the stationary distribution Qπ = 0 In this case π = (b/(a + b), a/(a + b)) tr The mean recurrence times are µ ii = a + b, i = 1, 2 ab 9

10 Lecture 5: Continuous-Time Markov Chains (CTMCs), Interevent Time To generate sample paths, we must know the time between jumps and the state to which the process jumps The Markov assumption implies the interevent time is exponentially distributed because the exponential distribution has the memoryless property Let T i be the continuous random variable for the time until the i + 1st event See Figure 5 Theorem 5 (Interevent Time) Assume j n p jn( t) = α(n) t + o( t) Then the cumulative distribution function for the interevent time T i is F i (t) = 1 exp( α(n)t) with mean and variance µ Ti = 1 α(n) and σ 2 T i = 1 [α(n)] 2 Theorem 6 (Interevent Time Simulation) Let U U[0, 1] be the uniform distribution on [0,1] and T i the continuous random variable for interevent time with state space [0, ) Then T i = Fi 1 (U) = ln(u) α(n) Simple Birth and Death Markov Chain: In the simple birth and death process, an event can be a birth or a death Let X(0) = N The infinitesimal transition probabilities are p i+j,i ( t) = Prob{ X(t) = j X(t) = i} µi t + o( t), j = 1 λi t + o( t), j = 1 = 1 (λ + µ)i t + o( t), j = 0 o( t), j 1, 0, 1 Use two random numbers, u 1 and u 2, from the uniform distribution U(0, 1) to determine the interevent time and the state to which the process jumps In MATLAB, indices begin from 1, so instead of writing t(0), we use t(1) Consider the simple birth and death chain, in a MATLAB program, t(1) = 0, and the time to the next event is t(2) = t(1)+ln(u 1 )/(α(n)), where α(n) = λn+µn, given the process is in state n Since there are two events, to determine whether there is a birth or a death, the unit interval is divided into two subintervals, one subinterval has probability λ/(λ + µ) and the other has probability µ/(λ + µ) Generate a uniform random number u 2 If u 2 < λ/(λ + µ), then this random number lies in the first subinterval and there is a birth, otherwise if u 2 > λ/(λ + µ), the random number lies in the second subinterval and there is a death This concept can be easily extended to k > 2 events In a MATLAB program with k events the unit interval must be divided into k subintervals, each with predetermined probability for i = 1,, k that depends on the current state and the transition probabilities For example, suppose there are four events with the following rates a i (n), i = 1, 2, 3, 4 which depend on the current state n The probabilities of these four events are a i (n)/a(n), a(n) = i a i(n), i = 1, 2, 3, 4 The subinterval [0, 1] is subdivided into four subintervals with the following endpoints: Therefore, in a MATLAB program: 0, a 1 a, a 1 + a 2, a a 1 + a 2 + a 3, 1 a if u2<a1/a, then event 1 occurs elseif u2>=a1/a & u2<(a1+a2)/a, then event 2 occurs elseif u2>=(a1+a2)/2 & u2<(a1+a2+a3)/a, then event 3 occurs else u2>=(a1+a2+a3)/a, then event 4 occurs The subintervals change each time the process changes state n If the number of events are large, deciding which event occurs can become quite lengthy and there are ways to speed up the process of selecting a particular event 10

11 %MatLab program: simple birth and death process clear all x0=5; b=1; d=05; % initial and parameter values for j=1:3 % Three sample paths clear x t n=1; t(1,j)=0; x(n)=x0; % starting values while x(n)>0 & x(n)<50; % continue until the process hits zero or reaches 50 u1=rand; u2=rand; % two uniform random numbers t(n+1,j)=-log(u1)/(b*x(n)+d*x(n))+t(n,j); if u2< b/(b+d); x(n+1)=x(n)+1; else x(n+1)=x(n)-1; end n=n+1; end s=stairs(t(:,j),x, r-, Linewidth,2); hold on end xlabel( Time ); ylabel( Population size ); hold off Figure 7: Three sample paths of the simple birth and death process, X(0) = 5, λ = b = 1, µ = d = Population size Time 11

12 Lecture 6: Continuous-Time Birth and Death Processes: The generator matrix Q for a general birth and death process is either λ 0 µ λ 0 λ 1 µ 1 µ 2 0 Q = 0 λ 1 λ 2 µ 2 µ 3, 0 0 λ 2 λ 3 µ 3 or for a finite state space, λ 0 µ λ 0 λ 1 µ 1 µ λ 1 λ 2 µ 2 0 Q = µ N µ N The simple birth, simple death, simple birth and death, and simple birth, death, and immigration processes are linear in the rates, λ i = λi + ν and µ i = µi Applying the forward Kolmogorov differential equations, dp = Qp, first-order partial differential equations for the pgf and mgf can be derived Applying the method of characteristics to the first-order dt partial differential equations, explicit expressions can be found for the pgf and mgf of these processes Denote the pgf for these simple process as follows (note change in notation z and θ because t is time): M(θ, t) = P(z, t) = p i (t)z i i=0 p i (t)e iθ = P(e θ, t) i=0 Simple Birth, Death, and Immigration Process: Let X(0) = N The infinitesimal transition probabilities are p i+j,i ( t) = Prob{ X(t) = j X(t) = i} µi t + o( t), j = 1 (ν + λi) t + o( t), j = 1 = 1 [ν + (λ + µ)i] t + o( t), j = 0 o( t), j 1, 0, 1 The forward Kolmogorov differential equations are dp i dt dp 0 dt = [λ(i 1) + ν]p i 1 + µ(i + 1)p i+1 (λi + µi + ν)p i = νp 0 + µp 1 for i = 1, 2, with initial condition X(0) = N Applying the generating function technique, it follows that the mgf M(θ, t) is a solution of the following first-order partial differential equation M t = [ ] M λ(e θ 1) + µ(e θ 1) θ + ν(eθ 1)M 12

13 with initial condition M(θ, 0) = e Nθ The preceding differential equation is first-order because the rates are linear The mgf is the solution of this first-order partial differential equation, M(θ, t) = (λ µ)ν/λ [ µ(e (λ µ)t 1) e θ (µe (λ µ)t λ) ] N [ (λe (λ µ)t µ) λ(e (λ µ)t 1)e θ] N+ν/λ The moments E(X n (t)) of the probability distribution X(t) can be found by differentiating the mgf with respect to θ and evaluating at θ = 0: E(X n (t)) = n M(θ, t) θ n θ=0 Table 3: Mean, variance, and pgf for the simple birth, simple death, and simple birth and death processes, where X(0) = N and ρ = e (λ µ)t, λ µ Simple Simple Simple Birth Death Birth and Death m(t) Ne λt Ne µt Ne (λ µ)t σ 2 (t) Ne 2λt (1 e λt ) Ne µt (1 e µt ) N λ + µ ρ(ρ 1) λ µ (pz) N (1 z(1 p)) N (1 p + pz) N P(z, t) Negative binomial p = e λt Binomial b(n,p) p = e µt ( ρ 1 ) N (λz µ) µ(z 1) ρ 1 (λz µ) λ(z 1) Table 4: Mean, variance, and pgf X(0) = N and ρ = e (λ µ)t, λ µ for the simple birth and death with immigration process, where m(t) Simple Birth and Death with Immigration ρ[n(λ µ) + ν] ν λ µ σ 2 (t) N (λ2 µ 2 )ρ[ρ 1] µ + ρ(λρ µ λ) (λ µ) 2 + ν (λ µ) 2 P(z, t) (λ µ) ν/λ [µ(ρ 1) z(µρ λ)] N [λρ µ λ(ρ 1)z] N+ν/λ For each of the preceding simple birth and death processes, the probability of extinction p 0 (t) can be found by evaluating the pgf at z = 0, P(0, t), and the probability of ultimate extinction by taking the limit: lim t p 0 (t) Except for the stochastic process with immigration, the other birth and death 13

14 processes either hit zero or approach infinity For example, given X(0) = N, for the simple birth process, p 0 (t) = 0, simple death process p 0 (t) = (1 e µt ) N and for the simple birth and death process: lim p 0(t) = t { ( µ ) N, λ λ > µ 1, λ µ dn Logistic Growth Model: (1 dt = rn n ) What is the birth rate and what is the death rate? There K are an infinite number of choices for stochastic birth and death rates that yield the same deterministic logistic growth model The birth and death rates should have the form so that λ n µ n = rn(1 n/k) leads to λ i = b 1 i + b 2 i 2 > 0 and µ i = d 1 i + d 2 i 2 > 0 dn dt = (b 1 d 1 )n + (b 2 d 2 )n 2 = rn r K n2, b 1 d 1 = r > 0 and b 1 d 1 d 2 b 2 = K > 0 Define the birth and death rates, λ i and µ i, as follows: (a) λ i = i and µ i = i2, i = 0, 1, 2, 10 (b) λ i = i i2, i = 0, 1,, and µ i = i2, i = 0, 1, 2, 0, i > In both cases the deterministic model is dn (1 dt = n n ), 10 where r = 1 and K = 10 In the deterministic model, solutions approach the carrying capacity K = 10 Three sample paths for models (a) and (b) are graphed in Figure 8 20 (a) 20 (b) Population size Population size Time Time Figure 8: Three sample paths of the stochastic logistic model for cases (a) and (b) with X(0) = 10 For these stochastic logistic models, lim p 0(t) = 1 t 14

15 For a population of finite size N, we will show that the mean time to extinction is a solution of the following matrix equation, τ Q = 1 where Q is the truncated generator matrix without state zero, τ = (τ 1,, τ N ), τ k is the mean time until extinction from state k, τ = 1 Q 1 Let λ i = i i2 N, i = 0, 1,, N and µ i = i2, i = 0, 1, 2,, N 0, i > N N The intrinsic growth rate r = 1 and the carrying capacity K = N/2 The expected time to extinction can be calculated for X(0) = m, m = 1, 2,, N by solving τ Q = 1 When N = 10, the carrying capacity is K = 5, and when N = 20, the carrying capacity is K = 10 The time units depend on the particular problem If the time is measured in days, then for the first example, extinction occurs, on the average, in less than one year but for the second example, the mean time to extinction is over 300 years Hence, for even larger values of K, the mean time to extinction will take much longer; the convergence of p 0 (t) to 1 is slow x 105 Expected time to extinction K = 5 Expected time to extinction K = Initial population size Initial population size Figure 9: Expected time until extinction in the stochastic logistic model with K = 5 and K = 10 15

16 Lecture 7 Biological Applications of CTMC Models: Let P(z, t) be the pgf for X(t), and f(t) be the offspring pgf, P(z, t) = p 0 (t) + p 1 (t)z + p 2 (t)z 2 + f(z) = p 0 + p 1 z + p 2 z 2 + As in discrete-time branching processes, the pgf P(z, t + t) is a composition of pgfs At time t = 0, P(z, 0) = z or X(0) = 1 The same theorem applies to the continuous-time Markov branching process as in discrete-time Theorem 7 Suppose {X(t) : t [0, )} is a nonexplosive, continuous-time Markov branching process with X(0) = 1 Assume f is the pgf of the offspring distribution, where m = f (1) and P(z, t) is the pgf of X(t) [f(0) = p 0 > 0 and p 0 + p 1 < 1] If m 1, then lim Prob{X(t) = 0} = lim p 0(t) = 1 t t and if m > 1, then there exists a q satisfying f(q) = q such that lim Prob{X(t) = 0} = lim p 0(t) = q < 1 t t Simple Birth and Death Process: Let λ > 0 be the per capita birth rate and µ > 0 be the per capita death rate of an individual, p i+1,i ( t) = λi t + o( t) and p i 1,i ( t) = µi t + o( t) The offspring pgf is f(z) = µ λ + µ + λ λ + µ z2 An individual either dies or survives and gives birth There is an important difference in this offspring pgf and the one for the discrete-time model For this pgf, a single offspring does not replace the parent but adds to the population size Also, m = f (1) = 2λ/(µ + λ) > 1 if λ > µ In this case, the fixed point q of f(q) = q is µ/λ The pgf of X(t) for the simple birth and death process is given in Table 3 SIS Epidemic Model ds dt = β SI + γi N di dt = β SI γi = N (β SN ) ( γ I = β γ β I ) I N because S + I = N The value R 0 = β is called the basic reproduction number This system has two γ equilibria, ( S, Ī) = (N, 0) and (N/R 0, N(1 1/R 0 ) if R 0 > 1 If R 0 > 1, then there is an outbreak and solutions approach the endemic equilibrium and if R 0 1, there is no outbreak For the CTMC SIS epidemic model, let Prob{ I(t) = j I(t) = i} β i(n i) t + o( t), j = 1 N γi t [ + o( t), ] j = 1 = β 1 i(n i) + γi t N +o( t), j = 0 o( t), j 1, 0, 1, 16

17 where i {0, 1,, N} The SIS epidemic model is a birth and death process with λ i = β N i(n i) and µ i = γi, for i = 0, 1,, N There is a single absorbing state at zero, lim t p 0 (t) = 1 However, before complete extinction, which may take a long time, I(t) behaves like a simple birth and death process Consider only the state I when S N, then di = βi γi = (β γ)i dt Thus, if N is large, I behaves like a simple birth and death process, where β = λ and γ = µ Thus, using branching process theory or the simple birth and death process we can approximate the dynamics at the beginning of an epidemic What is the probability that I hits zero (no outbreak)? The probability of absorption from branching process theory tells us that the probability of absorption as t (no outbreak) given I(0) = 1 is µ/λ = γ/β = 1/R 0 if R 0 > 1 If I(0) = k, then the probability of no outbreak is (1/R 0 ) k The process either hits zero rapidly or grows Even though this is an asymptotic approximation, it is a good for prediction of probability of no outbreak See Exercise 15 for the SIR epidemic model Lectures 8 & 9 Stochastic Epidemic Models, DTMC, CTMC and SDE Models See Chapter on Stochastic Epidemic Models 17

Probability Distributions

Probability Distributions Lecture : Background in Probability Theory Probability Distributions The probability mass function (pmf) or probability density functions (pdf), mean, µ, variance, σ 2, and moment generating function (mgf)

More information

6 Continuous-Time Birth and Death Chains

6 Continuous-Time Birth and Death Chains 6 Continuous-Time Birth and Death Chains Angela Peace Biomathematics II MATH 5355 Spring 2017 Lecture notes follow: Allen, Linda JS. An introduction to stochastic processes with applications to biology.

More information

AARMS Homework Exercises

AARMS Homework Exercises 1 For the gamma distribution, AARMS Homework Exercises (a) Show that the mgf is M(t) = (1 βt) α for t < 1/β (b) Use the mgf to find the mean and variance of the gamma distribution 2 A well-known inequality

More information

Lecture 4a: Continuous-Time Markov Chain Models

Lecture 4a: Continuous-Time Markov Chain Models Lecture 4a: Continuous-Time Markov Chain Models Continuous-time Markov chains are stochastic processes whose time is continuous, t [0, ), but the random variables are discrete. Prominent examples of continuous-time

More information

An Introduction to Stochastic Epidemic Models

An Introduction to Stochastic Epidemic Models An Introduction to Stochastic Epidemic Models Linda J. S. Allen Department of Mathematics and Statistics Texas Tech University Lubbock, Texas 79409-1042, U.S.A. linda.j.allen@ttu.edu 1 Introduction The

More information

2 Discrete-Time Markov Chains

2 Discrete-Time Markov Chains 2 Discrete-Time Markov Chains Angela Peace Biomathematics II MATH 5355 Spring 2017 Lecture notes follow: Allen, Linda JS. An introduction to stochastic processes with applications to biology. CRC Press,

More information

Stochastic modelling of epidemic spread

Stochastic modelling of epidemic spread Stochastic modelling of epidemic spread Julien Arino Centre for Research on Inner City Health St Michael s Hospital Toronto On leave from Department of Mathematics University of Manitoba Julien Arino@umanitoba.ca

More information

1.1 Review of Probability Theory

1.1 Review of Probability Theory 1.1 Review of Probability Theory Angela Peace Biomathemtics II MATH 5355 Spring 2017 Lecture notes follow: Allen, Linda JS. An introduction to stochastic processes with applications to biology. CRC Press,

More information

AN INTRODUCTION TO STOCHASTIC EPIDEMIC MODELS-PART I

AN INTRODUCTION TO STOCHASTIC EPIDEMIC MODELS-PART I AN INTRODUCTION TO STOCHASTIC EPIDEMIC MODELS-PART I Linda J. S. Allen Department of Mathematics and Statistics Texas Tech University Lubbock, Texas U.S.A. 2008 Summer School on Mathematical Modeling of

More information

Derivation of Itô SDE and Relationship to ODE and CTMC Models

Derivation of Itô SDE and Relationship to ODE and CTMC Models Derivation of Itô SDE and Relationship to ODE and CTMC Models Biomathematics II April 23, 2015 Linda J. S. Allen Texas Tech University TTU 1 Euler-Maruyama Method for Numerical Solution of an Itô SDE dx(t)

More information

Part I Stochastic variables and Markov chains

Part I Stochastic variables and Markov chains Part I Stochastic variables and Markov chains Random variables describe the behaviour of a phenomenon independent of any specific sample space Distribution function (cdf, cumulative distribution function)

More information

8. Statistical Equilibrium and Classification of States: Discrete Time Markov Chains

8. Statistical Equilibrium and Classification of States: Discrete Time Markov Chains 8. Statistical Equilibrium and Classification of States: Discrete Time Markov Chains 8.1 Review 8.2 Statistical Equilibrium 8.3 Two-State Markov Chain 8.4 Existence of P ( ) 8.5 Classification of States

More information

Stochastic modelling of epidemic spread

Stochastic modelling of epidemic spread Stochastic modelling of epidemic spread Julien Arino Department of Mathematics University of Manitoba Winnipeg Julien Arino@umanitoba.ca 19 May 2012 1 Introduction 2 Stochastic processes 3 The SIS model

More information

Markov Chains CK eqns Classes Hitting times Rec./trans. Strong Markov Stat. distr. Reversibility * Markov Chains

Markov Chains CK eqns Classes Hitting times Rec./trans. Strong Markov Stat. distr. Reversibility * Markov Chains Markov Chains A random process X is a family {X t : t T } of random variables indexed by some set T. When T = {0, 1, 2,... } one speaks about a discrete-time process, for T = R or T = [0, ) one has a continuous-time

More information

Statistics 150: Spring 2007

Statistics 150: Spring 2007 Statistics 150: Spring 2007 April 23, 2008 0-1 1 Limiting Probabilities If the discrete-time Markov chain with transition probabilities p ij is irreducible and positive recurrent; then the limiting probabilities

More information

Part IA Probability. Definitions. Based on lectures by R. Weber Notes taken by Dexter Chua. Lent 2015

Part IA Probability. Definitions. Based on lectures by R. Weber Notes taken by Dexter Chua. Lent 2015 Part IA Probability Definitions Based on lectures by R. Weber Notes taken by Dexter Chua Lent 2015 These notes are not endorsed by the lecturers, and I have modified them (often significantly) after lectures.

More information

Recap. Probability, stochastic processes, Markov chains. ELEC-C7210 Modeling and analysis of communication networks

Recap. Probability, stochastic processes, Markov chains. ELEC-C7210 Modeling and analysis of communication networks Recap Probability, stochastic processes, Markov chains ELEC-C7210 Modeling and analysis of communication networks 1 Recap: Probability theory important distributions Discrete distributions Geometric distribution

More information

IEOR 6711, HMWK 5, Professor Sigman

IEOR 6711, HMWK 5, Professor Sigman IEOR 6711, HMWK 5, Professor Sigman 1. Semi-Markov processes: Consider an irreducible positive recurrent discrete-time Markov chain {X n } with transition matrix P (P i,j ), i, j S, and finite state space.

More information

4 Branching Processes

4 Branching Processes 4 Branching Processes Organise by generations: Discrete time. If P(no offspring) 0 there is a probability that the process will die out. Let X = number of offspring of an individual p(x) = P(X = x) = offspring

More information

SMSTC (2007/08) Probability.

SMSTC (2007/08) Probability. SMSTC (27/8) Probability www.smstc.ac.uk Contents 12 Markov chains in continuous time 12 1 12.1 Markov property and the Kolmogorov equations.................... 12 2 12.1.1 Finite state space.................................

More information

MARKOV PROCESSES. Valerio Di Valerio

MARKOV PROCESSES. Valerio Di Valerio MARKOV PROCESSES Valerio Di Valerio Stochastic Process Definition: a stochastic process is a collection of random variables {X(t)} indexed by time t T Each X(t) X is a random variable that satisfy some

More information

THE QUEEN S UNIVERSITY OF BELFAST

THE QUEEN S UNIVERSITY OF BELFAST THE QUEEN S UNIVERSITY OF BELFAST 0SOR20 Level 2 Examination Statistics and Operational Research 20 Probability and Distribution Theory Wednesday 4 August 2002 2.30 pm 5.30 pm Examiners { Professor R M

More information

Stochastic process. X, a series of random variables indexed by t

Stochastic process. X, a series of random variables indexed by t Stochastic process X, a series of random variables indexed by t X={X(t), t 0} is a continuous time stochastic process X={X(t), t=0,1, } is a discrete time stochastic process X(t) is the state at time t,

More information

Continuous-Time Markov Chain

Continuous-Time Markov Chain Continuous-Time Markov Chain Consider the process {X(t),t 0} with state space {0, 1, 2,...}. The process {X(t),t 0} is a continuous-time Markov chain if for all s, t 0 and nonnegative integers i, j, x(u),

More information

Lecture 5: Moment generating functions

Lecture 5: Moment generating functions Lecture 5: Moment generating functions Definition 2.3.6. The moment generating function (mgf) of a random variable X is { x e tx f M X (t) = E(e tx X (x) if X has a pmf ) = etx f X (x)dx if X has a pdf

More information

1 Types of stochastic models

1 Types of stochastic models 1 Types of stochastic models Models so far discussed are all deterministic, meaning that, if the present state were perfectly known, it would be possible to predict exactly all future states. We have seen

More information

Irreducibility. Irreducible. every state can be reached from every other state For any i,j, exist an m 0, such that. Absorbing state: p jj =1

Irreducibility. Irreducible. every state can be reached from every other state For any i,j, exist an m 0, such that. Absorbing state: p jj =1 Irreducibility Irreducible every state can be reached from every other state For any i,j, exist an m 0, such that i,j are communicate, if the above condition is valid Irreducible: all states are communicate

More information

Probability Models. 4. What is the definition of the expectation of a discrete random variable?

Probability Models. 4. What is the definition of the expectation of a discrete random variable? 1 Probability Models The list of questions below is provided in order to help you to prepare for the test and exam. It reflects only the theoretical part of the course. You should expect the questions

More information

Outlines. Discrete Time Markov Chain (DTMC) Continuous Time Markov Chain (CTMC)

Outlines. Discrete Time Markov Chain (DTMC) Continuous Time Markov Chain (CTMC) Markov Chains (2) Outlines Discrete Time Markov Chain (DTMC) Continuous Time Markov Chain (CTMC) 2 pj ( n) denotes the pmf of the random variable p ( n) P( X j) j We will only be concerned with homogenous

More information

Lecture Notes 7 Random Processes. Markov Processes Markov Chains. Random Processes

Lecture Notes 7 Random Processes. Markov Processes Markov Chains. Random Processes Lecture Notes 7 Random Processes Definition IID Processes Bernoulli Process Binomial Counting Process Interarrival Time Process Markov Processes Markov Chains Classification of States Steady State Probabilities

More information

Birth and Death Processes. Birth and Death Processes. Linear Growth with Immigration. Limiting Behaviour for Birth and Death Processes

Birth and Death Processes. Birth and Death Processes. Linear Growth with Immigration. Limiting Behaviour for Birth and Death Processes DTU Informatics 247 Stochastic Processes 6, October 27 Today: Limiting behaviour of birth and death processes Birth and death processes with absorbing states Finite state continuous time Markov chains

More information

Lectures on Probability and Statistical Models

Lectures on Probability and Statistical Models Lectures on Probability and Statistical Models Phil Pollett Professor of Mathematics The University of Queensland c These materials can be used for any educational purpose provided they are are not altered

More information

TMA4265 Stochastic processes ST2101 Stochastic simulation and modelling

TMA4265 Stochastic processes ST2101 Stochastic simulation and modelling Norwegian University of Science and Technology Department of Mathematical Sciences Page of 7 English Contact during examination: Øyvind Bakke Telephone: 73 9 8 26, 99 4 673 TMA426 Stochastic processes

More information

Continuous Distributions

Continuous Distributions A normal distribution and other density functions involving exponential forms play the most important role in probability and statistics. They are related in a certain way, as summarized in a diagram later

More information

STAT STOCHASTIC PROCESSES. Contents

STAT STOCHASTIC PROCESSES. Contents STAT 3911 - STOCHASTIC PROCESSES ANDREW TULLOCH Contents 1. Stochastic Processes 2 2. Classification of states 2 3. Limit theorems for Markov chains 4 4. First step analysis 5 5. Branching processes 5

More information

STAT/MATH 395 A - PROBABILITY II UW Winter Quarter Moment functions. x r p X (x) (1) E[X r ] = x r f X (x) dx (2) (x E[X]) r p X (x) (3)

STAT/MATH 395 A - PROBABILITY II UW Winter Quarter Moment functions. x r p X (x) (1) E[X r ] = x r f X (x) dx (2) (x E[X]) r p X (x) (3) STAT/MATH 395 A - PROBABILITY II UW Winter Quarter 07 Néhémy Lim Moment functions Moments of a random variable Definition.. Let X be a rrv on probability space (Ω, A, P). For a given r N, E[X r ], if it

More information

UNIVERSITY OF LONDON IMPERIAL COLLEGE LONDON

UNIVERSITY OF LONDON IMPERIAL COLLEGE LONDON UNIVERSITY OF LONDON IMPERIAL COLLEGE LONDON BSc and MSci EXAMINATIONS (MATHEMATICS) MAY JUNE 23 This paper is also taken for the relevant examination for the Associateship. M3S4/M4S4 (SOLUTIONS) APPLIED

More information

Stat410 Probability and Statistics II (F16)

Stat410 Probability and Statistics II (F16) Stat4 Probability and Statistics II (F6 Exponential, Poisson and Gamma Suppose on average every /λ hours, a Stochastic train arrives at the Random station. Further we assume the waiting time between two

More information

Markov chains. 1 Discrete time Markov chains. c A. J. Ganesh, University of Bristol, 2015

Markov chains. 1 Discrete time Markov chains. c A. J. Ganesh, University of Bristol, 2015 Markov chains c A. J. Ganesh, University of Bristol, 2015 1 Discrete time Markov chains Example: A drunkard is walking home from the pub. There are n lampposts between the pub and his home, at each of

More information

Part IA Probability. Theorems. Based on lectures by R. Weber Notes taken by Dexter Chua. Lent 2015

Part IA Probability. Theorems. Based on lectures by R. Weber Notes taken by Dexter Chua. Lent 2015 Part IA Probability Theorems Based on lectures by R. Weber Notes taken by Dexter Chua Lent 2015 These notes are not endorsed by the lecturers, and I have modified them (often significantly) after lectures.

More information

STATS 3U03. Sang Woo Park. March 29, Textbook: Inroduction to stochastic processes. Requirement: 5 assignments, 2 tests, and 1 final

STATS 3U03. Sang Woo Park. March 29, Textbook: Inroduction to stochastic processes. Requirement: 5 assignments, 2 tests, and 1 final STATS 3U03 Sang Woo Park March 29, 2017 Course Outline Textbook: Inroduction to stochastic processes Requirement: 5 assignments, 2 tests, and 1 final Test 1: Friday, February 10th Test 2: Friday, March

More information

HANDBOOK OF APPLICABLE MATHEMATICS

HANDBOOK OF APPLICABLE MATHEMATICS HANDBOOK OF APPLICABLE MATHEMATICS Chief Editor: Walter Ledermann Volume II: Probability Emlyn Lloyd University oflancaster A Wiley-Interscience Publication JOHN WILEY & SONS Chichester - New York - Brisbane

More information

Continuous time Markov chains

Continuous time Markov chains Continuous time Markov chains Alejandro Ribeiro Dept. of Electrical and Systems Engineering University of Pennsylvania aribeiro@seas.upenn.edu http://www.seas.upenn.edu/users/~aribeiro/ October 16, 2017

More information

Lecture 21. David Aldous. 16 October David Aldous Lecture 21

Lecture 21. David Aldous. 16 October David Aldous Lecture 21 Lecture 21 David Aldous 16 October 2015 In continuous time 0 t < we specify transition rates or informally P(X (t+δ)=j X (t)=i, past ) q ij = lim δ 0 δ P(X (t + dt) = j X (t) = i) = q ij dt but note these

More information

Math Homework 5 Solutions

Math Homework 5 Solutions Math 45 - Homework 5 Solutions. Exercise.3., textbook. The stochastic matrix for the gambler problem has the following form, where the states are ordered as (,, 4, 6, 8, ): P = The corresponding diagram

More information

Chapter 5. Continuous-Time Markov Chains. Prof. Shun-Ren Yang Department of Computer Science, National Tsing Hua University, Taiwan

Chapter 5. Continuous-Time Markov Chains. Prof. Shun-Ren Yang Department of Computer Science, National Tsing Hua University, Taiwan Chapter 5. Continuous-Time Markov Chains Prof. Shun-Ren Yang Department of Computer Science, National Tsing Hua University, Taiwan Continuous-Time Markov Chains Consider a continuous-time stochastic process

More information

CDA5530: Performance Models of Computers and Networks. Chapter 3: Review of Practical

CDA5530: Performance Models of Computers and Networks. Chapter 3: Review of Practical CDA5530: Performance Models of Computers and Networks Chapter 3: Review of Practical Stochastic Processes Definition Stochastic ti process X = {X(t), t T} is a collection of random variables (rvs); one

More information

Figure 10.1: Recording when the event E occurs

Figure 10.1: Recording when the event E occurs 10 Poisson Processes Let T R be an interval. A family of random variables {X(t) ; t T} is called a continuous time stochastic process. We often consider T = [0, 1] and T = [0, ). As X(t) is a random variable

More information

Review 1: STAT Mark Carpenter, Ph.D. Professor of Statistics Department of Mathematics and Statistics. August 25, 2015

Review 1: STAT Mark Carpenter, Ph.D. Professor of Statistics Department of Mathematics and Statistics. August 25, 2015 Review : STAT 36 Mark Carpenter, Ph.D. Professor of Statistics Department of Mathematics and Statistics August 25, 25 Support of a Random Variable The support of a random variable, which is usually denoted

More information

Continuous Time Markov Chains

Continuous Time Markov Chains Continuous Time Markov Chains Stochastic Processes - Lecture Notes Fatih Cavdur to accompany Introduction to Probability Models by Sheldon M. Ross Fall 2015 Outline Introduction Continuous-Time Markov

More information

Sampling Distributions

Sampling Distributions In statistics, a random sample is a collection of independent and identically distributed (iid) random variables, and a sampling distribution is the distribution of a function of random sample. For example,

More information

STAT 3610: Review of Probability Distributions

STAT 3610: Review of Probability Distributions STAT 3610: Review of Probability Distributions Mark Carpenter Professor of Statistics Department of Mathematics and Statistics August 25, 2015 Support of a Random Variable Definition The support of a random

More information

Lecture 20: Reversible Processes and Queues

Lecture 20: Reversible Processes and Queues Lecture 20: Reversible Processes and Queues 1 Examples of reversible processes 11 Birth-death processes We define two non-negative sequences birth and death rates denoted by {λ n : n N 0 } and {µ n : n

More information

{σ x >t}p x. (σ x >t)=e at.

{σ x >t}p x. (σ x >t)=e at. 3.11. EXERCISES 121 3.11 Exercises Exercise 3.1 Consider the Ornstein Uhlenbeck process in example 3.1.7(B). Show that the defined process is a Markov process which converges in distribution to an N(0,σ

More information

Markov processes and queueing networks

Markov processes and queueing networks Inria September 22, 2015 Outline Poisson processes Markov jump processes Some queueing networks The Poisson distribution (Siméon-Denis Poisson, 1781-1840) { } e λ λ n n! As prevalent as Gaussian distribution

More information

Things to remember when learning probability distributions:

Things to remember when learning probability distributions: SPECIAL DISTRIBUTIONS Some distributions are special because they are useful They include: Poisson, exponential, Normal (Gaussian), Gamma, geometric, negative binomial, Binomial and hypergeometric distributions

More information

Let (Ω, F) be a measureable space. A filtration in discrete time is a sequence of. F s F t

Let (Ω, F) be a measureable space. A filtration in discrete time is a sequence of. F s F t 2.2 Filtrations Let (Ω, F) be a measureable space. A filtration in discrete time is a sequence of σ algebras {F t } such that F t F and F t F t+1 for all t = 0, 1,.... In continuous time, the second condition

More information

The SIS and SIR stochastic epidemic models revisited

The SIS and SIR stochastic epidemic models revisited The SIS and SIR stochastic epidemic models revisited Jesús Artalejo Faculty of Mathematics, University Complutense of Madrid Madrid, Spain jesus_artalejomat.ucm.es BCAM Talk, June 16, 2011 Talk Schedule

More information

1 Solution to Problem 2.1

1 Solution to Problem 2.1 Solution to Problem 2. I incorrectly worked this exercise instead of 2.2, so I decided to include the solution anyway. a) We have X Y /3, which is a - function. It maps the interval, ) where X lives) onto

More information

Econ 508B: Lecture 5

Econ 508B: Lecture 5 Econ 508B: Lecture 5 Expectation, MGF and CGF Hongyi Liu Washington University in St. Louis July 31, 2017 Hongyi Liu (Washington University in St. Louis) Math Camp 2017 Stats July 31, 2017 1 / 23 Outline

More information

1 Probability and Random Variables

1 Probability and Random Variables 1 Probability and Random Variables The models that you have seen thus far are deterministic models. For any time t, there is a unique solution X(t). On the other hand, stochastic models will result in

More information

Chapter 5. Chapter 5 sections

Chapter 5. Chapter 5 sections 1 / 43 sections Discrete univariate distributions: 5.2 Bernoulli and Binomial distributions Just skim 5.3 Hypergeometric distributions 5.4 Poisson distributions Just skim 5.5 Negative Binomial distributions

More information

Quantitative Model Checking (QMC) - SS12

Quantitative Model Checking (QMC) - SS12 Quantitative Model Checking (QMC) - SS12 Lecture 06 David Spieler Saarland University, Germany June 4, 2012 1 / 34 Deciding Bisimulations 2 / 34 Partition Refinement Algorithm Notation: A partition P over

More information

Lecture 11: Introduction to Markov Chains. Copyright G. Caire (Sample Lectures) 321

Lecture 11: Introduction to Markov Chains. Copyright G. Caire (Sample Lectures) 321 Lecture 11: Introduction to Markov Chains Copyright G. Caire (Sample Lectures) 321 Discrete-time random processes A sequence of RVs indexed by a variable n 2 {0, 1, 2,...} forms a discretetime random process

More information

Introduction to Queuing Networks Solutions to Problem Sheet 3

Introduction to Queuing Networks Solutions to Problem Sheet 3 Introduction to Queuing Networks Solutions to Problem Sheet 3 1. (a) The state space is the whole numbers {, 1, 2,...}. The transition rates are q i,i+1 λ for all i and q i, for all i 1 since, when a bus

More information

LTCC. Exercises. (1) Two possible weather conditions on any day: {rainy, sunny} (2) Tomorrow s weather depends only on today s weather

LTCC. Exercises. (1) Two possible weather conditions on any day: {rainy, sunny} (2) Tomorrow s weather depends only on today s weather 1. Markov chain LTCC. Exercises Let X 0, X 1, X 2,... be a Markov chain with state space {1, 2, 3, 4} and transition matrix 1/2 1/2 0 0 P = 0 1/2 1/3 1/6. 0 0 0 1 (a) What happens if the chain starts in

More information

The story of the film so far... Mathematics for Informatics 4a. Continuous-time Markov processes. Counting processes

The story of the film so far... Mathematics for Informatics 4a. Continuous-time Markov processes. Counting processes The story of the film so far... Mathematics for Informatics 4a José Figueroa-O Farrill Lecture 19 28 March 2012 We have been studying stochastic processes; i.e., systems whose time evolution has an element

More information

Exponential Distribution and Poisson Process

Exponential Distribution and Poisson Process Exponential Distribution and Poisson Process Stochastic Processes - Lecture Notes Fatih Cavdur to accompany Introduction to Probability Models by Sheldon M. Ross Fall 215 Outline Introduction Exponential

More information

Assignment 3 with Reference Solutions

Assignment 3 with Reference Solutions Assignment 3 with Reference Solutions Exercise 3.: Poisson Process Given are k independent sources s i of jobs as shown in the figure below. The interarrival time between jobs for each source is exponentially

More information

BMIR Lecture Series on Probability and Statistics Fall 2015 Discrete RVs

BMIR Lecture Series on Probability and Statistics Fall 2015 Discrete RVs Lecture #7 BMIR Lecture Series on Probability and Statistics Fall 2015 Department of Biomedical Engineering and Environmental Sciences National Tsing Hua University 7.1 Function of Single Variable Theorem

More information

Moments. Raw moment: February 25, 2014 Normalized / Standardized moment:

Moments. Raw moment: February 25, 2014 Normalized / Standardized moment: Moments Lecture 10: Central Limit Theorem and CDFs Sta230 / Mth 230 Colin Rundel Raw moment: Central moment: µ n = EX n ) µ n = E[X µ) 2 ] February 25, 2014 Normalized / Standardized moment: µ n σ n Sta230

More information

Discrete Distributions

Discrete Distributions Chapter 2 Discrete Distributions 2.1 Random Variables of the Discrete Type An outcome space S is difficult to study if the elements of S are not numbers. However, we can associate each element/outcome

More information

Chapter 6: Random Processes 1

Chapter 6: Random Processes 1 Chapter 6: Random Processes 1 Yunghsiang S. Han Graduate Institute of Communication Engineering, National Taipei University Taiwan E-mail: yshan@mail.ntpu.edu.tw 1 Modified from the lecture notes by Prof.

More information

LECTURE #6 BIRTH-DEATH PROCESS

LECTURE #6 BIRTH-DEATH PROCESS LECTURE #6 BIRTH-DEATH PROCESS 204528 Queueing Theory and Applications in Networks Assoc. Prof., Ph.D. (รศ.ดร. อน นต ผลเพ ม) Computer Engineering Department, Kasetsart University Outline 2 Birth-Death

More information

12 Markov chains The Markov property

12 Markov chains The Markov property 12 Markov chains Summary. The chapter begins with an introduction to discrete-time Markov chains, and to the use of matrix products and linear algebra in their study. The concepts of recurrence and transience

More information

LIST OF FORMULAS FOR STK1100 AND STK1110

LIST OF FORMULAS FOR STK1100 AND STK1110 LIST OF FORMULAS FOR STK1100 AND STK1110 (Version of 11. November 2015) 1. Probability Let A, B, A 1, A 2,..., B 1, B 2,... be events, that is, subsets of a sample space Ω. a) Axioms: A probability function

More information

APM 541: Stochastic Modelling in Biology Discrete-time Markov Chains. Jay Taylor Fall Jay Taylor (ASU) APM 541 Fall / 92

APM 541: Stochastic Modelling in Biology Discrete-time Markov Chains. Jay Taylor Fall Jay Taylor (ASU) APM 541 Fall / 92 APM 541: Stochastic Modelling in Biology Discrete-time Markov Chains Jay Taylor Fall 2013 Jay Taylor (ASU) APM 541 Fall 2013 1 / 92 Outline 1 Motivation 2 Markov Processes 3 Markov Chains: Basic Properties

More information

Statistics 992 Continuous-time Markov Chains Spring 2004

Statistics 992 Continuous-time Markov Chains Spring 2004 Summary Continuous-time finite-state-space Markov chains are stochastic processes that are widely used to model the process of nucleotide substitution. This chapter aims to present much of the mathematics

More information

Northwestern University Department of Electrical Engineering and Computer Science

Northwestern University Department of Electrical Engineering and Computer Science Northwestern University Department of Electrical Engineering and Computer Science EECS 454: Modeling and Analysis of Communication Networks Spring 2008 Probability Review As discussed in Lecture 1, probability

More information

The Transition Probability Function P ij (t)

The Transition Probability Function P ij (t) The Transition Probability Function P ij (t) Consider a continuous time Markov chain {X(t), t 0}. We are interested in the probability that in t time units the process will be in state j, given that it

More information

Data analysis and stochastic modeling

Data analysis and stochastic modeling Data analysis and stochastic modeling Lecture 7 An introduction to queueing theory Guillaume Gravier guillaume.gravier@irisa.fr with a lot of help from Paul Jensen s course http://www.me.utexas.edu/ jensen/ormm/instruction/powerpoint/or_models_09/14_queuing.ppt

More information

CDA6530: Performance Models of Computers and Networks. Chapter 3: Review of Practical Stochastic Processes

CDA6530: Performance Models of Computers and Networks. Chapter 3: Review of Practical Stochastic Processes CDA6530: Performance Models of Computers and Networks Chapter 3: Review of Practical Stochastic Processes Definition Stochastic process X = {X(t), t2 T} is a collection of random variables (rvs); one rv

More information

The Wright-Fisher Model and Genetic Drift

The Wright-Fisher Model and Genetic Drift The Wright-Fisher Model and Genetic Drift January 22, 2015 1 1 Hardy-Weinberg Equilibrium Our goal is to understand the dynamics of allele and genotype frequencies in an infinite, randomlymating population

More information

SIMILAR MARKOV CHAINS

SIMILAR MARKOV CHAINS SIMILAR MARKOV CHAINS by Phil Pollett The University of Queensland MAIN REFERENCES Convergence of Markov transition probabilities and their spectral properties 1. Vere-Jones, D. Geometric ergodicity in

More information

STAT/MATH 395 PROBABILITY II

STAT/MATH 395 PROBABILITY II STAT/MATH 395 PROBABILITY II Chapter 6 : Moment Functions Néhémy Lim 1 1 Department of Statistics, University of Washington, USA Winter Quarter 2016 of Common Distributions Outline 1 2 3 of Common Distributions

More information

Stat 475 Life Contingencies I. Chapter 2: Survival models

Stat 475 Life Contingencies I. Chapter 2: Survival models Stat 475 Life Contingencies I Chapter 2: Survival models The future lifetime random variable Notation We are interested in analyzing and describing the future lifetime of an individual. We use (x) to denote

More information

Quasi-Stationary Distributions in Linear Birth and Death Processes

Quasi-Stationary Distributions in Linear Birth and Death Processes Quasi-Stationary Distributions in Linear Birth and Death Processes Wenbo Liu & Hanjun Zhang School of Mathematics and Computational Science, Xiangtan University Hunan 411105, China E-mail: liuwenbo10111011@yahoo.com.cn

More information

Mathematical Methods for Computer Science

Mathematical Methods for Computer Science Mathematical Methods for Computer Science Computer Science Tripos, Part IB Michaelmas Term 2016/17 R.J. Gibbens Problem sheets for Probability methods William Gates Building 15 JJ Thomson Avenue Cambridge

More information

Linear-fractional branching processes with countably many types

Linear-fractional branching processes with countably many types Branching processes and and their applications Badajoz, April 11-13, 2012 Serik Sagitov Chalmers University and University of Gothenburg Linear-fractional branching processes with countably many types

More information

PROBABILITY: LIMIT THEOREMS II, SPRING HOMEWORK PROBLEMS

PROBABILITY: LIMIT THEOREMS II, SPRING HOMEWORK PROBLEMS PROBABILITY: LIMIT THEOREMS II, SPRING 218. HOMEWORK PROBLEMS PROF. YURI BAKHTIN Instructions. You are allowed to work on solutions in groups, but you are required to write up solutions on your own. Please

More information

P 2 (t) = o(t). e λt. 1. keep on adding independent, exponentially distributed (with mean 1 λ ) inter-arrival times, or

P 2 (t) = o(t). e λt. 1. keep on adding independent, exponentially distributed (with mean 1 λ ) inter-arrival times, or POISSON PROCESS PP): State space consists of all non-negative integers, time is continuous. The process is time homogeneous, with independent increments and P 0 t) = λ t+ot) P t) = λ t+ot) P 2 t) = ot).

More information

1 Continuous-time chains, finite state space

1 Continuous-time chains, finite state space Université Paris Diderot 208 Markov chains Exercises 3 Continuous-time chains, finite state space Exercise Consider a continuous-time taking values in {, 2, 3}, with generator 2 2. 2 2 0. Draw the diagramm

More information

A review of Continuous Time MC STA 624, Spring 2015

A review of Continuous Time MC STA 624, Spring 2015 A review of Continuous Time MC STA 624, Spring 2015 Ruriko Yoshida Dept. of Statistics University of Kentucky polytopes.net STA 624 1 Continuous Time Markov chains Definition A continuous time stochastic

More information

System Simulation Part II: Mathematical and Statistical Models Chapter 5: Statistical Models

System Simulation Part II: Mathematical and Statistical Models Chapter 5: Statistical Models System Simulation Part II: Mathematical and Statistical Models Chapter 5: Statistical Models Fatih Cavdur fatihcavdur@uludag.edu.tr March 20, 2012 Introduction Introduction The world of the model-builder

More information

1 IEOR 4701: Continuous-Time Markov Chains

1 IEOR 4701: Continuous-Time Markov Chains Copyright c 2006 by Karl Sigman 1 IEOR 4701: Continuous-Time Markov Chains A Markov chain in discrete time, {X n : n 0}, remains in any state for exactly one unit of time before making a transition (change

More information

3 Continuous Random Variables

3 Continuous Random Variables Jinguo Lian Math437 Notes January 15, 016 3 Continuous Random Variables Remember that discrete random variables can take only a countable number of possible values. On the other hand, a continuous random

More information

TCOM 501: Networking Theory & Fundamentals. Lecture 6 February 19, 2003 Prof. Yannis A. Korilis

TCOM 501: Networking Theory & Fundamentals. Lecture 6 February 19, 2003 Prof. Yannis A. Korilis TCOM 50: Networking Theory & Fundamentals Lecture 6 February 9, 003 Prof. Yannis A. Korilis 6- Topics Time-Reversal of Markov Chains Reversibility Truncating a Reversible Markov Chain Burke s Theorem Queues

More information

Bayesian Methods with Monte Carlo Markov Chains II

Bayesian Methods with Monte Carlo Markov Chains II Bayesian Methods with Monte Carlo Markov Chains II Henry Horng-Shing Lu Institute of Statistics National Chiao Tung University hslu@stat.nctu.edu.tw http://tigpbp.iis.sinica.edu.tw/courses.htm 1 Part 3

More information

Midterm 2 Review. CS70 Summer Lecture 6D. David Dinh 28 July UC Berkeley

Midterm 2 Review. CS70 Summer Lecture 6D. David Dinh 28 July UC Berkeley Midterm 2 Review CS70 Summer 2016 - Lecture 6D David Dinh 28 July 2016 UC Berkeley Midterm 2: Format 8 questions, 190 points, 110 minutes (same as MT1). Two pages (one double-sided sheet) of handwritten

More information