Probability Distributions
|
|
- Vanessa Lloyd
- 5 years ago
- Views:
Transcription
1 Lecture 1: Background in Probability Theory Probability Distributions The probability mass function (pmf) or probability density functions (pdf), mean, µ, variance, σ 2, and moment generating function (mgf) M(t) for some well-known discrete and continuous probability distributions Discrete Distributions Discrete Uniform: f(x) = { 1, n x = 1, 2,, n, 0, otherwise µ = 1 + n, σ 2 = n , M(t) = e(n+1)t e t n(e t 1), t 0 Geometric: f(x) = { p(1 p) x, x = 0, 1, 2,, 0, otherwise, where 0 < p < 1 µ = 1 p p, σ2 = 1 p p 2, M(t) = Binomial b(n, p): f(x) = ( ) n 0 < p < 1 The notation = x p 1 (1 p)e t ( ) n p x (1 p) n x, x = 0, 1, 2,, n, 0, otherwise, n! x!(n x)! where n is a positive integer and µ = np, σ 2 = np(1 p), M(t) = (1 p + pe t ) n where n is a positive inte- ( ) x + n 1 p Negative Binomial: f(x) = n 1 n (1 p) x, x = 0, 1, 2,, 0, otherwise, ger and 0 < p < 1 µ = n(1 p), σ 2 n(1 p) = p p 2, M(t) = p n [1 (1 p)e t ] n λ x e λ Poisson, P o(λ): f(x) =, x = 0, 1, 2,, x! 0, otherwise, µ = λ, σ 2 = λ, M(t) = e λ(et 1) where λ is a positive constant Continuous Distributions Uniform U(a, b): f(x) = 1 b a, a x b, 0, otherwise, where a < b are constants µ = a + b 2, σ2 = (b a)2, M(t) = ebt e at 12 t(b a), t 0 1
2 Gamma: f(x) = 0 1 Γ(α)β α xα 1 e x/β, x 0, 0, x < 0, 1 β α xα 1 e x/β dx For a positive integer n, Γ(n) = (n 1)! µ = αβ, σ 2 = αβ 2, M(t) = 1 (1 βt) α, t < 1/β where α and β are positive constants and Γ(α) = Exponential: f(x) = { λe λx, x 0, 0, x < 0, where λ is a positive constant µ = 1 λ, σ2 = 1 λ 2, M(t) = λ λ t, t < λ Normal, N(µ, σ 2 ): f(x) = 1 ( ) σ 2π exp (x µ)2, < x <, where µ and σ are constants 2σ 2 E(X) = µ, V ar(x) = σ 2, M(t) = e µt+σ2 t 2 /2 The probability generating function (pgf) is P X (t) = E(t X ), moment generating function (mgf) is M X (t) = E(e tx ) and cumulant generating function (cgf) is K X (t) = ln(m X (t)) The mean µ X = E(X) and variance σ 2 X = E(X2 ) E 2 (X) computed from the pgf, P X (t), mgf, M X (t) and cgf, K X (t): µ X = P X(1) = M X(0) = K X(0) and P X (1) + P X (1) [P X (1)]2 σx 2 = M X (0) [M X (0)]2 K X (0) Continuous-time Markov Chain Example, Simple Birth Process: The per capita rate of birth is λ = Population size Time Figure 1: Three sample paths (stochastic realizations) of a simple birth process, X(0) = 1; n(t) = e t is the dashed curve See Table 1 2
3 Table 1: For two stochastic realizations, the times at which a birth occurs for a simple birth process Realization 1 Realization 2 Size X(t) Event Time t Size X(t) Event Time t
4 Lecture 2: Discrete-Time Markov Chains (DTMCs) Transition probabilities: p ji (n) = Prob{X n+1 = j X n = i} If p ji (n) does not depend on n, then the process is said to be time homogeneous The transition matrix of a DTMC {X n } n=0 with state space {1, 2, } and one-step transition probabilities, {p ij} i,j=1, is denoted as P = (p ij ), where p 11 p 12 p 13 p 21 p 22 p 23 P = p 31 p 32 p 33 The column sums equal one, a stochastic matrix, j=1 p ji = 1 The n-step transition matrix P n = (p (n) The probabilities p i (n) = Prob{X n = i}, p i (n + 1) = j=1 p ijp j (n), i = 1, 2 p(n + 1) = P p(n) Random walk model with absorbing barriers See the directed graph in Figure 2 and the corresponding (N + 1) (N + 1) transition matrix: 1 q q p p P = q p p 1 The Markov chain, graphed in Figure 2, has three communication classes: {0}, {1, 2,, N 1}, and {N} The Markov chain is reducible States 0 and N are absorbing; the remaining states are transient ij ) N Figure 2: Probability of moving to right is p and to the left is q, p + q = 1 Boundaries 0 and N are absorbing, p 00 = 1 = p NN (random walk with absorbing barriers or gambler s ruin problem) A DTMC is irreducible if its digraph is strongly connected Otherwise it is called reducible An irreducible DTMC can be positive recurrent or null recurrent or transient It may also be classified as periodic or aperiodic Recurrence is defined for each state i in the chain Recurrence means for each state i if the process leaves state i it will return to state i at some future time If not, the state is transient A state i is positive recurrent if the mean recurrence time (µ ii mean return time) is finite A state i with an infinite mean recurrence time is called null recurrent be a recurrent, ir- Theorem 1 (Basic Limit Theorem for aperiodic Markov chains) Let {X n } n=0 reducible, and aperiodic DTMC with transition matrix P = (p ij ) Then lim n p(n) ij = 1, µ ii where µ ii is the mean recurrence time for state i and i and j are any states of the chain [If µ ii =, then lim n p (n) ij = 0] 4
5 Theorem 2 (Basic Limit Theorem for periodic Markov chains) Let {X n } n=0 be a recurrent, irreducible, and d-periodic DTMC, d > 1, with transition matrix P = (p ij ) Then lim n p(nd) ii = d µ ii and p (m) ii = 0 if m is not a multiple of d, where µ ii is the mean recurrence time for state i [If µ ii =, then lim n p (nd) ii = 0] Random Walk with Absorbing Boundaries or Gambler s Ruin Let s return to the random walk model with absorbing boundaries This DTMC does not satisfy either of the Basic Limit Theorems because it is a reducible Markov chain with three communicating classes, two absorbing states {0} and {N} and the transient class {1, 2,, N 1} Interesting questions about problems with absorbing states such as an extinction state in population dynamics are (1) what is the probability of absorption? and (2) what is the expected time until absorption? We will investigate this question with the random walk model and with genetics of inbreeding for a finite DTMC Later, we will investigate the probability of extinction in a population model via a branching process Consider the random walk model with absorbing boundaries at 0 and N In the gambler s ruin problem, state 0 is ruin and state N is winning the maximum amount of money In either case, after hitting 0 or N, you stop playing the game Let a k be the probability of ruin (absorption at k = 0) beginning with a capital of k: a k = pa k+1 + qa k 1 (1) for 1 k N 1 Equation (1) is a second-order difference equation in a k The difference equation can be written as pa k+1 a k + qa k 1 = 0, k = 1,, N 1 (2) This method of deriving equation (1) is referred to as a first-step analysis (Taylor and Karlin, 1998) This difference equation (2) can be solved with the following boundary conditions Alternately, it will be shown that a 0 = 1 and a N = 0 (a 1,, a N 1 ) = (p 01, p 02,, p 0,N 1 )(I T ) 1 In general, the difference equation for probability of absorption has the form: a k = N p ik a i To find the probability of absorption into state N, the boundary conditions are a 0 = 0 and a N = 1 k=0 Expected Time until Absorption: Let τ k be the expected time until absorption at either 0 or N in the gambler s ruin problem The difference equation for τ k has the following form: for k = 1, 2,, N 1 or τ k = p(1 + τ k+1 ) + q(1 + τ k 1 ), pτ k+1 τ k + qτ k 1 = 1, (3) The boundary conditions are τ 0 = 0 = τ N In the gambler s ruin problem, this equation can be solved explicitly But in general, N N τ k = p ik (1 + τ i ) or τ k p ik τ i = 1 i=0 5 i=0
6 It will be shown that (τ 1, τ 2,, τ N 1 ) = 1(I T ) 1, 1 = (1, 1,, 1) Table 2: Gambler s ruin problem with a beginning capital of k = 50 and a total capital of N = 100 Prob a 50 b 50 τ 50 A 50 (1) B 50 (1) q = q = q = q = Capital Games Figure 3: Three sample paths for the gambler s ruin problem when N = 100, k = 50, and q = 055 Example of Genetics of Inbreeding: Two alleles A and a There are six possible breeding pairs which denote the six states of the DTMC, 1: AA AA, 2 aa aa, 3 Aa Aa, 4 Aa aa, 5 AA aa, 6 AA Aa Inbreeding of the first two types results in offspring of the same genotypes and inbreeding in the next generation will be of the same type; they are absorbing states The remaining states, 3,4,5,6 are transient The transition matrix has the following form: 1 f 0 1/ / /16 1/4 0 0 ( ) P = 0 0 1/4 1/4 1 1/4 I A 0 0 1/4 1/2 0 0 = O T 0 0 1/ / /2 Probability of absorption into states 1 or 2 from states 3, 4, 5, 6 is computed from the fundamental matrix, (I T ) 1, ( lim P n I (A + AT + AT = 2 ) ( ) + ) I A(I T ) 1 = n O O O O Thus, A(I T ) 1 is the probability of absorption into states 1 or 2 (fixation) from states 3, 4, 5 or 6 Let 1 = (1, 1,, 1), then 1(I T ) 1 is the mean time until absorption from states 3, 4, 5, or 6 6
7 Lecture 3: Discrete-Time Branching Processes Figure 4: Sample path of a branching process {X n } n=0 In the first generation, four individuals are born, X 1 = 4 The four individuals in generation one give birth to three, zero, four, and one individuals, respectively, making a total of eight individuals in generation two, X 2 = 8 Generating functions rather than transition matrices are useful in analysis of branching processes Offspring pgf: f(t) = p k t k k=0 Recall f(1) = 1 and m = f (1) is the mean number of offspring The branching process is called subcritical if m < 1, critical if m = 1, and supercritical if m > 1 Theorem 3 (Branching Process Theorem) Let X 0 = 1 Assume f(0) = p 0 > 0 and p 0 + p 1 < 1 (i) If m 1, then lim n Prob{X n = 0} = 1 (ii) If m > 1, then lim n Prob{X n = 0} = q, where q = f(q) is the unique fixed point in the interval (0, 1) To verify this theorem, we show that given X 0 = 1, then P Xn (t) = f n 1 (f(t)) = f n (t), the pgf of X n is the n-fold composition of the offspring pgf Then p 0 (n) = P Xn (0) = f n (0) and p 0 (n) = f(p 0 (n 1)) The sequence {p 0 (n)} is an increasing function, bounded above by one and therefore, has a limit q This limit is a fixed point of f, f(q) = q If X 0 = N and m > 1, then lim n p 0 (n) = lim n Prob{X n = 0} = q N The conditional expectation E(X n+1 X n ) = mx n, E(X n+1 ) = me(x n ) If the growth rate varies with n, with a mean m n, then E(X n+1 ) = m n E(X n ) Multitype Branching Process, n different types: (X 1,, X n ) The offspring random variable of type i is Y i Offspring pgfs: f i (t 1,, n) = Expectation matrix M = (m ij ), where s n s 1 P i (s 1,, s n )t s1 1 tsn n P i (s 1,, s n ) = Prob{Y 1 = s 1,, Y n = s n } m ji = f t j t1 =1,,t n=1 Assume M is irreducible Denote the spectral radius of M as ρ(m), the maximum modulus of the eigenvalues of MThe multitype branching process is called subcritical if ρ(m) < 1, critical if ρ(m) = 1 and supercritical if ρ(m) > 1 See Chapter on Branching Processes 7
8 Lecture 4: Continuous-Time Markov Chains (CTMCs), Introduction Discrete random variable X(t), t [0, ) Probabilities p i (t) = Prob{X(t) = i} Transition probability: p ji (t, s) = Prob{X(t) = j X(s) = i}, s < t We will assume time-homogenous transition probabilities p ji (t, s) = p ji (t s) The transition matrix is a stochastic matrix: where p 00 (t) p 01 (t) P (t) = (p ij (t)) = p 10 (t) p 11 (t), p ji ( t) = δ ji + q ji t + o( t) is an infinitesimal transition probability The infinitesimal generator matrix: q 00 q 01 Q = (q ij ) = q 10 q 11 P ( t) I, Q = lim t 0 t The column sums of Q equal zero dp (t) Forward Kolmogorov differential equations: = QP (t) dt dp (t) Backward Kolmogorov differential equations: = P (t)q dt In physics and chemistry, the forward Kolmogorov differential equations are often referred to as the Master equation The embedded DTMC is used to define irreducible, recurrent, and transient states or chains for the associated CTMC Let Y n denote the random variable for the state of a CTMC {X(t) : t [0, )} at the time of the nth jump, Y n = X(W n ) (See Figure 5) The set of discrete random variables {Y n } 0 is the embedded Markov chain T0 T1 T2 T3 0 W1 W2 W3 W4 Figure 5: Sample path of a CTMC, illustrating waiting times {W i } and interevent times, {T i } A CTMC is irreducible, recurrent or transient if the corresponding embedded Markov chain has these properties Some differences in the dynamics of a CTMC as opposed to a DTMC are the possibility of a finite-time blow up in a CTMC (explosive process) and the fact that CTMC are not periodic See Figure 6 The embedded MC cannot be used to classify chains as positive recurrent or null recurrent This latter classification depends on the mean recurrence time µ ii 8
9 0 W1 W2 W3 W4 W Figure 6: One sample path of a continuous time Markov chain that is explosive Theorem 4 (Basic Limit Theorem for CTMCs) If the CTMC {X(t) : t [0, )} is nonexplosive and irreducible, then for all i and j, lim p ij(t) = 1, (4) t q ii µ ii where µ ii is the mean recurrence time, 0 < µ ii In particular, a finite, irreducible CTMC is nonexplosive and the limit (4) exists and is positive If the DTMC is nonexplosive and positive recurrent, it has a limiting positive stationary distribution π satisfying Qπ = 0, where π i = 1 q ii µ ii Poisson process with X(0) = 0, p i+1,i ( t) = λ t + o( t) and p i (t) = e λt (λt) i /i! has generator matrix λ 0 0 λ λ 0 Q = 0 λ λ The associated embedded Markov chain has a transition matrix T = The Poisson process is transient A finite CTMC with two states {1, 2} The generator matrix ( ) a b Q =, a, b > 0 a b The CTMC is irreducible and positive recurrent The limiting stationary distribution can be found from the forward Kolmogorov differential equations by solving for the stationary distribution Qπ = 0 In this case π = (b/(a + b), a/(a + b)) tr The mean recurrence times are µ ii = a + b, i = 1, 2 ab 9
10 Lecture 5: Continuous-Time Markov Chains (CTMCs), Interevent Time To generate sample paths, we must know the time between jumps and the state to which the process jumps The Markov assumption implies the interevent time is exponentially distributed because the exponential distribution has the memoryless property Let T i be the continuous random variable for the time until the i + 1st event See Figure 5 Theorem 5 (Interevent Time) Assume j n p jn( t) = α(n) t + o( t) Then the cumulative distribution function for the interevent time T i is F i (t) = 1 exp( α(n)t) with mean and variance µ Ti = 1 α(n) and σ 2 T i = 1 [α(n)] 2 Theorem 6 (Interevent Time Simulation) Let U U[0, 1] be the uniform distribution on [0,1] and T i the continuous random variable for interevent time with state space [0, ) Then T i = Fi 1 (U) = ln(u) α(n) Simple Birth and Death Markov Chain: In the simple birth and death process, an event can be a birth or a death Let X(0) = N The infinitesimal transition probabilities are p i+j,i ( t) = Prob{ X(t) = j X(t) = i} µi t + o( t), j = 1 λi t + o( t), j = 1 = 1 (λ + µ)i t + o( t), j = 0 o( t), j 1, 0, 1 Use two random numbers, u 1 and u 2, from the uniform distribution U(0, 1) to determine the interevent time and the state to which the process jumps In MATLAB, indices begin from 1, so instead of writing t(0), we use t(1) Consider the simple birth and death chain, in a MATLAB program, t(1) = 0, and the time to the next event is t(2) = t(1)+ln(u 1 )/(α(n)), where α(n) = λn+µn, given the process is in state n Since there are two events, to determine whether there is a birth or a death, the unit interval is divided into two subintervals, one subinterval has probability λ/(λ + µ) and the other has probability µ/(λ + µ) Generate a uniform random number u 2 If u 2 < λ/(λ + µ), then this random number lies in the first subinterval and there is a birth, otherwise if u 2 > λ/(λ + µ), the random number lies in the second subinterval and there is a death This concept can be easily extended to k > 2 events In a MATLAB program with k events the unit interval must be divided into k subintervals, each with predetermined probability for i = 1,, k that depends on the current state and the transition probabilities For example, suppose there are four events with the following rates a i (n), i = 1, 2, 3, 4 which depend on the current state n The probabilities of these four events are a i (n)/a(n), a(n) = i a i(n), i = 1, 2, 3, 4 The subinterval [0, 1] is subdivided into four subintervals with the following endpoints: Therefore, in a MATLAB program: 0, a 1 a, a 1 + a 2, a a 1 + a 2 + a 3, 1 a if u2<a1/a, then event 1 occurs elseif u2>=a1/a & u2<(a1+a2)/a, then event 2 occurs elseif u2>=(a1+a2)/2 & u2<(a1+a2+a3)/a, then event 3 occurs else u2>=(a1+a2+a3)/a, then event 4 occurs The subintervals change each time the process changes state n If the number of events are large, deciding which event occurs can become quite lengthy and there are ways to speed up the process of selecting a particular event 10
11 %MatLab program: simple birth and death process clear all x0=5; b=1; d=05; % initial and parameter values for j=1:3 % Three sample paths clear x t n=1; t(1,j)=0; x(n)=x0; % starting values while x(n)>0 & x(n)<50; % continue until the process hits zero or reaches 50 u1=rand; u2=rand; % two uniform random numbers t(n+1,j)=-log(u1)/(b*x(n)+d*x(n))+t(n,j); if u2< b/(b+d); x(n+1)=x(n)+1; else x(n+1)=x(n)-1; end n=n+1; end s=stairs(t(:,j),x, r-, Linewidth,2); hold on end xlabel( Time ); ylabel( Population size ); hold off Figure 7: Three sample paths of the simple birth and death process, X(0) = 5, λ = b = 1, µ = d = Population size Time 11
12 Lecture 6: Continuous-Time Birth and Death Processes: The generator matrix Q for a general birth and death process is either λ 0 µ λ 0 λ 1 µ 1 µ 2 0 Q = 0 λ 1 λ 2 µ 2 µ 3, 0 0 λ 2 λ 3 µ 3 or for a finite state space, λ 0 µ λ 0 λ 1 µ 1 µ λ 1 λ 2 µ 2 0 Q = µ N µ N The simple birth, simple death, simple birth and death, and simple birth, death, and immigration processes are linear in the rates, λ i = λi + ν and µ i = µi Applying the forward Kolmogorov differential equations, dp = Qp, first-order partial differential equations for the pgf and mgf can be derived Applying the method of characteristics to the first-order dt partial differential equations, explicit expressions can be found for the pgf and mgf of these processes Denote the pgf for these simple process as follows (note change in notation z and θ because t is time): M(θ, t) = P(z, t) = p i (t)z i i=0 p i (t)e iθ = P(e θ, t) i=0 Simple Birth, Death, and Immigration Process: Let X(0) = N The infinitesimal transition probabilities are p i+j,i ( t) = Prob{ X(t) = j X(t) = i} µi t + o( t), j = 1 (ν + λi) t + o( t), j = 1 = 1 [ν + (λ + µ)i] t + o( t), j = 0 o( t), j 1, 0, 1 The forward Kolmogorov differential equations are dp i dt dp 0 dt = [λ(i 1) + ν]p i 1 + µ(i + 1)p i+1 (λi + µi + ν)p i = νp 0 + µp 1 for i = 1, 2, with initial condition X(0) = N Applying the generating function technique, it follows that the mgf M(θ, t) is a solution of the following first-order partial differential equation M t = [ ] M λ(e θ 1) + µ(e θ 1) θ + ν(eθ 1)M 12
13 with initial condition M(θ, 0) = e Nθ The preceding differential equation is first-order because the rates are linear The mgf is the solution of this first-order partial differential equation, M(θ, t) = (λ µ)ν/λ [ µ(e (λ µ)t 1) e θ (µe (λ µ)t λ) ] N [ (λe (λ µ)t µ) λ(e (λ µ)t 1)e θ] N+ν/λ The moments E(X n (t)) of the probability distribution X(t) can be found by differentiating the mgf with respect to θ and evaluating at θ = 0: E(X n (t)) = n M(θ, t) θ n θ=0 Table 3: Mean, variance, and pgf for the simple birth, simple death, and simple birth and death processes, where X(0) = N and ρ = e (λ µ)t, λ µ Simple Simple Simple Birth Death Birth and Death m(t) Ne λt Ne µt Ne (λ µ)t σ 2 (t) Ne 2λt (1 e λt ) Ne µt (1 e µt ) N λ + µ ρ(ρ 1) λ µ (pz) N (1 z(1 p)) N (1 p + pz) N P(z, t) Negative binomial p = e λt Binomial b(n,p) p = e µt ( ρ 1 ) N (λz µ) µ(z 1) ρ 1 (λz µ) λ(z 1) Table 4: Mean, variance, and pgf X(0) = N and ρ = e (λ µ)t, λ µ for the simple birth and death with immigration process, where m(t) Simple Birth and Death with Immigration ρ[n(λ µ) + ν] ν λ µ σ 2 (t) N (λ2 µ 2 )ρ[ρ 1] µ + ρ(λρ µ λ) (λ µ) 2 + ν (λ µ) 2 P(z, t) (λ µ) ν/λ [µ(ρ 1) z(µρ λ)] N [λρ µ λ(ρ 1)z] N+ν/λ For each of the preceding simple birth and death processes, the probability of extinction p 0 (t) can be found by evaluating the pgf at z = 0, P(0, t), and the probability of ultimate extinction by taking the limit: lim t p 0 (t) Except for the stochastic process with immigration, the other birth and death 13
14 processes either hit zero or approach infinity For example, given X(0) = N, for the simple birth process, p 0 (t) = 0, simple death process p 0 (t) = (1 e µt ) N and for the simple birth and death process: lim p 0(t) = t { ( µ ) N, λ λ > µ 1, λ µ dn Logistic Growth Model: (1 dt = rn n ) What is the birth rate and what is the death rate? There K are an infinite number of choices for stochastic birth and death rates that yield the same deterministic logistic growth model The birth and death rates should have the form so that λ n µ n = rn(1 n/k) leads to λ i = b 1 i + b 2 i 2 > 0 and µ i = d 1 i + d 2 i 2 > 0 dn dt = (b 1 d 1 )n + (b 2 d 2 )n 2 = rn r K n2, b 1 d 1 = r > 0 and b 1 d 1 d 2 b 2 = K > 0 Define the birth and death rates, λ i and µ i, as follows: (a) λ i = i and µ i = i2, i = 0, 1, 2, 10 (b) λ i = i i2, i = 0, 1,, and µ i = i2, i = 0, 1, 2, 0, i > In both cases the deterministic model is dn (1 dt = n n ), 10 where r = 1 and K = 10 In the deterministic model, solutions approach the carrying capacity K = 10 Three sample paths for models (a) and (b) are graphed in Figure 8 20 (a) 20 (b) Population size Population size Time Time Figure 8: Three sample paths of the stochastic logistic model for cases (a) and (b) with X(0) = 10 For these stochastic logistic models, lim p 0(t) = 1 t 14
15 For a population of finite size N, we will show that the mean time to extinction is a solution of the following matrix equation, τ Q = 1 where Q is the truncated generator matrix without state zero, τ = (τ 1,, τ N ), τ k is the mean time until extinction from state k, τ = 1 Q 1 Let λ i = i i2 N, i = 0, 1,, N and µ i = i2, i = 0, 1, 2,, N 0, i > N N The intrinsic growth rate r = 1 and the carrying capacity K = N/2 The expected time to extinction can be calculated for X(0) = m, m = 1, 2,, N by solving τ Q = 1 When N = 10, the carrying capacity is K = 5, and when N = 20, the carrying capacity is K = 10 The time units depend on the particular problem If the time is measured in days, then for the first example, extinction occurs, on the average, in less than one year but for the second example, the mean time to extinction is over 300 years Hence, for even larger values of K, the mean time to extinction will take much longer; the convergence of p 0 (t) to 1 is slow x 105 Expected time to extinction K = 5 Expected time to extinction K = Initial population size Initial population size Figure 9: Expected time until extinction in the stochastic logistic model with K = 5 and K = 10 15
16 Lecture 7 Biological Applications of CTMC Models: Let P(z, t) be the pgf for X(t), and f(t) be the offspring pgf, P(z, t) = p 0 (t) + p 1 (t)z + p 2 (t)z 2 + f(z) = p 0 + p 1 z + p 2 z 2 + As in discrete-time branching processes, the pgf P(z, t + t) is a composition of pgfs At time t = 0, P(z, 0) = z or X(0) = 1 The same theorem applies to the continuous-time Markov branching process as in discrete-time Theorem 7 Suppose {X(t) : t [0, )} is a nonexplosive, continuous-time Markov branching process with X(0) = 1 Assume f is the pgf of the offspring distribution, where m = f (1) and P(z, t) is the pgf of X(t) [f(0) = p 0 > 0 and p 0 + p 1 < 1] If m 1, then lim Prob{X(t) = 0} = lim p 0(t) = 1 t t and if m > 1, then there exists a q satisfying f(q) = q such that lim Prob{X(t) = 0} = lim p 0(t) = q < 1 t t Simple Birth and Death Process: Let λ > 0 be the per capita birth rate and µ > 0 be the per capita death rate of an individual, p i+1,i ( t) = λi t + o( t) and p i 1,i ( t) = µi t + o( t) The offspring pgf is f(z) = µ λ + µ + λ λ + µ z2 An individual either dies or survives and gives birth There is an important difference in this offspring pgf and the one for the discrete-time model For this pgf, a single offspring does not replace the parent but adds to the population size Also, m = f (1) = 2λ/(µ + λ) > 1 if λ > µ In this case, the fixed point q of f(q) = q is µ/λ The pgf of X(t) for the simple birth and death process is given in Table 3 SIS Epidemic Model ds dt = β SI + γi N di dt = β SI γi = N (β SN ) ( γ I = β γ β I ) I N because S + I = N The value R 0 = β is called the basic reproduction number This system has two γ equilibria, ( S, Ī) = (N, 0) and (N/R 0, N(1 1/R 0 ) if R 0 > 1 If R 0 > 1, then there is an outbreak and solutions approach the endemic equilibrium and if R 0 1, there is no outbreak For the CTMC SIS epidemic model, let Prob{ I(t) = j I(t) = i} β i(n i) t + o( t), j = 1 N γi t [ + o( t), ] j = 1 = β 1 i(n i) + γi t N +o( t), j = 0 o( t), j 1, 0, 1, 16
17 where i {0, 1,, N} The SIS epidemic model is a birth and death process with λ i = β N i(n i) and µ i = γi, for i = 0, 1,, N There is a single absorbing state at zero, lim t p 0 (t) = 1 However, before complete extinction, which may take a long time, I(t) behaves like a simple birth and death process Consider only the state I when S N, then di = βi γi = (β γ)i dt Thus, if N is large, I behaves like a simple birth and death process, where β = λ and γ = µ Thus, using branching process theory or the simple birth and death process we can approximate the dynamics at the beginning of an epidemic What is the probability that I hits zero (no outbreak)? The probability of absorption from branching process theory tells us that the probability of absorption as t (no outbreak) given I(0) = 1 is µ/λ = γ/β = 1/R 0 if R 0 > 1 If I(0) = k, then the probability of no outbreak is (1/R 0 ) k The process either hits zero rapidly or grows Even though this is an asymptotic approximation, it is a good for prediction of probability of no outbreak See Exercise 15 for the SIR epidemic model Lectures 8 & 9 Stochastic Epidemic Models, DTMC, CTMC and SDE Models See Chapter on Stochastic Epidemic Models 17
Probability Distributions
Lecture : Background in Probability Theory Probability Distributions The probability mass function (pmf) or probability density functions (pdf), mean, µ, variance, σ 2, and moment generating function (mgf)
More information6 Continuous-Time Birth and Death Chains
6 Continuous-Time Birth and Death Chains Angela Peace Biomathematics II MATH 5355 Spring 2017 Lecture notes follow: Allen, Linda JS. An introduction to stochastic processes with applications to biology.
More informationAARMS Homework Exercises
1 For the gamma distribution, AARMS Homework Exercises (a) Show that the mgf is M(t) = (1 βt) α for t < 1/β (b) Use the mgf to find the mean and variance of the gamma distribution 2 A well-known inequality
More informationLecture 4a: Continuous-Time Markov Chain Models
Lecture 4a: Continuous-Time Markov Chain Models Continuous-time Markov chains are stochastic processes whose time is continuous, t [0, ), but the random variables are discrete. Prominent examples of continuous-time
More informationAn Introduction to Stochastic Epidemic Models
An Introduction to Stochastic Epidemic Models Linda J. S. Allen Department of Mathematics and Statistics Texas Tech University Lubbock, Texas 79409-1042, U.S.A. linda.j.allen@ttu.edu 1 Introduction The
More information2 Discrete-Time Markov Chains
2 Discrete-Time Markov Chains Angela Peace Biomathematics II MATH 5355 Spring 2017 Lecture notes follow: Allen, Linda JS. An introduction to stochastic processes with applications to biology. CRC Press,
More informationStochastic modelling of epidemic spread
Stochastic modelling of epidemic spread Julien Arino Centre for Research on Inner City Health St Michael s Hospital Toronto On leave from Department of Mathematics University of Manitoba Julien Arino@umanitoba.ca
More information1.1 Review of Probability Theory
1.1 Review of Probability Theory Angela Peace Biomathemtics II MATH 5355 Spring 2017 Lecture notes follow: Allen, Linda JS. An introduction to stochastic processes with applications to biology. CRC Press,
More informationMarkov Chains. X(t) is a Markov Process if, for arbitrary times t 1 < t 2 <... < t k < t k+1. If X(t) is discrete-valued. If X(t) is continuous-valued
Markov Chains X(t) is a Markov Process if, for arbitrary times t 1 < t 2
More informationAN INTRODUCTION TO STOCHASTIC EPIDEMIC MODELS-PART I
AN INTRODUCTION TO STOCHASTIC EPIDEMIC MODELS-PART I Linda J. S. Allen Department of Mathematics and Statistics Texas Tech University Lubbock, Texas U.S.A. 2008 Summer School on Mathematical Modeling of
More informationDerivation of Itô SDE and Relationship to ODE and CTMC Models
Derivation of Itô SDE and Relationship to ODE and CTMC Models Biomathematics II April 23, 2015 Linda J. S. Allen Texas Tech University TTU 1 Euler-Maruyama Method for Numerical Solution of an Itô SDE dx(t)
More informationPart I Stochastic variables and Markov chains
Part I Stochastic variables and Markov chains Random variables describe the behaviour of a phenomenon independent of any specific sample space Distribution function (cdf, cumulative distribution function)
More information8. Statistical Equilibrium and Classification of States: Discrete Time Markov Chains
8. Statistical Equilibrium and Classification of States: Discrete Time Markov Chains 8.1 Review 8.2 Statistical Equilibrium 8.3 Two-State Markov Chain 8.4 Existence of P ( ) 8.5 Classification of States
More informationStochastic modelling of epidemic spread
Stochastic modelling of epidemic spread Julien Arino Department of Mathematics University of Manitoba Winnipeg Julien Arino@umanitoba.ca 19 May 2012 1 Introduction 2 Stochastic processes 3 The SIS model
More informationMarkov Chains CK eqns Classes Hitting times Rec./trans. Strong Markov Stat. distr. Reversibility * Markov Chains
Markov Chains A random process X is a family {X t : t T } of random variables indexed by some set T. When T = {0, 1, 2,... } one speaks about a discrete-time process, for T = R or T = [0, ) one has a continuous-time
More informationStatistics 150: Spring 2007
Statistics 150: Spring 2007 April 23, 2008 0-1 1 Limiting Probabilities If the discrete-time Markov chain with transition probabilities p ij is irreducible and positive recurrent; then the limiting probabilities
More informationPart IA Probability. Definitions. Based on lectures by R. Weber Notes taken by Dexter Chua. Lent 2015
Part IA Probability Definitions Based on lectures by R. Weber Notes taken by Dexter Chua Lent 2015 These notes are not endorsed by the lecturers, and I have modified them (often significantly) after lectures.
More informationRecap. Probability, stochastic processes, Markov chains. ELEC-C7210 Modeling and analysis of communication networks
Recap Probability, stochastic processes, Markov chains ELEC-C7210 Modeling and analysis of communication networks 1 Recap: Probability theory important distributions Discrete distributions Geometric distribution
More informationIEOR 6711, HMWK 5, Professor Sigman
IEOR 6711, HMWK 5, Professor Sigman 1. Semi-Markov processes: Consider an irreducible positive recurrent discrete-time Markov chain {X n } with transition matrix P (P i,j ), i, j S, and finite state space.
More information4 Branching Processes
4 Branching Processes Organise by generations: Discrete time. If P(no offspring) 0 there is a probability that the process will die out. Let X = number of offspring of an individual p(x) = P(X = x) = offspring
More informationSMSTC (2007/08) Probability.
SMSTC (27/8) Probability www.smstc.ac.uk Contents 12 Markov chains in continuous time 12 1 12.1 Markov property and the Kolmogorov equations.................... 12 2 12.1.1 Finite state space.................................
More informationMARKOV PROCESSES. Valerio Di Valerio
MARKOV PROCESSES Valerio Di Valerio Stochastic Process Definition: a stochastic process is a collection of random variables {X(t)} indexed by time t T Each X(t) X is a random variable that satisfy some
More informationTHE QUEEN S UNIVERSITY OF BELFAST
THE QUEEN S UNIVERSITY OF BELFAST 0SOR20 Level 2 Examination Statistics and Operational Research 20 Probability and Distribution Theory Wednesday 4 August 2002 2.30 pm 5.30 pm Examiners { Professor R M
More informationStochastic process. X, a series of random variables indexed by t
Stochastic process X, a series of random variables indexed by t X={X(t), t 0} is a continuous time stochastic process X={X(t), t=0,1, } is a discrete time stochastic process X(t) is the state at time t,
More informationContinuous-Time Markov Chain
Continuous-Time Markov Chain Consider the process {X(t),t 0} with state space {0, 1, 2,...}. The process {X(t),t 0} is a continuous-time Markov chain if for all s, t 0 and nonnegative integers i, j, x(u),
More informationLecture 5: Moment generating functions
Lecture 5: Moment generating functions Definition 2.3.6. The moment generating function (mgf) of a random variable X is { x e tx f M X (t) = E(e tx X (x) if X has a pmf ) = etx f X (x)dx if X has a pdf
More information1 Types of stochastic models
1 Types of stochastic models Models so far discussed are all deterministic, meaning that, if the present state were perfectly known, it would be possible to predict exactly all future states. We have seen
More informationIrreducibility. Irreducible. every state can be reached from every other state For any i,j, exist an m 0, such that. Absorbing state: p jj =1
Irreducibility Irreducible every state can be reached from every other state For any i,j, exist an m 0, such that i,j are communicate, if the above condition is valid Irreducible: all states are communicate
More informationProbability Models. 4. What is the definition of the expectation of a discrete random variable?
1 Probability Models The list of questions below is provided in order to help you to prepare for the test and exam. It reflects only the theoretical part of the course. You should expect the questions
More informationOutlines. Discrete Time Markov Chain (DTMC) Continuous Time Markov Chain (CTMC)
Markov Chains (2) Outlines Discrete Time Markov Chain (DTMC) Continuous Time Markov Chain (CTMC) 2 pj ( n) denotes the pmf of the random variable p ( n) P( X j) j We will only be concerned with homogenous
More informationLecture Notes 7 Random Processes. Markov Processes Markov Chains. Random Processes
Lecture Notes 7 Random Processes Definition IID Processes Bernoulli Process Binomial Counting Process Interarrival Time Process Markov Processes Markov Chains Classification of States Steady State Probabilities
More informationBirth and Death Processes. Birth and Death Processes. Linear Growth with Immigration. Limiting Behaviour for Birth and Death Processes
DTU Informatics 247 Stochastic Processes 6, October 27 Today: Limiting behaviour of birth and death processes Birth and death processes with absorbing states Finite state continuous time Markov chains
More informationLectures on Probability and Statistical Models
Lectures on Probability and Statistical Models Phil Pollett Professor of Mathematics The University of Queensland c These materials can be used for any educational purpose provided they are are not altered
More informationTMA4265 Stochastic processes ST2101 Stochastic simulation and modelling
Norwegian University of Science and Technology Department of Mathematical Sciences Page of 7 English Contact during examination: Øyvind Bakke Telephone: 73 9 8 26, 99 4 673 TMA426 Stochastic processes
More informationContinuous Distributions
A normal distribution and other density functions involving exponential forms play the most important role in probability and statistics. They are related in a certain way, as summarized in a diagram later
More informationSTAT STOCHASTIC PROCESSES. Contents
STAT 3911 - STOCHASTIC PROCESSES ANDREW TULLOCH Contents 1. Stochastic Processes 2 2. Classification of states 2 3. Limit theorems for Markov chains 4 4. First step analysis 5 5. Branching processes 5
More informationSTAT/MATH 395 A - PROBABILITY II UW Winter Quarter Moment functions. x r p X (x) (1) E[X r ] = x r f X (x) dx (2) (x E[X]) r p X (x) (3)
STAT/MATH 395 A - PROBABILITY II UW Winter Quarter 07 Néhémy Lim Moment functions Moments of a random variable Definition.. Let X be a rrv on probability space (Ω, A, P). For a given r N, E[X r ], if it
More informationUNIVERSITY OF LONDON IMPERIAL COLLEGE LONDON
UNIVERSITY OF LONDON IMPERIAL COLLEGE LONDON BSc and MSci EXAMINATIONS (MATHEMATICS) MAY JUNE 23 This paper is also taken for the relevant examination for the Associateship. M3S4/M4S4 (SOLUTIONS) APPLIED
More informationStat410 Probability and Statistics II (F16)
Stat4 Probability and Statistics II (F6 Exponential, Poisson and Gamma Suppose on average every /λ hours, a Stochastic train arrives at the Random station. Further we assume the waiting time between two
More informationMarkov chains. 1 Discrete time Markov chains. c A. J. Ganesh, University of Bristol, 2015
Markov chains c A. J. Ganesh, University of Bristol, 2015 1 Discrete time Markov chains Example: A drunkard is walking home from the pub. There are n lampposts between the pub and his home, at each of
More informationPart IA Probability. Theorems. Based on lectures by R. Weber Notes taken by Dexter Chua. Lent 2015
Part IA Probability Theorems Based on lectures by R. Weber Notes taken by Dexter Chua Lent 2015 These notes are not endorsed by the lecturers, and I have modified them (often significantly) after lectures.
More informationSTATS 3U03. Sang Woo Park. March 29, Textbook: Inroduction to stochastic processes. Requirement: 5 assignments, 2 tests, and 1 final
STATS 3U03 Sang Woo Park March 29, 2017 Course Outline Textbook: Inroduction to stochastic processes Requirement: 5 assignments, 2 tests, and 1 final Test 1: Friday, February 10th Test 2: Friday, March
More informationHANDBOOK OF APPLICABLE MATHEMATICS
HANDBOOK OF APPLICABLE MATHEMATICS Chief Editor: Walter Ledermann Volume II: Probability Emlyn Lloyd University oflancaster A Wiley-Interscience Publication JOHN WILEY & SONS Chichester - New York - Brisbane
More informationContinuous time Markov chains
Continuous time Markov chains Alejandro Ribeiro Dept. of Electrical and Systems Engineering University of Pennsylvania aribeiro@seas.upenn.edu http://www.seas.upenn.edu/users/~aribeiro/ October 16, 2017
More informationLecture 21. David Aldous. 16 October David Aldous Lecture 21
Lecture 21 David Aldous 16 October 2015 In continuous time 0 t < we specify transition rates or informally P(X (t+δ)=j X (t)=i, past ) q ij = lim δ 0 δ P(X (t + dt) = j X (t) = i) = q ij dt but note these
More informationMath Homework 5 Solutions
Math 45 - Homework 5 Solutions. Exercise.3., textbook. The stochastic matrix for the gambler problem has the following form, where the states are ordered as (,, 4, 6, 8, ): P = The corresponding diagram
More informationChapter 5. Continuous-Time Markov Chains. Prof. Shun-Ren Yang Department of Computer Science, National Tsing Hua University, Taiwan
Chapter 5. Continuous-Time Markov Chains Prof. Shun-Ren Yang Department of Computer Science, National Tsing Hua University, Taiwan Continuous-Time Markov Chains Consider a continuous-time stochastic process
More informationCDA5530: Performance Models of Computers and Networks. Chapter 3: Review of Practical
CDA5530: Performance Models of Computers and Networks Chapter 3: Review of Practical Stochastic Processes Definition Stochastic ti process X = {X(t), t T} is a collection of random variables (rvs); one
More informationFigure 10.1: Recording when the event E occurs
10 Poisson Processes Let T R be an interval. A family of random variables {X(t) ; t T} is called a continuous time stochastic process. We often consider T = [0, 1] and T = [0, ). As X(t) is a random variable
More informationReview 1: STAT Mark Carpenter, Ph.D. Professor of Statistics Department of Mathematics and Statistics. August 25, 2015
Review : STAT 36 Mark Carpenter, Ph.D. Professor of Statistics Department of Mathematics and Statistics August 25, 25 Support of a Random Variable The support of a random variable, which is usually denoted
More informationContinuous Time Markov Chains
Continuous Time Markov Chains Stochastic Processes - Lecture Notes Fatih Cavdur to accompany Introduction to Probability Models by Sheldon M. Ross Fall 2015 Outline Introduction Continuous-Time Markov
More informationSampling Distributions
In statistics, a random sample is a collection of independent and identically distributed (iid) random variables, and a sampling distribution is the distribution of a function of random sample. For example,
More informationSTAT 3610: Review of Probability Distributions
STAT 3610: Review of Probability Distributions Mark Carpenter Professor of Statistics Department of Mathematics and Statistics August 25, 2015 Support of a Random Variable Definition The support of a random
More informationLecture 20: Reversible Processes and Queues
Lecture 20: Reversible Processes and Queues 1 Examples of reversible processes 11 Birth-death processes We define two non-negative sequences birth and death rates denoted by {λ n : n N 0 } and {µ n : n
More information{σ x >t}p x. (σ x >t)=e at.
3.11. EXERCISES 121 3.11 Exercises Exercise 3.1 Consider the Ornstein Uhlenbeck process in example 3.1.7(B). Show that the defined process is a Markov process which converges in distribution to an N(0,σ
More informationMarkov processes and queueing networks
Inria September 22, 2015 Outline Poisson processes Markov jump processes Some queueing networks The Poisson distribution (Siméon-Denis Poisson, 1781-1840) { } e λ λ n n! As prevalent as Gaussian distribution
More informationThings to remember when learning probability distributions:
SPECIAL DISTRIBUTIONS Some distributions are special because they are useful They include: Poisson, exponential, Normal (Gaussian), Gamma, geometric, negative binomial, Binomial and hypergeometric distributions
More informationLet (Ω, F) be a measureable space. A filtration in discrete time is a sequence of. F s F t
2.2 Filtrations Let (Ω, F) be a measureable space. A filtration in discrete time is a sequence of σ algebras {F t } such that F t F and F t F t+1 for all t = 0, 1,.... In continuous time, the second condition
More informationThe SIS and SIR stochastic epidemic models revisited
The SIS and SIR stochastic epidemic models revisited Jesús Artalejo Faculty of Mathematics, University Complutense of Madrid Madrid, Spain jesus_artalejomat.ucm.es BCAM Talk, June 16, 2011 Talk Schedule
More information1 Solution to Problem 2.1
Solution to Problem 2. I incorrectly worked this exercise instead of 2.2, so I decided to include the solution anyway. a) We have X Y /3, which is a - function. It maps the interval, ) where X lives) onto
More informationEcon 508B: Lecture 5
Econ 508B: Lecture 5 Expectation, MGF and CGF Hongyi Liu Washington University in St. Louis July 31, 2017 Hongyi Liu (Washington University in St. Louis) Math Camp 2017 Stats July 31, 2017 1 / 23 Outline
More information1 Probability and Random Variables
1 Probability and Random Variables The models that you have seen thus far are deterministic models. For any time t, there is a unique solution X(t). On the other hand, stochastic models will result in
More informationChapter 5. Chapter 5 sections
1 / 43 sections Discrete univariate distributions: 5.2 Bernoulli and Binomial distributions Just skim 5.3 Hypergeometric distributions 5.4 Poisson distributions Just skim 5.5 Negative Binomial distributions
More informationQuantitative Model Checking (QMC) - SS12
Quantitative Model Checking (QMC) - SS12 Lecture 06 David Spieler Saarland University, Germany June 4, 2012 1 / 34 Deciding Bisimulations 2 / 34 Partition Refinement Algorithm Notation: A partition P over
More informationLecture 11: Introduction to Markov Chains. Copyright G. Caire (Sample Lectures) 321
Lecture 11: Introduction to Markov Chains Copyright G. Caire (Sample Lectures) 321 Discrete-time random processes A sequence of RVs indexed by a variable n 2 {0, 1, 2,...} forms a discretetime random process
More informationIntroduction to Queuing Networks Solutions to Problem Sheet 3
Introduction to Queuing Networks Solutions to Problem Sheet 3 1. (a) The state space is the whole numbers {, 1, 2,...}. The transition rates are q i,i+1 λ for all i and q i, for all i 1 since, when a bus
More informationLTCC. Exercises. (1) Two possible weather conditions on any day: {rainy, sunny} (2) Tomorrow s weather depends only on today s weather
1. Markov chain LTCC. Exercises Let X 0, X 1, X 2,... be a Markov chain with state space {1, 2, 3, 4} and transition matrix 1/2 1/2 0 0 P = 0 1/2 1/3 1/6. 0 0 0 1 (a) What happens if the chain starts in
More informationThe story of the film so far... Mathematics for Informatics 4a. Continuous-time Markov processes. Counting processes
The story of the film so far... Mathematics for Informatics 4a José Figueroa-O Farrill Lecture 19 28 March 2012 We have been studying stochastic processes; i.e., systems whose time evolution has an element
More informationExponential Distribution and Poisson Process
Exponential Distribution and Poisson Process Stochastic Processes - Lecture Notes Fatih Cavdur to accompany Introduction to Probability Models by Sheldon M. Ross Fall 215 Outline Introduction Exponential
More informationAssignment 3 with Reference Solutions
Assignment 3 with Reference Solutions Exercise 3.: Poisson Process Given are k independent sources s i of jobs as shown in the figure below. The interarrival time between jobs for each source is exponentially
More informationBMIR Lecture Series on Probability and Statistics Fall 2015 Discrete RVs
Lecture #7 BMIR Lecture Series on Probability and Statistics Fall 2015 Department of Biomedical Engineering and Environmental Sciences National Tsing Hua University 7.1 Function of Single Variable Theorem
More informationMoments. Raw moment: February 25, 2014 Normalized / Standardized moment:
Moments Lecture 10: Central Limit Theorem and CDFs Sta230 / Mth 230 Colin Rundel Raw moment: Central moment: µ n = EX n ) µ n = E[X µ) 2 ] February 25, 2014 Normalized / Standardized moment: µ n σ n Sta230
More informationDiscrete Distributions
Chapter 2 Discrete Distributions 2.1 Random Variables of the Discrete Type An outcome space S is difficult to study if the elements of S are not numbers. However, we can associate each element/outcome
More informationChapter 6: Random Processes 1
Chapter 6: Random Processes 1 Yunghsiang S. Han Graduate Institute of Communication Engineering, National Taipei University Taiwan E-mail: yshan@mail.ntpu.edu.tw 1 Modified from the lecture notes by Prof.
More informationLECTURE #6 BIRTH-DEATH PROCESS
LECTURE #6 BIRTH-DEATH PROCESS 204528 Queueing Theory and Applications in Networks Assoc. Prof., Ph.D. (รศ.ดร. อน นต ผลเพ ม) Computer Engineering Department, Kasetsart University Outline 2 Birth-Death
More information12 Markov chains The Markov property
12 Markov chains Summary. The chapter begins with an introduction to discrete-time Markov chains, and to the use of matrix products and linear algebra in their study. The concepts of recurrence and transience
More informationLIST OF FORMULAS FOR STK1100 AND STK1110
LIST OF FORMULAS FOR STK1100 AND STK1110 (Version of 11. November 2015) 1. Probability Let A, B, A 1, A 2,..., B 1, B 2,... be events, that is, subsets of a sample space Ω. a) Axioms: A probability function
More informationAPM 541: Stochastic Modelling in Biology Discrete-time Markov Chains. Jay Taylor Fall Jay Taylor (ASU) APM 541 Fall / 92
APM 541: Stochastic Modelling in Biology Discrete-time Markov Chains Jay Taylor Fall 2013 Jay Taylor (ASU) APM 541 Fall 2013 1 / 92 Outline 1 Motivation 2 Markov Processes 3 Markov Chains: Basic Properties
More informationStatistics 992 Continuous-time Markov Chains Spring 2004
Summary Continuous-time finite-state-space Markov chains are stochastic processes that are widely used to model the process of nucleotide substitution. This chapter aims to present much of the mathematics
More informationNorthwestern University Department of Electrical Engineering and Computer Science
Northwestern University Department of Electrical Engineering and Computer Science EECS 454: Modeling and Analysis of Communication Networks Spring 2008 Probability Review As discussed in Lecture 1, probability
More informationThe Transition Probability Function P ij (t)
The Transition Probability Function P ij (t) Consider a continuous time Markov chain {X(t), t 0}. We are interested in the probability that in t time units the process will be in state j, given that it
More informationData analysis and stochastic modeling
Data analysis and stochastic modeling Lecture 7 An introduction to queueing theory Guillaume Gravier guillaume.gravier@irisa.fr with a lot of help from Paul Jensen s course http://www.me.utexas.edu/ jensen/ormm/instruction/powerpoint/or_models_09/14_queuing.ppt
More informationCDA6530: Performance Models of Computers and Networks. Chapter 3: Review of Practical Stochastic Processes
CDA6530: Performance Models of Computers and Networks Chapter 3: Review of Practical Stochastic Processes Definition Stochastic process X = {X(t), t2 T} is a collection of random variables (rvs); one rv
More informationThe Wright-Fisher Model and Genetic Drift
The Wright-Fisher Model and Genetic Drift January 22, 2015 1 1 Hardy-Weinberg Equilibrium Our goal is to understand the dynamics of allele and genotype frequencies in an infinite, randomlymating population
More informationSIMILAR MARKOV CHAINS
SIMILAR MARKOV CHAINS by Phil Pollett The University of Queensland MAIN REFERENCES Convergence of Markov transition probabilities and their spectral properties 1. Vere-Jones, D. Geometric ergodicity in
More informationSTAT/MATH 395 PROBABILITY II
STAT/MATH 395 PROBABILITY II Chapter 6 : Moment Functions Néhémy Lim 1 1 Department of Statistics, University of Washington, USA Winter Quarter 2016 of Common Distributions Outline 1 2 3 of Common Distributions
More informationStat 475 Life Contingencies I. Chapter 2: Survival models
Stat 475 Life Contingencies I Chapter 2: Survival models The future lifetime random variable Notation We are interested in analyzing and describing the future lifetime of an individual. We use (x) to denote
More informationQuasi-Stationary Distributions in Linear Birth and Death Processes
Quasi-Stationary Distributions in Linear Birth and Death Processes Wenbo Liu & Hanjun Zhang School of Mathematics and Computational Science, Xiangtan University Hunan 411105, China E-mail: liuwenbo10111011@yahoo.com.cn
More informationMathematical Methods for Computer Science
Mathematical Methods for Computer Science Computer Science Tripos, Part IB Michaelmas Term 2016/17 R.J. Gibbens Problem sheets for Probability methods William Gates Building 15 JJ Thomson Avenue Cambridge
More informationLinear-fractional branching processes with countably many types
Branching processes and and their applications Badajoz, April 11-13, 2012 Serik Sagitov Chalmers University and University of Gothenburg Linear-fractional branching processes with countably many types
More informationPROBABILITY: LIMIT THEOREMS II, SPRING HOMEWORK PROBLEMS
PROBABILITY: LIMIT THEOREMS II, SPRING 218. HOMEWORK PROBLEMS PROF. YURI BAKHTIN Instructions. You are allowed to work on solutions in groups, but you are required to write up solutions on your own. Please
More informationP 2 (t) = o(t). e λt. 1. keep on adding independent, exponentially distributed (with mean 1 λ ) inter-arrival times, or
POISSON PROCESS PP): State space consists of all non-negative integers, time is continuous. The process is time homogeneous, with independent increments and P 0 t) = λ t+ot) P t) = λ t+ot) P 2 t) = ot).
More information1 Continuous-time chains, finite state space
Université Paris Diderot 208 Markov chains Exercises 3 Continuous-time chains, finite state space Exercise Consider a continuous-time taking values in {, 2, 3}, with generator 2 2. 2 2 0. Draw the diagramm
More informationA review of Continuous Time MC STA 624, Spring 2015
A review of Continuous Time MC STA 624, Spring 2015 Ruriko Yoshida Dept. of Statistics University of Kentucky polytopes.net STA 624 1 Continuous Time Markov chains Definition A continuous time stochastic
More informationSystem Simulation Part II: Mathematical and Statistical Models Chapter 5: Statistical Models
System Simulation Part II: Mathematical and Statistical Models Chapter 5: Statistical Models Fatih Cavdur fatihcavdur@uludag.edu.tr March 20, 2012 Introduction Introduction The world of the model-builder
More information1 IEOR 4701: Continuous-Time Markov Chains
Copyright c 2006 by Karl Sigman 1 IEOR 4701: Continuous-Time Markov Chains A Markov chain in discrete time, {X n : n 0}, remains in any state for exactly one unit of time before making a transition (change
More information3 Continuous Random Variables
Jinguo Lian Math437 Notes January 15, 016 3 Continuous Random Variables Remember that discrete random variables can take only a countable number of possible values. On the other hand, a continuous random
More informationTCOM 501: Networking Theory & Fundamentals. Lecture 6 February 19, 2003 Prof. Yannis A. Korilis
TCOM 50: Networking Theory & Fundamentals Lecture 6 February 9, 003 Prof. Yannis A. Korilis 6- Topics Time-Reversal of Markov Chains Reversibility Truncating a Reversible Markov Chain Burke s Theorem Queues
More informationBayesian Methods with Monte Carlo Markov Chains II
Bayesian Methods with Monte Carlo Markov Chains II Henry Horng-Shing Lu Institute of Statistics National Chiao Tung University hslu@stat.nctu.edu.tw http://tigpbp.iis.sinica.edu.tw/courses.htm 1 Part 3
More informationMidterm 2 Review. CS70 Summer Lecture 6D. David Dinh 28 July UC Berkeley
Midterm 2 Review CS70 Summer 2016 - Lecture 6D David Dinh 28 July 2016 UC Berkeley Midterm 2: Format 8 questions, 190 points, 110 minutes (same as MT1). Two pages (one double-sided sheet) of handwritten
More information