= p(t)(1 λδt + o(δt)) (from axioms) Now p(0) = 1, so c = 0 giving p(t) = e λt as required. 5 For a non-homogeneous process we have

Similar documents
UNIVERSITY OF LONDON IMPERIAL COLLEGE LONDON

6 Stationary Distributions

THE QUEEN S UNIVERSITY OF BELFAST

Solution: (Course X071570: Stochastic Processes)

Continuous-Time Markov Chain

Continuous Time Processes

Birth and Death Processes. Birth and Death Processes. Linear Growth with Immigration. Limiting Behaviour for Birth and Death Processes

STAT STOCHASTIC PROCESSES. Contents

CDA5530: Performance Models of Computers and Networks. Chapter 3: Review of Practical

Statistics 150: Spring 2007

4 Branching Processes

Continuous Time Markov Chains

Quantitative Model Checking (QMC) - SS12

1 Random Variables and Probability Distributions

TMA4265 Stochastic processes ST2101 Stochastic simulation and modelling

MATH 829: Introduction to Data Mining and Analysis Consistency of Linear Regression

EXAM IN COURSE TMA4265 STOCHASTIC PROCESSES Wednesday 7. August, 2013 Time: 9:00 13:00

Chapter 7: Special Distributions

The Transition Probability Function P ij (t)

Stat-491-Fall2014-Assignment-III

= P{X 0. = i} (1) If the MC has stationary transition probabilities then, = i} = P{X n+1

Approximating Deterministic Changes to Ph(t)/Ph(t)/1/c and Ph(t)/M(t)/s/c Queueing Models

6 Continuous-Time Birth and Death Chains

Let (Ω, F) be a measureable space. A filtration in discrete time is a sequence of. F s F t

Stochastic process. X, a series of random variables indexed by t

STAT 380 Continuous Time Markov Chains

STOCHASTIC PROCESSES Basic notions

MS&E 321 Spring Stochastic Systems June 1, 2013 Prof. Peter W. Glynn Page 1 of 10. x n+1 = f(x n ),

Lecture 4a: Continuous-Time Markov Chain Models

Solutions For Stochastic Process Final Exam

1 Probability Spaces and Random Variables

Markov Processes and Queues

MATH37012 Week 10. Dr Jonathan Bagley. Semester

Irreducibility. Irreducible. every state can be reached from every other state For any i,j, exist an m 0, such that. Absorbing state: p jj =1

Probability Distributions

Module 8. Lecture 3: Markov chain

Examination paper for TMA4265 Stochastic Processes

Stat 516, Homework 1

Stochastic Processes

EXAMINATIONS OF THE ROYAL STATISTICAL SOCIETY

Example: physical systems. If the state space. Example: speech recognition. Context can be. Example: epidemics. Suppose each infected

Manual for SOA Exam MLC.

Chapter 5. Continuous-Time Markov Chains. Prof. Shun-Ren Yang Department of Computer Science, National Tsing Hua University, Taiwan

Lecture 21. David Aldous. 16 October David Aldous Lecture 21

ELEMENTS OF PROBABILITY THEORY

8. Statistical Equilibrium and Classification of States: Discrete Time Markov Chains

ECE353: Probability and Random Processes. Lecture 18 - Stochastic Processes

Outlines. Discrete Time Markov Chain (DTMC) Continuous Time Markov Chain (CTMC)

Homework Solution 4 for APPM4/5560 Markov Processes

Probability Distributions

Markov Chains and MCMC

Statistics 253/317 Introduction to Probability Models. Winter Midterm Exam Monday, Feb 10, 2014

Approximate Counting and Markov Chain Monte Carlo

Markov Processes Hamid R. Rabiee

CDA6530: Performance Models of Computers and Networks. Chapter 3: Review of Practical Stochastic Processes

ECE 6960: Adv. Random Processes & Applications Lecture Notes, Fall 2010

Combining Logistic Regression with Kriging for Mapping the Risk of Occurrence of Unexploded Ordnance (UXO)

SMSTC (2007/08) Probability.

Question Points Score Total: 70

Statistics 992 Continuous-time Markov Chains Spring 2004

Applied Stochastic Processes

Markov Chain Monte Carlo The Metropolis-Hastings Algorithm

Positive and null recurrent-branching Process

ECE534, Spring 2018: Solutions for Problem Set #5

Markov chains. 1 Discrete time Markov chains. c A. J. Ganesh, University of Bristol, 2015

Data analysis and stochastic modeling

Stochastic Modelling Unit 1: Markov chain models

STA 624 Practice Exam 2 Applied Stochastic Processes Spring, 2008

Analysis of some entrance probabilities for killed birth-death processes

Variance reduction. Michel Bierlaire. Transport and Mobility Laboratory. Variance reduction p. 1/18

Mathematical Methods for Computer Science

THE ROYAL STATISTICAL SOCIETY GRADUATE DIPLOMA

(b) What is the variance of the time until the second customer arrives, starting empty, assuming that we measure time in minutes?

Part I Stochastic variables and Markov chains

8 STOCHASTIC PROCESSES

Markov Chains CK eqns Classes Hitting times Rec./trans. Strong Markov Stat. distr. Reversibility * Markov Chains

Chapter 6: Random Processes 1

Markov Chains. X(t) is a Markov Process if, for arbitrary times t 1 < t 2 <... < t k < t k+1. If X(t) is discrete-valued. If X(t) is continuous-valued

Be sure this exam has 9 pages including the cover. The University of British Columbia

Fig 1: Stationary and Non Stationary Time Series

Assignment 3 with Reference Solutions

UCSD ECE250 Handout #27 Prof. Young-Han Kim Friday, June 8, Practice Final Examination (Winter 2017)

1 Types of stochastic models

Problems 5: Continuous Markov process and the diffusion equation

MATH 56A: STOCHASTIC PROCESSES CHAPTER 1

Poisson processes and their properties

IEOR 6711, HMWK 5, Professor Sigman

Generating Functions. STAT253/317 Winter 2013 Lecture 8. More Properties of Generating Functions Random Walk w/ Reflective Boundary at 0

Problem Set 8

A review of Continuous Time MC STA 624, Spring 2015

A note on adiabatic theorem for Markov chains and adiabatic quantum computation. Yevgeniy Kovchegov Oregon State University

On random walks. i=1 U i, where x Z is the starting point of

Module 9: Stationary Processes

IEOR 6711: Stochastic Models I, Fall 2003, Professor Whitt. Solutions to Final Exam: Thursday, December 18.

Lecture 11: Introduction to Markov Chains. Copyright G. Caire (Sample Lectures) 321

q P (T b < T a Z 0 = z, X 1 = 1) = p P (T b < T a Z 0 = z + 1) + q P (T b < T a Z 0 = z 1)

Stochastic modelling of epidemic spread

Before you begin read these instructions carefully.

Kolmogorov Equations and Markov Processes

Section 27. The Central Limit Theorem. Po-Ning Chen, Professor. Institute of Communications Engineering. National Chiao Tung University

Transcription:

. (a (i I: P(exactly event occurs in [t, t + δt = λδt + o(δt, [o(δt/δt 0 as δt 0]. II: P( or more events occur in [t, t + δt = o(δt. III: Occurrence of events after time t is indeendent of occurrence of events before t. Let X(t be the number of realizations by time t and let (t = P(X(t = 0 seen (t + δt = P(0 realizations in [0, t and 0 realizations in [t, t + δt (t + δt (t δt d(t dt = (t( λδt + o(δt (from axioms = (tλ + o(δt δt = (tλ log((t = λt + c (b (iii Now (0 =, so c = 0 giving (t = e λt as required. 5 For a non-homogeneous rocess we have So, μ(t = (t = e μ(t where μ(t = t 0 t 0 λ(u du + sin(u du = [u cos(u] t 0 = t cos(t +, method seen giving (t = ex(cos(t t. Let X(t, t be the number of questions answered in in [t, t, then method seen E(X(t, t + δt = δt + o(δt. + t Let D(t = number questions answered by t in deterministic model. Then D(t + δt = D(t + δt + o(δt + t D(t + δt D(t = δt + t + o(δt δt dd = dt + t D(t = log( + t + c and since D(0 = 0, we have c = 0 D(t = log( + t 6 MS/MSSolutions Page of 9

. (a (i Let Π(s be the gf for X, and let μ = E(X and μ n = E(Z n, so art seen μ = Π ( μ n = Π n( where Π n(s is the gf of Z n = Y +... + Y Zn with Y i being the number of offsring of individual i in generation n. From standard gf results, we have Π n (s = Π n [Π(s] Π n(s = Π n [Π(s] Π (s Π n( = Π n [Π(] Π ( = Π n (Π ( so μ = μ n μ = μ n μ =... = μ n. as E(X = we have E(Z n = n =. 5 Let σ = var(x and let σ n = var(z n. Π n(s = Π n [Π(s] Π (s Π n(s = Π n [Π(s] Π (s + Π n [Π(s] Π (s ( Now Π( =, Π ( = μ, Π ( = σ μ + μ. Also, since σ n = Π n( + μ n μ n, we have From (, Π n( = σ n μ n + μ n and Π n ( = σ n μ n + μ n. Π n( = Π n (Π ( + Π n (Π ( σ n μ n + μ n = (σ n μ n + μ n μ + μ n (σ μ + μ σ n = μ σ n + μ n σ Leading to σ n = μ n σ ( + μ + μ +... + μ n So, as μ = we have var(z n = σ n = nσ. 7 MS/MSSolutions Page of 9

(b (i We have Giving, Π(s = P(X = is i = α + ( αs. i=0 Π (s = ( αs So μ = Π ( = ( α. Let P(ultimate extinction = θ, then. μ θ = ultimate extinction certain.. μ > θ < ultimate extinction not certain. μ > when ( α > α <. So, when α < ultimate extinction is not certain. θ = smallest ositive solution of θ = Π(θ, and θ = : We know that θ = is a solution: θ = α + ( αθ θ θ + = 0 θ θ + = 0 θ θ + = (θ (θ + θ roots of θ + θ are ± 5, therefore method seen Probability of extinction = 5 MS/MSSolutions Page of 9

. (a (i Show that P(ever reach 0 starts at i = P(ever reach i starts at i P(ever reach i starts at i. P(ever reach 0 starts at = A A... A }{{} (satial homogeneity i times = (A i condition on first ste: A = P(visit origin start from = qp(visit origin start from 0 + P(visit origin start from (iii A A + q = 0 = q + A Solving A A + q = 0 gives A = or A = Now look for solutions in [0, ]: also, take ositive solutions: + q ± < 0, + q + + q + q 9 A = (Noting that when = /, A =. { + + q if > / if / unseen 5 MS/MSSolutions Page 5 of 9

(b (i Let X t be the value of the share on day t. Let Y t be the change in value of the share on day t. Then, X t = X 0 + Y +... + Y t. X 0 = 00 Y i = 0. with robability 0.5 0. with robability 0. 0 with robability 0. Giving E(Y i = 0. 0.5 0. 0. = 0. 0.06 = 0.. var(y i = E(Yi E (Y i E(Yi = 0. 0.5 + (0. 0. = 0.6 0.5 + 0.0 0. = 0.08 + 0.0 = 0.09 var(y i = 0.09 0. = 0.09 0.096 = 0.07. E(X t = 00 + 0.t = μ t For large t var(x t = 0.07t = σ t 5 X t X 0 = t Y i N(0.0t, 0.07t i= X t X 0 0.t 0.07t N(0, ( 0 (00 + 0. 65 P(X 65 > 0 = Φ 0.07 65 ( 0 μ 65 = Φ σ 65 MS/MSSolutions Page 6 of 9

. (a (i A Markov chain is irreducible if it has only one communicating class, i.e. there is a ath of non-zero robability from state i to state j and back again for all i, j in the samle sace. A Markov chain is aeriodic if all states have eriod, i.e. if where (n ij gcd{n : (n ij > 0} =. is the robability of going from i to j in n stes. (b (i State sace = {0,,, } (Number of balls in the first urn. 0 0 0 0 0 0 P = 0 0 0 0 0 transition diagram: seen 0 (iii Irreducible, finite state sace there is a unique stationary distribution. (iv Need aeriodicity for this distribution to also be limiting. Here the Markov chain is eriodic with eriod, so the stationary distribution is not limiting. (v Find stationary distribution, π from, π = πp, i=0 π i = and π 0 = π π 0 + π = π + π = π = π π = π 0 π = π 0 π = π 0 π 0 + π + π + π = π 0 + π 0 + π 0 + π 0 = π 0 = 8 Giving, π = ( 8, 8, 8, 8 Mean recurrence time to state 0, μ 0 is given by: μ 0 = π 0 = 8 MS/MSSolutions Page 7 of 9 5

5. (a (i Define seen = d dt P (t t=0 the transition rate matrix with elements q ij, and let P (t have element ij (t, i, j in the samle sace. The forward differential equations are given by: d P (t = P (t dt d dt ij(t = k ik (tq jk i, j. π i =, π = 0 (or π = P (tπ, t. i (b (i Let state 0 good mood, and state bad mood. 0 (δt = P(0 in [t, t + δt = α δt + o(δt 0 (δt = P( 0 in [t, t + δt = β δt + o(δt We have, by definition, { + δt qii + o(δt i = j ij (δt = δt q ij + o(δt i j small δt So, as the rows of sum to zero, we have ( α α = β β assume true for n = k, let n = k +. unseen = k+ = k = (( k from assumtion = ( k ( α + αβ α αβ αβ β αβ + β ( α α = ( α β β β k+ = ( k ( = ( k = ( α β true for n = k+ if true for n = k, as true for n = ( = ( α β 0, result follows by induction. MS/MSSolutions Page 8 of 9

ex(t = I + = I + n= t n n! (n (t( n n=0 n! = I + {ex(t( } = I + α + β + ex(t( (iii Backward differential equations: d dt P (t = d { I + dt α + β + = ex(t( { P (t = I + α + β + } ex(t( ex(t( = + α + β + ex(t( ( ( = + + ex(t( α + β = ex(t( } i.e. P (t satisfies the backward differential equations: d dt P (t = P (t. (iv stationary distribution satisfies ( α α (π 0 π β β = (0 0 π 0 + π =. απ 0 + βπ = 0 and π 0 + π = απ 0 + β( π 0 = 0 π 0 = β α + β π = β α + β = α α + β giving π = ( β α + β, α α + β MS/MSSolutions Page 9 of 9