Let (Ω, F) be a measureable space. A filtration in discrete time is a sequence of. F s F t

Similar documents
Markov Chains. X(t) is a Markov Process if, for arbitrary times t 1 < t 2 <... < t k < t k+1. If X(t) is discrete-valued. If X(t) is continuous-valued

MARKOV PROCESSES. Valerio Di Valerio

Lecture 4a: Continuous-Time Markov Chain Models

2. Transience and Recurrence

IEOR 6711, HMWK 5, Professor Sigman

8. Statistical Equilibrium and Classification of States: Discrete Time Markov Chains

Random Times and Their Properties

Stochastic Processes

Markov chains. Randomness and Computation. Markov chains. Markov processes

ELEMENTS OF PROBABILITY THEORY

Markov Chains, Stochastic Processes, and Matrix Decompositions

Statistics 150: Spring 2007

Lecture 5. If we interpret the index n 0 as time, then a Markov chain simply requires that the future depends only on the present and not on the past.

Markov Chains. Arnoldo Frigessi Bernd Heidergott November 4, 2015

Markov Processes Hamid R. Rabiee

Solutions For Stochastic Process Final Exam

Stochastic process. X, a series of random variables indexed by t

Applied Stochastic Processes

Basic Definitions: Indexed Collections and Random Functions

(b) What is the variance of the time until the second customer arrives, starting empty, assuming that we measure time in minutes?

Notes on uniform convergence

MASSACHUSETTS INSTITUTE OF TECHNOLOGY 6.436J/15.085J Fall 2008 Lecture 8 10/1/2008 CONTINUOUS RANDOM VARIABLES

POISSON PROCESSES 1. THE LAW OF SMALL NUMBERS

Random point patterns and counting processes

Discrete Markov Chain. Theory and use

Statistics 992 Continuous-time Markov Chains Spring 2004

Universal examples. Chapter The Bernoulli process

CDA6530: Performance Models of Computers and Networks. Chapter 3: Review of Practical Stochastic Processes

6. Brownian Motion. Q(A) = P [ ω : x(, ω) A )

Markov Chains (Part 3)

Markov Chains and Stochastic Sampling

The exponential distribution and the Poisson process

Recap. Probability, stochastic processes, Markov chains. ELEC-C7210 Modeling and analysis of communication networks

STOCHASTIC PROCESSES Basic notions

Lecture 1: Brief Review on Stochastic Processes

Lecture 7. We can regard (p(i, j)) as defining a (maybe infinite) matrix P. Then a basic fact is

Stochastic Processes

Stat 516, Homework 1


MATH 56A: STOCHASTIC PROCESSES CHAPTER 6

Ergodic Theorems. Samy Tindel. Purdue University. Probability Theory 2 - MA 539. Taken from Probability: Theory and examples by R.

Continuous time Markov chains

Examination paper for TMA4265 Stochastic Processes

Homework 4 due on Thursday, December 15 at 5 PM (hard deadline).

25.1 Ergodicity and Metric Transitivity

STAT/MATH 395 A - PROBABILITY II UW Winter Quarter Moment functions. x r p X (x) (1) E[X r ] = x r f X (x) dx (2) (x E[X]) r p X (x) (3)

CS 798: Homework Assignment 3 (Queueing Theory)

Treball final de grau GRAU DE MATEMÀTIQUES Facultat de Matemàtiques Universitat de Barcelona MARKOV CHAINS

Lecture 20 : Markov Chains

On asymptotic behavior of a finite Markov chain

Question Points Score Total: 70

Markov Chains Handout for Stat 110

Insert your Booktitle, Subtitle, Edition

Manual for SOA Exam MLC.

CDA5530: Performance Models of Computers and Networks. Chapter 3: Review of Practical

JOSEPH B. KADANE AND JIASHUN JIN, IN HONOR OF ARNOLD ZELLNER

MATH 56A: STOCHASTIC PROCESSES CHAPTER 1

A Concise Course on Stochastic Partial Differential Equations

IEOR 6711: Stochastic Models I, Fall 2003, Professor Whitt. Solutions to Final Exam: Thursday, December 18.

MS&E 321 Spring Stochastic Systems June 1, 2013 Prof. Peter W. Glynn Page 1 of 10

Lectures on Markov Chains

PROBABILITY: LIMIT THEOREMS II, SPRING HOMEWORK PROBLEMS

4.7.1 Computing a stationary distribution

Stochastic Proceses Poisson Process

Random Process. Random Process. Random Process. Introduction to Random Processes

Math Homework 5 Solutions

Point Process Control

Lecture 10. Theorem 1.1 [Ergodicity and extremality] A probability measure µ on (Ω, F) is ergodic for T if and only if it is an extremal point in M.

Functional Limit theorems for the quadratic variation of a continuous time random walk and for certain stochastic integrals

Discrete time Markov chains. Discrete Time Markov Chains, Limiting. Limiting Distribution and Classification. Regular Transition Probability Matrices

STAT STOCHASTIC PROCESSES. Contents

INTRODUCTION TO MARKOV CHAINS AND MARKOV CHAIN MIXING

Stochastic modelling of epidemic spread

Brownian Motion. 1 Definition Brownian Motion Wiener measure... 3

Stochastic Processes. Theory for Applications. Robert G. Gallager CAMBRIDGE UNIVERSITY PRESS

SMSTC (2007/08) Probability.

Figure 10.1: Recording when the event E occurs

Tyler Hofmeister. University of Calgary Mathematical and Computational Finance Laboratory

Inference for Stochastic Processes

Advanced Computer Networks Lecture 2. Markov Processes

Midterm 2 Review. CS70 Summer Lecture 6D. David Dinh 28 July UC Berkeley

Stochastic process for macro

Disjointness and Additivity

ADVANCED PROBABILITY: SOLUTIONS TO SHEET 1

Verona Course April Lecture 1. Review of probability

Probability. Lecture Notes. Adolfo J. Rumbos

Filtrations, Markov Processes and Martingales. Lectures on Lévy Processes and Stochastic Calculus, Braunschweig, Lecture 3: The Lévy-Itô Decomposition

Random Process Lecture 1. Fundamentals of Probability

Stochastic Models: Markov Chains and their Generalizations

Part I Stochastic variables and Markov chains

Things to remember when learning probability distributions:

Markov chains. 1 Discrete time Markov chains. c A. J. Ganesh, University of Bristol, 2015

MA3H2 Markov Processes and Percolation theory. Stefan Adams

Random variables. DS GA 1002 Probability and Statistics for Data Science.

Stochastic Processes

MS&E 321 Spring Stochastic Systems June 1, 2013 Prof. Peter W. Glynn Page 1 of 7

DISCRETE STOCHASTIC PROCESSES Draft of 2nd Edition

Essentials on the Analysis of Randomized Algorithms

STAT331 Lebesgue-Stieltjes Integrals, Martingales, Counting Processes

Ergodic Properties of Markov Processes

Transcription:

2.2 Filtrations Let (Ω, F) be a measureable space. A filtration in discrete time is a sequence of σ algebras {F t } such that F t F and F t F t+1 for all t = 0, 1,.... In continuous time, the second condition is replaced by F s F t for all s t. 3 Markov processes The idea of a Markov process is to capture the idea of a short-memory stochastic process: once its current state is known, past history is irrelevant from the point of view of predicting its future. Definition. Let (Ω, F) be a measurable space and let (P, F) be, respectively, a probability measure on and a filtration of this space. Let X be a stochastic process in discrete time on (Ω, F). Then X is called a (P, F)-Markov process if 1. X is F adapted, and 11

2. For each t Z + and each Borel set B B (R) P (X t+1 B F t ) = P (X t+1 B σ (X t )). (9) Sometimes when the probability measure and filtration are understood, we will talk simply of a Markov process. Remark. Often the filtration F is taken to be that generated by the process X itself. Proposition. Let (Ω, F, P, F) be a filtered probability space and let X be a (P, F)- Markov process. Let t, k Z +. Let f : R R be a Borel function such that f (X t+k ) is integrable. Then, E [f (X t+k ) F t ] = E [f (X t+k ) σ (X t )] (10) and hence there is a Borel function g : Z + Z + R R such that, for each t, E [f (X t+k ) F t ] = g (t, k, X t ). (11) Proof. To show it for k = 1, use that f (X t+1 ) is a σ (X t+1 ) measurable random variable and hence is the limit of a monotone increasing sequence of σ (X t+1 ) simple random variables. But such random variables are linear combinations of indicator functions of sets X t+1 (B) with B a Borel set. This completes the proof for k = 1. To prove it for arbitrary positive k, use induction. To prove it for k + 1 assuming it true for k, use the law of iterated expectations. The vector case is a simple extension of the scalar case. However, it is important that the definition of a vector Markov process is not that each component is Markov. 12

Instead, we require that all the relevant (for the future of X) bits of information in F t are in the σ algebra generated by all the stochastic variables in X t, i.e. σ (X t ) is defined as the single σ algebra σ ({X 1,t, X 2,t,..., X n,t }). This means that, for each Borel function f : R n R m, E [f (X t+k ) F t ] = E [f (X t+k ) σ (X t )] (12) and hence that there is a Borel function g : Z + R n R m such that, for each t, E [f (X t+1 ) F t ] = g (t, X t ). (13) (The case k = 1 is so important that we stress it here by ignoring greater values of k.) 3.0.1 Probability transition functions and time homogeneity Definition. Let (Ω, F, P, F) be a filtered probability space and let X be a (P, F)- Markov process. Then, for each t = 0, 1, 2... its probability transition function Q t : R B (R) [0, 1] is defined via Q t (x, B) = P (X t B X t = x). (14) Note that any Markov process has a sequence of probability transition functions. Note also that for each fixed t and x, Q t+1 (x, ) is a probability measure on B (R). Meanwhile, if we fix B, Q t+1 (X t ( ), B) is a random variable. Indeed, it is the conditional probability of X t+1 B given X t, i.e. Q t+1 (X t, B) = E I X [ ] t+1 σ (B) (Xt ). Moreover, the conditional expectation of any σ (X t+1 )-measurable random variable (given X t ) is an integral with respect to the measure Q t+1 in the following sense. 13

Proposition. Let (Ω, F, P, F) be a filtered probability space and let X be a (P, F)-Markov process. Let Q t be its probability transition functions and let Z L 1 (Ω, σ (X t+1 ), P ). Then, for each t = 0, 1,... E [Z X t ] = R f (y) Q t+1 (X t, dy) (15) or, put differently, we have for each t = 0, 1,... and each x, Proof. E [Z X t = x] = R f (y) Q t+1 (x, dy). (16) We will show it first for an indicator variable Z = I X t+1 (A) where A B (R). Then f (y) = I A (y). We now need to show that the random variable R f (y) Q t+1 (X t, dy) qualifies as the conditional expectation E [Z X t ]. Clearly it is σ (X t ) measurable. But does it integrate to the right thing? Well, let σ (X t ) and recall that, by ( ) definition, Q t+1 (X t, A) = E I X (A) σ (X t). Hence t+1 R f (y) Q t+1 (X t, dy) P (dω) = R I A (y) Q t+1 (X t, dy) P (dω) = = Q t+1 (X t, A) P (dω) = E ( ) I X (A) σ (X t) P (dω) = t+1 I X t+1 (A)P (dω). (17) Meanwhile, since Z = I X t+1 (A) we obviously have ZP (dω) = I X t+1 (A)P (dω). (18) 14

To show the theorem for an arbitrary Z L 1 (Ω, σ (X t+1 ), P ), use the Monotone Convergence Theorem. We now use the probability transition function to define a time homogeneous Markov process. Definition. Let (Ω, F, P, F) be a filtered probability space and let X be a (P, F)- Markov process. Let Q t t=1 be its probability transition functions. If there is a Q such that Q t = Q for all t = 1, 2,... then X is called a time homogeneous Markov process. Proposition. Let (Ω, F, P, F) be a filtered probability space and let X be a time homogeneous (P, F)-Markov process. For any nonnegative integers k, t, let Y t+k L 1 (Ω, σ (X t+k ), P ). Then for each k = 0, 1,... there is a Borel function g k : R R such that, for each t = 0, 1,... E [Y t+k F t ] = g k (X t ). (19) In particular, there is a Borel function h such that, for each t = 0, 1,... E [Y t+1 F t ] = h (X t ). (20) 3.1 Finite state Markov chains in discrete time This is perhaps the simplest class of Markov processes. Let (Ω, F, P ) be a probability space and let X = {x 1, x 2,..., x n } be a finite set. X : Z + X be a stochastic process. Denote by µ t the vector of probabilities that X t = x i and suppose there is 15

a sequence of matrices Γ t such that µ t+1 = Γ t µ t. (21) Then X is said to be a finite state Markov chain in discrete time. Call Γ t the probability transition matrix. The interpretation of (21) is the following. P (X t+1 = x i X t = x j ) = Γ t (i, j). If Γ t = Γ we call the process and time homogenenous or stationary. If Γ is sufficiently well behaved, then X has a unique stationary distribution. Definition. Let X be a stationary finite state Markov chain and let T be the time of the first visit to state j after t = 0. Then state j is called recurrent (opposite: transient) if P ({T < } X 0 = x j ) = 1. Definition. The j is called periodic with period δ > 1 if δ is the largest integer for which P ({T = nδ for some n 1} X 0 = j) = 1. If there is no such δ > 1, then j is called aperiodic. Definition. A Markov chain is said to be aperiodic if all its states are aperiodic. Definition. A state j can be reached from i if there exists an integer n 0 such that Γ n (i, j) > 0. 16

where by Γ n (i, j) we mean the (i, j) element of the matrix Γ n. Definition. A set of states is said to be closed if no state outside it can be reached from any state in it. Definition. A set of states is said to be ergodic if it is closed and no proper subset is closed. Definition. A Markov chain is called irreducible if its only closed set is the set of all states. Theorem. Let X be a finite state stationary Markov chain with transition matrix Γ and suppose X irreducible and aperiodic. Then µ = Γµ µ 1 = 1 has a unique solution µ and this µ has the property that for all µ 0 such that µ 1 = 1. Proof. See Cinlar (1975). lim t Γt µ 0 = µ 3.2 Finite state Markov chains in continuous time Let (Ω, F, P ) be a probability space and let X = {x 1, x 2,..., x n } be a finite set. Let X : R + X be a stochastic process. Denote by µ(t) the vector of probabilities 17

that X(t) = x i and suppose there is a matrix-valued function Γ(t) such that µ(t) satisfies µ(t) = Γ(t)µ(t). Then we call X a finite-state continuous time Markov chain and If Γ(t) = Γ we call the process and time homogenenous or stationary. 3.3 Poisson processes 3.3.1 The Poisson distribution Intuitively, the Poisson distribution comes from taking the limit of a sum of Bernoulli random variables. Definition. A random variable X is said to be Bernoulli if there is a real number 0 p 1 such that P ({X = 1}) = p and P ({X = 0}) = 1 p. Definition. A random variable Y is said to be binomial distribution if there is an integer n and a real number 0 p 1 such that n ( ) n P ({Y = k}) = p k (1 p) n k k k=0 where the binomial coefficient is defined as follows. ( ) n n! = k k!(n k)! 18

Proposition. If {X i, i = 1,..., n} is a collection of independent Bernoulli random variables, then Y defined via is binomially distributed. Y = n i=1 X i Proposition. Let X be binomial with parameters n, p. Then Proof. Exercise. E[X] = np. Now imagine that we are on a fishing expedition. For some reason we dip the fishing pole into the water 10 times for one minute at a time. The probability of catching a fish during any one dipping is p, independently of whether I caught a fish in any previous dipping. Then the total number of fish caught is a binomial variable. But what if I dip 20 times for half a minute, or 40 times for 15 seconds etc. Assume that the probability of catching a fish during a half minute dip is p/2 and similarly for shorter dips. What happens in the limit? As we take limits, let the expected total number of fish caught λ = np be constant and let n. (It follows that p 0.) Then it turns out that the distribution of the total number of fish caught tends to the Poisson distribution with parameter λ. Definition. A random variable X is said to be Poisson distributed if there is a real number λ 0 such that λ λk P ({X = k}) = e k!. Proposition Suppose two independent random variables X and Y are Poisson 19

distributed with parameters λ and µ, respectively. Then Z = X + Y is Poisson distributed with parameter λ + µ. Proof. Use the characteristic function. Definition. A continuous time stochastic process is a function X(t) where for each fixed t 0, X(t) is a random variable. Definition. Let (Ω, F, P, F) be a filtered probability space. A stochastic process {N(t, ω); t 0} is said to be a (P, F) Poisson process with intensity λ if N is F adapted. The trajectories of N are (with probability one) right continuous and piecewise continuous. N(0) = 0. N(t) = 0 or 1 (with probability one) where N(t) = N(t) N(t ). For all s t, N(t) N(s) is independent of F s. N(t) N(s) is Poisson distributed with parameter λ(t s), that is P (N(t) N(s) = k F s )) = P (N(t) N(s) = k) = e λ(t s) λk (t s) k. k! Proposition. The time between jumps is is exponentially distributed. More precisely, let τ be defined via τ = inf{n(t) > 0}. t 0 20

Then P ({τ < t}) = F (t) = 1 e λt and the corresponding probability density function is f(t) = F (t) = λe λt. Proof. Exercise. 21

References Chung, K. L. (2001). A Course in Probability Theory, Third Edition. Academic Press. Cinlar, E. (1975). Introduction to Stochastic Processes. Prentice Hall. Williams, D. (1991). Probability with Martingales. Cambridge University Press. 22