Markov Processes Hamid R. Rabiee

Similar documents
Markov Chains (Part 3)

Markov Chains (Part 4)

Lesson Plan. AM 121: Introduction to Optimization Models and Methods. Lecture 17: Markov Chains. Yiling Chen SEAS. Stochastic process Markov Chains

ISM206 Lecture, May 12, 2005 Markov Chain

Markov Chains Absorption (cont d) Hamid R. Rabiee

Markov Chains. As part of Interdisciplinary Mathematical Modeling, By Warren Weckesser Copyright c 2006.

8. Statistical Equilibrium and Classification of States: Discrete Time Markov Chains

Markov Chains Absorption Hamid R. Rabiee

Chapter 16 focused on decision making in the face of uncertainty about one future

Lecture 20 : Markov Chains

IEOR 6711: Professor Whitt. Introduction to Markov Chains

Probability, Random Processes and Inference

Markov Chains. X(t) is a Markov Process if, for arbitrary times t 1 < t 2 <... < t k < t k+1. If X(t) is discrete-valued. If X(t) is continuous-valued

STOCHASTIC PROCESSES Basic notions

1.3 Convergence of Regular Markov Chains

MATH 56A: STOCHASTIC PROCESSES CHAPTER 1

Markov Chains CK eqns Classes Hitting times Rec./trans. Strong Markov Stat. distr. Reversibility * Markov Chains

Outlines. Discrete Time Markov Chain (DTMC) Continuous Time Markov Chain (CTMC)

2. Transience and Recurrence

Markov chains. Randomness and Computation. Markov chains. Markov processes

Lecture 11: Introduction to Markov Chains. Copyright G. Caire (Sample Lectures) 321

Markov Chains Handout for Stat 110

Statistics 150: Spring 2007

ISE/OR 760 Applied Stochastic Modeling

CS145: Probability & Computing Lecture 18: Discrete Markov Chains, Equilibrium Distributions

ISyE 6650 Probabilistic Models Fall 2007

Lecture Notes 7 Random Processes. Markov Processes Markov Chains. Random Processes

Classification of Countable State Markov Chains

MAS275 Probability Modelling Exercises

Irreducibility. Irreducible. every state can be reached from every other state For any i,j, exist an m 0, such that. Absorbing state: p jj =1

Markov Chains, Random Walks on Graphs, and the Laplacian

Markov Chains Introduction

Discrete time Markov chains. Discrete Time Markov Chains, Limiting. Limiting Distribution and Classification. Regular Transition Probability Matrices

1 Ways to Describe a Stochastic Process

Treball final de grau GRAU DE MATEMÀTIQUES Facultat de Matemàtiques Universitat de Barcelona MARKOV CHAINS

Markov Chains on Countable State Space

MATH 56A: STOCHASTIC PROCESSES CHAPTER 2

Recap. Probability, stochastic processes, Markov chains. ELEC-C7210 Modeling and analysis of communication networks

TMA 4265 Stochastic Processes Semester project, fall 2014 Student number and

Markov Chains and Stochastic Sampling

Lecture 9 Classification of States

CHAPTER 6. Markov Chains

Lecture 5: Random Walks and Markov Chain

Transience: Whereas a finite closed communication class must be recurrent, an infinite closed communication class can be transient:

Stochastic process. X, a series of random variables indexed by t

Interlude: Practice Final

2 Discrete-Time Markov Chains

Markov Chain Model for ALOHA protocol

T 1. The value function v(x) is the expected net gain when using the optimal stopping time starting at state x:

INTRODUCTION TO MARKOV CHAINS AND MARKOV CHAIN MIXING

Discrete Markov Chain. Theory and use

Probability & Computing

The Theory behind PageRank

The cost/reward formula has two specific widely used applications:

Markov Model. Model representing the different resident states of a system, and the transitions between the different states

MARKOV PROCESSES. Valerio Di Valerio

30 Classification of States

Markov Chains. Chapter 16. Markov Chains - 1

Powerful tool for sampling from complicated distributions. Many use Markov chains to model events that arise in nature.

STOCHASTIC MODELS LECTURE 1 MARKOV CHAINS. Nan Chen MSc Program in Financial Engineering The Chinese University of Hong Kong (ShenZhen) Sept.

Countable state discrete time Markov Chains

APPM 4/5560 Markov Processes Fall 2019, Some Review Problems for Exam One

Social network analysis: social learning

Let (Ω, F) be a measureable space. A filtration in discrete time is a sequence of. F s F t

Math Homework 5 Solutions

Chapter 11 Advanced Topic Stochastic Processes

Chapter 35 out of 37 from Discrete Mathematics for Neophytes: Number Theory, Probability, Algorithms, and Other Stuff by J. M. Cargal.

Markov Chains. Andreas Klappenecker by Andreas Klappenecker. All rights reserved. Texas A&M University

Lecture 6: Entropy Rate

Definition A finite Markov chain is a memoryless homogeneous discrete stochastic process with a finite number of states.

18.175: Lecture 30 Markov chains

Stochastic Processes

Lecture #5. Dependencies along the genome

LIMITING PROBABILITY TRANSITION MATRIX OF A CONDENSED FIBONACCI TREE

1 Gambler s Ruin Problem

MATH 56A SPRING 2008 STOCHASTIC PROCESSES

Quantitative Evaluation of Emedded Systems Solution 2: Discrete Time Markov Chains (DTMC)

STAT STOCHASTIC PROCESSES. Contents

P i [B k ] = lim. n=1 p(n) ii <. n=1. V i :=

1 Random Walks and Electrical Networks

COMS 4721: Machine Learning for Data Science Lecture 20, 4/11/2017

Markov Chain Monte Carlo

18.600: Lecture 32 Markov Chains

18.440: Lecture 33 Markov Chains

Stochastic Models: Markov Chains and their Generalizations

Today. Next lecture. (Ch 14) Markov chains and hidden Markov models

6 Markov Chain Monte Carlo (MCMC)

Lecture notes for Analysis of Algorithms : Markov decision processes

25.1 Markov Chain Monte Carlo (MCMC)

4.7.1 Computing a stationary distribution

The Markov Chain Monte Carlo Method

2 DISCRETE-TIME MARKOV CHAINS

Markov Chains and Pandemics

MATH 56A SPRING 2008 STOCHASTIC PROCESSES 65

= P{X 0. = i} (1) If the MC has stationary transition probabilities then, = i} = P{X n+1

Essentials on the Analysis of Randomized Algorithms

Homework set 2 - Solutions

Stochastic modelling of epidemic spread

Midterm 2 Review. CS70 Summer Lecture 6D. David Dinh 28 July UC Berkeley

Note that in the example in Lecture 1, the state Home is recurrent (and even absorbing), but all other states are transient. f ii (n) f ii = n=1 < +

Transcription:

Markov Processes Hamid R. Rabiee

Overview Markov Property Markov Chains Definition Stationary Property Paths in Markov Chains Classification of States Steady States in MCs. 2

Markov Property A discrete process has the Markov property if given its value at time t, the value at time t+ is independent of values at times before t. That is: Pr X t+ = x t+ X t = x t, X t = x t,, X = x = Pr X t+ = x t+ X t = x t For all t, x t+, x t, x t, x t 2,, x. 3

Stationary Property A Markov Process is called stationary if Pr X t+ = u X t = v = Pr X = u X 0 = v for all t. The evolution of stationary processes don t change over time. For defining the complete joint distribution of a stationary Markov Process it is sufficient to define Pr X = u X 0 = v and Pr X 0 = v for all u and v. (why?) We will mainly consider stationary Markov processes here. 4

Example (Coin Tossing Game) Consider a single player game in which at every step a biased coin is tossed and according to the result, the score will be increased or decreased by one point. The game ends if either the score reaches 00 (winning) or -00 (losing). Score of the player at each step t 0 is a random variable and the sequence of scores as the game progresses forms a random process X 0, X,, X t. 5

Example (Coin Tossing Game) It is easy to verify that X is a stationary Markov chain: At the end of each step the score solely depends on the current score s c and the result of tossing the coin (which is independent of time and previous tosses). Stating this mathematically (for s c { 00,00}): Pr X t+ = s X t = s c, X t = s t,, X 0 = 0 = p ; s = s c + p ; s = s c 0 ; otherwise = Pr X t+ = s X t = s c = Pr X = s X 0 = s c If value of p was subject to change in time, the process would not be stationary. In the formulation we would have p t instead of p. Independent of t and s 0,, s t 6

Example (Tracking) Assume we want to track a target in XY plane, and so we need to model its movement. Assuming that current position and speed of the target describes everything about its future movements, we can consider the 4- dimensional state: S t = (X t, Y t, X t, Y t ) It is common to consider linear relation between S t and S t with a Gaussian additive noise: S t = A t S t + v t ; v t ~N(0, Σ) Or equivalently: f St S t s t s t = N(s t ; A t s t, Σ) There exists efficient algorithms to work with these linear- Gaussian models. 7

Example (Tracking) Considering the kinematics relations of a moving object we can define linear relation as: 0 Δt 0 0 0 Δt A t = 0 0 0 0 0 0 This approach can not model Acceleration the speed is changed only because of noise. Independent of t (stationary) We can model arbitrary speed change by appropriately setting the 3 rd and 4 th rows of the matrix at each time. An example of a non stationary Markov process. Another approach is to extend the states to 6- dimensions containing X t and Y t. 8

Example (Speech) A basic model for speech signal considers the value at time t to be a linear combination of its d previous values with a Gaussian additive noise: S t = d i= α i S t i S t is not a Markov process. + v t ; v t ~N(0, Σ) 9

Example (Speech) We can make a stationary Markov process by considering the d-dimensional state U t = S t, S t, S t d T : Where: A = U t = AU t + v k B α α 2 α d 0 0 0 0, B = Equivalently: f Ut U t u t u t = N u t ; A t u t, Σ I i d u t i = u t i Which (x) i is the ith dimension of vector x and I is the indicator function (used for guaranteeing consistency). Note that we have repeated X t in d consecutive states of U and there should be no inconsistency between those values. 0 0

Markov Process Types Two types of Markov processes based on domain of X t values: Discrete Continuous Discrete Markov processes are called Markov Chains (MC). In this session we will focus on stationary MCs.

Transition matrix According to the Markov property and stationary property, at each time step the process moves according to a fixed transition matrix: P X t+ = j X t = i = p ij Stochastic matrix: Rows sum up to Double stochastic matrix: Rows and columns sum up to 2

State Graph It is convenient to visualize a stationary Markov Chain with a transition diagram: A node represents a possible value of X t. At each time t, we say the process is in state s if X t =s. Each edge represents the probability of going from one state to another (we omit edges with zero weight). We should also specify the vector of initial probabilities π = π,, π n where π i = Pr(X 0 = i). So a stationary discrete process could be described as a person walking randomly on a graph (considering each step to depend only on the state he is currently in). The result path is called a Random Walk. 3

Example The transition diagram of the coin tossing game: p p p p -00-99 -98 99 00 -p -p -p -p -p We modeled winning and losing by states which when we get into, we never get out. Note that if the process was not stationary we were not able to visualize it in this way: For example consider the case that p is changing in time. 4

Example (Modeling Weather) Example: Each day is sunny or rainy. If a day is rainy, the next day is rainy with probability α (and sunny with probability α). If the day is sunny, the next day is rainy with probability β (and sunny with probability β). S = {rainy, sunny}, P = α α β β α α β R S β 5

Paths If the state space is {0,} we have: pr x 2 = 0 x 0 = 0 = Pr x = x 0 = 0 p x 2 = 0 x = +Pr x = 0 x 0 = 0 p x 2 = 0 x = 0 (n) Define p ij as the probability that starting from state i, the process stops at state j after n time steps. (2) = r pik p kj p ij k= 6

Paths Theorem: (n) p ij = r k= (n ) pik pkj To simplify the notation we define the matrix P (n) such that (n) the value at the i th row and j th column is p ij. Corollary : P (n) can be calculated by: P (n) = P n Corollary 2: If the process starts at time 0 with distribution π on the states then after n steps the distribution is πp n. 7

Absorbing Markov Chain An absorbing state is one in which the probability that the process remains in that state once it enters the state is (i.e., p ii = ). A Markov chain is absorbing if it has at least one absorbing state, and if from every state it is possible to go to an absorbing state (not necessarily in one step). An absorbing Markov chain will eventually enter one of the absorbing states and never leave it. Example: The 00 and -00 states in coin tossing game Note: After playing ling enough, the player will either win or lose (with probability ). p p p p -00-99 -98 99 00 -p -p -p -p -p 8

Absorption theorem In an absorbing Markov chain the probability that the process will be absorbed is. Proof: From each non-absorbing state s j it is possible to reach an absorbing state starting from s j. Therefore there exists p and m, such that the probability of not absorbing after m steps is at most p, in 2m steps at most p 2, etc. Since the probability of not being absorbed is monotonically decreasing, we have: lim n Pr(not absorbed after n steps) =0 9

Classification of States Accessibility: State j is said to be accessible from state i if starting in i it is possible that the process will ever enter state j: (P n ) ij > 0. Communication: Two states i and j that are accessible to each other are said to communicate. Every node communicates with itself: p ii 0 = Pr X 0 = i X 0 = i = Communication is an equivalence relation: It divides the state space up into a number of separate classes in which every pair of states communicate. (why?) The Markov chain is said to be irreducible if it has only one class. 20

Transient and Recurrent states For any state i we let f i denote the probability that, starting in state i, the process will ever reenter state i : f i = Pr n: X n = i X 0 = i) State i is said to be recurrent if f i = and transient if f i <. Theorem : State i is recurrent if and only if, starting in state i, the expected number of time periods that the process is in state i is infinite: Corollary : A transient state will only be visited a finite number of times Proof: E size n: X n = i X 0 = i = k= k Pr(size n: X n = i = k X 0 = i) = + prob size n: X n = i = X 0 = i < prob size n: X n = i = X 0 = i = 0 2

Transient and Recurrent states Theorem 2: State i is recurrent iff n=(p n ) ii =. Look at the text book for proof. Corollary 2: A finite state Markov chain has at least one recurrent state. If all states are transient there will be a finite number of steps that after that the process should not be in any state (which is a contradiction). 22

Ergodic states If state i is recurrent, then it is said to be positive recurrent if, starting in i, the expected time until the process returns to state i is finite. In a finite-state MC, all recurrent states are positive recurrent. State i is said to have period d if p n ii=0 whenever n is not divisible by d, and d is the largest integer with this property. Equivalently: d = gcd n: Pr X n = i X 0 = i > 0} A state with period is said to be aperiodic. We call an MC aperiodic if all its states are aperiodic. A state i is said to be ergodic if it is aperiodic and positive recurrent. Period, recurrence and positive recurrence are all class properties. They are shared between states of a class. 23

Example 0.25 Classes: {},{2,3},{4,5},{6},{7.8} Recurrent states: 6, 7, 8 => all positive Periodic states: 2, 3, 7, 8 (with period 2) Absorbing state(s): 6 Ergodic state(s): 6 0.75 0.25 4 2 0.25 0.75 0.75 5 0.25 3 0.25 0.25 6 7 8 As time goes to infinity, what is the probability of being in each class? Answer: The process will be in transient classes {},{2,3},{4,5} with probability 0. Problem is symmetric for entering classes {6} and {7,8} as their only input edge is one from 5 with equal probabilities 0.25, and once it enters them, there is no way out. So at infinity probability of being in each of these two classes is 0.5. 0.5 0.75 24

Example 0.25 Classes: {},{2,3},{4,5},{6},{7.8} Recurrent states: 6, 7, 8 => all positive Periodic states: 2, 3, 7, 8 (with period 2) Absorbing state(s): 6 Ergodic state(s): 6 0.75 0.25 4 2 0.25 0.75 0.75 5 0.25 3 0.25 0.25 6 7 8 If the process is absorbed in {7,8} (which could be considered as an absorbing super state) what will happen after that? It will alternate between 7 and 8 to the end. So at time t probability of being in 7 (or 8) will depend on the parity of t. In general finding the exact behavior of non-ergodic states as t is not easy. (try it!) We will talk about the ergodic cases in next slides. 0.5 0.75 25

Steady State Theorem: For an irreducible ergodic Markov chain lim n P n ij exists and is independent of i. Furthermore, letting π j = lim n P n ij Then π = π, π d t is unique nonnegative solution of π = π P d j= π j If the ergodicity condition is removed, lim n P n ij does not exist in general but the given equations yet have a unique solution π = π, π d t in which π j will be equal to the long run proportion of time that the Markov chain is in state j. = 26

Example Consider the weather model example discussed before. We want to see how will the weather be when time goes to infinity: P = α α β π 0 = απ 0 + βπ π = α π 0 + β π π 0 + π = β One of these equations is redundant. (why?) Which yields that π 0 = β and π +β α = α. +β α Exercise: In each of the following cases investigate the existence of solution and its meaning: ) α=0 and β = 2) α= and β = 0 27

Acknowledge Thanks to Hassan Hafez and Abbas Hosseini for preparation of slides 28

References Ross, Sheldon M. Introduction to probability models. 29