The cost/reward formula has two specific widely used applications:

Similar documents
At the boundary states, we take the same rules except we forbid leaving the state space, so,.

Examples of Countable State Markov Chains Thursday, October 16, :12 PM

Classification of Countable State Markov Chains

Countable state discrete time Markov Chains

Homework 3 posted, due Tuesday, November 29.

Transience: Whereas a finite closed communication class must be recurrent, an infinite closed communication class can be transient:

So in terms of conditional probability densities, we have by differentiating this relationship:

Stochastic2010 Page 1

Let's contemplate a continuous-time limit of the Bernoulli process:

No class on Thursday, October 1. No office hours on Tuesday, September 29 and Thursday, October 1.

Homework 4 due on Thursday, December 15 at 5 PM (hard deadline).

Birth-death chain models (countable state)

Markov Chains (Part 4)

Discrete time Markov chains. Discrete Time Markov Chains, Limiting. Limiting Distribution and Classification. Regular Transition Probability Matrices

2. Transience and Recurrence

Markov Processes Hamid R. Rabiee

Math Homework 5 Solutions

Stochastic Models: Markov Chains and their Generalizations

Homework 2 will be posted by tomorrow morning, due Friday, October 16 at 5 PM.

Markov Chains. X(t) is a Markov Process if, for arbitrary times t 1 < t 2 <... < t k < t k+1. If X(t) is discrete-valued. If X(t) is continuous-valued

IEOR 6711: Professor Whitt. Introduction to Markov Chains

Mathematical Foundations of Finite State Discrete Time Markov Chains

Markov Chains Absorption Hamid R. Rabiee

MATH 56A SPRING 2008 STOCHASTIC PROCESSES 65

Continuing the calculation of the absorption probability for a branching process. Homework 3 is due Tuesday, November 29.

Mathematical Framework for Stochastic Processes

Markov Chains and Stochastic Sampling

0.1 Naive formulation of PageRank

= P{X 0. = i} (1) If the MC has stationary transition probabilities then, = i} = P{X n+1

FDST Markov Chain Models

I will post Homework 1 soon, probably over the weekend, due Friday, September 30.

Birth and Death Processes. Birth and Death Processes. Linear Growth with Immigration. Limiting Behaviour for Birth and Death Processes

Markov Chains Handout for Stat 110

Markov Chains (Part 3)

Discrete time Markov chains. Discrete Time Markov Chains, Definition and classification. Probability axioms and first results

This is now an algebraic equation that can be solved simply:

4.7.1 Computing a stationary distribution

8. Statistical Equilibrium and Classification of States: Discrete Time Markov Chains

Markov Chains Absorption (cont d) Hamid R. Rabiee

Readings: Finish Section 5.2

Online Social Networks and Media. Link Analysis and Web Search

(implicitly assuming time-homogeneity from here on)

stochnotes Page 1

1/2 1/2 1/4 1/4 8 1/2 1/2 1/2 1/2 8 1/2 6 P =

Statistics 150: Spring 2007

Lecture 10: Powers of Matrices, Difference Equations

MARKOV PROCESSES. Valerio Di Valerio

INTRODUCTION TO MARKOV CHAINS AND MARKOV CHAIN MIXING

Reinforcement Learning

88 CONTINUOUS MARKOV CHAINS

Reading: Karlin and Taylor Ch. 5 Resnick Ch. 3. A renewal process is a generalization of the Poisson point process.

Detailed Balance and Branching Processes Monday, October 20, :04 PM

Finite-Horizon Statistics for Markov chains

Erik A. van Doorn Department of Applied Mathematics University of Twente Enschede, The Netherlands

Lecture 20 : Markov Chains

Online Social Networks and Media. Link Analysis and Web Search

MATH 564/STAT 555 Applied Stochastic Processes Homework 2, September 18, 2015 Due September 30, 2015

Topics in Probability Theory and Stochastic Processes Steven R. Dunbar. Waiting Time to Absorption

Markov Chains CK eqns Classes Hitting times Rec./trans. Strong Markov Stat. distr. Reversibility * Markov Chains

Note special lecture series by Emmanuel Candes on compressed sensing Monday and Tuesday 4-5 PM (room information on rpinfo)

Markov chains (week 6) Solutions

Lecture 11: Introduction to Markov Chains. Copyright G. Caire (Sample Lectures) 321

IEOR 6711, HMWK 5, Professor Sigman

Probability, Random Processes and Inference

Example: physical systems. If the state space. Example: speech recognition. Context can be. Example: epidemics. Suppose each infected

MATH 56A: STOCHASTIC PROCESSES CHAPTER 2

Markov Chains, Stochastic Processes, and Matrix Decompositions

Matrix analytic methods. Lecture 1: Structured Markov chains and their stationary distribution

ISM206 Lecture, May 12, 2005 Markov Chain

2 Discrete-Time Markov Chains

TMA 4265 Stochastic Processes Semester project, fall 2014 Student number and

CS145: Probability & Computing Lecture 18: Discrete Markov Chains, Equilibrium Distributions

P(X 0 = j 0,... X nk = j k )

STOCHASTIC PROCESSES Basic notions

Evolutionary dynamics on graphs

Markov Chains, Random Walks on Graphs, and the Laplacian

First Passage Time Calculations

Social network analysis: social learning

1.3 Convergence of Regular Markov Chains

IEOR 6711: Stochastic Models I. Solutions to Homework Assignment 9


DATA MINING LECTURE 13. Link Analysis Ranking PageRank -- Random walks HITS

Section Notes 9. Midterm 2 Review. Applied Math / Engineering Sciences 121. Week of December 3, 2018

Stochastic Processes

STA 624 Practice Exam 2 Applied Stochastic Processes Spring, 2008

ECE-517: Reinforcement Learning in Artificial Intelligence. Lecture 4: Discrete-Time Markov Chains

EECS 1028 M: Discrete Mathematics for Engineers

P i [B k ] = lim. n=1 p(n) ii <. n=1. V i :=

12 Markov chains The Markov property

MATH 56A: STOCHASTIC PROCESSES CHAPTER 1

Lectures on Markov Chains

Lecture 21. David Aldous. 16 October David Aldous Lecture 21

1 Ways to Describe a Stochastic Process

Stochastic processes. MAS275 Probability Modelling. Introduction and Markov chains. Continuous time. Markov property

MARKOV MODEL WITH COSTS In Markov models we are often interested in cost calculations.

Chapter 5. Continuous-Time Markov Chains. Prof. Shun-Ren Yang Department of Computer Science, National Tsing Hua University, Taiwan

Earlier this week we defined Boolean atom: A Boolean atom a is a nonzero element of a Boolean algebra, such that ax = a or ax = 0 for all x.

Note that in the example in Lecture 1, the state Home is recurrent (and even absorbing), but all other states are transient. f ii (n) f ii = n=1 < +

Markov Chains. Sarah Filippi Department of Statistics TA: Luke Kelly

Lecture 7. µ(x)f(x). When µ is a probability measure, we say µ is a stationary distribution.

Transcription:

Applications of Absorption Probability and Accumulated Cost/Reward Formulas for FDMC Friday, October 21, 2011 2:28 PM No class next week. No office hours either. Next class will be 11/01. The cost/reward formula has two specific widely used applications: Take the cost/reward function to be f=1. Then w i will be the expected number of epochs spent in a transient state before reaching a recurrent state, given that the MC starts in a transient state i. Take the cost reward function to be: Where k is a transient state. Then w i will be the expected number of visits to state k, starting from state i. A simple trick will make the absorption probability and cost/reward calculations apply to many more questions than one might have naively thought: Starting from state i, what is the expected number of epochs until another state j is visited? (first passage time problem) Starting from state i, how many epochs is the Markov chain expected to be in state k before state j is visited. Starting from state i, what is the probability that state k is visited before state j? First let's consider the case of an irreducible Markov chain. For the first question, modify the Markov chain by making state j absorbing (probability one to remain in Stochastic11 Page 1

that state). Then j will be recurrent, and all other states of the Markov chain will become transient. Now one can apply to this modified Markov chain the formula for the expected time to reach a recurrent state. For the second question, make the same modification, and now use the cost/reward function For the third question, make both states j and k absorbing, then use absorption probability formula. If the Markov chain is reducible, then one uses similar ideas, but have to be careful about whether some answers may be infinite. Example Birth-death chain (biased random walk) The Markov chain dynamics are prescribed by the above diagram, giving the probabilities for transitions to the same or neighboring state during an epoch. Birth-death chains also arise in: Stochastic11 Page 2

Biased random walk model (with inhomogeneities) Atomic physics models, including defects in lattices Chemical reactions Financial price models Molecular motor models with multiple motors bound to a cargo But we'll just approach the birth-death chains with a generic analytical approach without consideration of a specific application. We'll in particular take the special case of absorbing boundary conditions. We'll take this as the original Markov chain, but it could also have arisen from a modification to a Markov chain without absorbing boundary conditions, but where we are trying to compute statistics about what happens before I reach one of the endpoints. The stochastic update rule is easy to write down, but we have developed the formulas to more naturally refer to the probability transition matrix, so that's how we'll encode the Markov chain. Let's consider how to compute analytically the answers Stochastic11 Page 3

to two basic questions involving this Markov chain: If the MC starts in state I, what is the probability that it will become absorbed in state 0 (or M). What is the expected number of epochs until the Markov chain becomes absorbed at an endpoint? Optionally, we can prepare for these questions by forming the canonical decomposition of the probability transition matrix: Using our notation from last time: This would be a good way to compute the answers Stochastic11 Page 4

numerically using numerical linear algebra. But birth-death chains are simple enough that we can actually obtain the answers analytically, and to do so, it's more convenient to work with the recursion equations: We now substitute in the structure of the probability transition matrix to write this system of equations for U ij explicitly. For For i=1, For i=m-1, Stochastic11 Page 5

For i=m-1, Now the problem boils down to solving a system of linear equations with a tridiagonal structure. The calculation is a bit technical though not difficult, once the right ideas are introduced. First we'll strengthen the recursive structure by noting that we can make the equations for i=1,m-1 look the same as for all other states, if we simply define: By introducing this extended definition of U, we have to solve the following equations: These are just linear difference equations, whose solution technique can be found, for example, in Lawler Ch. 0. Because our coefficients are fairly arbitrary (though they have a tridiagonal structure), simple approaches don't necessarily work so well. Note the equations for j=0 and j=m decouple so we can just solve them separately. But we can solve the j=0 equations because Stochastic11 Page 6

separately. But we can solve the j=0 equations because We can show that these equations have a special structure that can be summed "by quadrature." if we note by using We'll now apply a trick that one would use differential equations y'' = a(x) y'(x). Solve by introducing the function v=y', and then solve first order equation for v. By induction: How do we get a boundary condition so we can solve for V 1? Check that by using telescoping sum ideas: Stochastic11 Page 7

Sum over these to get U i. Stochastic11 Page 8

Similar techniques yield the mean time until absorption via analytical formulas; see the texts. Stochastic11 Page 9