Lecture Notes: Markov chains Tuesday, September 16 Dannie Durand

Similar documents
Lecture Notes: Markov chains

No class on Thursday, October 1. No office hours on Tuesday, September 29 and Thursday, October 1.

CS145: Probability & Computing Lecture 18: Discrete Markov Chains, Equilibrium Distributions

Outlines. Discrete Time Markov Chain (DTMC) Continuous Time Markov Chain (CTMC)

ISM206 Lecture, May 12, 2005 Markov Chain

Markov Model. Model representing the different resident states of a system, and the transitions between the different states

0.1 Naive formulation of PageRank

Markov Chains Handout for Stat 110

Discrete Markov Chain. Theory and use

At the boundary states, we take the same rules except we forbid leaving the state space, so,.

1 Ways to Describe a Stochastic Process

CMPSCI 240: Reasoning Under Uncertainty

1 Random Walks and Electrical Networks

IEOR 6711: Professor Whitt. Introduction to Markov Chains

Definition A finite Markov chain is a memoryless homogeneous discrete stochastic process with a finite number of states.

Readings: Finish Section 5.2

So in terms of conditional probability densities, we have by differentiating this relationship:

Discrete time Markov chains. Discrete Time Markov Chains, Limiting. Limiting Distribution and Classification. Regular Transition Probability Matrices

6.842 Randomness and Computation February 24, Lecture 6

Examples of Countable State Markov Chains Thursday, October 16, :12 PM

Random Walk on a Graph

Lecture 20 : Markov Chains

( ) is symmetric about the y - axis.

8. Statistical Equilibrium and Classification of States: Discrete Time Markov Chains

2 Discrete-Time Markov Chains

MAS275 Probability Modelling Exercises

Markov Processes Hamid R. Rabiee

Markov Chains, Stochastic Processes, and Matrix Decompositions

MASSACHUSETTS INSTITUTE OF TECHNOLOGY 6.436J/15.085J Fall 2008 Recitation 13 10/31/2008. Markov Chains

Prime numbers and Gaussian random walks

Lecture 5: Random Walks and Markov Chain

Markov Chains. X(t) is a Markov Process if, for arbitrary times t 1 < t 2 <... < t k < t k+1. If X(t) is discrete-valued. If X(t) is continuous-valued

ECE-517: Reinforcement Learning in Artificial Intelligence. Lecture 4: Discrete-Time Markov Chains

9.5 HONORS Determine Odd and Even Functions Graphically and Algebraically

MAA704, Perron-Frobenius theory and Markov chains.

Link Analysis. Leonid E. Zhukov

6.842 Randomness and Computation March 3, Lecture 8

Probability, Random Processes and Inference

Random walks, Markov chains, and how to analyse them

STA 4273H: Statistical Machine Learning


Time Reversibility and Burke s Theorem

Lecture #5. Dependencies along the genome

Markov Chains. Andreas Klappenecker by Andreas Klappenecker. All rights reserved. Texas A&M University

Markov chains. Randomness and Computation. Markov chains. Markov processes

Some Definition and Example of Markov Chain

Chapter 5. Continuous-Time Markov Chains. Prof. Shun-Ren Yang Department of Computer Science, National Tsing Hua University, Taiwan

Machine Learning CPSC 340. Tutorial 12

MARKOV PROCESSES. Valerio Di Valerio

Countable state discrete time Markov Chains

Introduction to Computational Biology Lecture # 14: MCMC - Markov Chain Monte Carlo

Chapter 16 focused on decision making in the face of uncertainty about one future

Lecture 11: Introduction to Markov Chains. Copyright G. Caire (Sample Lectures) 321

STA 414/2104: Machine Learning

Lecture 2: September 8

Lab 2 Worksheet. Problems. Problem 1: Geometry and Linear Equations

SMSTC (2007/08) Probability.

Lecture Notes 7 Random Processes. Markov Processes Markov Chains. Random Processes

Math Models of OR: Some Definitions

Irreducibility. Irreducible. every state can be reached from every other state For any i,j, exist an m 0, such that. Absorbing state: p jj =1

INTRODUCTION TO MARKOV CHAINS AND MARKOV CHAIN MIXING

Algorithms and Theory of Computation. Lecture 22: NP-Completeness (2)

Columbus State Community College Mathematics Department. CREDITS: 5 CLASS HOURS PER WEEK: 5 PREREQUISITES: MATH 2173 with a C or higher

Lecture 14: Random Walks, Local Graph Clustering, Linear Programming

Information Theory. Lecture 5 Entropy rate and Markov sources STEFAN HÖST

Using Markov Chains To Model Human Migration in a Network Equilibrium Framework

Stochastic Modelling Unit 1: Markov chain models

Recap. Probability, stochastic processes, Markov chains. ELEC-C7210 Modeling and analysis of communication networks

(b) What is the variance of the time until the second customer arrives, starting empty, assuming that we measure time in minutes?

MATH3200, Lecture 31: Applications of Eigenvectors. Markov Chains and Chemical Reaction Systems

Chapter 10 Markov Chains and Transition Matrices

Numerical Solution for Two Dimensional Laplace Equation with Dirichlet Boundary Conditions

Further Applications of Advanced Mathematics (FP3) THURSDAY 14 JUNE 2007

Convex Optimization CMU-10725

LIMITING PROBABILITY TRANSITION MATRIX OF A CONDENSED FIBONACCI TREE

Markov chains and the number of occurrences of a word in a sequence ( , 11.1,2,4,6)

each nonabsorbing state to each absorbing state.

TCOM 501: Networking Theory & Fundamentals. Lecture 6 February 19, 2003 Prof. Yannis A. Korilis

Markov Chains. As part of Interdisciplinary Mathematical Modeling, By Warren Weckesser Copyright c 2006.

Powerful tool for sampling from complicated distributions. Many use Markov chains to model events that arise in nature.

Section Notes 9. Midterm 2 Review. Applied Math / Engineering Sciences 121. Week of December 3, 2018

Mathematical Methods for Computer Science

Markov Chains and Stochastic Sampling

Transience: Whereas a finite closed communication class must be recurrent, an infinite closed communication class can be transient:

Introduction to MCMC. DB Breakfast 09/30/2011 Guozhang Wang

Markov Chains, Random Walks on Graphs, and the Laplacian

1.3 Convergence of Regular Markov Chains

6.207/14.15: Networks Lectures 4, 5 & 6: Linear Dynamics, Markov Chains, Centralities

Markov Chains. Contents

Lecture 1: Brief Review on Stochastic Processes

2. Transience and Recurrence

MATH 446/546 Test 2 Fall 2014

LEARNING DYNAMIC SYSTEMS: MARKOV MODELS

8 Wyner Honors Algebra II Fall 2013

Compartmental modeling

Asymptotic properties of imprecise Markov chains

MATH 564/STAT 555 Applied Stochastic Processes Homework 2, September 18, 2015 Due September 30, 2015

n α 1 α 2... α m 1 α m σ , A =

Probabilistic Model Checking Michaelmas Term Dr. Dave Parker. Department of Computer Science University of Oxford

Lesson Plan. AM 121: Introduction to Optimization Models and Methods. Lecture 17: Markov Chains. Yiling Chen SEAS. Stochastic process Markov Chains

Transcription:

Computational Genomics and Molecular Biology, Lecture Notes: Markov chains Tuesday, September 6 Dannie Durand In the last lecture, we introduced Markov chains, a mathematical formalism for modeling how a random variable progresses over time. We introduced the following notation for describing the properties of a Markov chain: A Markov chain has states E, E,... E s corresponding to the range of the associated random variable. ϕ j (t) is the probability that the chain is in state E j at time t. The vector ϕ(t) = (ϕ (t),... ϕ s (t)) is the state probability distribution at time t. π = ϕ() is the initial state probability distribution. P is the transition probability matrix. P jk gives the probability of making a transition to state E k at time t +, given that the chain was in state E j at time t. The rows of this matrix sum to one ( k P jk = ). The state probability distribution at time t+ is given by ϕ(t+) = ϕ(t) P. The probability of being in state E j at t + is ϕ j (t + ) = j ϕ j (t)p jk () The Markov property states that Markov chains are memoryless. The probability that the chain is in state E k at time t +, depends only on ϕ(t) and is independent of ϕ(t ), ϕ(t ), ϕ(t 3)... In this course, we will focus on discrete, finite, time-homogeneous Markov chains. These are models with a finite number of states, in which time (or space) is split into discrete steps. A Markov chain is time-homogeneous if the transition matrix does not change over time. Absorbing states Last Thursday, we considered a Markov chain to model the position of a drunk moving back and forth on a railroad track on top of a mesa. When the drunk reaches either end of the railway (either the th or the th tie), he falls off the mesa. In the Markov model, states E and E are absorbing states: Once the system enters one of these states, it remains in that state forever, because P = P =. Our model of the drunk is an example of a random walk with absorbing boundaries.

Computational Genomics and Molecular Biology, The transition matrix of a Markov chain can be represented as a graph, where the nodes represent states and the edges represent transitions with non-zero probability. For example, the random walk with absorbing boundaries can be modeled like this: Periodic Markov chains In order to save the drunk from an early death, we introduced a random walk with reflecting boundaries. At each step, the drunk moves to the left or to the right with equal probability. When the drunk reaches one of the boundary states (E or E ), he returns to the adjacent state (E or E 3 ) at the next step, with probability one. This yields the following transition probability matrix: 3 3 and can be represented graphically like this:

Computational Genomics and Molecular Biology, 3 The initial state probability distribution and the state distribution for the first two time steps are the same in both random walk models, namely ϕ() = (,,,, ) ϕ() = (,,,, ) ϕ() = (,,,, ). We calculate the state probability distribution at t = 3 by multiplying the vector ϕ() times the matrix P : ϕ(3) = ϕ() P = (,,,, ) P = (,,,, ). This demonstrates that ϕ(3) = ϕ(). Similarly, ϕ() = ϕ() as can be seen from the following calculation: ϕ() = ϕ(3) P = (,,,, ) P = (,,,, ). From this we can see that the probability state distribution will be (,,,, ) at all odd time steps and (,,,, ) at all even time steps. Thus, the random walk with reflecting boundaries is a periodic Markov chain. A Markov chain is periodic if there is some state that can only be visited in multiples of m time steps, where m >. We do not require periodic Markov chains for modeling sequence evolution and will only consider aperiodic Markov chains going forward. Stationary distributions A state probability distribution, ϕ, that satisfies the equation ϕ = ϕ P () is called a stationary distribution. A key question for a given Markov chain is whether such a stationary distribution exists. Equation is equivalent to a system of s equations in s unknowns.

Computational Genomics and Molecular Biology, One way to determine the steady state distribution is to solve that system of equations. The stationary distribution can also be obtained using matrix algebra, but that approach is beyond the scope of this course. The random walk with reflecting boundaries clearly does not have a stationary distribution, since every state with non-zero probability at time t has zero probability at time t +. The random walk with absorbing boundaries does have a stationary distribution, but it is not unique. For example, both (,,,, ) and (,,,, ) are stationary distributions of the random walk with absorbing boundaries. For the rest of this course, we will concern ourselves only with aperiodic Markov chains that do not have absorbing states. In fact, we will make an even stronger assumption and restrict our consideration to Markov chains in which every state is connected to every other state via a series of zero or more states. If a finite Markov chain is aperiodic and connected in this way, it has a unique stationary distribution. We will not attempt to prove this or even to state the theorem in a rigorous way. That is beyond the scope of this class. For those who are interested, a very nice treatment can be found in Chapter 5 of Probability Theory and its Applications ( Volume I ) by William Feller (John Wiley & Sons). As an example of a Markov chain with a unique stationary distribution, we introduced a random walk that has neither absorbing, nor reflecting boundaries. In this third random walk model, if the drunk is in one of the boundary states (E or E ) at time t, then at time t + he either remains in the boundary state or returns to the adjacent state (E or E 3 ). In the model we considered in class, we assigned a probability of.5 to each of these events, resulting in the following state transition matrix, P : 3 3 We can determine the stationary state distribution for this random walk model by substitution this

Computational Genomics and Molecular Biology, 5 transition matrix into Equation. The probability of being in state E is ϕ = ϕ j P j j= = ϕ P + ϕ P + ϕ P + ϕ 3 P 3 + ϕ P = ϕ + ϕ, since P, P 3 and P are all equal to zero. The other steady state probabilities are derived similarly, yielding ϕ = ϕ + ϕ (3) ϕ = ϕ + ϕ () ϕ = ϕ + ϕ 3 (5) ϕ 3 = ϕ + ϕ (6) ϕ = ϕ 3 + ϕ. (7) In addition, the probability that the system is in some state is unity, imposing an additional constraint: ϕ + ϕ + ϕ + ϕ 3 + ϕ =. (8) The model has a stationary distribution if the above equations have a solution. In class, we showed that Equations 3-7 reduce to ϕ = ϕ = ϕ = ϕ 3 = ϕ. Applying the constraint in Equation 8, we see that the solution to the above equations must be ϕ = (.,.,.,.,.). If we know the stationary state distribution, or have an educated guess, we can verify that it indeed satisfies Equation. For example, it is easy to verify that (.,.,.,.,.) P = (.,.,.,.,.).