Markov Chains and Transition Probabilities

Similar documents
Finite Math - J-term Section Systems of Linear Equations in Two Variables Example 1. Solve the system

Matrix Basic Concepts

The probability of going from one state to another state on the next trial depends only on the present experiment and not on past history.

Math "Matrix Approach to Solving Systems" Bibiana Lopez. November Crafton Hills College. (CHC) 6.3 November / 25

Math 166: Topics in Contemporary Mathematics II

Matrices: 2.1 Operations with Matrices

Markov Model. Model representing the different resident states of a system, and the transitions between the different states

REPLACE ONE ROW BY ADDING THE SCALAR MULTIPLE OF ANOTHER ROW

Definition 2.3. We define addition and multiplication of matrices as follows.

ANALYTICAL MATHEMATICS FOR APPLICATIONS 2018 LECTURE NOTES 3

Markov Chains, Stochastic Processes, and Matrix Decompositions

Chapter 2. Ma 322 Fall Ma 322. Sept 23-27

CS100: DISCRETE STRUCTURES. Lecture 3 Matrices Ch 3 Pages:

Finite-Horizon Statistics for Markov chains

Lectures on Probability and Statistical Models

Lecture 20 : Markov Chains

Lecture 4: Gaussian Elimination and Homogeneous Equations

Definition A finite Markov chain is a memoryless homogeneous discrete stochastic process with a finite number of states.

Chapter 10 Markov Chains and Transition Matrices

ECE 541 Project Report: Modeling the Game of RISK Using Markov Chains

Mathematics 13: Lecture 10

Lecture 12: Solving Systems of Linear Equations by Gaussian Elimination

Stochastic processes. MAS275 Probability Modelling. Introduction and Markov chains. Continuous time. Markov property

PH1105 Lecture Notes on Linear Algebra.

Introduction to Stochastic Processes

18.440: Lecture 33 Markov Chains

Markov Chains Absorption Hamid R. Rabiee

Chapter 4 Systems of Linear Equations; Matrices

2.1 Matrices. 3 5 Solve for the variables in the following matrix equation.

Review of Linear Algebra

= P{X 0. = i} (1) If the MC has stationary transition probabilities then, = i} = P{X n+1

7.3. Determinants. Introduction. Prerequisites. Learning Outcomes

Markov Chains. As part of Interdisciplinary Mathematical Modeling, By Warren Weckesser Copyright c 2006.

Chapter 35 out of 37 from Discrete Mathematics for Neophytes: Number Theory, Probability, Algorithms, and Other Stuff by J. M. Cargal.

RISKy Business: An In-Depth Look at the Game RISK

Matrices. Chapter Definitions and Notations

TMA Calculus 3. Lecture 21, April 3. Toke Meier Carlsen Norwegian University of Science and Technology Spring 2013

n α 1 α 2... α m 1 α m σ , A =

Matrices. In this chapter: matrices, determinants. inverse matrix

Systems of Linear Equations. By: Tri Atmojo Kusmayadi and Mardiyana Mathematics Education Sebelas Maret University

Exercise Set Suppose that A, B, C, D, and E are matrices with the following sizes: A B C D E

The Boundary Problem: Markov Chain Solution

Quantum Computing Lecture 2. Review of Linear Algebra

1111: Linear Algebra I

Lecture Notes 7 Random Processes. Markov Processes Markov Chains. Random Processes

Undergraduate Mathematical Economics Lecture 1

Homework 1 Elena Davidson (B) (C) (D) (E) (F) (G) (H) (I)

Lecture 11: Introduction to Markov Chains. Copyright G. Caire (Sample Lectures) 321

QUEUING MODELS AND MARKOV PROCESSES

Linear Algebra I Lecture 8

The Leslie Matrix. The Leslie Matrix (/2)

William Stallings Copyright 2010

MTH 215: Introduction to Linear Algebra

No class on Thursday, October 1. No office hours on Tuesday, September 29 and Thursday, October 1.

Pre-Calculus I. For example, the system. x y 2 z. may be represented by the augmented matrix

Countable state discrete time Markov Chains

Linear Algebra, Vectors and Matrices

22.3. Repeated Eigenvalues and Symmetric Matrices. Introduction. Prerequisites. Learning Outcomes

Linear Algebra review Powers of a diagonalizable matrix Spectral decomposition

Markov Decision Processes

Chapter 3. Determinants and Eigenvalues

Chapter 2. Matrix Arithmetic. Chapter 2

Linear Algebra Practice Problems

7.6 The Inverse of a Square Matrix

Systems of Linear Equations

8. Statistical Equilibrium and Classification of States: Discrete Time Markov Chains

IEOR 6711: Professor Whitt. Introduction to Markov Chains

MATH.2720 Introduction to Programming with MATLAB Vector and Matrix Algebra

1 Ways to Describe a Stochastic Process

Chapter 1: Systems of linear equations and matrices. Section 1.1: Introduction to systems of linear equations

Multiplying matrices by diagonal matrices is faster than usual matrix multiplication.

AMS526: Numerical Analysis I (Numerical Linear Algebra)

Advanced Engineering Mathematics Prof. Pratima Panigrahi Department of Mathematics Indian Institute of Technology, Kharagpur

(b) What is the variance of the time until the second customer arrives, starting empty, assuming that we measure time in minutes?

Image Registration Lecture 2: Vectors and Matrices

Linear equations in linear algebra

Markov Chains (Part 3)

MAT 1302B Mathematical Methods II

Linear Algebra and Matrix Inversion

1111: Linear Algebra I

Algebra & Trig. I. For example, the system. x y 2 z. may be represented by the augmented matrix

Linear Algebra review Powers of a diagonalizable matrix Spectral decomposition

We use the overhead arrow to denote a column vector, i.e., a number with a direction. For example, in three-space, we write

Graduate Mathematical Economics Lecture 1

is a 3 4 matrix. It has 3 rows and 4 columns. The first row is the horizontal row [ ]

7.5 Operations with Matrices. Copyright Cengage Learning. All rights reserved.

On asymptotic behavior of a finite Markov chain

Advanced Engineering Mathematics Prof. Pratima Panigrahi Department of Mathematics Indian Institute of Technology, Kharagpur

Linear Equations in Linear Algebra

Linear Algebra (Review) Volker Tresp 2017

Markov Processes Hamid R. Rabiee

Recall, we solved the system below in a previous section. Here, we learn another method. x + 4y = 14 5x + 3y = 2

Markov Processes. Stochastic process. Markov process

CHAPTER 6. Markov Chains

Lecture 10: Markov Chains

Information Theory. Lecture 5 Entropy rate and Markov sources STEFAN HÖST

Repeated Eigenvalues and Symmetric Matrices

Elementary maths for GMT

Math 166: Topics in Contemporary Mathematics II

5.1 Introduction to Matrices

Transcription:

Hinthada University Research Journal 215, Vol. 6, No. 1 3 Markov Chains and Transition Probabilities Ko Ko Oo Abstract Markov chain is widely applicable to the study of many real-world phenomene. We represent the probability vectors, stochastic matrices, Markov chains, higher transition probabilities, stationary distribution, and absorbing states in this paper. Keywords: Stochastic matrices, Markov chains, higher transition probabilities. Introduction We come into contact with our first random process (stochastic process), known as a Markov chain, which is widely applicable to the study of many real-world phenomena. Applications to genetics and production processes are presented. Probability Vectors A row vector u = (u 1, u 2,, u n ) is called a probability vector if its components are non-negative and their sum is 1. Consider the following vectors: u = ( 3 4,, 1 4, 1 2 ), v = ( 3 4, 1 2,, 1 4 ), and w = ( 1 4, 1 4,, 1 2 ). Then: Remark: u is not a probability vector since its third component is negative; v is not a probability vector since the sum of its components is greater than 1; w is a probability vector. Since the sum of the components of a probability vector is one, an arbitrary probability vector with n components can be represented in terms of n-1 unknowns as follows: (x 1, x 2,, x n-1, 1 x 1 x 2 x n-1 ). In particular, arbitrary probability vectors with 2 and 3 components can be represented, respectively, in the form (x, 1 x) and (x, y, 1 x y). Stochastic and Regular Stochastic Matrices A square matrix P = (p ij ) is called a stochastic matrix (Hus, 1997) if each of its rows is a probability vector, that is, if each entry of P is non-negative and the sum of the entries in every row is 1. Theorem 1 If A and B are stochastic matrices, then the product AB is a stochastic matrix. Therefore, in particular, all powers A n are stochastic matrix. Lecturer, Department of Mathematics, Hinthada University

Hinthada University Research Journal 215, Vol. 6, No. 1 31 Definition: A stochastic matrix P is said to be regular if all the entries of some power P m are positive. 1 The stochastic matrix A = is regular since 2 2 1 A 2 = 2 2 Consider the stochastic matrix B = 1 = 2 2 1. 2 2 2 2 1 3 4 4 1 1 1 Here B 2 = 3 1, B 3 = 7 1, B 4 = 15 1,. 4 4 8 8 16 16 is positive in every entry. In fact every power B m will have 1 and in the first row, hence B is not regular. Fixed Points of Square Matrices A non-zero row vector u = (u 1, u 2,, u n ) is called a fixed points of a square matrix (Kandasamy et al., 21). A if u is left fixed, that is, is not changed, when multiplied by A : u A = u. We do not include the zero vector as a fixed point of a matrix since it is always left fixed by every matrix A: A =. Theorem 2 If u is a fixed vector of a matrix A, then for any real number the scalar multiple u is also a fixed vector of A. Fixed Points and Regular Stochastic Matrices The main relationship between regular stochastic matrices and fixed points is contained in the following theorem. Theorem 3 Let P be a regular stochastic matrix. Then : (i) (ii) (iii) P has a unique fixed probability vector t, and the components of t are all positive; the sequence P, P 2, P 3, of powers of P approaches the matrix T whose row are each the fixed point t; if p is any probability vector, then the sequence of vectors pp, pp 2, pp 3, approaches the fixed point t; Note: P n approaches T means that each entry of P n approaches the corresponding entry of T, and pp n approaches t means that each component of pp n approaches the corresponding component of t. Stochastic matrix P is said to be regular if all the entries of some power P m are positive.

32 Hinthada University Research Journal 215, Vol. 6, No. 1 Markov Chains We now consider a sequence of trials whose outcomes, say, X 1, X 2,, satisfy the following two properties : (i) Each outcome belongs to a finite set of outcomes (a 1, a 2,, a m ) called the state space of the system; if the outcomes on the n th trial is a i, then we say that the system is in state a i at time n or at the n th step. (ii) The outcome of any trial depends at most upon the outcome of the immediately preceding trial and not upon any other previous outcomes; with each pair of states (a i, a j ) there is given the probability p ij that a j occurs immediately after a i occurs. Such a stochastic process is called a (finite) Markov chain (Natarajan & Tamilarsi, 27). The numbers p ij, called the transition probabilities, can be arranged in a matrix P = called the transition matrix. p11 p12 p1m p21 p22 p2m pm1 pm1 pmm Thus with each state a i there corresponds the i th row (p i1, p i2,, p im ) of the transition matrix P; if the system is in state a i, then this row vector represents the probabilities of all the possible outcomes of the next trial and so it is a probability vector. Theorem 4 The transition matrix P of a Markov chain is a stochastic matrix. Higher Transition Probabilities The entry p ij in the transition matrix P of a Markov chain is the probability that the system changes from the state a i to the state a j in one step: a i a j. The entry p ij (n) in the transition matrix P (n) of a Markov chain is the probability that the system changes from the state a i to the state a j in exactly n steps : a i a k a l a j. Theorem 5 Let P be the transition matrix of a Markov chain process. Then the n-step transition matrix is equal to the n th power of P : P (n) = P n. Now suppose that, at some arbitrary time, the probability that the system is in state a i is p i ; we denote these probabilities by the probability vector we denote these probabilities by the probability vector p = (p 1, p 2,, p m ) which is called the probability distribution of the system at that time. In particular, we shall let p = (p 1, p 2,, p m ) denote the initial probability distribution, that is, the distribution when the process begins, and we shall let p n = (p 1 n, p 2 n,, p m n ) denote the n th step probability distribution, that is, the distribution after the first n steps.

Hinthada University Research Journal 215, Vol. 6, No. 1 33 Theorem 6 Let P be the transition matrix of a Markov chain process. If p = (p i ) is the probability distribution of the system at some arbitrary time, then pp is the probability distribution of the system one step later and pp n is the probability distribution of the system n steps later. In particular, p 1 = p P, p 2 = p 1 P, p 3 = p 2 P,, p n = p P n. Stationary Distribution of Regular Markov Chains Suppose that a Markov chain is regular, that is, that its transition matrix P is regular. By Theorem 3, the sequence of n-step transition matrices P n approaches the matrix T whose rows are each the unique fixed probability vector t of P; hence the probability p ij n that a j occurs for sufficiently large n is independent of the original state a i and it approaches the component t j of t. Theorem 7 Let the transition matrix P of a Markov chain be regular. Then, in the long run, the probability that any state a j occurs is approximately equal to the component t j of the unique fixed probability vector t of P. Thus we see that the effect of the initial state or the initial probability distribution of the process wears off as the number of steps of the process increase. Furthermore, every sequence of probability distribution approaches the fixed probability vector t of P, called the stationary distribution of the Markov chain. Absorbing States A state a i of a Markov chain is called absorbing if the system remains in the state a i once it enters there. Thus a state a i is absorbing if and only if the i th row of the transition matrix P has a 1 on the main diagonal and zeros everywhere else. Suppose the following matrix is the transition matrix of a Markov chain: P = a1 a2 a3 a4 a5 1 1 4 4 4 4 a 1 a2 1 a3 1 2 4 4. a4 1 a5 1 The states a 2 and a 5 are each absorbing, since each of the second and fifth rows has a 1 on the main diagonal. Acknowledgement I would like to express my sincere gratitude to Dr. Aung Win, Acting Rector, Hinthada University, and Dr. Theingi Shwe, Pro-Rector, Hithada University, for their kind permission to submit this paper. Special thanks are also due to Daw Aye Aye, Professor and Head of the Department of Mathematics, Hinthada University, for her encouragement and comments.

34 Hinthada University Research Journal 215, Vol. 6, No. 1 References Hsu, H. P., 1997. Probability, Random Variables & Random Processes, (S.O.S), Mc-Graw-Hill Company, Inc., New York, USA. Kandasamy, P., Thilagavathi, K., and K. Gunavathi, 21. Probability, Random Variables, & Random Processes, Reprint, S-Chand & Company Limited, New Delhi. Natarajan, A. M., and A. Tamilarsi, 27. Probability Random Processes and Queuing Theory, Reprint, New Age International Publishing Company, New Delhi.