PRODUCTS OF INDECOMPOSABLE, APERIODIC, STOCHASTIC MATRICES1 J. WOLFOWITZ

Similar documents
On Backward Product of Stochastic Matrices

Statistics 992 Continuous-time Markov Chains Spring 2004

A SIMPLIFIED FORM FOR NEARLY REDUCIBLE AND NEARLY DECOMPOSABLE MATRICES D. J. HARTFIEL

Chapter 1: Systems of linear equations and matrices. Section 1.1: Introduction to systems of linear equations

MATH Mathematics for Agriculture II

Alternative Characterization of Ergodicity for Doubly Stochastic Chains

MEAN VALUE METHODS IN ITERATION

SIMILAR MARKOV CHAINS

ON THE MAXIMUM OF A NORMAL STATIONARY STOCHASTIC PROCESS 1 BY HARALD CRAMER. Communicated by W. Feller, May 1, 1962

On asymptotic behavior of a finite Markov chain

Lecture 11: Introduction to Markov Chains. Copyright G. Caire (Sample Lectures) 321

Markov Chains, Stochastic Processes, and Matrix Decompositions

Markov chains. Randomness and Computation. Markov chains. Markov processes

Section 9.2: Matrices.. a m1 a m2 a mn

Lecture notes for Analysis of Algorithms : Markov decision processes

Chapter 2. Ma 322 Fall Ma 322. Sept 23-27

8. Statistical Equilibrium and Classification of States: Discrete Time Markov Chains

Using Markov Chains To Model Human Migration in a Network Equilibrium Framework

Stochastic Processes

Lecture 5: Random Walks and Markov Chain

Asynchronous Implementation of Distributed Coordination Algorithms Chen, Yao; Xia, Weiguo; Cao, Ming; Lu, Jinhu

IEOR 6711: Stochastic Models I Fall 2013, Professor Whitt Lecture Notes, Thursday, September 5 Modes of Convergence

ELA

Markov Chains and Stochastic Sampling

Let's contemplate a continuous-time limit of the Bernoulli process:

Injective semigroup-algebras

Markov decision processes and interval Markov chains: exploiting the connection

MATH 56A: STOCHASTIC PROCESSES CHAPTER 1

Let (Ω, F) be a measureable space. A filtration in discrete time is a sequence of. F s F t

Key words. Feedback shift registers, Markov chains, stochastic matrices, rapid mixing

REPRESENTING HOMOLOGY AUTOMORPHISMS OF NONORIENTABLE SURFACES

(1) 0 á per(a) Í f[ U, <=i

Rafał Kapica, Janusz Morawiec CONTINUOUS SOLUTIONS OF ITERATIVE EQUATIONS OF INFINITE ORDER

Math 3013 Problem Set 4

On Distributed Coordination of Mobile Agents with Changing Nearest Neighbors

Section 9.2: Matrices. Definition: A matrix A consists of a rectangular array of numbers, or elements, arranged in m rows and n columns.

IN THIS paper we investigate the diagnosability of stochastic

ISyE 6650 Probabilistic Models Fall 2007

A NEW PROOF OF THE WIENER HOPF FACTORIZATION VIA BASU S THEOREM

ON CONSTRUCTING NEARLY DECOMPOSABLE MATRICES D. J. HARTFIEL

SMSTC (2007/08) Probability.

PRESENT STATE AND FUTURE PROSPECTS OF STOCHASTIC PROCESS THEORY

Packing-Dimension Profiles and Fractional Brownian Motion

Notes on Row Reduction

A Single-letter Upper Bound for the Sum Rate of Multiple Access Channels with Correlated Sources

Perron Frobenius Theory

Chapter 5. Linear Algebra. A linear (algebraic) equation in. unknowns, x 1, x 2,..., x n, is. an equation of the form

Section 1.5. Solution Sets of Linear Systems

MARKOV PROCESSES. Valerio Di Valerio

ON THE SHORTEST PATH THROUGH A NUMBER OF POINTS

103 Some problems and results in elementary number theory. By P. ERDÖS in Aberdeen (Scotland). Throughout this paper c, c,... denote absolute constant

Primes in the Semigroup of Non-Negative Matrices

Invertible Matrices over Idempotent Semirings

Mock Exam 3. 1 Hong Kong Educational Publishing Company. Section A. 1. Reference: HKDSE Math M Q1 (a) (1 + 2x) 2 (1 - x) n

Irreducibility. Irreducible. every state can be reached from every other state For any i,j, exist an m 0, such that. Absorbing state: p jj =1

A proof of the Jordan normal form theorem

ON CONVERGENCE OF STOCHASTIC PROCESSES

10. Linear Systems of ODEs, Matrix multiplication, superposition principle (parts of sections )

Reaching a Consensus in a Dynamically Changing Environment A Graphical Approach

A matrix over a field F is a rectangular array of elements from F. The symbol

This paper was published in Connections in Discrete Mathematics, S. Butler, J. Cooper, and G. Hurlbert, editors, Cambridge University Press,

ON ORDER-CONVERGENCE

642:550, Summer 2004, Supplement 6 The Perron-Frobenius Theorem. Summer 2004

Systems of Linear Equations and Matrices

Reinforcement Learning

CHAINS IN PARTIALLY ORDERED SETS

THE OHIO JOURNAL OF SCIENCE

Inverses of Square Matrices

Systems of Linear Equations and Matrices

KOREKTURY wim11203.tex

On integral representations of q-gamma and q beta functions

2 DISCRETE-TIME MARKOV CHAINS

SOME APPLICATIONS OF INFORMATION THEORY TO CELLULAR AUTOMATA

1 Γ x. x V (Γ\X) Vol (Γ\\X)

A finite basis theorem for product varieties of groups

Solving Homogeneous Systems with Sub-matrices

arxiv: v1 [math.pr] 26 Mar 2008

1 - Systems of Linear Equations

Entropy Rate of Stochastic Processes

Recap. Probability, stochastic processes, Markov chains. ELEC-C7210 Modeling and analysis of communication networks

Linear Algebra Homework and Study Guide

MATH 56A SPRING 2008 STOCHASTIC PROCESSES 65

Solutions to Exam I MATH 304, section 6

Markov Chain Monte Carlo

Each copy of any part of a JSTOR transmission must contain the same copyright notice that appears on the screen or printed page of such transmission.

Transience: Whereas a finite closed communication class must be recurrent, an infinite closed communication class can be transient:

THE NUMBER OF LOCALLY RESTRICTED DIRECTED GRAPHS1

Research Collection. The relation between block length and reliability for a cascade of AWGN links. Conference Paper. ETH Library

MATH 152 Exam 1-Solutions 135 pts. Write your answers on separate paper. You do not need to copy the questions. Show your work!!!

Entropy and Ergodic Theory Notes 21: The entropy rate of a stationary process

A Criterion for the Stochasticity of Matrices with Specified Order Relations

MATH 56A: STOCHASTIC PROCESSES CHAPTER 6

Models of Computation,

Definition 6.1 (p.277) A positive integer n is prime when n > 1 and the only positive divisors are 1 and n. Alternatively

Markov Chains. X(t) is a Markov Process if, for arbitrary times t 1 < t 2 <... < t k < t k+1. If X(t) is discrete-valued. If X(t) is continuous-valued

Section Notes 9. Midterm 2 Review. Applied Math / Engineering Sciences 121. Week of December 3, 2018

MATH 564/STAT 555 Applied Stochastic Processes Homework 2, September 18, 2015 Due September 30, 2015

Stochastic process. X, a series of random variables indexed by t

Solving a linear equation in a set of integers II

An Oscillation Theorem for a Sturm - Liouville Eigenvalue Problem

Transcription:

PRODUCTS OF INDECOMPOSABLE, APERIODIC, STOCHASTIC MATRICES1 J. WOLFOWITZ 1. Introduction. A finite square matrix P ={/»«} is called stochastic if pij^q for all i, j, and Hipa=l for all i. A stochastic matrix P is called indecomposable and aperiodic (SIA) if Q = lim P" exists and all the rows of Q are the same. SIA matrices are defined differently in books on probability theory; see, for example, [l] or [2]. The latter definition is more intuitive, takes longer to state, is easier to verify, and explains why the probabilist is interested in SIA matrices. A theorem in probability theory or matrix theory then says that the customary definition is equivalent to the one we have given. The latter is brief and emphasizes the property which will interest us in this note. We define Ô(P) by 5(P) = max max pixi pitj \. i fi.ti Thus S(P) measures, in a certain sense, how different the rows of P are. If the rows of P are identical, 5(P) =0 and conversely. Let Ai,, At be any square matrices of the same order. By a word (in the A's) of length t we mean the product of t A's (repetitions permitted). The object of this note is to prove the following: Theorem. Let Ai,, A* be square stochastic matrices of the same order such that any word in the A's is SIA. For any e > 0 there exists an integer vie) such that any word B {in the A's) of length n^v(e) satisfies «(B) < e. In words, the result is that any sufficiently long word in the A's has all its rows approximately the same. It is sufficient to require that any word in the A's be indecomposable. (This means, in the terminology of [l], that it contains at most one closed set of states other than the set of all states, or, in the Presented to the Society, June 6, 1962; received by the editors June 7, 1962. 1 Research under contract with the Office of Scientific Research of the U. S. Air Force. 733

734 J. WOLFOWITZ [October terminology of [2], that it contains only one ergodic class.) For, as pointed out by Thomasian [ó], if C is a word which is indecomposable and has period r>l, C is decomposable. The theorem has applications to coding (information) theory (see [3]) and obvious applications to the study of nonhomogeneous Markov chains. For 2X2 matrices A the theorem is trivial. For then the determinant of each matrix is less than one in absolute value. The determinant of a long enough word has a value close to zero, so that the rows of the word are almost identical. It would be desirable to have the hypothesis of the theorem require only that the A's be SIA, but it is easy to prove by a counter-example that this would be insufficient. Indeed (e.g., [4, Equation 4(b)]), one can construct two SIA matrices Ai and A2 such that A1A2 is decomposable. Hence for no n does the word (AiA2)n have all its rows approximately the same. Thus the condition of the theorem is not only sufficient but also necessary. Thomasian [ó] has proved a result which will be described at the end of 2 and which implies an algorithm such that, in a bounded number of arithmetical operations, starting with Ai,, At, one can determine whether every word in the A's (of whatever length) is SIA. The number of arithmetical operations required by his method may still be very large. At the end of 2 we make a suggestion which will usually substantially reduce the amount of calculation required. A result related to ours has been proved by Sarymsakov [5]. Lemma 1 of his paper (which he considers its most important result) is Lemma 4 of the present paper, with our / replaced by two less than the order of the A's. Sarymsakov's conditions are only on the individual A's (and not on all the words in the A's) which is a pleasant feature. However, his conditions are rather strong and not obviously meaningful in a probabilistic sense. Essentially they are that each A remain SIA under permutation of rows and columns, the permutation of rows being independent of the permutation of the columns. As an example of the implications of this consider the matrix D, where fo 1 01 D- 0 0 1 [X X x) Here the x's stand for positive elements whose actual value is immaterial. D is obviously SIA. It does not fulfill the above described conditions of [5], as may be seen by putting the first column in third position.

1963] INDECOMPOSABLE, APERIODIC, STOCHASTIC MATRICES 735 2. Proof of the theorem. Henceforth, all matrices under discussion will be square and stochastic without any further mention of this fact. Capital Roman letters will always denote matrices. We will say that Pi and P2 are of the same type, Pi~P2, if they have zero elements and positive elements in the same places. Whether or not a matrix is SIA depends solely on its type. Define X(P) = 1 - min H min (ph, phi). If X(P) <1 we will call P a scrambling matrix. X(P) = 0 if and only if 5(P)=0 and conversely. X(P)<1 implies that, for every pair of rows ii and i2, there exists a column b (which may depend on *j and i2) such that piib>0 and > 2,>0, and conversely. The introduction of X is due to Hajnal [4]. Russian writers (e.g., [5]) use a related coefficient of "ergodicity." We note the following for future reference : Lemma 1. // one or more matrices in a product of matrices is scrambling, so is the product. This is Lemma 2 of [4], Lemma 2. For any k This is Theorem 2 of [4]. We now prove 5(PiP2 Pt) ^ II X(Pi). =i Lemma 3. If F2 is an SIA matrix and (2.1) PiP2~Pi, then Pi is a scrambling matrix. Proof. From (2.1) we obtain that, for every positive integral», (2.2) PiP;~Pi, because P1P2 ~ P1P2 ~ Pi. From the definition of an SIA matrix it follows that, for n sum-

736 J. WOLFOWITZ [October ciently large, PS is a scrambling matrix. The desired result now follows from (2.2) and Lemma 1. The conclusion of this lemma could easily be strengthened, but the present lemma will suffice for our purposes. Let t be the number of different types of all SIA matrices of the same order as the A's. Lemma 4. All words in the A's of length ^i + 1 are scrambling matrices. Proof. Consider any word in the A's of length t + l, say Bi-B2 Bt+i. From the definition of t it follows that there exist integers a and b, 0<a<b^t+l such that (2.3) Bi-B2 B0~Bi-B2 B,,. It follows from Lemma 3 that Bi B2 Ba is a scrambling matrix. The lemma now follows from Lemma 1. Lemma 5. There exists a constant d,0^d<l, with the following property: Let C be any word of length (t + l) in the A's. Then X(C) d. Proof. There are a finite number of words of length ( + 1) in the A's. By Lemma 4 each of these words has a X which is less than one. Let d be the largest of these values of X. Proof of the theorem. Let h be such that dh <. The theorem now follows from Lemma 2, with p(e) = *(/ + 1). The result of Thomasian's referred to earlier is as follows: Let T(k) be the set of types of all words in the A's of length ^k. If, for some k, T(k) =T(k-\-l) and all the types in T(k) are those of SIA matrices, then all words in the A's are SIA; the converse also holds. Let 5 be the order of the A's. If all words in the A's are SIA then the above property must hold for k ^ 2*'. (Actually it must hold for k^t.) Since any word which has a scrambling matrix as a factor is SIA, we can at once disregard all A's which are scrambling and consider the classes T only for those A's which are not scrambling. Thus if all the A's are scrambling no computations are necessary. In computing the T's for the A's which are not scrambling we need not consider further any word, a factor of which is scrambling. This procedure

1963] INDECOMPOSABLE, APERIODIC, STOCHASTIC MATRICES 737 will generally considerably reduce the amount of computation required. It is easy to verify whether or not a given matrix is scrambling. The only place where we make use of the fact that there are only finitely many A's is in Lemma 5. If there are infinitely many A's but the conclusion of Lemma 5 holds then the theorem also holds. Some of the results on Markov chains in the literature are immediate consequences of this remark. In particular also, Lemma 4 always holds, even if there are infinitely many A's. References 1. W. Feller, An introduction to probability theory and its applications, 2nd ed., Vol. 1, Wiley, New York, 1957. 2. J. L. Doob, Stochastic processes, Wiley, New York, 1953. 3. J. Wolfowitz, Strong converse of the coding theorem for indecomposable channels, Sankhya (1963). See also the abstract by the same title in Ann. Math. Statist. 33 (1962), 1212. 4. J. Hajnal, Weak ergodicity in non-homogeneous Markov chains, Proc. Cambridge Philos. Soc. 54 (1958), 233-246. 5. T. A. Sarymsakov, Inhomogeneous Markov chains, Teor. Verojatnost. i Primenen. 6 (1961), 194-201. 6. A. J. Thomasian, A finite criterion for indecomposable channels, Ann. Math. Statist. 34 (1963), 337-338. Cornell University