STOCHASTIC COMPLEMENTATION, UNCOUPLING MARKOV CHAINS, AND THE THEORY OF NEARLY REDUCIBLE SYSTEMS

Similar documents
SENSITIVITY OF THE STATIONARY DISTRIBUTION OF A MARKOV CHAIN*

AN ELEMENTARY PROOF OF THE SPECTRAL RADIUS FORMULA FOR MATRICES

Eventually reducible matrix, eventually nonnegative matrix, eventually r-cyclic

Detailed Proof of The PerronFrobenius Theorem

Markov Chains and Spectral Clustering

Chapter 3 Transformations

Linear Algebra. The analysis of many models in the social sciences reduces to the study of systems of equations.

Markov Chains and Stochastic Sampling

Markov Chains, Stochastic Processes, and Matrix Decompositions

Using Markov Chains To Model Human Migration in a Network Equilibrium Framework

Z-Pencils. November 20, Abstract

Rectangular Systems and Echelon Forms

Lecture Summaries for Linear Algebra M51A

On the mathematical background of Google PageRank algorithm

Central Groupoids, Central Digraphs, and Zero-One Matrices A Satisfying A 2 = J

EE731 Lecture Notes: Matrix Computations for Signal Processing

In particular, if A is a square matrix and λ is one of its eigenvalues, then we can find a non-zero column vector X with

Finite-Horizon Statistics for Markov chains

Geometric Mapping Properties of Semipositive Matrices

Foundations of Matrix Analysis

Numerical Analysis Lecture Notes

Intrinsic products and factorizations of matrices

On Backward Product of Stochastic Matrices

Scientific Computing

Jim Lambers MAT 610 Summer Session Lecture 2 Notes

Definition A finite Markov chain is a memoryless homogeneous discrete stochastic process with a finite number of states.

Updating PageRank. Amy Langville Carl Meyer

Stochastic modelling of epidemic spread

MATH36001 Perron Frobenius Theory 2015

Introduction to Real Analysis Alternative Chapter 1

MATRIX ALGEBRA AND SYSTEMS OF EQUATIONS. + + x 1 x 2. x n 8 (4) 3 4 2

Boolean Inner-Product Spaces and Boolean Matrices

LECTURE NOTES ELEMENTARY NUMERICAL METHODS. Eusebius Doedel

Perron eigenvector of the Tsetlin matrix

Direct Methods for Solving Linear Systems. Simon Fraser University Surrey Campus MACM 316 Spring 2005 Instructor: Ha Le

Review Questions REVIEW QUESTIONS 71

Discrete Applied Mathematics

AMS 209, Fall 2015 Final Project Type A Numerical Linear Algebra: Gaussian Elimination with Pivoting for Solving Linear Systems

Key words. Strongly eventually nonnegative matrix, eventually nonnegative matrix, eventually r-cyclic matrix, Perron-Frobenius.

MAA704, Perron-Frobenius theory and Markov chains.

Solving Homogeneous Systems with Sub-matrices

Linear Algebra March 16, 2019

Updating Markov Chains Carl Meyer Amy Langville

642:550, Summer 2004, Supplement 6 The Perron-Frobenius Theorem. Summer 2004

No class on Thursday, October 1. No office hours on Tuesday, September 29 and Thursday, October 1.

Statistics 992 Continuous-time Markov Chains Spring 2004

Kernels of Directed Graph Laplacians. J. S. Caughman and J.J.P. Veerman

PROOF OF TWO MATRIX THEOREMS VIA TRIANGULAR FACTORIZATIONS ROY MATHIAS

Preliminary/Qualifying Exam in Numerical Analysis (Math 502a) Spring 2012

Block-tridiagonal matrices

1 Basic Combinatorics

MAT 2037 LINEAR ALGEBRA I web:

Interlacing Inequalities for Totally Nonnegative Matrices

= W z1 + W z2 and W z1 z 2

Math Camp Lecture 4: Linear Algebra. Xiao Yu Wang. Aug 2010 MIT. Xiao Yu Wang (MIT) Math Camp /10 1 / 88

Algebraic Methods in Combinatorics

A Generalized Eigenmode Algorithm for Reducible Regular Matrices over the Max-Plus Algebra

Numerical Linear Algebra

Linear Algebra Massoud Malek

Chapter 1: Systems of linear equations and matrices. Section 1.1: Introduction to systems of linear equations

MAT Linear Algebra Collection of sample exams

Institute for Advanced Computer Studies. Department of Computer Science. On Markov Chains with Sluggish Transients. G. W. Stewart y.

Math 471 (Numerical methods) Chapter 3 (second half). System of equations

14.2 QR Factorization with Column Pivoting

Principal Component Analysis

Chapter 16 focused on decision making in the face of uncertainty about one future

The Solution of Linear Systems AX = B

Modeling and Stability Analysis of a Communication Network System

Direct Methods for Solving Linear Systems. Matrix Factorization

Numerical Analysis: Solving Systems of Linear Equations

Matrices and Matrix Algebra.

RATIONAL REALIZATION OF MAXIMUM EIGENVALUE MULTIPLICITY OF SYMMETRIC TREE SIGN PATTERNS. February 6, 2006

Comparison of perturbation bounds for the stationary distribution of a Markov chain

LU Factorization. Marco Chiarandini. DM559 Linear and Integer Programming. Department of Mathematics & Computer Science University of Southern Denmark

Conditioning of the Entries in the Stationary Vector of a Google-Type Matrix. Steve Kirkland University of Regina

The initial involution patterns of permutations

1 Multiply Eq. E i by λ 0: (λe i ) (E i ) 2 Multiply Eq. E j by λ and add to Eq. E i : (E i + λe j ) (E i )

Linear Algebra (part 1) : Vector Spaces (by Evan Dummit, 2017, v. 1.07) 1.1 The Formal Denition of a Vector Space

YOUNG TABLEAUX AND THE REPRESENTATIONS OF THE SYMMETRIC GROUP

Positive entries of stable matrices

On hitting times and fastest strong stationary times for skip-free and more general chains

Chapter 4 No. 4.0 Answer True or False to the following. Give reasons for your answers.

Definition 2.3. We define addition and multiplication of matrices as follows.

Lecture notes for Analysis of Algorithms : Markov decision processes

Linearizing Symmetric Matrix Polynomials via Fiedler pencils with Repetition

EXAMPLES OF CLASSICAL ITERATIVE METHODS

Section 3.3. Matrix Rank and the Inverse of a Full Rank Matrix

Standard forms for writing numbers

Linear Algebra: Matrix Eigenvalue Problems

MTH Linear Algebra. Study Guide. Dr. Tony Yee Department of Mathematics and Information Technology The Hong Kong Institute of Education

Sign Patterns with a Nest of Positive Principal Minors

Stochastic modelling of epidemic spread

Bare-bones outline of eigenvalue theory and the Jordan canonical form

The Cayley-Hamilton Theorem and the Jordan Decomposition

Linear Algebra and Eigenproblems

Abed Elhashash and Daniel B. Szyld. Report August 2007

MARKOV CHAIN MONTE CARLO

Pivoting. Reading: GV96 Section 3.4, Stew98 Chapter 3: 1.3

Introduction to Search Engine Technology Introduction to Link Structure Analysis. Ronny Lempel Yahoo Labs, Haifa

Lecture 20 : Markov Chains

Transcription:

STOCHASTIC COMLEMENTATION, UNCOULING MARKOV CHAINS, AND THE THEORY OF NEARLY REDUCIBLE SYSTEMS C D MEYER Abstract A concept called stochastic complementation is an idea which occurs naturally, although not always explicitly, in the theory and application of finite Markov chains This paper brings this idea to the forefront with an explicit definition and a development of some of its properties Applications of stochastic complementation are explored with respect to problems involving uncoupling procedures in the theory of Markov chains Furthermore, the role of stochastic complementation in the development of the classical Simon Ando theory of nearly reducible system is presented Key words Markov chains, stationary distributions, stochastic matrix, stochastic complementation, nearly reducible systems, Simon Ando theory AMS(MOS subject classifications 65U05, 60-02, 60J10, 60J20, 15-02, 15A51, 90A14 1 Introduction Although not always given an explicit name, a quantity which we shall refer to as a stochastic complement arises very naturally in the consideration of finite Markov chains This concept has heretofore not been focused upon as an entity unto itself, and a detailed study of the properties of stochastic complementation has not yet been given The purpose of the first part of this exposition is to explicitly publicize the utility of this concept and to present a more complete and unified discussion of the important properties of stochastic complementation A focal point of the development concerns the problem of determining the stationary distribution of an irreducible Markov chain by uncoupling the chain into several smaller independent chains In particular, if m m is the transition matrix for an m- state, homogeneous, irreducible Markov chain C, the stationary distribution problem concerns the determination of the unique vector π 1 m which satisfies π = π, π i > 0, m π i =1 i=1 For chains with relatively few states, this is not a difficult problem to solve using adaptations of standard techniques for solving systems of linear equations However, there exist many applications for which the number of states is too large to be comfortably handled by standard methods For large-scale problems, it is only natural to attempt to somehow uncouple the original m-state chain C into two or more smaller chains say C 1, C 2,, C k containing r 1,r 2,,r k states, respectively, where k i=1 r i = m Ideally, this sequence of smaller chains should have the following properties: Each smaller chain C i should be irreducible whenever the original chain C is irreducible so that each C i has a unique stationary distribution vector s i It should be possible to determine the s i s completely independent of each other For modern multiprocessor computer architectures it is desirable to be able to execute the computation of the s i s in parallel Received by the editors January 26, 1988; accepted for publication (in revised form February 2, 1989 This work was supported by National Science Foundation grant DMS- 8521154 North Carolina State University, Department of Mathematics; Center for Research in Scientific Computation, Box 8205, Raleigh, North Carolina 27695-8205 1

2 C D MEYER Finally, it must be possible to easily couple the smaller stationary distribution vectors s i back together in order to produce the stationary distribution vector π for the larger chain C The second part of this paper is dedicated to showing how to accomplish the above three goals, and it is indicated how this can lead to fully parallel algorithms for the determination of the stationary distribution vector of the original chain Finally, in the third part of this survey, the application of stochastic complementation to the classical Simon Ando theory developed in [18] for nearly completely reducible chains is presented It is demonstrated how to apply the concept of stochastic complementation in order to develop the theory for nearly completely reducible systems in a unified, clear, and simple manner while simultaneously sharpening some results and generalizing others 2 Stochastic complementation The purpose of this section is to introduce the concept of a stochastic complement in an irreducible stochastic matrix and to develop some of the basic properties of stochastic complementation These ideas will be the cornerstone for all subsequent discussions Unless otherwise stated, will denote an m m, irreducible, stochastic matrix which will be partitioned as 11 12 1k = 21 22 2k k1 k2 kk where all diagonal blocks are square It is well known that I is a singular M-matrix of rank m 1 and that every principal submatrix of I of order m 1 or smaller is a nonsingular M-matrix (see [2, p 156] In particular, if i denotes the principal submatrix of obtained by deleting the ith row and ith column of blocks from the partitioned form of, then each I i is a nonsingular M-matrix Therefore, (21 (I i 1 0, so the indicated inverses in the following definition are well defined Definition 21 Let be an m m irreducible stochastic matrix with a k-level partition 11 12 1k = 21 22 2k k1 k2 kk in which all diagonal blocks are square For a given index i, let i denote the principal block submatrix of obtained by deleting the ith row and ith column of blocks from, and let i and i designate i =( i1 i2 i,i 1 i,i+1 ik

and STOCHASTIC COMLEMENTATION 3 i = 1i i 1,i i+1,i ki That is, i is the ith row of blocks with ii removed, and i is the ith column of blocks with ii removed The stochastic complement of ii in is defined to be the matrix S ii = ii + i (I i 1 i For example, the stochastic complement of 22 in = 11 12 13 21 22 23 31 32 33 is given by ( I 11 S 22 = 22 +( 21 23 13 31 I 33 1 ( 12 32 The reason for the terminology stochastic complement stems from the fact that all stochastic complements are stochastic matrices (see Theorem 21 together with the observation that although stochastic complementation is not the same as the well known concept of Schur complementation, there is nevertheless a direct connection in the case of a two-level partition 11 = 12 21 22 If is an irreducible stochastic matrix with square diagonal blocks, then the stochastic complement of 11 is given by and the stochastic complement of 22 is S 11 = 11 + 12 (I 22 1 21, S 22 = 22 + 21 (I 11 1 12, and it is easy to see that the stochastic complement of 11 is in fact [I Schur Complement(I 22 ] in the matrix I, while the stochastic complement of 22 is [I Schur Complement(I 11 ] in the matrix I The concept of stochastic complementation arises very naturally in the consideration of finite Markov chains, and more generally, in the theory of nonnegative matrices

4 C D MEYER Consequently, it is not surprising that the concept of stochastic complementation although not always known by that name appears at least implicitly, if not explicitly, in a variety of places For example, stochastic complementation is a special case of a more general concept known as erron complementation which has been studied in [17] Simple forms of stochastic complementation can be found either explicitly or implicitly in a variety of Markov chain applications such as those in [1], [3] [6], [8] [10], [12] [16], [20], [21], in addition to several others However, it seems that stochastic complementation has heretofore not been focused upon as an entity unto itself, and a detailed study of the properties of stochastic complement matrices has not yet been given art of the purpose of this paper is to explicitly publicize the utility of this concept and to present a more complete and unified discussion of the important properties of stochastic complementation The following technical lemma is needed to help develop the subsequent theory Its proof is a straightforward application of the standard features of permutation matrices, and consequently the proof is omitted Lemma 21 For an m m irreducible stochastic matrix 11 12 1k = 21 22 2k k1 k2 kk in which all diagonal blocks are square, let Q be the permutation matrix associated with an interchange of the first and ith block rows (or block columns, and let be the matrix = QQ If is partitioned into a 2 2 block matrix 11 12 = where 11 = ii, 21 22 then the stochastic complement of ii in is given by (22 S ii = S 11 = 11 + 12 (I 22 1 21 We are now in a position to prove that every stochastic complement is indeed a stochastic matrix Theorem 21 If 11 12 1k = 21 22 2k k1 k2 kk is an irreducible stochastic matrix, then each stochastic complement is also a stochastic matrix S ii = ii + i (I i 1 i

STOCHASTIC COMLEMENTATION 5 roof For a given index i, assume that has been permuted and repartitioned as described in Lemma 21 so that we may consider 11 12 = where 11 = ii 21 22 According to (22, the stochastic complement of ii in is the same as the stochastic complement of 11 in that is, S ii = S 11 Since each principal submatrix of I (and I of order m 1 or less is a nonsingular M-matrix, it follows from (21 ( that I 1 22 0, and hence S 11 = 11 + 12 (I 22 1 21 0 Note S 11 need not be strictly positive an example is given after this proof To see that the row sums of S 11 are each 1, let 1 e =, 1 and allow the dimension of e to be defined by the context in which it appears The fact that the row sums in are all 1 can be expressed by writing (23 11 e + 12 e = e and (24 Equation (24 implies e = and this together with (23 yields 21 e + 22 e = e ( I 22 1 21 e, S 11 e = 11 e + 12 (I 22 1 21 e = 11 e + 12 e = e Consequently, S 11 = S ii must be stochastic To see that a stochastic complement need not be strictly positive, consider a 4 4 irreducible stochastic matrix whose partitions and zero pattern are shown below = + + 0 0 + + + + + + + + + + + + For this configuration, S 11 = 11 + 12 (I 22 1 21 = + + 0 + + + + 0 + [+](+ + + + + + + = + + 0 + + + + + +

6 C D MEYER Although stochastic complements can have zero entries, the zeros are always in just the right places so as to guarantee that each S ii is an irreducible matrix However, before this can be established, it is necessary to observe some additional facts concerning stochastic complementation Theorem 22 Suppose that 11 12 1k = 21 22 2k k1 k2 kk is an irreducible stochastic matrix in which each ii is square, and let π =(π (1 π (2 π (k be the conformably partitioned stationary distribution vector for If each π (i is normalized in order to produce the probability vector (25 then s i = π(i π (i e, s i S ii = s i for each i =1, 2,,k That is, s i is a stationary distribution vector for the stochastic complement S ii roof Assume that has been permuted and repartitioned as described in Lemma 21 so that 11 12 = QQ = 21 22 Use the fact that 0 = π(i implies 0 = πq 2 (I Q =(π (i π (2 π (1 π (k (I together with the equation ( I 11 ( ( 12 21 I I 0 I S11 22 (I 22 1 = 12 21 I 0 I 22 in order to conclude that π (i (I S 11 =0 The desired result now follows from Lemma 21 because S ii = S 11 Although Theorem 22 establishes that each s i is a stationary distribution vector for S ii, nothing proven to this point allows for the conclusion that the s i s are unique However, this will follow once it is established that each S ii inherits the property of irreducibility from the original matrix Theorem 23 If 11 12 1k = 21 22 2k k1 k2 kk

STOCHASTIC COMLEMENTATION 7 is an irreducible stochastic matrix, then each stochastic complement is also an irreducible stochastic matrix S ii = ii + i (I i 1 i roof Assume that has been permuted and repartitioned as described in Lemma 21 so that 11 12 = 21 22 and (26 S ii = S 11 The fact that each S ii is stochastic was established in Theorem 21 (26, we need only to prove irreducibility of S 11 By noting that ( ( I 12 I 1 22 0 I ( I 11 12 21 I 22 ( I S11 0 = 0 I 22 By virtue of ( I 0 I 1 22 21 I, it follows that Rank I = Rank (I S 11 + Rank (I 22 Suppose that, 11, and 22 have dimensions m m, r r, and q q, respectively, with r+q = m It is ( well known (see [2, p 156] that being irreducible and stochastic implies that Rank I = m 1 and Rank (I 22 = q Consequently, Rank (I S 11 = m 1 q = r 1, and therefore I S 11 has a one-dimensional nullspace The left-hand nullspace of I S 11 is spanned by the strictly positive row vector s 1 = s i defined in (25, and the right-hand nullspace of I S 11 is spanned by the strictly positive column vector e containing r 1 s Since s 1 e = 1, the spectral projector associated with the eigenvalue λ = 1 for S 11 must be R r r = e s 1 > 0 Because every stochastic matrix is Cesàro summable to the spectral projector associated with the unit eigenvalue (see [11], it follows that (27 lim n I + S 11 + S 2 11 + + n S n 1 11 = R > 0

8 C D MEYER It is now evident that S 11 cannot be reducible otherwise S 11 could be permuted to a form A B 0 C and the limit in (27 would necessarily contain zero entries The proof given above is algebraic in nature, but because irreducibility is a strictly combinatorial concept, one might wonder if a strictly combinatorial proof is possible The answer is yes, and the details may be obtained by restricting the discussion given in [17] to the special case of stochastic matrices Furthermore, it seems reasonable that there should be a probabilistic argument, and indeed there is Irreducibility is a direct consequence of the probabilistic interpretation of stochastic complementation as explained in the next section 3 The probabilistic interpretation of stochastic complementation Assume again that for a given index i, the original transition matrix has been permuted and repartitioned as described in Lemma 21 so that ( A B A = 11 12 B 21 22 where 11 = ii, and S ii = S 11 = 11 + 12 (I 22 1 21 Consider a new process called the reduced chain which is defined by observing the old process only when it is in a state belonging to the subclass A That is, transitions to states in B are masked out, so a direct path A k = A j in the reduced chain corresponds in the old process to either a direct path A k A j or a detour A k B A j passing through B In loose terms, one can say that the reduced chain is derived from the original chain by turning off the meter whenever the old process is in B If the original chain is irreducible, then it is clear that the reduced chain is also irreducible For example, if A 1 B 2 A 3 B 4 B 5 A 6 A 7 is a sequence of direct paths from A 1 to A 7 in the original chain, then A 1 = A 3 = A 6 = A 7 is a sequence of direct paths from A 1 to A 7 in the reduced chain For the reduced chain, the one-step transition probability of moving from A k to A j is the probability in the original process of moving directly from A k to A j plus the probability of moving directly from A k to some state in B, and then eventually moving back to A, hitting A j first upon return The probability of moving directly from A k to A j in the original chain is [ ] q kj = 11, kj

STOCHASTIC COMLEMENTATION 9 and the probability of moving directly from A k to B h Bis [ ] q kh = 12 The probability of moving from B h to A such that A j is the first state entered upon return to A is [ ( ] q hj = I 1 22 21 As explained in [12], this expression for q hj is obtained by considering the states in A to be absorbing states and applying the theory of absorbing chains Consequently, the one-step transition probabilities in the reduced chain are q kj + [ q kh q hj = 11 + 12 (I ] 1 22 21 =[S ii ] kj B h B In other words, the stochastic complement S ii represents the transition matrix for the reduced chain which is obtained from the original chain by masking out transitions to states in B 4 Uncoupling Markov chains For an m-state irreducible Markov chain C with transition matrix, the object of uncoupling is to somehow decompose the chain C into two or more smaller chains say C 1, C 2,, C k containing r 1,r 2,,r k states, respectively, where k i=1 r i = m This sequence of smaller chains is required to possess the following properties: Each smaller chain C i should be irreducible whenever the original chain C is irreducible so as to guarantee that each C i has a unique stationary distribution vector s i It should be possible to determine the s i s completely independent of each other For modern multiprocessor computer architectures it is desirable to be able to execute the computation of the s i s in parallel Finally, it must be possible to easily couple the smaller stationary distribution vectors s i back together in order to produce the stationary distribution vector π for the larger chain C The purpose of this section is to show how the concept of stochastic complementation can be used to achieve these goals As in the previous section, will denote an m m irreducible stochastic matrix with a k-level partition 11 12 1k = 21 22 2k k1 k2 kk in which all diagonal blocks ii are square, and the stationary distribution vector π for will be partitioned as kh hj π =(π (1 π (2 π (k where the size of π (i corresponds to the order of ii We know from the results of Theorem 23 that each stochastic complement S ii = ii + i (I i 1 i kj

10 C D MEYER is an irreducible stochastic matrix, and consequently, each S ii possesses a unique stationary distribution vector s i Theorem 22 shows that each s i and π (i differ only by a positive scalar multiple so that it is indeed possible to combine the s i s with coupling factors {ξ 1,ξ 2,,ξ k } in order to produce π =(ξ 1 s 1 ξ 2 s 2 ξ k s k However, the expressions for the coupling factors ξ i given in Theorem 22 have the form ξ i = h π (i h At first glance, this might seem to place the issue in a hopeless circle because prior knowledge of π, in the form of the sums h π(i h, is necessary in order to reconstruct π from the s i s Fortunately, there is an elegant way around this dilemma The following theorem shows that the coupling factors ξ i are easily determined without prior knowledge of the π (i s this will become the key feature in the uncouplingcoupling technique Theorem 41 (The coupling theorem If is an m m irreducible stochastic matrix partitioned as 11 12 1k = 21 22 2k k1 k2 kk with square diagonal blocks, then the stationary distribution vector for is given by π =(ξ 1 s 1 ξ 2 s 2 ξ k s k where s i is the unique stationary distribution vector for the stochastic complement and where S ii = ii + i (I i 1 i, ξ =(ξ 1 ξ 2 ξ k is the unique stationary distribution vector for the k k irreducible stochastic matrix C whose entries are defined by c ij s i ij e The matrix C is hereafter referred to as the coupling matrix, and the scalars ξ i are called the coupling factors roof First prove that the coupling matrix C is stochastic and irreducible By its definition, it is clear that C 0 To see that C is stochastic, use the fact that k j=1 ije = e and write k c ij = j=1 k k s i ij e = s i ij e = s i e =1 j=1 j=1

STOCHASTIC COMLEMENTATION 11 To show that C is irreducible, note that because ij 0, s i > 0, and e > 0, it must be the case that c ij = 0 if and only if ij = 0 Since is irreducible, this implies that C must also be irreducible otherwise, if C could be permuted to a block triangular form, then so could Let π =(π (1 π (2 π (k denote the partitioned stationary distribution vector for where the sizes of the π (i s correspond to the sizes of the ii s, respectively, and define ξ to be the vector where each component is defined by ξ =(ξ 1 ξ 2 ξ k ξ i = π (i e = h π (i h According to Theorem 22, (41 ξ i s i = π (i, and this together with the fact that k i=1 π(i ij = π (j yields ( k k k (ξc j = ξ i c ij = π (i ij e = π (i ij e = π (j e = ξ j i=1 i=1 Consequently, ξc = ξ so that ξ is a stationary vector for C It is clear that ξ must also be a probability vector because k ξ i = i=1 i=1 i=1 k m π (i e = π j =1 Therefore, since C is irreducible, ξ must be the unique stationary distribution vector for C, and the desired conclusion that j=1 π =(ξ 1 s 1 ξ 2 s 2 ξ k s k follows from (41 The results of Theorem 41 can be viewed as an exact aggregation technique The case of a two-level partition is of special interest because, as the following corollary shows, it is particularly easy to uncouple and couple the stationary distribution vector for these situations Corollary 41 If is an m m irreducible stochastic matrix partitioned as 11 = 12 21 22 where 11 and 22 are square, then the stationary distribution vector for is given by π =(ξ 1 s 1 ξ 2 s 2

12 C D MEYER where s 1 and s 2 are the unique stationary distribution vectors for the stochastic complements S 11 = 11 + 12 (I 22 1 21 and S 22 = 22 + 21 (I 11 1 12, respectively, and where the coupling factors ξ 1 and ξ 2 are given by ξ 1 = s 2 21 e s 1 12 e + s 2 21 e and ξ 2 =1 ξ 1 = s 1 12 e s 1 12 e + s 2 21 e In this corollary, the coupling factors ξ i are described in terms of the off-diagonal blocks 12 and 21, but these quantities can also be expressed using only the diagonal blocks in by replacing the terms 12 e and 21 e with e 11 e and e 22 e, respectively, so as to produce ξ 1 = 1 s 2 22 e 2 s 1 11 e s 2 22 e and ξ 2 =1 ξ 1 = 1 s 1 11 e 2 s 1 11 e s 2 22 e There is always a balancing act to be performed when uncoupling a Markov chain using stochastic complementation as described in Theorem 41 As k increases and the partition of becomes finer, the sizes of the stochastic complements become smaller thus making it easier to determine each of the stationary distribution vectors s i, but the order of the matrix inversion embedded in each stochastic complement becomes larger, and the size of the coupling matrix C k k becomes larger In the two extreme cases when k = m or k = 1, there is no uncoupling of the chain whatsoever if k = m, then C =, and if k = 1, then S 11 = One must therefore choose the partition which best suits the needs of the underlying application For example, if computation utilizing a particular multiprocessor computer is the goal, then the specific nature of the hardware and associated software may dictate the partitioning strategy Rather than performing a single uncoupling-coupling operation to a high level partition of, an alternate strategy is to execute a divide-and-conquer procedure using only two-level partitions at each stage Starting with an irreducible stochastic matrix of size m m, partition roughly in half as 11 = 12 21 22 to produce two stochastic complements, S 11 and S 22, which are each irreducible stochastic matrices of order approximately m/2 The stochastic complements S 11 and S 22 may in turn be partitioned roughly in half so as to produce four stochastic complements say (S 11 11,(S 11 22,(S 22 11, and (S 22 22 each of which is of order approximately m/4 This process can continue until all stochastic complements are sufficiently small in size so that each easily yields a stationary distribution vector The small stationary distribution vectors corresponding to the small stochastic complements are then successively coupled according to the rules given in Corollary 41 until the stationary distribution vector π for the original chain is produced Using this divide-and-conquer strategy, there are only two coupling factors to determine at each coupling step, and they are almost trivial to compute Furthermore, the order of the largest inversion imbedded in any stochastic complement never exceeds m/2 For example, consider the following 8 8 irreducible stochastic matrix

STOCHASTIC COMLEMENTATION 13 0 5 0 1 2 0 1 1 2 0 5 0 0 2 0 1 1 2 0 5 1 0 1 0 0 4 2 0 1 1 0 2 = 1 0 5 0 1 1 0 2 0 5 0 2 1 0 2 0 5 0 2 1 0 1 0 1 0 2 4 0 2 0 1 1 For the indicated partition, the stochastic complements of are given by 1 0899 5566 2329 1205 2327 1311 5885 0476 S 11 = 1665 2194 0986 5155 0443 5156 4102 0299 and 2832 2409 1110 3649 2576 2395 2819 2210 S 22 = 2511 2823 1329 3337, 3633 1683 2006 2678 with coupling vector ξ (0 =(6852 3148 Now, the two stochastic complements of S 11 are 1744 8256 (S 11 11 = and (S 4041 5959 11 22 = with coupling vector ξ (1 =(4548 5452, while the two stochastic complements of S 22 are 5776 4224 (S 22 11 = and (S 5512 4488 22 22 = with coupling vector ξ (2 =(5219 4781 ( 4278 5722 9056 0944 ( 3468 6532 3955 6045 1 Numbers have been rounded so as to display four digits behind the decimal point

14 C D MEYER The stationary distribution vector for (S 11 11 is given by s (1 1 =(3286 6714, and the stationary distribution vector for (S 11 22 is s (1 2 =(6128 3872, so that the stationary distribution vector for S 11 is s 1 = ( ξ (1 1 s(1 1 ξ (1 2 s(1 2 =(4548 ( 3286 6714 5452 ( 6128 3872 =(1495 3054 3341 2111 Similarly, the stationary distribution vector for (S 22 11 is given by s (2 1 =(5661 4339, and the stationary distribution vector for (S 22 22 is s (2 2 =(3771 6229, so that the stationary distribution vector for S 22 is s 2 = ( ξ (2 1 s(2 1 ξ (2 2 s(2 2 =(5219 ( 5661 4339 4781 ( 3771 6229 =(2955 2264 1803 2978 Therefore, the stationary distribution vector for must be π = ( ξ (0 1 s 1 ξ (0 2 s 2 ( = 6852 ( 1495 3054 3341 2111 3148 ( 2955 2264 1803 2978 = (1024 2092 2289 1446 0930 0713 0568 0938 In addition to the divide-and-conquer process illustrated above, there are several other variations and hybrid techniques (eg, iterative methods which are possible, and it is clear that the remarks of this section can serve as the basis for fully parallel algorithms for computing the stationary distribution vector for an irreducible chain R B Mattingly has conducted some detailed experiments along these lines in which he implemented the divide-and-conquer technique described above on a Sequent Balance 21000 a shared memory machine with 24 tightly coupled processors Exploring the numerical details and implementation of Mattingly s work would lead this exposition too far astray, but the interested reader can consult [14] or [15] In brief, Mattingly was able to achieve moderate speedups with the stochastic complementation divide-and-conquer technique eg, a speedup of approximately 85 with 16 processors was obtained Although the operation count for straightforward elimination is less than that for the divide-and-conquer technique, elimination methods do not parallelize as well as the divide-and-conquer technique, so, as the number of

STOCHASTIC COMLEMENTATION 15 processors was increased, the actual running times of the divide-and-conquer technique were competitive with optimized elimination schemes But perhaps even more significant is the fact that the divide-and-conquer technique can be extremely stable for ill-conditioned chains ie, irreducible stochastic matrices which have a cluster of eigenvalues very near the unit eigenvalue For Mattingly s ill-conditioned test matrices, standard elimination methods returned only one or two correct digits whereas the divide-and-conquer scheme based on stochastic complementation returned results which were correct to the machine precision This in part is due to the spectrum splitting effect which stochastic complementation provides these results are discussed in 6 of this exposition which concerns the spectral properties of stochastic complementation 5 rimitivity issues rimitivity is, of course, an important issue because for an irreducible stochastic matrix, the limit lim n k exists if and only if is primitive If is not primitive, then we are necessarily restricted to investigating limiting behavior in the weak (the Cesàro sense, and consequently it is worthwhile to make some observations concerning the degree to which primitivity or lack of it in a partitioned stochastic matrix is inherited by the smaller stochastic complements S ii The first observation to make is that being a primitive matrix is not sufficient to guarantee that all stochastic complements are primitive For example, the matrix 0 1 0 1 1 = 2 0 2 1 0 0 is irreducible and primitive because 5 > 0 However, for the indicated partition, the stochastic complement S 11 = 0 1 1 0 is not primitive Although primitivity of a stochastic complement is not inherited from the matrix itself, the following theorem shows that primitivity is inherited from the diagonal blocks of Theorem 51 If ii is primitive, then the corresponding stochastic complement S ii must also be primitive roof Since ii 0 and i (I i 1 i 0, it follows that for each positive integer n, S n ii = [ ii + i (I i 1 i ] n = n ii + N where N 0 Therefore, S n ii > 0 whenever n ii > 0 While stochastic complements need not be primitive, most of them are The next theorem explains why those stochastic complements which are not primitive must come from rather special stochastic matrices Theorem 52 If ii has at least one nonzero diagonal entry, then the corresponding stochastic complement S ii must be primitive

16 C D MEYER roof If ii has at least one nonzero diagonal entry, then so does S ii Furthermore, Theorems 21 and 23 guarantee that each S ii is always nonnegative and irreducible, and it is well known [2, p 34] that an irreducible nonnegative matrix with a positive trace must be primitive, so each S ii must be primitive The converse of Theorem 51 as well as the converse of Theorem 52 is false The stochastic matrix 0 1 0 0 0 0 1 1 0 0 2 2 0 0 0 0 0 0 1 0 (51 = 0 0 0 0 1 0 0 0 0 0 0 1 1 0 0 0 0 0 is irreducible, and the stochastic complements corresponding to the indicated partition are S 11 = 0 1 0 1 1 2 0 2 and S 22 = 0 1 0 0 0 1 1 0 0 1 2 1 2 0 Notice that each stochastic complement is primitive, but, 11, and 22 are not primitive The preceding example indicates another advantage which stochastic complementation can provide The matrix in (51 is not primitive because it is the transition matrix of a periodic chain, and hence the associated stationary distribution vector π cannot be computed by the power method ie, by the simple iteration π n+1 = π n where π 0 is arbitrary However, the chain uncouples into two aperiodic chains in the sense that S 11 and S 22 are both primitive, and therefore each S ii yields a stationary distribution vector s i by means of two straightforward iterations s (n+1 1 = s (n 1 S 11 and s (n+1 2 = s (n 2 S 22 where each s (0 i is arbitrary Take note of the fact that the two iterations represented here are completely independent of each other, and consequently they can be implemented simultaneously By using the coupling factors described in Corollary 41, it is easy to couple the two limiting distributions s 1 and s 2 in order to produce the stationary distribution vector π for the larger periodic chain When these observations are joined with the results of 6 of this exposition pertaining to the spectrum splitting effects which stochastic complementation can afford, it will become even more apparent that stochastic complementation has the potential to become a valuable computational technique for use with multiprocessor machines 6 Nearly completely reducible chains and spectral properties Consider an m-state irreducible chain C and k subclasses Q 1, Q 2,, Q k which partition the state space The chain C as well as an associated transition matrix is considered to be nearly completely reducible 2 when the Q i s are only very weakly coupled together 2 Gantmacher s [7] terminology completely reducible and the associated phrase nearly completely reducible are adopted in this exposition Some authors use the alternate terminology nearly completely decomposable

STOCHASTIC COMLEMENTATION 17 In this case, the states can be ordered so as to make the transition matrix have a k-level partition 11 12 1k = 21 22 2k (61 k1 k2 kk in which the magnitude of each off-diagonal block ij is very small relative to 1 We begin by investigating the limiting behavior of the chain as the off-diagonal blocks in (61 tend toward zero To this end, allow the off-diagonal blocks to vary independently, and assume the diagonal block ii in each row depends continuously on the ij s (the off-diagonal blocks in the corresponding row while always maintaining the irreducible and stochastic structure of the larger matrix The following result gives an indication of why stochastic complementation is useful in understanding the nature of a nearly completely reducible chain For vectors x and matrices A, the norms defined by x = max i x i and A = max i a ij will be employed throughout Theorem 61 If is an irreducible stochastic matrix with a k-level partition as indicated in (61, and if j S ii = ii + i (I i 1 i, i =1, 2,,k are the associated stochastic complements, then (62 S ii ii = i, and (63 lim S ii = ii i 0 Moreover, if S is the completely reducible stochastic matrix then S = S 11 0 0 0 S 22 0 0 0 S kk, (64 S = 2 max i i roof Let e denote a column of 1 s whose size is determined by the context in which it appears, and begin by observing that all matrices which are involved are nonnegative so that i (I i 1 i = i (I i 1 i e and i = i e

18 C D MEYER Since is stochastic, it must be the case that which in turn implies Therefore, i e + i e = e, (I i 1 i e = e S ii ii = i (I i 1 i = i (I i 1 i e = i e = i, which is the desired conclusion of (62, and the statement (63 that S ii ii as i 0 follows To establish (64, use the fact that ii S ii 0 in order to write (65 S = max ( i1 i2 ii S ii ik i = max i1 e + i2 e + +(S ii ii e + + ik e i Now observe that ii e + i e = e implies (S ii ii e = e ii e = i e, and use this in (65 together with the fact that to produce k ij e = i e j=1 j i S = max i1 e + + i e + + ik e i = max i If δ denotes the expression 2 i = 2 max i i δ = 2 max i i, then it is clear that 0 δ 2, and is completely reducible if and only if δ =0 This together with the result (64 motivates the following terminology Definition 61 For an m m irreducible stochastic matrix with a k-level partition 11 12 1k = 21 22 2k, k1 k2 kk the number δ = 2 max i i is called the deviation from complete reducibility

STOCHASTIC COMLEMENTATION 19 It is worthwhile to explore the spectral properties of stochastic complementation as the deviation from complete reducibility tends toward 0 If is nearly completely reducible, then each ii is nearly stochastic (in particular, Theorem 61 guarantees ii S ii, and continuity of the eigenvalues insures that must necessarily have at least k 1 nonunit eigenvalues clustered near the simple eigenvalue λ = 1 This means that is necessarily badly conditioned in the sense that if the sequence { n } n=1 converges, then it must converge very slowly For the time being, suppose has exactly k 1 nonunit eigenvalues clustered near λ = 1 Since each stochastic complement S ii is an irreducible stochastic matrix, the erron Frobenius theorem guarantees that the unit eigenvalue of each S ii is simple By virtue of Theorem 61 and continuity of the eigenvalues, it follows that the nonunit eigenvalues of each S ii must necessarily be rather far removed from the unit eigenvalue of S ii otherwise the spectrum σ( of would contain a cluster of at least k nonunit eigenvalues positioned near λ =1 In other words, S as δ 0, and the cluster consisting of the k 1 nonunit eigenvalues together with λ = 1 itself in σ( must split and map to the k unit eigenvalues in σ(s one unit eigenvalue in each σ(s ii This splitting effect is pictorially illustrated below in Fig 1 for a 12 12 matrix with the three-level partition = 11 12 13 21 22 23 and S = S 11 0 0 0 S 22 0 31 32 33 0 0 S 33 ( * * * * * * * * * * * * * * * * * * * * * * (S 11 (S 22 (S 33 Fig 1 The conclusion to be derived from this discussion is summarized in the following statement Theorem 62 If the underlying transition matrix 11 12 1k = 21 22 2k k1 k2 kk of a nearly completely reducible Markov chain has exactly k 1 nonunit eigenvalues clustered near λ =1, then the process of uncoupling the chain into k smaller chains by

20 C D MEYER using the method of stochastic complementation forces the spectrum of to naturally split apart so that each of the smaller chains has a well-conditioned transition matrix S ii in the sense that each S ii has no eigenvalues near its unit eigenvalue When the states in a nearly completely reducible chain are naturally ordered, and when the transition matrix is partitioned in the natural way according to the closely coupled subclasses, then the desirable spectrum splitting effect described above almost always occurs However, there are pathological cases when this effect is not achieved If, for a k-level partition of, the spectrum σ( contains more than k 1 nonunit eigenvalues clustered near λ = 1, then continuity dictates that some S ii must necessarily have a nonunit eigenvalue near its unit eigenvalue This can occur even for naturally partitioned matrices For example, consider 11 = 12 21 22 in which both 12 and 21 are extremely small in magnitude, and assume neither 11 nor 22 are nearly completely reducible Furthermore, suppose that 11 is a very small perturbation of 0 1 0 0 0 0 1 0 C n n = 0 0 0 1 1 0 0 0 where n is quite large Since σ(c consists of the nth roots of unity ω h = e (2πi/nh, h =0, 1,,n 1, it is clear that ω 1 = e 2πi/n must be very close to ω 0 = 1 whenever n is large Therefore, 11 must have a nonunit eigenvalue near λ = 1 and, by virtue of Theorem 61, the same must hold for S 11 thereby making S 11 badly conditioned Needless to say, pathological cases of this nature are rare in practical work As illustrated with the matrix given in (51, it is possible for stochastic complementation to uncouple a periodic chain into two aperiodic chains, thereby allowing the straightforward power method to be used where it ordinarily would not be applicable A similar situation can hold for nearly completely reducible chains The standard power method (66 π n+1 = π n where π 0 is arbitrary applied to the transition matrix of an aperiodic chain which is nearly completely reducible will fail not in theory, but in practice because the existence of eigenvalues of which are near λ = 1 causes (66 to converge too slowly However, if the chain is not pathological in nature (ie, if does not have more than k 1 eigenvalues near λ = 1, then the spectrum splitting effect described earlier insures that each of the k sequences (67 s (n+1 i = s (n i S ii where s (0 i is arbitrary, i =1, 2 k will exhibit rapid convergence Recall from Theorems 51 and 52 that the i th sequence in (67 converges if and only if ii is primitive, and this is guaranteed if ii has at

STOCHASTIC COMLEMENTATION 21 least one nonzero diagonal entry Again take note of the fact that these k sequences in (67 are completely independent of each other, so each iteration can be executed in parallel after which the limiting vectors s i = lim n s(n i are easily coupled according to the rules of Theorem 41 in order to produce the stationary distribution vector for as π =(ξ 1 s 1 ξ 2 s 2 ξ k s k 7 The Simon Ando theory for nearly completely reducible systems Simon and Ando [18] provided the classical theory for nearly completely reducible systems, and Courtois [3] (along with others who followed his pioneering work applied the theory and helped develop numerical aspects associated with queueing networks The contribution of Simon and Ando [18] was to provide mathematical arguments for what had previously been a rather heuristic theory concerning nearly completely reducible systems The major conclusion of this theory is that if the off-diagonal blocks ij in the transition matrix 11 12 1k = 21 22 2k k1 k2 kk of a finite aperiodic chain have sufficiently small magnitudes, then closely coupled subclasses associated with the diagonal blocks ii must each tend toward a local equilibrium long before a global equilibrium for the entire system is attained In the short run, the chain behaves as though it were completely reducible with each subclass tending toward a local equilibrium independent of the evolution in other subclasses After this short-run stabilization, the local equilibria are approximately maintained in a relative sense within each subclass while the entire chain evolves toward its global equilibrium The approach and substance of the theorems of Simon and Ando have become accepted as the theoretical basis for aggregation techniques and algorithms Although the conclusions of Simon and Ando [18] are extremely important, their mathematical development utilizes some rather cumbersome notation and proofs At times, the arguments are difficult to appreciate, and they do not fully illuminate the basic underlying mechanisms This is corroborated by the fact that although the conclusions of the Simon Ando theory are fundamental to the development of material in his text, Courtois [3] chooses to omit the Simon Ando proofs on the grounds that they have little relevance to subsequent developments The purpose of the latter portion of this exposition is to apply the concept of stochastic complementation in order to develop the theory for nearly completely reducible systems in a unified, clear, and simple manner while simultaneously sharpening some results and generalizing others The discussion is divided into three parts the short-run dynamics, the middle-run dynamics, and the long-run dynamics 8 Short-run dynamics The object of the short-run analysis is to examine the behavior of the distribution π n = π 0 n where π 0 is arbitrary,

22 C D MEYER for smaller values of n The following theorem provides the basis for doing this Theorem 81 For an m m irreducible stochastic matrix with a k-level partition 11 12 1k = 21 22 2k, k1 k2 kk let S be the completely reducible stochastic matrix S 11 0 0 0 S S = 22 0, 0 0 S kk and assume that each stochastic complement S ii is primitive 3 Let the eigenvalues of S m m be ordered as λ 1 = λ 2 = = λ k =1> λ k+1 λ k+2 λ m, and assume that S is similar to a diagonal matrix 4 Z 1 SZ = Ik k 0 0 D λ k+1 0 where D = 0 λ m If S denotes the limit es 1 0 0 0 es S = lim n Sn = 2 0 0 0 es k where s i is the stationary distribution vector for S ii, then (81 n S nδ + κ λ k+1 n where δ is the deviation from complete reducibility as defined in Definition 61, and where κ is the constant κ = Z Z 1 Moreover, if the (row vector norm defined by r 1 = i r i is used, then for every n =1, 2,, the difference between π n and the limit s = lim π 0S n n can be measured by (82 π n s 1 nδ + κ λ k+1 n 3 The primitivity assumption is included here for clarity of exposition because it insures that limits exist in the strong sense Although Theorems 51 and 52 show that almost all stochastic complements are primitive, it will later be argued that primitivity is not needed to reach the same general conclusions of this theorem 4 This assumption is also for clarity of exposition it will later be indicated how the results of this theorem can be preserved without this diagonalizability assumption

STOCHASTIC COMLEMENTATION 23 roof Begin by observing that the identity n S n = S n 1 ( S+S n 2 ( S + + S( S n 2 +( S n 1 is valid for all n =1, 2, Use this together with (64 and the fact that j = S j = 1 for every j in order to conclude that (83 By writing n S n nδ S n I 0 = Z 0 D n Z 1 I 0 and lim n Sn = Z Z 1, 0 0 it is clear that S n S 0 0 = Z 0 D n Z 1 Taking norms produces (84 S n S Z λ k+1 n Z 1 = κ λ k+1 n Now use (84 together with (83 to write n S = n S n +S n S n S n + S n S nδ +κ λ k+1 n, which is the desired conclusion (81 The second conclusion (82 follows by noting that the inequality ra 1 r 1 A holds for row vectors r and square matrices A so that π n s 1 = π 0 ( n S 1 n S nδ + κ λ k+1 n The results of Theorem 81 motivate the following definition Definition 81 For each ɛ>0, there is an associated short-run stabilization interval I(ɛ which is defined to be the set { } I(ɛ = n n S <ɛ The set E(ɛ = { n } δn + κ λ k+1 n <ɛ is referred to as the estimated short-run stabilization interval, and the function defined by f(x =δx + κ λ k+1 x is called the estimating function Notice that Theorem 81 insures E(ɛ I(ɛ

24 C D MEYER Examining the characteristics of the graph of the estimating function f(x provides an indication of the nature of the short-run stabilization interval I(ɛ Refer to Fig 2 and notice that as long as ɛ is greater than the minimum value of f, which occurs at δ ln κ ln λ k+1 x min =, ln λ k+1 the estimated short-run stabilization interval E(ɛ is nonempty Because f(x is asymptotic to the line y = δx, the length of the interval E(ɛ increases as δ 0, ie, E(ɛ grows as the ij s 0 in the underlying partitioned matrix y = f(x =δx + κ λ k+1 x y = ɛ y = t δ Fig 2 To appreciate the relationship between the short-run stabilization interval I(ɛ and its estimate E(ɛ, consider the following example in which is the nearly completely reducible matrix E(ɛ (85 6999 3000 0001 0 2000 7996 0 0004 = 0003 0 7997 2000 0 0007 1000 8993 Let g(n be the function defined by g(n = n S, and let f(n be the estimating function f(n =δn + κ λ k+1 n

STOCHASTIC COMLEMENTATION 25 For the matrix (85, δ = 0014, κ = 8657, and λ k+1 = 6996 (rounded to four significant digits The graph of g(n, shown in Fig 3, illustrates how rapidly powers of initially approach the matrix S and then how n gradually pulls away from S 03 02 f(n g(n 01 004 00 0 10 20 30 40 50 n Fig 3 For this example (which is not pathological, the graphs in Fig 3 show that the estimating function f(n gives a good indication of the evolution of the process in the sense that f(n is essentially parallel with g(n That is, f(n decreases, stabilizes, and then increases more or less in concert with g(n Although necessarily conservative, the estimating function f(n is not overly pessimistic in estimating the length of the short-run stabilization period For example, if ɛ is taken to be ɛ = 04, then as depicted in Fig 3 the associated short-run stabilization interval is I(04 = [11, 35] while the estimated short-run stabilization interval is E(04 = [18, 28] Corollary 81 If, for a given ɛ>0, n lies in the interval defined by (86 then ln ɛ/2κ ln λ k+1 <n< ɛ 2δ, n S <ɛ, and for every π 0, π n s 1 <ɛ

26 C D MEYER More precisely, { n ln ɛ/2κ ln λ k+1 <n< ɛ } E(ɛ I(ɛ 2δ roof The fact that λ k+1 < 1 means ln λ k+1 < 0 so that the left-hand inequality in (86 implies ln λ k+1 n < ln ɛ/2κ, and hence κ λ k+1 n <ɛ/2 Similarly, the right-hand inequality in (86 says that nδ < ɛ/2, and the desired conclusions follow from Theorem 81 To understand the significance of a short-run stabilization interval, suppose there are m i states in the ith subclass Q i (ie, ii is m i m i, and observe that if s i is the stationary distribution vector for S ii, then for an initial distribution π 0 partitioned as π 0 = ( π (1 0 π (2 0 π (k 0 in which π (i 0 contains m i components, the limiting distribution s is given by es 1 0 0 0 es s = π 0 S = π 0 2 0 =(ν 1s 1 ν 2 s 2 ν k s k 0 0 es k where the coefficient ν i associated with the ith subclass is given by mi ν i = π (i 0 e = Suppose that n belongs to a short-run stabilization interval I(ɛ Roughly speaking, this means that n simultaneously satisfies the two conditions h=1 ( π (i 0 h (87 λ k+1 n << 1 and n<< 1 δ, and for such values of n, Theorem 81 guarantees that π n s =(ν 1 s 1 ν 2 s 2 ν k s k where the ν i s are constant they depend only on π 0 Therefore, each closely coupled subclass Q i will approximately be in a local equilibrium which is defined by the stationary distribution vector s i as long as n satisfies (87 In other words, the chain initially evolves in such a way that each closely coupled subclass begins to approach a short-run equilibrium defined by the s i s completely independent of the evolution in all other subclasses As the chain continues to evolve, there will exist a period (the short-run stabilization interval I in which each subclass and hence the entire chain is approximately stable As time proceeds beyond I, the chain will move away from s, the approximate point of short-run stability, and it will evolve along a path en route toward global stability

STOCHASTIC COMLEMENTATION 27 The existence of an approximate short-run local equilibrium is a major conclusion of the classical Simon Ando theory, but the description of the short-run stabilization interval and the estimate given in Definition 81 or (87 is new The fact that the approximate short-run local equilibrium can be described by the stationary distribution vectors of the individual stochastic complements also appears to be new The preceding remarks are formally summarized in the following theorem Theorem 82 (Short-run dynamics For an irreducible stochastic matrix with a k-level partition 11 12 1k = 21 22 2k k1 k2 kk in which the stochastic complements (S ii mi m i are each primitive and diagonalizable, let π 0 = ( π (1 0 π (2 0 π (k 0 be an arbitrary initial distribution, and let { } k λ k+1 = max λ λ σ(s ii λ 1 If the magnitudes of the off-diagonal blocks ij are small enough to guarantee the existence of values of n such that i=1 (88 λ k+1 n << 1 and n<< 1 δ, then as long as n satisfies (88, it must be the case that and S 11 0 0 0 S n 22 0 0 0 S kk = es 1 0 0 0 es 2 0 0 0 es k π n = π 0 n ( ν 1 s 1 ν 2 s 2 ν k s k where s i is the stationary distribution vector for S ii, and where mi ν i = π (i 0 e = Before turning to the middle-run and long-run dynamics, observe that the assumption of Theorems 81 and 82 that S is diagonalizable is not necessary it merely makes the conclusions easier to grasp In the more general case, there exists a nonsingular matrix Z such that Z 1 Ik k 0 SZ = 0 J h=1 ( π (i 0 h

28 C D MEYER where J is in Jordan canonical form, and the spectral radius of J is ρ(j = λ k+1 It is well known (see [19] or [11] that for every square matrix A and each ɛ>0, there exists a matrix norm such that ρ(a A ρ(a+ɛ Clearly, this norm can be made to be meaningful for row vectors r simply by concatenating enough zero rows to r so as to fill out a square matrix This means that for every n and for each ɛ>0, there is a norm such that and (82 can be replaced by λ k+1 n J n λ k+1 n + ɛ, (89 π n s nδ + κ ( λ k+1 n + ɛ where κ = π 0 Z Z 1 Although the analysis is more involved, (89 used in place of (82 leads to the same general conclusions given in Theorems 81 and 82 9 Middle-run dynamics According to the short-run dynamics, n S and π n s =(ν 1 s 1 ν 2 s 2 ν k s k as long as n belongs to a short-run stabilization interval characterized by (88 As the chain continues to evolve, and as n advances beyond the short-run stabilization interval, n moves away from S and tends toward (assuming is primitive, while the distribution π n diverges away from s and moves along a path en route to the global stationary distribution vector π However, after the short-run stabilization period has ended, components of π n belonging to the same subclass Q i will nevertheless continue to remain in approximate equilibrium relative to each other The following theorems precisely articulate the sense in which this occurs Theorem 91 For an m m irreducible stochastic matrix 11 12 1k = 21 22 2k k1 k2 kk with stochastic complements S ii which are each primitive, let s i be the stationary distribution vector for S ii, and let π n = π 0 n where π 0 is an arbitrary initial distribution If ɛ>0 is a number such that the associated short-run stabilization interval I(ɛ is nonempty, then for each integer n beyond I(ɛ, there exist scalars β i (which vary with n such that the vector satisfies the inequality v n =(β 1 s 1 β 2 s 2 β k s k π n v n 1 <ɛ