Comparison of perturbation bounds for the stationary distribution of a Markov chain

Similar documents
SENSITIVITY OF THE STATIONARY DISTRIBUTION OF A MARKOV CHAIN*

The Kemeny Constant For Finite Homogeneous Ergodic Markov Chains

On The Inverse Mean First Passage Matrix Problem And The Inverse M Matrix Problem

}, (n 0) be a finite irreducible, discrete time MC. Let S = {1, 2,, m} be its state space. Let P = [p ij. ] be the transition matrix of the MC.

The effect on the algebraic connectivity of a tree by grafting or collapsing of edges

Linear Algebra and its Applications

Yimin Wei a,b,,1, Xiezhang Li c,2, Fanbin Bu d, Fuzhen Zhang e. Abstract

Perron eigenvector of the Tsetlin matrix

RITZ VALUE BOUNDS THAT EXPLOIT QUASI-SPARSITY

for an m-state homogeneous irreducible Markov chain with transition probability matrix

Another algorithm for nonnegative matrices

Intrinsic products and factorizations of matrices

Perturbation results for nearly uncoupled Markov. chains with applications to iterative methods. Jesse L. Barlow. December 9, 1992.

CONVERGENCE ANALYSIS OF A PAGERANK UPDATING ALGORITHM BY LANGVILLE AND MEYER

Analysis of Google s PageRank

On some matrices related to a tree with attached graphs

The Second Eigenvalue of the Google Matrix

Ann. Funct. Anal. 5 (2014), no. 2, A nnals of F unctional A nalysis ISSN: (electronic) URL:

Censoring Technique in Studying Block-Structured Markov Chains

CONDITIONING PROPERTIES OF THE STATIONARY DISTRIBUTION FOR A MARKOV CHAIN STEVE KIRKLAND

Lecture 11: Introduction to Markov Chains. Copyright G. Caire (Sample Lectures) 321

On Distributed Coordination of Mobile Agents with Changing Nearest Neighbors

Harnack s theorem for harmonic compact operator-valued functions

ELA

Conditioning of the Entries in the Stationary Vector of a Google-Type Matrix. Steve Kirkland University of Regina

The Distribution of Mixing Times in Markov Chains

Linear estimation in models based on a graph

Markov Chains and Spectral Clustering

Mathematical Properties & Analysis of Google s PageRank

SOMEWHAT STOCHASTIC MATRICES

Interval solutions for interval algebraic equations

Utilizing Network Structure to Accelerate Markov Chain Monte Carlo Algorithms

Using Markov Chains To Model Human Migration in a Network Equilibrium Framework

Linear Algebra and its Applications

Minimum number of non-zero-entries in a 7 7 stable matrix

A Cycle-Based Bound for Subdominant Eigenvalues of Stochastic Matrices

Spectrally arbitrary star sign patterns

Bounds for Levinger s function of nonnegative almost skew-symmetric matrices

Minimizing the Laplacian eigenvalues for trees with given domination number

Updating PageRank. Amy Langville Carl Meyer

SENSITIVITY OF HIDDEN MARKOV MODELS

Modified Gauss Seidel type methods and Jacobi type methods for Z-matrices

On the eigenvalues of specially low-rank perturbed matrices

CLASSIFICATION OF TREES EACH OF WHOSE ASSOCIATED ACYCLIC MATRICES WITH DISTINCT DIAGONAL ENTRIES HAS DISTINCT EIGENVALUES

Some additive results on Drazin inverse

Calculating Web Page Authority Using the PageRank Algorithm

Absolute value equations

Subset selection for matrices

Updating Markov Chains Carl Meyer Amy Langville

On factor width and symmetric H -matrices

Two Characterizations of Matrices with the Perron-Frobenius Property

Inequalities for the spectra of symmetric doubly stochastic matrices

Some bounds for the spectral radius of the Hadamard product of matrices

A note on estimates for the spectral radius of a nonnegative matrix

A property concerning the Hadamard powers of inverse M-matrices

OVERCOMING INSTABILITY IN COMPUTING THE FUNDAMENTAL MATRIX FOR A MARKOV CHAIN

Applied Mathematics Letters. Stationary distribution, ergodicity and extinction of a stochastic generalized logistic system

On Finding Optimal Policies for Markovian Decision Processes Using Simulation

Finite-Horizon Statistics for Markov chains

SIMILAR MARKOV CHAINS

An estimation of the spectral radius of a product of block matrices

Spectral Properties of Matrix Polynomials in the Max Algebra

Graphs with given diameter maximizing the spectral radius van Dam, Edwin

Properties for the Perron complement of three known subclasses of H-matrices

OPTIMAL SCALING FOR P -NORMS AND COMPONENTWISE DISTANCE TO SINGULARITY

Series Expansions in Queues with Server

Tridiagonal test matrices for eigenvalue computations: two-parameter extensions of the Clement matrix

The Laplacian spectrum of a mixed graph

STOCHASTIC COMPLEMENTATION, UNCOUPLING MARKOV CHAINS, AND THE THEORY OF NEARLY REDUCIBLE SYSTEMS

ELA THE MINIMUM-NORM LEAST-SQUARES SOLUTION OF A LINEAR SYSTEM AND SYMMETRIC RANK-ONE UPDATES

Hierarchical open Leontief models

A Note on Eigenvalues of Perturbed Hermitian Matrices

The Kemeny constant of a Markov chain

Key words. Feedback shift registers, Markov chains, stochastic matrices, rapid mixing

Majorizations for the Eigenvectors of Graph-Adjacency Matrices: A Tool for Complex Network Design

Diagonal matrix solutions of a discrete-time Lyapunov inequality

Linear Algebra and its Applications

On the second Laplacian eigenvalues of trees of odd order

Two Results About The Matrix Exponential

A Geometric Interpretation of the Metropolis Hastings Algorithm

On matrix equations X ± A X 2 A = I

Linear Algebra and its Applications

The minimum rank of matrices and the equivalence class graph

QUALITATIVE CONTROLLABILITY AND UNCONTROLLABILITY BY A SINGLE ENTRY

On the Perturbation of the Q-factor of the QR Factorization

Kernels of Directed Graph Laplacians. J. S. Caughman and J.J.P. Veerman

1.3 Convergence of Regular Markov Chains

MATH36001 Perron Frobenius Theory 2015

The maximal determinant and subdeterminants of ±1 matrices

Appendix A: Matrices

The skew-symmetric orthogonal solutions of the matrix equation AX = B

= P{X 0. = i} (1) If the MC has stationary transition probabilities then, = i} = P{X n+1

Cospectral graphs and the generalized adjacency matrix

Chapter 16 focused on decision making in the face of uncertainty about one future

Classical Complexity and Fixed-Parameter Tractability of Simultaneous Consecutive Ones Submatrix & Editing Problems

6.842 Randomness and Computation March 3, Lecture 8

The Optimal Stopping of Markov Chain and Recursive Solution of Poisson and Bellman Equations

RESEARCH ARTICLE. An extension of the polytope of doubly stochastic matrices

Persistence and global stability in discrete models of Lotka Volterra type

Linear Algebra and its Applications

Transcription:

Linear Algebra and its Applications 335 (00) 37 50 www.elsevier.com/locate/laa Comparison of perturbation bounds for the stationary distribution of a Markov chain Grace E. Cho a, Carl D. Meyer b,, a Mathematics and Engineering Analysis, The Boeing Company, P.O. Box 3707, MC 7L-, Seattle, WA 984-07, USA b Department of Mathematics, North Carolina State University, P.O. Box 805, Raleigh, NC 7695-805, USA Received 8 November 999; accepted November 000 Submitted by V. Mehrmann Abstract The purpose of this paper is to review and compare the existing perturbation bounds for the stationary distribution of a finite, irreducible, homogeneous Markov chain. 00 Elsevier Science Inc. All rights reserved. AMS classification: 65F35; 60J0; 5A5; 5A; 5A8 Keywords: Markov chains; Stationary distribution; Stochastic matrix; Group inversion; Sensitivity analysis; Perturbation theory; Condition numbers. Introduction Let P be the transition probability matrix of an n state finite, irreducible, homogeneous Markov chain. The stationary distribution vector of P is the unique positive vector π T satisfying π T P = π T, n π =. = Corresponding author. Tel.: +-99-55-384. E-mail addresses: grace.e.cho@boeing.com (G.E. Cho), meyer@math.ncsu.edu (C.D. Meyer). URL: http://www4.ncsu.edu/ ehcho/home.htmlb (G.E. Cho), http://meyer.math.ncsu.edu/ (C.D. Meyer). Supported in part by National Science Foundation grants CCR-973856 and DMS-9748. 004-3795/0/$ - see front matter 00 Elsevier Science Inc. All rights reserved. PII:S004-3795(0)0030-

38 G.E. Cho, C.D. Meyer / Linear Algebra and its Applications 335 (00) 37 50 Suppose P is perturbed to a matrix P, that is the transition probability matrix of an n state finite, irreducible, homogeneous Markov chain as well. Denoting the stationary distribution vector of P by π, the goal is to describe the change π π in the stationary distribution in terms of the change E P P in the transition probability matrix. For suitable norms, π π κ E for various different condition numbers κ. We review eight existing condition numbers κ,...,κ 8. Most of the condition numbers we consider are expressed in terms of either the fundamental matrix of the underlying Markov chain or the group inverse of I P. The condition number κ 8 is expressed in terms of mean first passage times, providing a qualitative interpretation of error bound. In Section 4, we compare the condition numbers.. Notation Throughout the article the matrix P denotes the transition probability matrix of an n state finite, irreducible, homogeneous Markov chain C and π denotes the stationary distribution vector. Then π T P = π T, π > 0, π T e =, where e is the column vector of all ones. The perturbed matrix P = P E is the transition probability matrix of another n state finite, irreducible, homogeneous Markov chain C with the stationary distribution vector π: π T P = π T, π >0, π T e =. The identity matrix of size n is denoted by I. For a matrix B,the(i, ) component of B is denoted by b i. The -norm v of a vector v is the absolute entry sum, and the -norm B of a matrix B is its maximum absolute row sum. 3. Condition numbers of a Markov chain The norm-wise perturbation bounds we review in this section are of the following form: π π p κ l E q, where (p, q) = (, ) or (, ), depending on l.

G.E. Cho, C.D. Meyer / Linear Algebra and its Applications 335 (00) 37 50 39 Most of the perturbation bounds we will consider are in terms of one the two matrices related to the chain C: thefundamental matrix and the group inverse of A I P.Thefundamental matrix of the chain C is defined by Z (A + eπ T ). The group inverse of A is the unique square matrix A # satisfying AA # A = A, A # AA # = A #, AA # = A # A. The lists of the condition numbers κ l and the references are as follows: κ = Z Schweitzer [7] κ = A # Meyer [] κ 3 = max (a # min i a i # ) Haviv and van Heyden [5] Kirkland et al. [9] κ 4 = max i, a i # Funderlic and Meyer [4] κ 5 = τ (P ) Seneta [] κ 6 = τ (A # ) = τ (Z) Seneta [] κ 7 = min A () Ipsen and Meyer [6] [ ] Kirkland et al. [9] κ 8 = max maxi/= m i Cho and Meyer [] m where m i,i/=, is the mean first passage time from state i to state, andm is the the mean return time state. 3.. Schweitzer [7] Kemeny and Snell [8] call Z the fundamental matrix of the chain because most of the questions concerning the chain can be answered in terms of Z. For instance, the stationary distribution vector of the perturbed matrix P can be expressed in terms of π and Z: π T = π T (I + EZ) and π T π T = π T EZ. (3.) Eq. (3.), by Schweitzer [7], gives the first perturbation bound: π π Z E.

40 G.E. Cho, C.D. Meyer / Linear Algebra and its Applications 335 (00) 37 50 We define κ Z. 3.. Meyer [] The second matrix related to C is the group inverse of A. In his papers [,], Meyer showed that the group inverse A # can be used in a similar way Z is used, and, since Z = A # + eπ T, all relevant information is contained in A #, and the term eπ T is redundant [4]. In fact, in the place of (3.), we have π T = π T (I + EA # ) and π T π T = π T EA #, (3.) and the resulting perturbation bound is [] π π A # E. We define the second condition number κ A #. 3.3. Haviv and van Heyden [5] and Kirkland et al. [9] The perturbation bound in this section is derived from (3.) with a use of the following lemma: Lemma 3.. For any vector d and for any vector c such that c T e = 0, c T max i, d i d d c. The resulting perturbation bound is π π max (a # min i a i # ) (cf. Refs. [5,9].) We define κ 3 max (a # min i a i # ). E.

G.E. Cho, C.D. Meyer / Linear Algebra and its Applications 335 (00) 37 50 4 3.4. Funderlic and Meyer [4] Eq. (3.) provides a component-wise bound for the stationary distribution vector: π π max i a # i E, whichleadsusto π π max i, a # i E. Funderlic and Meyer [4] called the number κ 4 max i, a # i the chain condition number. The behaviour of κ 4 is provided by Meyer [3,4]. He showed that the size of κ 4 is primarily governed by how close the subdominant eigenvalues of the chain are to. (This is not true for arbitrary matrices. For an example, see [4, p.76].) To be precise, denoting the eigenvalues of P by, λ,...,λ n, the lower bound and the upper bound of κ 4 are given by max a # (n ) i < n min i λ i i, i ( λ i). (3.3) Hence, if the chain is well-conditioned, then all subdominant eigenvalues must be well separated from, and if all subdominant eigenvalues are well separated from, then the chain must be well-conditioned. In [3,4], it is indicated that the upper bound (n )/ i ( λ i) in (3.3) is a rather conservative estimate of κ 4. If no single eigenvalue of P is extremely close to, but enough eigenvalues are within range of, then (n )/ i ( λ i) is large, even if the chain is not too badly condition. Seneta [3] provides a condition number and its bounds to overcome this problem. (See Section 3.6.) 3.5. Seneta [] The condition numbers κ, κ,andκ 4 are in terms of matrix norms. Seneta [], however, proposed the ergodicity coefficient instead of the matrix norm. The ergodicity coefficient τ (B) of a matrix B with equal row sums b is defined by τ (B) sup v T B. v = v T e=0

4 G.E. Cho, C.D. Meyer / Linear Algebra and its Applications 335 (00) 37 50 Note that τ (B) is an ordinary norm of B on the hyperspace H n {v : v R n,v T e = 0} of R n. eigenvalues are to b. For more discussion and study, we refer readers to [,3,7,0,5,6,8 3,5,6]. For the stochastic matrix P, the ergodicity coefficient satisfies 0 τ (P ). In case of τ (P ) <, we have a perturbation bound in terms of the ergodicity coefficient of P []: π π τ (P ) E. (3.4) (For the case τ (P ) =, see [].) We denote κ 5 τ (P ). 3.6. Seneta [] In the previous parts we noted that the group inverse A # of A = I P can be used in place of Kemeny and Snell s fundamental matrix Z. In fact, if we use ergodicity coefficients as a measure of sensitivity of the stationary distribution, then Z and A # give exactly the same information: κ 6 τ (A # ) = τ (Z), which is the condition number in the perturbation bound given by Seneta []: π π τ (A # ) E (= τ (Z) E ). In Section 3.4, we observed that the size of the condition number κ 4 = max i, a # i is governed by the closeness of the subdominant eigenvalues of P to, giving the lower and upper bound for κ 4. However, the problem of overestimating the upper bound for κ 4 occurs if enough eigenvalues of P are within the range of, even if no single eigenvalue λ i of P is close to. The following bounds for κ 5 overcome this problem: min i λ i τ (A # ) i n λ i min i λ i. Unlike the upper bound (3.3) for κ 4, the far left upper bound for κ 6 takes only the closest eigenvalue to into account. Hence, it shows that as long as the closest eigenvalue of P is not close to, the chain is well-conditioned.

G.E. Cho, C.D. Meyer / Linear Algebra and its Applications 335 (00) 37 50 43 3.7. Ipsen and Meyer [6], and Kirkland et al. [9] Ipsen and Meyer [6] derived a set of perturbation bounds and showed that all stationary probabilities react in a uniform manner to perturbations in the transition probabilities. The main result of their paper is based on the following component-wise perturbation bounds: π π π π k π k min A () E, A () E, for all =,,...,n, for all k =,,...,n, (3.5) where A () is the principal submatrix of A obtained by deleting the th row and column from A. Kirkland et al. [9] improved the perturbation bound in (3.5) by a factor of : π π < min giving the condition number κ 7 min A () A (). E, 3.8. Cho and Meyer [] In the previous sections, we saw a number of perturbation bounds for the stationary distribution vector of an irreducible Markov chain. Unfortunately, they provide little qualitative information about the sensitivity of the underlying Markov chain. Moreover, the actual computation of the corresponding condition number is usually expensive relative to computation of the stationary distribution vector itself. In this section a perturbation bound is presented in terms of the structure of the underlying Markov chain. To be more precise, the condition number κ is in terms of mean first passage times. For an n-state irreducible Markov chain C, themean first passage time m i from state i to state ( /= i) is defined to be the expected number of steps to enter in state for the first time, starting in state i. Themean return time m of state is the expected number of steps to return to state for the first time, starting in state. Using the relationship m = π and a # i = a# π m i, i /=, we obtain the perturbation bound

44 G.E. Cho, C.D. Meyer / Linear Algebra and its Applications 335 (00) 37 50 π π [ ] max maxi/= m i E m (cf. []). We define κ 8 [ max maxi/= m i m ]. Viewing sensitivity in terms of mean first passage times can sometimes help practitioners decide whether or not to expect sensitivity in their Markov chain models merely by observing the structure of the chain without computing or estimating condition numbers. For example, consider chains consisting of a dominant central state with strong connections to and from all other states. Physical systems of this type have historically been called mammillary systems (see [4,7]). The simplest example of a mammillary Markov chain is one whose transition probability matrix has the form p 0 0 p P = 0 p 0 p........... 0 0 p k p k q q q k q (3.6) in which the p i s and q i s are not unduly small (say, p i >.5andq i /k). Mean first passage times m i, i =, in mammillary structures are never very large. For example, in the simple mammillary chain defined by (3.6) it can be demonstrated that p i when = k +, m i = +σ q p when i = k +, +σ q + p i p when i, = k +, where σ = k h= q h p h. Since each m, it is clear that κ 8 cannot be large, and consequently no stationary probability can be unduly sensitive to perturbations in P. It is apparent that similar remarks hold for more general mammillary structures as well. 4. Comparison of condition numbers In this section, we compare the condition numbers κ l. The norm-wise perturbation bounds in Section 3 are of the following form: π π p κ l E q,

G.E. Cho, C.D. Meyer / Linear Algebra and its Applications 335 (00) 37 50 45 where (p, q) = (, ),or(, ), depending on l.(forκ 4, a strict inequality holds.) The purpose of this section is simply to compare the condition numbers appearing in those norm-wise bounds. Therefore we use the form π π κ l E for all condition numbers although some of the bounds are tighter in this form. (See Remark 4..) Lemma 4.. (a) max max i,k a i # a# k =max (a # min i a i # ), (b) max (a # min i a i # ) (c) max i, a i # A#, max i, (d) max (a # min i a # i ) τ (A # ), max (a (e) τ (A # # max a i # ) i ) n, (f) τ (A # ) A #, τ (A # ) Z, τ (A # ) τ (P ), (g) A # Z A # +. a i # < max (a # min a i # ), i Proof. (a) Using a symmetric permutation, we may assume that a particular probability occurs in the last position of π. Partition A and π as follows: ( ) ( ) A(n) c π(n) A =, π =, π n d T a nn where A (n) is the principal matrix of A obtained by deleting the nth row and column from A. Since rank A = n, the relationship a nn = d T A (n) c holds, so that A # = (I eπt (n) )A (n) (I eπt (n) ) π n(i eπ(n) T )A (n) e. π(n) T A (n) (I eπt (n) ) π nπ(n) T A (n) e Hence π n e a in # = i TA (n) e + π nπ(n) T A (n) e, i /= n, (4.) π n π(n) T A (n) e, i = n,

46 G.E. Cho, C.D. Meyer / Linear Algebra and its Applications 335 (00) 37 50 where e i is the ith column of I. By M-matrix properties, it follows that A (n) > 0so that π n ei TA (n) e, π nπ(n) T A (n) e>0. Thus max a in # = i a# nn, and the result follows. (b) For each, a # > 0, by (4.). Since π T A # = 0 T, it follows that there exists k 0 such that a k # 0 < 0. Furthermore, let i 0 be such that max i a i # = a# i 0. Then a a i # 0 < i # 0 a# k 0 if a# i 0 0, a i # 0 a# otherwise. Thus, for all, max i so that a # i <max i,k a # i a# k max a i # <max max a i # i, i,k a# k. On the other hand, max max i,k a # i a# k max i, a # i +max k, a # k = max a i #. i, Now the assertion follows from (a). (c) It follows directly from the definitions of, and norm, and their relationship, (d) For a real number a, definea + max {a,0}. Then max max i,k a i # a# k =max i,k max i,k and the assertion follows by (a). (e) For any vector x R n, x n x, and the result follows, since max a i # a# k (a i # a# k )+ =τ (A # ) (by Seneta [9, p.39]),

G.E. Cho, C.D. Meyer / Linear Algebra and its Applications 335 (00) 37 50 47 τ (A # )= max i,k A# i A# k, and max max i,k a # i a# k = max i,k A# i A# k, where B i denotes the ith row of a matrix B. (f) Since for any matrix B with equal row sums, and τ (B)= sup v = v T e=0 v T B, y T B y B, for any vector y R n,wehave τ (B) B. The inequalities follow by this relationship together with the following facts, proven in []: τ (A # ) = τ (Z), and if τ (P ) <, then τ (A # ) τ (P ). (g) It follows by applying the triangle inequality to Z = A # + eπ T and A # = Z eπ T. The following list summarizes the relationship between the condition numbers: Relation between condition numbers κ 8 = κ 3 κ 4 < κ 3 κ 6 κ l for l =,, 5, κ 6 nκ 3, κ κ κ +.

48 G.E. Cho, C.D. Meyer / Linear Algebra and its Applications 335 (00) 37 50 Remark 4.. As remarked at the beginning of this section, some of the condition numbers provide tighter bounds than the form used in this section. To be more precise, for l =,, 5, 6, π π π π κ l. E E To give a fairer comparison to these condition numbers, note that (π π) T e = 0 implies π π (/) π π so that π π κ l E, where κ l = (/)κ l,forl =,, 5, 6. The comparison of these new condtion numbers κ l with κ 3 is as given above: κ 3 κ l, for l =,, 5, 6. 5. Concluding remarks We reviewed eight existing perturbation bounds, and the condition numbers are compared. The list at the end of the last section clarifies the relationships between seven condition numbers. Among these seven condition numbers, the smallest condition number is κ 3 max (a # min i a # i ) by Haviv and van Heyde [5] and Kirkland et al. [9] or equivalently, κ 8 [ ] max maxi/= m i m by Cho and Meyer []. The only condition number not included in the comparison is κ 7 = min A (). Is κ 3 a smaller condition number than κ 7? Since by (4.), a # min i a # i = π A (), The authors thank the referee for bringing this point to our attention.

G.E. Cho, C.D. Meyer / Linear Algebra and its Applications 335 (00) 37 50 49 for each, the question is whether max π A () min A () (5.) holds. In other words, is π always small enough so that max π A () is never larger than min A ()? This question leads us to the relationship between a stationary probability π and the corresponding A (). In a special case, the relationship is clear. Lemma 5.. If P is of rank, then A () = π, for all =,,...,n. The proof appears in [9, Observation 3.3]. This relationship for an arbitrary irreducible Markov chain does not hold. In general, however, A () tends to increase as π decreases, and vice versa. Notice that κ 8 is equal to κ 3. However, viewing sensitivity in terms of mean first passage times can sometimes help practitioner decide whether or not to expect sensitivity in their Markov chain models merely by observing the structure of the chain, thus obviating the need for computing or estimating condition numbers. References [] G. Cho, C. Meyer, Markov chain sensitivity measured by mean first passage times, Linear Algebra Appl. 36 (000) 8. [] R. Dobrushin, Central limit theorem for nonstationary Markov chains. I, Theory Probab. Appl. (956) 65 80. [3] R. Dobrushin, Central limit theorem for nonstationary Markov chains. II, Theory Probab. Appl. (956) 39 383. [4] R. Funderlic, C. Meyer, Sensitivity of the stationary distribution vector for an ergodic Markov chain. II, Linear Algebra Appl. 76 (986) 7. [5] M. Haviv, L. Van der Heyden, Perturbation bounds for the stationary probabilities of a finite Markov chain, Adv. Appl. Probab. 6 (984) 804 88. [6] I. Ipsen, C. Meyer, Uniform stability of Markov chains, SIAM J. Matrix Anal. Appl. 5 (994) 06 074. [7] D. Isaacson, R. Madsen, Markov Chains: Theory and Application, Wiley, New York, 976. [8] J. Kemeny, J. Snell, Finite Markov Chains, Van Nostrand, New York, 960. [9] S. Kirkland, M. Neumann, B. Shader, Applications of Paz s inequality to perturbation bounds for Markov chains, Linear Algebra Appl. 68 (998) 83 96. [0] A. Lešanovský, Coefficients of ergodicity generated by non-symmetrical vector norms, Czechoslovak Math. J. 40 (990) 84 94. [] C. Meyer, The role of the group generalized inverse in the theory of finite Markov chains, SIAM Rev. 7 (975) 443 464.

50 G.E. Cho, C.D. Meyer / Linear Algebra and its Applications 335 (00) 37 50 [] C. Meyer, The condition of a finite Markov chain and perturbation bounds for the limiting probabilities, SIAM J. Algebraic Discrete Methods (980) 73 83. [3] C. Meyer, The character of a finite Markov chain, in: C.D. Meyer, R.J. Plemmons (Eds.), Proceedings of the IMA Workshop on Linear Algebra, Markov Chains, and Queuing Models, Springer, Berlin, 993, pp. 47 58. [4] C. Meyer, Sensitivity of the stationary distribution of a Markov chain, SIAM J. Matrix Appl. 5 (994) 75 78. [5] A. Rhodius, On explicit forms for ergodic coefficients, Linear Algebra Appl. 94 (993) 7 83. [6] U. Rothblum, C. Tan, Upper bounds on the maximum modulus of subdominant eigenvlues of nonnegative matrices, Linear Algebra Appl. 66 (985) 45 86. [7] P. Schweitzer, Perturbation theory and finite Markov chains, J. Appl. Probab. 5 (968) 40 43. [8] E. Seneta, Coefficients of ergodicity: structure and applications, Adv. Appl. Probab. (979) 576 590. [9] E. Seneta, Non-negative Matrices and Markov Chains, Springer, New York, 98. [0] E. Seneta, Explicit forms for ergodicity coefficients and spectrum localization, Linear Algebra Appl. 60 (984) 87 97. [] E. Seneta, Perturbation of the stationary distribution measured by ergodicity coefficients, Adv. Appl. Probab. 0 (988) 8 30. [] E. Seneta, Sensitivity analysis, ergodicity coefficients, and rank-one updates for finite Markov chains, in: W.J. Stewart (Ed.), Numerical Solution of Markov Chains, Marcel Dekker, NY, 99, pp. 9. [3] E. Seneta, Sensitivity of finite Markov chains under perturbation, Statist. Probab. Lett. 7 (993) 63 68. [4] C. Sheppard, A. Householder, The mathematical basis of the interpretation of tracer experiments in closed steady-systems, J. Appl. Phys. (95) 50 50. [5] C. Tan, A functional form for a coefficient of ergodicity, J. Appl. Probab. 9 (98) 858 863. [6] C. Tan, Coefficients of ergodicity with respect to vector norms, J. Appl. Probab. 0 (983) 77 87. [7] R. Whittaker, Experiments with radisphosphorus tracer in aquarium microcosms, Ecological Monographs 3 (96) 57 88.