}, (n 0) be a finite irreducible, discrete time MC. Let S = {1, 2,, m} be its state space. Let P = [p ij. ] be the transition matrix of the MC.

Similar documents
The Role of Generalized Matrix Inverses in Markov Chains

Simple procedures for finding mean first passage times in Markov chains

Block designs and statistics

Chapter 6 1-D Continuous Groups

The Distribution of the Covariance Matrix for a Subset of Elliptical Distributions with Extension to Two Kurtosis Parameters

A note on the multiplication of sparse matrices

Generalized AOR Method for Solving System of Linear Equations. Davod Khojasteh Salkuyeh. Department of Mathematics, University of Mohaghegh Ardabili,

Comparison of perturbation bounds for the stationary distribution of a Markov chain

Physics 215 Winter The Density Matrix

ON THE TWO-LEVEL PRECONDITIONING IN LEAST SQUARES METHOD

The Fundamental Basis Theorem of Geometry from an algebraic point of view

Feature Extraction Techniques

The Distribution of Mixing Times in Markov Chains

13.2 Fully Polynomial Randomized Approximation Scheme for Permanent of Random 0-1 Matrices

Solutions of some selected problems of Homework 4

Study on Markov Alternative Renewal Reward. Process for VLSI Cell Partitioning

A Simplified Analytical Approach for Efficiency Evaluation of the Weaving Machines with Automatic Filling Repair

INNER CONSTRAINTS FOR A 3-D SURVEY NETWORK

A note on the realignment criterion

The Kemeny Constant For Finite Homogeneous Ergodic Markov Chains

Chaotic Coupled Map Lattices

RECOVERY OF A DENSITY FROM THE EIGENVALUES OF A NONHOMOGENEOUS MEMBRANE

Supplementary to Learning Discriminative Bayesian Networks from High-dimensional Continuous Neuroimaging Data

CSE525: Randomized Algorithms and Probabilistic Analysis May 16, Lecture 13

Slanted coupling of one-dimensional arrays of small tunnel junctions

Physics 139B Solutions to Homework Set 3 Fall 2009

Multi-Dimensional Hegselmann-Krause Dynamics

The Indefinite Admittance Matrix

NORMAL MATRIX POLYNOMIALS WITH NONSINGULAR LEADING COEFFICIENTS

Linear Transformations

Simulation of Discrete Event Systems

Standard & Canonical Forms

Topic 5a Introduction to Curve Fitting & Linear Regression

A Self-Organizing Model for Logical Regression Jerry Farlow 1 University of Maine. (1900 words)

Deflation of the I-O Series Some Technical Aspects. Giorgio Rampa University of Genoa April 2007

Reed-Muller Codes. m r inductive definition. Later, we shall explain how to construct Reed-Muller codes using the Kronecker product.

arxiv: v1 [math.pr] 17 May 2009

Stochastic vertex models and symmetric functions

Character analysis on linear elementary algebra with max-plus operation

A SIMPLE METHOD FOR FINDING THE INVERSE MATRIX OF VANDERMONDE MATRIX. E. A. Rawashdeh. 1. Introduction

DEPARTMENT OF ECONOMETRICS AND BUSINESS STATISTICS

A new type of lower bound for the largest eigenvalue of a symmetric matrix

The Simplex Method is Strongly Polynomial for the Markov Decision Problem with a Fixed Discount Rate

On a Multisection Style Binomial Summation Identity for Fibonacci Numbers

A Markov Framework for the Simple Genetic Algorithm

SENSITIVITY OF THE STATIONARY DISTRIBUTION OF A MARKOV CHAIN*

Introduction to Machine Learning. Recitation 11

+ -d-t-' )=1. = vpi. Aportaciones Matematicas Comunicaciones 17 (1996) 5-10.

Distributed Subgradient Methods for Multi-agent Optimization

COS 424: Interacting with Data. Written Exercises

Supporting Information for Supression of Auger Processes in Confined Structures

An Extension to the Tactical Planning Model for a Job Shop: Continuous-Time Control

On Lotka-Volterra Evolution Law

R. L. Ollerton University of Western Sydney, Penrith Campus DC1797, Australia

Pattern Recognition and Machine Learning. Artificial Neural networks

Lecture 13 Eigenvalue Problems

2 Q 10. Likewise, in case of multiple particles, the corresponding density in 2 must be averaged over all

Explicit solution of the polynomial least-squares approximation problem on Chebyshev extrema nodes

Polygonal Designs: Existence and Construction

AN APPLICATION OF CUBIC B-SPLINE FINITE ELEMENT METHOD FOR THE BURGERS EQUATION

Four-vector, Dirac spinor representation and Lorentz Transformations

Partial traces and entropy inequalities

STRONG LAW OF LARGE NUMBERS FOR SCALAR-NORMED SUMS OF ELEMENTS OF REGRESSIVE SEQUENCES OF RANDOM VARIABLES

Divisibility of Polynomials over Finite Fields and Combinatorial Applications

Pattern Recognition and Machine Learning. Artificial Neural networks

Support Vector Machine Classification of Uncertain and Imbalanced data using Robust Optimization

Polynomial Division By Convolution

NON-COMMUTATIVE GRÖBNER BASES FOR COMMUTATIVE ALGEBRAS

On The Inverse Mean First Passage Matrix Problem And The Inverse M Matrix Problem

Asynchronous Gossip Algorithms for Stochastic Optimization

Pseudo-marginal Metropolis-Hastings: a simple explanation and (partial) review of theory

Prepared with an evaluation version of PCT E X. Visit

3.3 Variational Characterization of Singular Values

The Hilbert Schmidt version of the commutator theorem for zero trace matrices

Intelligent Systems: Reasoning and Recognition. Artificial Neural Networks

Lectures 8 & 9: The Z-transform.

New upper bound for the B-spline basis condition number II. K. Scherer. Institut fur Angewandte Mathematik, Universitat Bonn, Bonn, Germany.

Data-Driven Imaging in Anisotropic Media

RESTARTED FULL ORTHOGONALIZATION METHOD FOR SHIFTED LINEAR SYSTEMS

Kinetic Theory of Gases: Elementary Ideas

Constrained Consensus and Optimization in Multi-Agent Networks arxiv: v2 [math.oc] 17 Dec 2008

International Scientific and Technological Conference EXTREME ROBOTICS October 8-9, 2015, Saint-Petersburg, Russia

AN ALGEBRAIC AND COMBINATORIAL APPROACH TO THE CONSTRUCTION OF EXPERIMENTAL DESIGNS

Pattern Recognition and Machine Learning. Artificial Neural networks

Kinetic Theory of Gases: Elementary Ideas

Perturbation results for nearly uncoupled Markov. chains with applications to iterative methods. Jesse L. Barlow. December 9, 1992.

= P{X 0. = i} (1) If the MC has stationary transition probabilities then, = i} = P{X n+1

G G G G G. Spec k G. G Spec k G G. G G m G. G Spec k. Spec k

Model Fitting. CURM Background Material, Fall 2014 Dr. Doreen De Leon

APPROXIMATION BY GENERALIZED FABER SERIES IN BERGMAN SPACES ON INFINITE DOMAINS WITH A QUASICONFORMAL BOUNDARY

On the summations involving Wigner rotation matrix elements

Vulnerability of MRD-Code-Based Universal Secure Error-Correcting Network Codes under Time-Varying Jamming Links

4 = (0.02) 3 13, = 0.25 because = 25. Simi-

Markov Processes. Fundamentals (1) State graph. Fundamentals (2) IEA Automation 2018

Feedforward Networks

STOCHASTIC PROCESSES Basic notions

Hee = ~ dxdy\jj+ (x) 'IJ+ (y) u (x- y) \jj (y) \jj (x), V, = ~ dx 'IJ+ (x) \jj (x) V (x), Hii = Z 2 ~ dx dy cp+ (x) cp+ (y) u (x- y) cp (y) cp (x),

Effective joint probabilistic data association using maximum a posteriori estimates of target states

The Solution of One-Phase Inverse Stefan Problem. by Homotopy Analysis Method

Measuring Temperature with a Silicon Diode

Transcription:

Abstract Questions are posed regarding the influence that the colun sus of the transition probabilities of a stochastic atrix (with row sus all one) have on the stationary distribution, the ean first passage ties and the Keeny constant of the associated irreducible discrete tie Markov chain. Soe new relationships, including soe inequalities, and partial answers to the questions, are given using a special generalized atrix inverse that has not previously been considered in the literature on Markov chains.

Outline 1. Introduction 2. Properties of the generalized inverse, H 3. Stationary distributions 4 Relationship between H and Z 5. Mean first passage ties 6. Keeny s constant 7. Exaples

Introduction Let {X n }, (n 0) be a finite irreducible, discrete tie MC. Let S = {1, 2,, } be its state space. Let P = [p ij ] be the transition atrix of the MC. {X n } has a unique stationary distribution { j }, j S. and finite ean first passage ties { ij }, (i, j) S S. P stochastic the row sus are all one. p j=1 ij = 1, i S. Let {c j } be the colun sus of the transition atrix. p i=1 ij c j =, j S.

Questions What influence does the sequence {c j } have on { j }? What influence does the sequence {c j } have on { ij }? Are there relationships connecting the {c j },{ j },{ ij }? Can we deduce bounds on the { j } and the { ij } involving the {c j }? What effect does the {c j } have on Keeny s constant K = j=1 j ij?

Technique We use the generalized atrix inverse H [ I P + ec T ] 1 where e is the colun vector of ones and c T = (c 1,c 2,..., c ) is the row vector of colun sus of P. Let = e T where T = ( 1, 2,..., ). H can also be expressed in ters of Keeny and Snell s fundaental atrix Z = [I P + ] 1, or in ters of Meyer s group g-inverse (I P) # = Z.

Previous results 1. Kirkland considers the subdoinant eigenvalue 2 associated with the set S(c) of stochastic atrices with colun su vector c T. The quantity (c) = ax{ 2 (A) A S(c)} is considered. The vectors c T such that (c) < 1 are identified. In those cases, nontrivial upper bounds on (c) and weak ergodocity results for forward products are provided.

Previous results 2. Kirkland considers an irreducible stochastic atrix P and studies the extent to which the colun su vector for P provides inforation on a certain condition nuber (P), which easures the sensitivity of the stationary distribution vector to perturbations in P.

Properties of H T t 0 and u T e 0 I P + tu T is non-singular. [I P + tu T ] 1 is a generalised inverse of I P. T e = 1 0 and c T e = 0 I P + ec T is non-singular. H = [I P + ec T ] 1 is a generalized inverse of I P.

Key properties of H If H = [I P + ec T ] 1 = [h ij ] then c T H = T. Thus j = c i=1 i h ij for all j S. Further He = e so that h ii h j=1 ij = 1 for all i S. Note also that c T He = 1.

Proof of key properties of H (I P + ec T )H = I H PH + ec T H = I. Preultiply by T. Since T P = T, c T H = T j = c i=1 i h ij for all j S. H(I P + ec T ) = I H HP + Hec T = I. Postultiply by e. Since Pe = e and c T e =, He = e h ii h j=1 ij = 1, for all i S.

Properties of the eleents of H Let e i T (e j ) be the i-th (j-th) eleentary row (col) vector. Let h j (c) He j denote the j-th colun of H. Let h i j = e T h j (c) be the su of the eleents of the j-th col. Let h i (r )T e i T H denote the i-th row of H. Let h ii = h i (r )T e be the su of the eleents of the i-th row. Let h rowsu = He = T Let h colsu h ij = e i T He j. (c) h j=1 j = [h 1i,h 2i,...,h i ] T, = e T H = (r )T h j = [h i1,h i2,...,h i ], j=1

Properties of the eleents of H (a) (Row properties) (r h )T (r i p )T i H = e T i T, (r h )T (r i h )T i P = e T i c T, and h ii = 1/. (b) (Colun properties) h (c) j Ph (c) j = e j j e, h (c) j Hp (c) j = e j (c j )e, and h i j = 1 ( 1) j.

Properties of the eleents of H (c) (Eleent properties) h ij = h ij = p ik h kj + ij j, k=1 h ik p kj + ij c j. k=1 (d) (Row and Colun su properties) h rowsu = e, T h colsu = e T ( 1) T. Explicit row and colun sus of the eleents of H. Explicit expressions for individual h ij not readily available.

Stationary distributions MC irreducible exists a unique stationary distribution { j }, j S. The stationary probabities found as the solution of the stationary equations: j = i p ij ( j S) with i = 1. i=1 We have shown j = i=1 c i h ij with c i =. i=1 i=1

Doubly stochastic atrices Let the stationary probability vector be T = ( 1, 2,..., ) so that T = T P with T e = 1. For doubly stochastic P, e T = e T P so that c T = e T, c = e = e. c i = 1 for all i S i = 1 for all i S.

Generalized inverses Keeny and Snell s fundaental atrix of ergodic MCs Z = [I P + ] 1. Meyer s group inverse of I P, (I P) # = Z. Z and (I P) # are generalized inverses of I P. If G is any generalized inverse of I P, (I P)G(I P) is invariant and = (I P) #.

Relationship between H and Z If H = [I P + ec T ] 1 and Z [I P + e T ] 1 then (a) Z = H + H, (b) H = Z + 1 1 ect Z, (c) (1+ ) = H + ec T Z, (d) (1+ ) T = T H + c T Z.

Eleental relationships If H = [h ij ] = [I P + ec T ] 1 and Z = [z ij ] = [I P + e T ] 1, then (a) z ij = h ij + j k h kj, k=1 (b) h ij = z ij + 1 j 1 z kj, k=1 c k (c) (1+ ) j = k h kj + c k z kj. k=1 k=1

Properties of Z and H For ergodic Markov chains the diagonal eleents of Z, z ij, are positive. Matlab exaples show that a siilar relationship holds for the diagonal eleents of H, h ij. (Forally established later.) z ij h ij = j k=1 k h kj = ( c k z kj j ), i.e. z ij h ij is independent of i, and thus = z jj h jj. Consequently, z jj z ij = h jj h ij for all i, j. k=1

Mean first passage ties For an irreducible finite MC with transition atrix P, let M = ij be the atrix of expected first passage ties fro state i to state j. M satisfies the atrix equation (I P)M = E PM d, where E = ee T = [1], M d = [ ij ij ] = ( d ) 1 D. If G is any g-inverse of I P, then M = [G E(G) d + I G + EG d ]D.

Mean first passage ties Under any of the following three equivalent conditions: (i) Ge = ge, g a constant, (ii) GE E(G) d D = 0, (iii) G E(G) d = 0, M = [I G + EG d ]D. H satisfies (i) ( He = (1/ )e and g = 1/ ) M = [I H + EH d ]D. Z satisfies (i) (since Ze = e and g = 1) M = [I Z + EZ d ]D.

Mean first passage ties Thus 1 1 =, i = j, j c i=1 i h ij ij = h jj h ij = h h jj ij, i j. j c i=1 i h ij Thus a knowledge of the {c i } and the {h ij } leads directly to expressions for the { ij }.

New relationships For all j {1,2,,}, c i ij =. i=1 ij a new connection between the {c j } and { ij }. i j i=1 c i ij = c j j 1+ h jj j, i=1 ij = i j = 1+ h jj j.

Expressions for { j } For all j {1,2,,}, c j j = = ij + c i ij i=1 i=1 1. ij + c i ij i j i j j = j = c + h j jj = 1+ c i ij i=1 h jj 1+ ij i=1 h jj, 1+ c i j i ij = h jj 1 1+ ij i j.

Positivity of the z jj and h jj T M = e T Z d D = ( z 11 1,...,z jj j,...,z ) z jj > 0 for all j. jj h jj = 1+ ij > 0 h jj > 0 for all j. i j c i Since j ij = h jj h ij = z jj z ij > 0 h jj > h ij. No surety regarding the sign of any of the {h ij } for i j.

Doubly stochastic atrices If c i = 1 for all i, then i j = 1+ 2 h jj = 2 z jj. Also i j ii z jj z ii h jj h ii.

Keeny s constant The expression K i = ij j=1 j is in fact independent of i {1,2,...,} K i = K, "Keeny's constant". K has any iportant interpretations in ters of properties of the Markov chain. K is used in the properties of ixing of Markov chains (expected tie to stationarity) K is used in bounding overall differences in the stationary probs of a MC subjected to perturbations.

Expressions for K If G = [g ij ] is any g-inverse of I P, then K = 1+ tr(g) tr(g) = 1+ (g j=1 jj K = 1 (1 ) + tr(h) = 1 (1 ) + K = tr(z) = z j=1 jj. g j i j ), h j=1 jj, For any irreducible -state MC, K + 1 2, tr(h) = h j=1 jj 1 2 + 1.

Doubly stochastic MCs If c i = 1 for all i, then ii = K = 1+ tr(h) = tr(z). Further, for all i, a new result. ii ( + 1) 2,

Exaple Two state MCs Let P = =, p11 p 12 1 a a p 21 p 22 b 1 b (0 a 1, 0 b 1). Let d = 1 a b. MC irreducible 1 d < 1. MC has a unique stationary probability vector T = ( 1, 2 ) = (b (a + b),a (a + b)) = (b (1 d),a (1 d)). 1< d < 1 MC is regular and the stationary distribution is the liiting distribution of the MC. d = 1 MC is irreducible periodic, period 2.

Exaple Two state MCs c T = (c 1,c 2 ) = (1 a + b,1+ a b). a and b specify all the transition probabilities. c 1 and c 2 do not uniquely specify the transition probabilities since c 1 + c 2 = 2. We cannot solve for a and b in ters of c 1 and c 2 Note c 2 c 1 = 2(a b).

Exaple Two state MCs H = [I P + ec T ] 1 = Z = [I P + e T ] 1 = M = 11 12 21 22 1 2(a + b) 1 a + b b + b 1+ a (1 b) (1 a) 1+ b a a + b a a a + b b a + b a + b a + b = (a + b) b 1 a 1 b (a + b) a...

Exaple Two state MCs K = 1+ 1 1.5, a + b K = 1.5 a = b = 1. c 1 c 2 b a 1 2 22 11 and h 11 h 22 a b 11 + 21 = i1 i2 = 12 + 22.

Exaple Three state MCs 1 p 2 p 3 p 2 p 3 P = p ij = q 1 1 q 1 q 3 q 3 r 1 r 2 1 r 1 r 2 Six constrained paraeters with 0 < p 2 + p 3 1, 0 < q 1 + q 3 1 and 0 < r 1 + r 2 1.. Let 1 q 3 r 1 + q 1 r 2 + q 1 r 1, 2 r 1 p 2 + r 2 p 3 + r 2 p 2, 3 p 2 q 3 + p 3 q 1 + p 3 q 3, 1 + 2 + 3.

Exaple Three state MCs MC is irreducible (and hence a stationary distribution exists) 1 > 0, 2 > 0, 3 > 0. Stationary distribution given by ( 1, 2, 3 ) = 1 ( 1, 2, 3 ).

Exaple Three state MCs Let 12 = p 3 + r 1 + r 2, 13 = p 2 + q 1 + q 3, 21 = q 3 + r 1 + r 2, 23 = q 1 + p 2 + p 3, 31 = r 2 + q 1 + q 3, 32 = r 1 + p 2 + p 3, Let = p 2 + p 3 + q 1 + q 3 + r 1 + r 2 = 12 + 13 = 21 + 23 = 31 + 32. M = 1 12 2 13 3 21 1 2 23 3 31 1 32 2 3

Exaple Three state MCs H = 1 3 + 1 3 1 2 3 1 2 3 1 2 3 c 2 21 + c 3 31 c 2 12 + c 3 ( 13 31 ) c 2 ( 12 21 ) c 3 13 c 1 21 + c 3 ( 23 32 ) c 1 12 + c 3 32 c 1 ( 21 12 ) c 3 23 c 1 31 + c 2 ( 32 23 ) c 1 ( 31 13 ) c 2 32 c 1 13 + c 2 23

Exaple Three state MCs Keeny s constant: K = 1 +. For all three-state irreducible MCs, K 2. K = 2 achieved in the inial period 3 case when p 2 = q 3 = r 1.

Exaple Three state MCs Under the iposition of colun totals with c 1 + c 2 + c 3 = 3, we can reduce the free paraeters to p 2,p 3,q 1,q 3,c 1 and c 2 by taking r 1 = c 1 1+ p 2 + p 3 q 1, r 2 = c 2 1 p 2 + q 1 + q 3. Let 1 q 1 + q 3 p 2, 2 p 2 + p 3 q 1,then 1 2 22 11 1 2 r 1 1 r 2 2, c 1 c 2 r 1 + 1 r 2 + 2. No universal inequalities connecting c 1 c 2 with 1 2.

Exaple Three state MCs The following table gives paraeter regions where the stated inequalities occur, in the case where r 1 > 0. c 1 c 2 c 2 c 1 1 2 1 in r 2 2,r r 2 r 1 + 2 r 2 r 1 + 2 1 r 2 2 1 r 1 r 1 1 r 2 2 r 2 2 2 1 r 1 r 2 r 1 + 2 ax r 2 2,r 1 r 2 r 1 + 2 1 1 r 2 2 r 1 1 r 1 + 1 r 2 + 2 r 2 + 2 r 1 + 1

Exaple Five state MC Five state irreducible MC fro Keeny and Snell [10] (p199). Rearrange the states so that the colun sus are ordered with c 1 c 2 c 3 c 4 c 5 with transition atrix P = 0.759 0.082 0.065 0.071 0.023 0.095 0.831 0.033 0.028 0.013 0.112 0.046 0.788 0.038 0.016 0.156 0.054 0.045 0.728 0.017 0.107 0.038 0.034 0.036 0.785 c T = (c 1,c 2,c 3,c 4,c 5 ) = (1.229, 1.051, 0.965, 0.910, 0.854) T = ( 1, 2, 3, 4, 5 ) = (0.3216, 0.2705, 0.1842, 0.1476, 0.0761) 1 2 3 4 5!. Not what we expected!

Exaple Five state MC 2.1984-0.5537-0.4911-0.3007-0.6530-0.8883 3.5691-0.9174-0.7613-0.8021 H = -0.6457-1.0375 3.2047-0.5873-0.7342-0.2485-0.8505-0.6652 2.6746-0.7104-0.7023-1.2092-0.8680-0.6157 3.5952 All the diagonal eleents of H,h jj, are positive (as expected) All the off-diagonal ters are negative. Each row su is 0.200. The colun sus are given as T h colsu = ( -0.2863, -0.0818, 0.2631, 0.4096, 0.6955) also ordered according to the order in c T.

Exaple Five state MC 3.1097 15.2435 20.0601 20.1581 55.7987 9.5987 3.6974 22.3742 23.2789 57.7567 M = 8.8444 17.0326 5.4278 22.1001 56.8645 7.6091 16.3412 21.0051 6.7752 56.5528 9.0204 17.6672 22.1062 22.2926 13.1345 The vector of row sus of M is ( i1, i2, i3, i4, i5 ) = (114.3702, 116.7059, 110.2695, 108.2834, 84.2210), leads to no ordered relationships. The vector of colun sus is ( 1i, 2i, 3i, 4i, 5i ) = (38.1824, 69.9820, 90.9734, 94.6048, 240.1074), with c i c j ii j i for i j. No general results of such a nature for general finite MCs.

Exaple Eight state MC 0.478 0.270 0 0 0.150 0 0.055 0.047 0.130 0.870 0 0 0 0 0 0 0.320 0 0.669 0.011 0 0 0 0 P = 0.088 0 0 0.912 0 0 0 0 0.150 0 0 0 0.740 0.110 0 0 0.300 0 0 0.011 0 0.689 0 0 0.260 0 0 0 0 0 0.740 0 0.600 0 0.400 0 0 0 0 0 (Funderlic and Meyer, exaple involving the analysis of radiophosphorous kinetics in an aquariu syste. ) States reordered so that P has colun sus with c i > c j for i < j. c T = (2.326, 1.140, 1.069, 0.934, 0.890, 0.799, 0.795, 0.047)

Exaple Eight state MC c T = (2.326, 1.140, 1.069, 0.934, 0.890, 0.799, 0.795, 0.047) T = (0.2378, 0.4938, 0.0135, 0.0078, 0.1372, 0.0485, 0.0503, 0.0112). Note, for exaple 1 < 2 even though c 1 > c 2. 1.036 1.056 0.352 1.403 0.170 0.261 0.163 0.043 0.792 4.949 0.456 1.463 0.885 0.634 0.550 0.043 0.228 0.623 2.623 1.052 0.296 0.426 0.334 0.005 H = 1.666 4.556 0.505 9.872 1.389 0.812 0.735 0.084 0.242 1.599 0.425 1.275 3.279 0.839 0.434 0.017 0.176 0.731 0.401 1.029 0.326 2.779 0.345 0.002 0.122 0.844 0.404 1.433 0.358 0.448 3.490 0.000 0.475 0.110 0.825 1.271 0.154 0.376 0.282 1.017 Diagonal eleents of H are positive. No obvious pattern for off-diagonal elents. Rows sus of H = 0.125, ( =1/8). Colun sus of H do not exhibit any pattern

Exaple Eight state MC M = 4.21 7.88 220.32 1454.40 22.66 62.66 72.62 87.13 7.69 2.03 228.01 1462.09 30.35 70.35 80.32 94.82 3.40 11.28 74.05 1409.09 26.06 66.06 76.02 90.53 11.36 19.25 231.68 128.99 34.03 74.02 83.99 98.49 5.38 13.26 225.69 1437.84 7.29 39.99 78.00 92.50 3.62 11.50 223.93 1406.17 26.23 20.61 76.24 90.74 3.85 11.73 224.16 1458.24 26.51 66.50 19.88 90.97 2.36 10.24 133.19 1437.27 25.02 65.02 74.98 89.49 No ordered relationship within the row sus of M No ordered relationship within the colun sus of M Keeny s constant = 29.9194.

Conjectures: c i c j for all i, j i j for all i, j h ij < 0 for all i j Valid in the two-state case and the special 5-state case. Not true in general. Exaple: 1/2 1/2 0 If P = 1/2 0 1/2 1/2 1/2 0 then H = 2 9 c T = (3/2,1,1/2), T = (1/2, 1/3, 1/6), h 23 > 0. 3 1/2 1 3/2 5/2 1/2 3/2 1/2 7/2 Are there any general inter-relationships? If h ij < 0 for all i j, does i j for all i, j? does c i c j for all i, j?. (Kirkland)

References - 1 [1] R. Funderlic, C.D. Meyer, Sensitivity of the stationary distribution vector for an ergodic Markov chain, Linear Algebra Appl. 76 (1986) 1-17. [2] J. J. Hunter, Generalized inverses and their application to applied probability probles, Linear Algebra Appl. 45 (1982) 157-198. [3] J. J. Hunter, Matheatical Techniques of Applied Probability, Volue 1, Discrete Tie Models: Basic Theory, Acadeic, New York, 1983. [4] J. J. Hunter, Matheatical Techniques of Applied Probability, Volue 2, Discrete Tie Models: Techniques and Applications, Acadeic, New York,1983. [5] J. J. Hunter, Paraetric Fors for Generalized inverses of Markovian Kernels and their Applications, Linear Algebra Appl. 127 (1990) 71-84. [6] J. J. Hunter, Mixing ties with applications to perturbed Markov chains, Linear Algebra Appl. 417 (2006) 108-123. [7] J. J. Hunter, Siple procedures for finding ean first passage ties in Markov chains, Asia-Pacific J of Operational Research, 24 (6), (2007), 813-829.

References - 2 [8] J. J. Hunter, Variances of First Passage Ties in a Markov chain with applications to Mixing Ties, Linear Algebra Appl. 429 (2008) 1135-1162. [9] J. J. Hunter, Soe stochastic properties of sei-agic and agic Markov chains, Linear Algebra Appl. 433 (2010), 893-907. [10] J. G. Keeny, J.L. Snell, Finite Markov Chains, Van Nostrand, New York, 1960. [11] S. Kirkland, Subdoinant eigenvalues for stochastic atrices with given colun sus, Electronic Journal of Linear Algebra, 18 (2009), 784-800, ISSN 1081-3810. [12] S. Kirkland, Colun sus and the conditioning of stationary distribution for a stochastic atrix, Operators and Matrices, 4 (2010), 431-443, ISSN 1846-3886. [13] C. D. Meyer Jr., The role of the group generalized inverse in the theory of finite Markov chains, SIAM Rev.17 (1975) 443-464.