Matrix analytic methods. Lecture 1: Structured Markov chains and their stationary distribution

Size: px
Start display at page:

Download "Matrix analytic methods. Lecture 1: Structured Markov chains and their stationary distribution"

Transcription

1 1/29 Matrix analytic methods Lecture 1: Structured Markov chains and their stationary distribution Sophie Hautphenne and David Stanford (with thanks to Guy Latouche, U. Brussels and Peter Taylor, U. Melbourne Université Libre de Bruxelles University of Western Ontario Cirm, Luminy. April 28, 2010

2 2/29 Matrix-Analytic Methods In the 1970s and 1980s Marcel Neuts proposed a class of techniques for analysing Markov chains with block-structured transition matrices that have become known as matrix-analytic methods. More recently, other classes of models have been analysed using similar techniques, such as fluid queues (see Lecture 2) branching processes (see Lecture 3) The interaction of mathematical analysis and physical insight has played an important role in the development of results in this area.

3 2/29 Matrix-Analytic Methods In the 1970s and 1980s Marcel Neuts proposed a class of techniques for analysing Markov chains with block-structured transition matrices that have become known as matrix-analytic methods. More recently, other classes of models have been analysed using similar techniques, such as fluid queues (see Lecture 2) branching processes (see Lecture 3) The interaction of mathematical analysis and physical insight has played an important role in the development of results in this area.

4 2/29 Matrix-Analytic Methods In the 1970s and 1980s Marcel Neuts proposed a class of techniques for analysing Markov chains with block-structured transition matrices that have become known as matrix-analytic methods. More recently, other classes of models have been analysed using similar techniques, such as fluid queues (see Lecture 2) branching processes (see Lecture 3) The interaction of mathematical analysis and physical insight has played an important role in the development of results in this area.

5 /29 Quasi-Birth-and-Death processes A QDB process is a two-dimensional Markov chain {(X k, ϕ k ), k N} where X k N is called the level, ϕ k {1, 2,, m} is called the phase, the only possible transitions from state (n, i) are to state (n + 1, j) (1 level up), (n, j) (the same level), (n 1, j) (1 level down)

6 Quasi-Birth-and-Death processes 4/29

7 5/29 Example: tandem queues λ µ 1 µ 2 C C < = maximum capacity of the second buffer n is number ahead of first server = level of the QBD j is number beyond first server = phase of the QBD.

8 6/29 Quasi-Birth-and-Death processes m m n n + 1 Transition probabilities: (A 1 ) ij to go up: (n, i) to (n+1, j) (A 0 ) ij to remain in same level: (n, i) to (n, j) (A 1 ) ij to go down: (n, i) to (n 1, j)

9 7/29 Stationary distribution Block-structured transition matrix: B A 1 0 A 1 A 0 A 1 P = 0 A 1 A 0... Transitions are homogeneous with respect to the levels π = [π 0, π 1, π 2, ] is the stationary probability vector of the QBD, where (π n ) i = lim k P[(X k, ϕ k ) = (n, i)]

10 /29 Stationary distribution Block-structured transition matrix: B A 1 0 A 1 A 0 A 1 P = 0 A 1 A 0... Transitions are homogeneous with respect to the levels π = [π 0, π 1, π 2, ] is the stationary probability vector of the QBD, where (π n ) i = lim k P[(X k, ϕ k ) = (n, i)]

11 /29 Stationary distribution Block-structured transition matrix: B A 1 0 A 1 A 0 A 1 P = 0 A 1 A 0... Transitions are homogeneous with respect to the levels π = [π 0, π 1, π 2, ] is the stationary probability vector of the QBD, where (π n ) i = lim k P[(X k, ϕ k ) = (n, i)]

12 8/29 Existence of the stationary distribution Let x be the solution to x(a 1 + A 0 + A 1 ) = x. Then the QBD is positive recurrent, null recurrent or transient according as x A 1 1 x A 1 1 is < 0, = 0 or > 0. The stationary probability vector π exists if and only if the QBD is positive recurrent.

13 8/29 Existence of the stationary distribution Let x be the solution to x(a 1 + A 0 + A 1 ) = x. Then the QBD is positive recurrent, null recurrent or transient according as x A 1 1 x A 1 1 is < 0, = 0 or > 0. The stationary probability vector π exists if and only if the QBD is positive recurrent.

14 8/29 Existence of the stationary distribution Let x be the solution to x(a 1 + A 0 + A 1 ) = x. Then the QBD is positive recurrent, null recurrent or transient according as x A 1 1 x A 1 1 is < 0, = 0 or > 0. The stationary probability vector π exists if and only if the QBD is positive recurrent.

15 9/29 Computation of the stationary distribution The stationary probability vector π satisfies π P = π, π 1 = 1 Partition the state space into E, E c [ ] [ P E P EE c π E π E c P E c E P E c ] [ = π E π E c ] One has that π E (P E + P EE c (I P E c ) 1 P E c E ) = π E π E c = π E P EE c (I P E c ) 1

16 9/29 Computation of the stationary distribution The stationary probability vector π satisfies π P = π, π 1 = 1 Partition the state space into E, E c [ ] [ P E P EE c π E π E c P E c E P E c ] [ = π E π E c ] One has that π E (P E + P EE c (I P E c ) 1 P E c E ) = π E π E c = π E P EE c (I P E c ) 1

17 9/29 Computation of the stationary distribution The stationary probability vector π satisfies π P = π, π 1 = 1 Partition the state space into E, E c [ ] [ P E P EE c π E π E c P E c E P E c ] [ = π E π E c ] One has that π E (P E + P EE c (I P E c ) 1 P E c E ) = π E π E c = π E P EE c (I P E c ) 1

18 Computation of the stationary distribution Let l(n) denote the level n of the QBD. Choose E = l(0), E c = {l(1), l(2), l(3),, } [ P = P E P E c E P EE c P E c ] = B A A 1 A 0 A 1 0 A 1 A 0 A 1 0 A 1 A 0. The equation π E c = π E P EE c (I P E c ) 1 becomes [ π 1 π 2 ] = π 0 [ A ] I A 0 A 1 A 1 I A 0 A 1 A 1 I A {z } N 10/29 We obtain for π 1 : π 1 = π 0 A 1 N 11.

19 Computation of the stationary distribution Let l(n) denote the level n of the QBD. Choose E = l(0), E c = {l(1), l(2), l(3),, } [ P = P E P E c E P EE c P E c ] = B A A 1 A 0 A 1 0 A 1 A 0 A 1 0 A 1 A 0. The equation π E c = π E P EE c (I P E c ) 1 becomes [ π 1 π 2 ] = π 0 [ A ] I A 0 A 1 A 1 I A 0 A 1 A 1 I A {z } N 10/29 We obtain for π 1 : π 1 = π 0 A 1 N 11.

20 Computation of the stationary distribution Let l(n) denote the level n of the QBD. Choose E = l(0), E c = {l(1), l(2), l(3),, } [ P = P E P E c E P EE c P E c ] = B A A 1 A 0 A 1 0 A 1 A 0 A 1 0 A 1 A 0. The equation π E c = π E P EE c (I P E c ) 1 becomes [ π 1 π 2 ] = π 0 [ A ] I A 0 A 1 A 1 I A 0 A 1 A 1 I A {z } N 10/29 We obtain for π 1 : π 1 = π 0 A 1 N 11.

21 Computation of the stationary distribution Let l(n) denote the level n of the QBD. Choose E = l(0), E c = {l(1), l(2), l(3),, } [ P = P E P E c E P EE c P E c ] = B A A 1 A 0 A 1 0 A 1 A 0 A 1 0 A 1 A 0. The equation π E c = π E P EE c (I P E c ) 1 becomes [ π 1 π 2 ] = π 0 [ A ] I A 0 A 1 A 1 I A 0 A 1 A 1 I A {z } N 10/29 We obtain for π 1 : π 1 = π 0 A 1 N 11.

22 Computation of the stationary distribution Let l(n) denote the level n of the QBD. Choose E = l(0), E c = {l(1), l(2), l(3),, } [ P = P E P E c E P EE c P E c ] = B A A 1 A 0 A 1 0 A 1 A 0 A 1 0 A 1 A 0. The equation π E c = π E P EE c (I P E c ) 1 becomes [ π 1 π 2 ] = π 0 [ A ] I A 0 A 1 A 1 I A 0 A 1 A 1 I A {z } N 10/29 We obtain for π 1 : π 1 = π 0 A 1 N 11.

23 11/29 Computation of the stationary distribution Then, for any n 1, choose E = {l(0), l(1),, l(n)}, E c = {l(n + 1), l(n + 2),, } By the same argument, we get n 0 π n+1 = π n A 1 N 11 = π 0 (A 1 N 11 ) n+1 (N 11 ) ij is the matrix of expected number of visits of state (n + 1, j), starting from (n + 1, i), before returning to {l(0), l(1),, l(n)}, for any n 0.

24 1/29 Computation of the stationary distribution Then, for any n 1, choose E = {l(0), l(1),, l(n)}, E c = {l(n + 1), l(n + 2),, } By the same argument, we get n 0 π n+1 = π n A 1 N 11 = π 0 (A 1 N 11 ) n+1 (N 11 ) ij is the matrix of expected number of visits of state (n + 1, j), starting from (n + 1, i), before returning to {l(0), l(1),, l(n)}, for any n 0.

25 1/29 Computation of the stationary distribution Then, for any n 1, choose E = {l(0), l(1),, l(n)}, E c = {l(n + 1), l(n + 2),, } By the same argument, we get n 0 π n+1 = π n A 1 N 11 = π 0 (A 1 N 11 ) n+1 (N 11 ) ij is the matrix of expected number of visits of state (n + 1, j), starting from (n + 1, i), before returning to {l(0), l(1),, l(n)}, for any n 0.

26 12/29 Stationary distribution In summarize, for all n 0, π n = π 0 R n, where R = A 1 N 11 and R ij is the expected number of visits to (1, j) starting from (0, i) before the first return to level 0. This is a matrix-geometric stationary distribution. To complete its characterization, it remains to characterize π 0 and R.

27 2/29 Stationary distribution In summarize, for all n 0, π n = π 0 R n, where R = A 1 N 11 and R ij is the expected number of visits to (1, j) starting from (0, i) before the first return to level 0. This is a matrix-geometric stationary distribution. To complete its characterization, it remains to characterize π 0 and R.

28 2/29 Stationary distribution In summarize, for all n 0, π n = π 0 R n, where R = A 1 N 11 and R ij is the expected number of visits to (1, j) starting from (0, i) before the first return to level 0. This is a matrix-geometric stationary distribution. To complete its characterization, it remains to characterize π 0 and R.

29 13/29 Characterization of π 0 Take again E = l(0). From π E (P E + P EE c (I P E c ) 1 P E c E ) = π E, we get π 0 (B + A 1 N 11 A 1 ) = π 0. From the normalization constraint π 1 = i=0 π i 1 = 1 we have π i 1 = π 0 i=0 i=0 R i 1 = π 0 (I R) 1 1 = 1. Thus π 0 satisfies { π0 (B + R A 1 ) = π 0 π 0 (I R) 1 1 = 1

30 13/29 Characterization of π 0 Take again E = l(0). From π E (P E + P EE c (I P E c ) 1 P E c E ) = π E, we get π 0 (B + A 1 N 11 A 1 ) = π 0. From the normalization constraint π 1 = i=0 π i 1 = 1 we have π i 1 = π 0 i=0 i=0 R i 1 = π 0 (I R) 1 1 = 1. Thus π 0 satisfies { π0 (B + R A 1 ) = π 0 π 0 (I R) 1 1 = 1

31 13/29 Characterization of π 0 Take again E = l(0). From π E (P E + P EE c (I P E c ) 1 P E c E ) = π E, we get π 0 (B + A 1 N 11 A 1 ) = π 0. From the normalization constraint π 1 = i=0 π i 1 = 1 we have π i 1 = π 0 i=0 i=0 R i 1 = π 0 (I R) 1 1 = 1. Thus π 0 satisfies { π0 (B + R A 1 ) = π 0 π 0 (I R) 1 1 = 1

32 13/29 Characterization of π 0 Take again E = l(0). From π E (P E + P EE c (I P E c ) 1 P E c E ) = π E, we get π 0 (B + A 1 N 11 A 1 ) = π 0. From the normalization constraint π 1 = i=0 π i 1 = 1 we have π i 1 = π 0 i=0 i=0 R i 1 = π 0 (I R) 1 1 = 1. Thus π 0 satisfies { π0 (B + R A 1 ) = π 0 π 0 (I R) 1 1 = 1

33 14/29 Matrix R and first passage probability matrix G Many expressions for R; for instance where R = A 1 [ I (A0 + A 1 G) ] 1 G ij is the probability of eventually moving to (0, j), starting from (1, i). (A 0 + A 1 G) ij is the probability of visiting (1, j) from (1, i) avoiding level 0. G is stochastic and it is the minimal nonnegative solution of X = A 1 + A 0 X + A 1 X 2

34 14/29 Matrix R and first passage probability matrix G Many expressions for R; for instance where R = A 1 [ I (A0 + A 1 G) ] 1 G ij is the probability of eventually moving to (0, j), starting from (1, i). (A 0 + A 1 G) ij is the probability of visiting (1, j) from (1, i) avoiding level 0. G is stochastic and it is the minimal nonnegative solution of X = A 1 + A 0 X + A 1 X 2

35 15/29 Formal manipulation X = A 1 + A 0 X + A 1 X 2 leads to X A 0 X A 1 X 2 = A 1 X = [ I (A 0 + A 1 X ) ] 1 A 1. (1) Other equations have been proposed, from decomposition A 1 + A 0 X + A 1 X 2 = F (X ) + G(X )X All may be solved by functional iteration, (1) is the most interesting for its efficiency and its ease of interpretation

36 15/29 Formal manipulation X = A 1 + A 0 X + A 1 X 2 leads to X A 0 X A 1 X 2 = A 1 X = [ I (A 0 + A 1 X ) ] 1 A 1. (1) Other equations have been proposed, from decomposition A 1 + A 0 X + A 1 X 2 = F (X ) + G(X )X All may be solved by functional iteration, (1) is the most interesting for its efficiency and its ease of interpretation

37 16/29 Interpretation of functional iterations Start with X 0 = 0 and iterate X n = [ I (A 0 + A 1 X n 1 ) ] 1 A 1 or X n = A 1 + A 0 X n + A 1 X n 1 X n n n + 1 n n 1 X n : process is allowed to move freely among n levels as n, restrictions disappear and X n G

38 17/29 Convergence rates Linear convergence with rates X n = [ I (A 0 + A 1 X n 1 ) ] 1 A 1 η = lim n G X n 1 n Actually, η is the Perron-Frobenius eigenvalue of R. When the QBD is positive recurrent, η < 1.

39 17/29 Convergence rates Linear convergence with rates X n = [ I (A 0 + A 1 X n 1 ) ] 1 A 1 η = lim n G X n 1 n Actually, η is the Perron-Frobenius eigenvalue of R. When the QBD is positive recurrent, η < 1.

40 8/29 Acceleration: keep it stochastic Instead of X 0 = 0, take X 0 = I with the same functional iteration X n = [ I (A 0 + A 1 X n 1) ] 1 A 1 The new sequence is well defined and X n corresponds to a new irreducible process which is forced to remain between levels 1 and n prove by induction that X n X n convergence rate strictly less than η, the P.F. eigenvalue of R. Total complexity O(m 3 ).

41 19/29 Newton s scheme X = A 1 + A 0 X + A 1 X 2 Newton s scheme: convergence to G, monotone and quadratic for any initial nonnegative matrix between 0 and G. Requires at each step the solution of T n H(T n 1 )A 1 T n H(T n 1 )A 1 = H(T n 1 )A 1, where H(X ) = [ I (A 0 + A 1 X ) ] 1 Sylvester equation, O(m 3 ) algorithm since 1992 (Gardiner, Laub, Amato and Moler) Total complexity O(m 6 ).

42 20/29 Logarithmic & cyclic reduction algorithms Two nearly identical, quadratically convergent, algorithms Logarithmic reduction: Latouche and Ramaswami, 1993 Cyclic reduction: Bini and Meini, 1995; slightly faster at each step, generalized to more complex Markov chains

43 21/29 Cyclic reduction The equation G = A 1 + A 0 G + A 1 G 2 is transformed into the linear system I A 0 A 1 0 A 1 I A 0 A 1. A 1 I A G G 2 G 3. = Apply an even-odd permutation to rows and columns. Perform one step of Gaussian elimination to remove the even-numbered blocks and obtain A

44 21/29 Cyclic reduction The equation G = A 1 + A 0 G + A 1 G 2 is transformed into the linear system I A 0 A 1 0 A 1 I A 0 A 1. A 1 I A G G 2 G 3. = Apply an even-odd permutation to rows and columns. Perform one step of Gaussian elimination to remove the even-numbered blocks and obtain A

45 21/29 Cyclic reduction The equation G = A 1 + A 0 G + A 1 G 2 is transformed into the linear system I A 0 A 1 0 A 1 I A 0 A 1. A 1 I A G G 2 G 3. = Apply an even-odd permutation to rows and columns. Perform one step of Gaussian elimination to remove the even-numbered blocks and obtain A

46 22/29 Cyclic reduction the same structure. I Â(1) 0 A (1) A (1) 1 I A (1) 0 A (1) 1 A (1) 1 I A (1) G G 3 G 5. = A Repeat as needed [ At nth step: G G 2n +1 G 2 2n +1 G 3 2n +1 ] T upper diagonal: A (n) 1 tends to 0 as n tends to. When A (n) 1 0, G (I Â(n) 0 ) 1 A 1.

47 22/29 Cyclic reduction the same structure. I Â(1) 0 A (1) A (1) 1 I A (1) 0 A (1) 1 A (1) 1 I A (1) G G 3 G 5. = A Repeat as needed [ At nth step: G G 2n +1 G 2 2n +1 G 3 2n +1 ] T upper diagonal: A (n) 1 tends to 0 as n tends to. When A (n) 1 0, G (I Â(n) 0 ) 1 A 1.

48 23/29 Other structured Markov chains GI /M/1-type Markov chains A GI /M/1-type Markov chain is a two-dimensional Markovian chain {(X k, ϕ k ), k N} where the only possible transitions from state (n, i) are to state (n + 1, j) (1 level up), (n, j) (the same level), (n l, j) (l levels down) for all 1 l n. Ã 0 A Ã 1 A 0 A 1 0 P = Ã 2 A 1 A 0 A 1 Ã 3 A 2 A 1 A

49 24/29 Solution for GI /M/1-type Markov Chains For a GI /M/1-type Markov chain, let x be the solution to [ ] x A 1 k = x. k=0 Then the chain is positive recurrent, null recurrent or transient according as [ ] x A 0 1 x (k 1)A 1 k 1 is < 0, = 0 or > 0. k=2 The stationary distribution π exists if and only if the chain is positive recurrent.

50 24/29 Solution for GI /M/1-type Markov Chains For a GI /M/1-type Markov chain, let x be the solution to [ ] x A 1 k = x. k=0 Then the chain is positive recurrent, null recurrent or transient according as [ ] x A 0 1 x (k 1)A 1 k 1 is < 0, = 0 or > 0. k=2 The stationary distribution π exists if and only if the chain is positive recurrent.

51 24/29 Solution for GI /M/1-type Markov Chains For a GI /M/1-type Markov chain, let x be the solution to [ ] x A 1 k = x. k=0 Then the chain is positive recurrent, null recurrent or transient according as [ ] x A 0 1 x (k 1)A 1 k 1 is < 0, = 0 or > 0. k=2 The stationary distribution π exists if and only if the chain is positive recurrent.

52 25/29 Stationary distribution of GI /M/1-type Markov Chains The stationary distribution π of a positive recurrent GI /M/1-type Markov chain of is such that π n = π 0 R n n 0, where the matrix R has exactly the same probabilistic interpretation as for a QBD, and the vector π 0 satisfies [ ] π 0 R k à 2 k = π 0, k=0 and a normalization constraint. This is the well-known Matrix-Geometric form of the stationary distribution.

53 25/29 Stationary distribution of GI /M/1-type Markov Chains The stationary distribution π of a positive recurrent GI /M/1-type Markov chain of is such that π n = π 0 R n n 0, where the matrix R has exactly the same probabilistic interpretation as for a QBD, and the vector π 0 satisfies [ ] π 0 R k à 2 k = π 0, k=0 and a normalization constraint. This is the well-known Matrix-Geometric form of the stationary distribution.

54 26/29 Matrix R for GI /M/1-type Markov chains The matrix R is the minimal nonnegative solution to the matrix equation X = X k+1 A k. k= 1 It is obtained using a duality property with M/G/1-type Markov chains: Determining R for a positive recurrent GI /M/1-type MC determining G for a transient M/G/1-type MC.

55 6/29 Matrix R for GI /M/1-type Markov chains The matrix R is the minimal nonnegative solution to the matrix equation X = X k+1 A k. k= 1 It is obtained using a duality property with M/G/1-type Markov chains: Determining R for a positive recurrent GI /M/1-type MC determining G for a transient M/G/1-type MC.

56 Other structured Markov chains M/G/1-type Markov chains An M/G/1-type Markov chain is a two-dimensional Markovian chain {(X k, ϕ k ), k N} where the only possible transitions from state (n, i) are to state (n + l, j) (l levels up) for all l 1, (n, j) (the same level), 7/29 (n 1, j) (1 level down). Ã 0 Ã 1 Ã 2 Ã 3 A 1 A 0 A 1 A 2 P = 0 A 1 A 0 A A 1 A M/G/1-type Markov chains do not have a matrix-geometric stationary distribution.

57 28/29 Matrix G for M/G/1-type Markov chains The matrix G is the minimal nonnegative solution to the matrix equation X = A k X k+1. k= 1 It may be obtained using for instance the generalization of the cyclic-reduction algorithm to M/G/1-type Markov chains.

58 8/29 Matrix G for M/G/1-type Markov chains The matrix G is the minimal nonnegative solution to the matrix equation X = A k X k+1. k= 1 It may be obtained using for instance the generalization of the cyclic-reduction algorithm to M/G/1-type Markov chains.

59 29/29 References Latouche and Ramaswami, Introduction to Matrix Analytic Methods in Stochastic Modeling, SIAM, 1999 Bini, Latouche and Meini, Numerical Methods for Structured Markov Chains, Oxford University Press, 2005 Latouche, Newton s iteration for non-linear equations in Markov chains, IMA Journal of Numerical Analysis, 1994, 14(4):

CHUN-HUA GUO. Key words. matrix equations, minimal nonnegative solution, Markov chains, cyclic reduction, iterative methods, convergence rate

CHUN-HUA GUO. Key words. matrix equations, minimal nonnegative solution, Markov chains, cyclic reduction, iterative methods, convergence rate CONVERGENCE ANALYSIS OF THE LATOUCHE-RAMASWAMI ALGORITHM FOR NULL RECURRENT QUASI-BIRTH-DEATH PROCESSES CHUN-HUA GUO Abstract The minimal nonnegative solution G of the matrix equation G = A 0 + A 1 G +

More information

Structured Markov chains solver: tool extension

Structured Markov chains solver: tool extension Structured Markov chains solver: tool extension D. A. Bini, B. Meini, S. Steffé Dipartimento di Matematica Università di Pisa, Pisa, Italy bini, meini, steffe @dm.unipi.it B. Van Houdt Department of Mathematics

More information

Modelling Complex Queuing Situations with Markov Processes

Modelling Complex Queuing Situations with Markov Processes Modelling Complex Queuing Situations with Markov Processes Jason Randal Thorne, School of IT, Charles Sturt Uni, NSW 2795, Australia Abstract This article comments upon some new developments in the field

More information

Multi Stage Queuing Model in Level Dependent Quasi Birth Death Process

Multi Stage Queuing Model in Level Dependent Quasi Birth Death Process International Journal of Statistics and Systems ISSN 973-2675 Volume 12, Number 2 (217, pp. 293-31 Research India Publications http://www.ripublication.com Multi Stage Queuing Model in Level Dependent

More information

The MATLAB toolbox SMCSolver for matrix-analytic methods

The MATLAB toolbox SMCSolver for matrix-analytic methods The MATLAB toolbox SMCSolver for matrix-analytic methods D. Bini, B. Meini, S. Steffe, J.F. Pérez, B. Van Houdt Dipartimento di Matematica, Universita di Pisa, Italy Department of Electrical and Electronics

More information

Censoring Technique in Studying Block-Structured Markov Chains

Censoring Technique in Studying Block-Structured Markov Chains Censoring Technique in Studying Block-Structured Markov Chains Yiqiang Q. Zhao 1 Abstract: Markov chains with block-structured transition matrices find many applications in various areas. Such Markov chains

More information

Matrix Analytic Methods for Stochastic Fluid Flows

Matrix Analytic Methods for Stochastic Fluid Flows Matrix Analytic Methods for Stochastic Fluid Flows V. Ramaswami, AT&T Labs, 2 Laurel Avenue D5-3B22, Middletown, NJ 7748 We present an analysis of stochastic fluid flow models along the lines of matrix-analytic

More information

Advanced Queueing Theory

Advanced Queueing Theory Advanced Queueing Theory 1 Networks of queues (reversibility, output theorem, tandem networks, partial balance, product-form distribution, blocking, insensitivity, BCMP networks, mean-value analysis, Norton's

More information

= P{X 0. = i} (1) If the MC has stationary transition probabilities then, = i} = P{X n+1

= P{X 0. = i} (1) If the MC has stationary transition probabilities then, = i} = P{X n+1 Properties of Markov Chains and Evaluation of Steady State Transition Matrix P ss V. Krishnan - 3/9/2 Property 1 Let X be a Markov Chain (MC) where X {X n : n, 1, }. The state space is E {i, j, k, }. The

More information

Examples of Countable State Markov Chains Thursday, October 16, :12 PM

Examples of Countable State Markov Chains Thursday, October 16, :12 PM stochnotes101608 Page 1 Examples of Countable State Markov Chains Thursday, October 16, 2008 12:12 PM Homework 2 solutions will be posted later today. A couple of quick examples. Queueing model (without

More information

Lecture 21. David Aldous. 16 October David Aldous Lecture 21

Lecture 21. David Aldous. 16 October David Aldous Lecture 21 Lecture 21 David Aldous 16 October 2015 In continuous time 0 t < we specify transition rates or informally P(X (t+δ)=j X (t)=i, past ) q ij = lim δ 0 δ P(X (t + dt) = j X (t) = i) = q ij dt but note these

More information

Statistics 150: Spring 2007

Statistics 150: Spring 2007 Statistics 150: Spring 2007 April 23, 2008 0-1 1 Limiting Probabilities If the discrete-time Markov chain with transition probabilities p ij is irreducible and positive recurrent; then the limiting probabilities

More information

Markov Chains and Stochastic Sampling

Markov Chains and Stochastic Sampling Part I Markov Chains and Stochastic Sampling 1 Markov Chains and Random Walks on Graphs 1.1 Structure of Finite Markov Chains We shall only consider Markov chains with a finite, but usually very large,

More information

Lecture 11: Introduction to Markov Chains. Copyright G. Caire (Sample Lectures) 321

Lecture 11: Introduction to Markov Chains. Copyright G. Caire (Sample Lectures) 321 Lecture 11: Introduction to Markov Chains Copyright G. Caire (Sample Lectures) 321 Discrete-time random processes A sequence of RVs indexed by a variable n 2 {0, 1, 2,...} forms a discretetime random process

More information

Markov Chains, Stochastic Processes, and Matrix Decompositions

Markov Chains, Stochastic Processes, and Matrix Decompositions Markov Chains, Stochastic Processes, and Matrix Decompositions 5 May 2014 Outline 1 Markov Chains Outline 1 Markov Chains 2 Introduction Perron-Frobenius Matrix Decompositions and Markov Chains Spectral

More information

Classification of Countable State Markov Chains

Classification of Countable State Markov Chains Classification of Countable State Markov Chains Friday, March 21, 2014 2:01 PM How can we determine whether a communication class in a countable state Markov chain is: transient null recurrent positive

More information

A TANDEM QUEUE WITH SERVER SLOW-DOWN AND BLOCKING

A TANDEM QUEUE WITH SERVER SLOW-DOWN AND BLOCKING Stochastic Models, 21:695 724, 2005 Copyright Taylor & Francis, Inc. ISSN: 1532-6349 print/1532-4214 online DOI: 10.1081/STM-200056037 A TANDEM QUEUE WITH SERVER SLOW-DOWN AND BLOCKING N. D. van Foreest

More information

Finite-Horizon Statistics for Markov chains

Finite-Horizon Statistics for Markov chains Analyzing FSDT Markov chains Friday, September 30, 2011 2:03 PM Simulating FSDT Markov chains, as we have said is very straightforward, either by using probability transition matrix or stochastic update

More information

Stationary Distribution of a Perturbed Quasi-Birth-and-Death Process

Stationary Distribution of a Perturbed Quasi-Birth-and-Death Process Université Libre de Bruxelles Faculté des Sciences Stationary Distribution of a Perturbed Quasi-Birth-and-Death Process Rapport d avancement des recherches 2009-2011 Sarah Dendievel Promoteur : Guy Latouche

More information

ON SOME BASIC PROPERTIES OF THE INHOMOGENEOUS QUASI-BIRTH-AND-DEATH PROCESS

ON SOME BASIC PROPERTIES OF THE INHOMOGENEOUS QUASI-BIRTH-AND-DEATH PROCESS Comm. Korean Math. Soc. 12 (1997), No. 1, pp. 177 191 ON SOME BASIC PROPERTIES OF THE INHOMOGENEOUS QUASI-BIRTH-AND-DEATH PROCESS KYUNG HYUNE RHEE AND C. E. M. PEARCE ABSTRACT. The basic theory of the

More information

Birth-death chain models (countable state)

Birth-death chain models (countable state) Countable State Birth-Death Chains and Branching Processes Tuesday, March 25, 2014 1:59 PM Homework 3 posted, due Friday, April 18. Birth-death chain models (countable state) S = We'll characterize the

More information

Solving Nonlinear Matrix Equation in Credit Risk. by Using Iterative Methods

Solving Nonlinear Matrix Equation in Credit Risk. by Using Iterative Methods Int. J. Contemp. Math. Sciences, Vol. 7, 2012, no. 39, 1921-1930 Solving Nonlinear Matrix Equation in Credit Risk by Using Iterative Methods a,b Gholamreza Farsadamanollahi 1, b Joriah Binti Muhammad and

More information

IEOR 6711: Stochastic Models I Professor Whitt, Thursday, November 29, Weirdness in CTMC s

IEOR 6711: Stochastic Models I Professor Whitt, Thursday, November 29, Weirdness in CTMC s IEOR 6711: Stochastic Models I Professor Whitt, Thursday, November 29, 2012 Weirdness in CTMC s Where s your will to be weird? Jim Morrison, The Doors We are all a little weird. And life is a little weird.

More information

Overload Analysis of the PH/PH/1/K Queue and the Queue of M/G/1/K Type with Very Large K

Overload Analysis of the PH/PH/1/K Queue and the Queue of M/G/1/K Type with Very Large K Overload Analysis of the PH/PH/1/K Queue and the Queue of M/G/1/K Type with Very Large K Attahiru Sule Alfa Department of Mechanical and Industrial Engineering University of Manitoba Winnipeg, Manitoba

More information

Cover Page. The handle holds various files of this Leiden University dissertation

Cover Page. The handle   holds various files of this Leiden University dissertation Cover Page The handle http://hdlhandlenet/1887/39637 holds various files of this Leiden University dissertation Author: Smit, Laurens Title: Steady-state analysis of large scale systems : the successive

More information

Departure Processes of a Tandem Network

Departure Processes of a Tandem Network The 7th International Symposium on perations Research and Its Applications (ISRA 08) Lijiang, China, ctober 31 Novemver 3, 2008 Copyright 2008 RSC & APRC, pp. 98 103 Departure Processes of a Tandem Network

More information

Simultaneous Transient Analysis of QBD Markov Chains for all Initial Configurations using a Level Based Recursion

Simultaneous Transient Analysis of QBD Markov Chains for all Initial Configurations using a Level Based Recursion Simultaneous Transient Analysis of QBD Markov Chains for all Initial Configurations using a Level Based Recursion J Van Velthoven, B Van Houdt and C Blondia University of Antwerp Middelheimlaan 1 B- Antwerp,

More information

Note that in the example in Lecture 1, the state Home is recurrent (and even absorbing), but all other states are transient. f ii (n) f ii = n=1 < +

Note that in the example in Lecture 1, the state Home is recurrent (and even absorbing), but all other states are transient. f ii (n) f ii = n=1 < + Random Walks: WEEK 2 Recurrence and transience Consider the event {X n = i for some n > 0} by which we mean {X = i}or{x 2 = i,x i}or{x 3 = i,x 2 i,x i},. Definition.. A state i S is recurrent if P(X n

More information

A FAST MATRIX-ANALYTIC APPROXIMATION FOR THE TWO CLASS GI/G/1 NON-PREEMPTIVE PRIORITY QUEUE

A FAST MATRIX-ANALYTIC APPROXIMATION FOR THE TWO CLASS GI/G/1 NON-PREEMPTIVE PRIORITY QUEUE A FAST MATRIX-ANAYTIC APPROXIMATION FOR TE TWO CASS GI/G/ NON-PREEMPTIVE PRIORITY QUEUE Gábor orváth Department of Telecommunication Budapest University of Technology and Economics. Budapest Pf. 9., ungary

More information

Discrete time Markov chains. Discrete Time Markov Chains, Limiting. Limiting Distribution and Classification. Regular Transition Probability Matrices

Discrete time Markov chains. Discrete Time Markov Chains, Limiting. Limiting Distribution and Classification. Regular Transition Probability Matrices Discrete time Markov chains Discrete Time Markov Chains, Limiting Distribution and Classification DTU Informatics 02407 Stochastic Processes 3, September 9 207 Today: Discrete time Markov chains - invariant

More information

Lecture 7. µ(x)f(x). When µ is a probability measure, we say µ is a stationary distribution.

Lecture 7. µ(x)f(x). When µ is a probability measure, we say µ is a stationary distribution. Lecture 7 1 Stationary measures of a Markov chain We now study the long time behavior of a Markov Chain: in particular, the existence and uniqueness of stationary measures, and the convergence of the distribution

More information

ECE-517: Reinforcement Learning in Artificial Intelligence. Lecture 4: Discrete-Time Markov Chains

ECE-517: Reinforcement Learning in Artificial Intelligence. Lecture 4: Discrete-Time Markov Chains ECE-517: Reinforcement Learning in Artificial Intelligence Lecture 4: Discrete-Time Markov Chains September 1, 215 Dr. Itamar Arel College of Engineering Department of Electrical Engineering & Computer

More information

STOCHASTIC PROCESSES Basic notions

STOCHASTIC PROCESSES Basic notions J. Virtamo 38.3143 Queueing Theory / Stochastic processes 1 STOCHASTIC PROCESSES Basic notions Often the systems we consider evolve in time and we are interested in their dynamic behaviour, usually involving

More information

Markov Chains, Random Walks on Graphs, and the Laplacian

Markov Chains, Random Walks on Graphs, and the Laplacian Markov Chains, Random Walks on Graphs, and the Laplacian CMPSCI 791BB: Advanced ML Sridhar Mahadevan Random Walks! There is significant interest in the problem of random walks! Markov chain analysis! Computer

More information

arxiv:math/ v1 [math.pr] 24 Mar 2005

arxiv:math/ v1 [math.pr] 24 Mar 2005 The Annals of Applied Probability 2004, Vol. 14, No. 4, 2057 2089 DOI: 10.1214/105051604000000477 c Institute of Mathematical Statistics, 2004 arxiv:math/0503555v1 [math.pr] 24 Mar 2005 SPECTRAL PROPERTIES

More information

Analysis of generalized QBD queues with matrix-geometrically distributed batch arrivals and services

Analysis of generalized QBD queues with matrix-geometrically distributed batch arrivals and services manuscript No. (will be inserted by the editor) Analysis of generalized QBD queues with matrix-geometrically distributed batch arrivals and services Gábor Horváth the date of receipt and acceptance should

More information

Finite queues at the limit of saturation

Finite queues at the limit of saturation Finite queues at the limit of saturation Miklós Telek Budapest University of Technology and Economics Department of Telecommunications 1521 Budapest, Hungary Email: telek@webspnhitbmehu Miklós Vécsei Budapest

More information

Cover Page. The handle holds various files of this Leiden University dissertation

Cover Page. The handle  holds various files of this Leiden University dissertation Cover Page The handle http://hdl.handle.net/1887/39637 holds various files of this Leiden University dissertation Author: Smit, Laurens Title: Steady-state analysis of large scale systems : the successive

More information

2. Transience and Recurrence

2. Transience and Recurrence Virtual Laboratories > 15. Markov Chains > 1 2 3 4 5 6 7 8 9 10 11 12 2. Transience and Recurrence The study of Markov chains, particularly the limiting behavior, depends critically on the random times

More information

Markov Chains CK eqns Classes Hitting times Rec./trans. Strong Markov Stat. distr. Reversibility * Markov Chains

Markov Chains CK eqns Classes Hitting times Rec./trans. Strong Markov Stat. distr. Reversibility * Markov Chains Markov Chains A random process X is a family {X t : t T } of random variables indexed by some set T. When T = {0, 1, 2,... } one speaks about a discrete-time process, for T = R or T = [0, ) one has a continuous-time

More information

No class on Thursday, October 1. No office hours on Tuesday, September 29 and Thursday, October 1.

No class on Thursday, October 1. No office hours on Tuesday, September 29 and Thursday, October 1. Stationary Distributions Monday, September 28, 2015 2:02 PM No class on Thursday, October 1. No office hours on Tuesday, September 29 and Thursday, October 1. Homework 1 due Friday, October 2 at 5 PM strongly

More information

Improved Newton s method with exact line searches to solve quadratic matrix equation

Improved Newton s method with exact line searches to solve quadratic matrix equation Journal of Computational and Applied Mathematics 222 (2008) 645 654 wwwelseviercom/locate/cam Improved Newton s method with exact line searches to solve quadratic matrix equation Jian-hui Long, Xi-yan

More information

1 Ways to Describe a Stochastic Process

1 Ways to Describe a Stochastic Process purdue university cs 59000-nmc networks & matrix computations LECTURE NOTES David F. Gleich September 22, 2011 Scribe Notes: Debbie Perouli 1 Ways to Describe a Stochastic Process We will use the biased

More information

Readings: Finish Section 5.2

Readings: Finish Section 5.2 LECTURE 19 Readings: Finish Section 5.2 Lecture outline Markov Processes I Checkout counter example. Markov process: definition. -step transition probabilities. Classification of states. Example: Checkout

More information

QUASI-BIRTH-AND-DEATH PROCESSES, LATTICE PATH COUNTING, AND HYPERGEOMETRIC FUNCTIONS

QUASI-BIRTH-AND-DEATH PROCESSES, LATTICE PATH COUNTING, AND HYPERGEOMETRIC FUNCTIONS September 17, 007 QUASI-BIRTH-AND-DEATH PROCESSES, LATTICE PATH COUNTING, AND HYPERGEOMETRIC FUNCTIONS J. S. H. van Leeuwaarden 1 M. S. Squillante E. M. M. Winands 3 Abstract This paper considers a class

More information

Recap. Probability, stochastic processes, Markov chains. ELEC-C7210 Modeling and analysis of communication networks

Recap. Probability, stochastic processes, Markov chains. ELEC-C7210 Modeling and analysis of communication networks Recap Probability, stochastic processes, Markov chains ELEC-C7210 Modeling and analysis of communication networks 1 Recap: Probability theory important distributions Discrete distributions Geometric distribution

More information

An Introduction to Entropy and Subshifts of. Finite Type

An Introduction to Entropy and Subshifts of. Finite Type An Introduction to Entropy and Subshifts of Finite Type Abby Pekoske Department of Mathematics Oregon State University pekoskea@math.oregonstate.edu August 4, 2015 Abstract This work gives an overview

More information

MATH 56A: STOCHASTIC PROCESSES CHAPTER 1

MATH 56A: STOCHASTIC PROCESSES CHAPTER 1 MATH 56A: STOCHASTIC PROCESSES CHAPTER. Finite Markov chains For the sake of completeness of these notes I decided to write a summary of the basic concepts of finite Markov chains. The topics in this chapter

More information

Perron Frobenius Theory

Perron Frobenius Theory Perron Frobenius Theory Oskar Perron Georg Frobenius (1880 1975) (1849 1917) Stefan Güttel Perron Frobenius Theory 1 / 10 Positive and Nonnegative Matrices Let A, B R m n. A B if a ij b ij i, j, A > B

More information

MATH 564/STAT 555 Applied Stochastic Processes Homework 2, September 18, 2015 Due September 30, 2015

MATH 564/STAT 555 Applied Stochastic Processes Homework 2, September 18, 2015 Due September 30, 2015 ID NAME SCORE MATH 56/STAT 555 Applied Stochastic Processes Homework 2, September 8, 205 Due September 30, 205 The generating function of a sequence a n n 0 is defined as As : a ns n for all s 0 for which

More information

Convex Optimization CMU-10725

Convex Optimization CMU-10725 Convex Optimization CMU-10725 Simulated Annealing Barnabás Póczos & Ryan Tibshirani Andrey Markov Markov Chains 2 Markov Chains Markov chain: Homogen Markov chain: 3 Markov Chains Assume that the state

More information

Numerical Methods in Matrix Computations

Numerical Methods in Matrix Computations Ake Bjorck Numerical Methods in Matrix Computations Springer Contents 1 Direct Methods for Linear Systems 1 1.1 Elements of Matrix Theory 1 1.1.1 Matrix Algebra 2 1.1.2 Vector Spaces 6 1.1.3 Submatrices

More information

A Heterogeneous Two-Server Queueing System with Balking and Server Breakdowns

A Heterogeneous Two-Server Queueing System with Balking and Server Breakdowns The Eighth International Symposium on Operations Research and Its Applications (ISORA 09) Zhangjiajie, China, September 20 22, 2009 Copyright 2009 ORSC & APORC, pp. 230 244 A Heterogeneous Two-Server Queueing

More information

At the boundary states, we take the same rules except we forbid leaving the state space, so,.

At the boundary states, we take the same rules except we forbid leaving the state space, so,. Birth-death chains Monday, October 19, 2015 2:22 PM Example: Birth-Death Chain State space From any state we allow the following transitions: with probability (birth) with probability (death) with probability

More information

An M/M/1 Queue in Random Environment with Disasters

An M/M/1 Queue in Random Environment with Disasters An M/M/1 Queue in Random Environment with Disasters Noam Paz 1 and Uri Yechiali 1,2 1 Department of Statistics and Operations Research School of Mathematical Sciences Tel Aviv University, Tel Aviv 69978,

More information

CS145: Probability & Computing Lecture 18: Discrete Markov Chains, Equilibrium Distributions

CS145: Probability & Computing Lecture 18: Discrete Markov Chains, Equilibrium Distributions CS145: Probability & Computing Lecture 18: Discrete Markov Chains, Equilibrium Distributions Instructor: Erik Sudderth Brown University Computer Science April 14, 215 Review: Discrete Markov Chains Some

More information

MARKOV PROCESSES. Valerio Di Valerio

MARKOV PROCESSES. Valerio Di Valerio MARKOV PROCESSES Valerio Di Valerio Stochastic Process Definition: a stochastic process is a collection of random variables {X(t)} indexed by time t T Each X(t) X is a random variable that satisfy some

More information

Chapter 1. Introduction. 1.1 Stochastic process

Chapter 1. Introduction. 1.1 Stochastic process Chapter 1 Introduction Process is a phenomenon that takes place in time. In many practical situations, the result of a process at any time may not be certain. Such a process is called a stochastic process.

More information

A New Look at Matrix Analytic Methods

A New Look at Matrix Analytic Methods Clemson University TigerPrints All Dissertations Dissertations 8-216 A New Look at Matrix Analytic Methods Jason Joyner Clemson University Follow this and additional works at: https://tigerprints.clemson.edu/all_dissertations

More information

57:022 Principles of Design II Final Exam Solutions - Spring 1997

57:022 Principles of Design II Final Exam Solutions - Spring 1997 57:022 Principles of Design II Final Exam Solutions - Spring 1997 Part: I II III IV V VI Total Possible Pts: 52 10 12 16 13 12 115 PART ONE Indicate "+" if True and "o" if False: + a. If a component's

More information

8. Statistical Equilibrium and Classification of States: Discrete Time Markov Chains

8. Statistical Equilibrium and Classification of States: Discrete Time Markov Chains 8. Statistical Equilibrium and Classification of States: Discrete Time Markov Chains 8.1 Review 8.2 Statistical Equilibrium 8.3 Two-State Markov Chain 8.4 Existence of P ( ) 8.5 Classification of States

More information

arxiv: v1 [math.na] 9 Apr 2010

arxiv: v1 [math.na] 9 Apr 2010 Quadratic Vector Equations Federico Poloni arxiv:1004.1500v1 [math.na] 9 Apr 2010 1 Introduction In this paper, we aim to study in an unified fashion several quadratic vector and matrix equations with

More information

Lecture 15 Perron-Frobenius Theory

Lecture 15 Perron-Frobenius Theory EE363 Winter 2005-06 Lecture 15 Perron-Frobenius Theory Positive and nonnegative matrices and vectors Perron-Frobenius theorems Markov chains Economic growth Population dynamics Max-min and min-max characterization

More information

QBD Markov Chains on Binomial-Like Trees and its Application to Multilevel Feedback Queues

QBD Markov Chains on Binomial-Like Trees and its Application to Multilevel Feedback Queues QBD Markov Chains on Binomial-Like Trees and its Application to Multilevel Feedback Queues B. Van Houdt, J. Van Velthoven and C. Blondia University of Antwerp, Middelheimlaan 1, B-2020 Antwerpen, Belgium

More information

IEOR 6711, HMWK 5, Professor Sigman

IEOR 6711, HMWK 5, Professor Sigman IEOR 6711, HMWK 5, Professor Sigman 1. Semi-Markov processes: Consider an irreducible positive recurrent discrete-time Markov chain {X n } with transition matrix P (P i,j ), i, j S, and finite state space.

More information

MAA704, Perron-Frobenius theory and Markov chains.

MAA704, Perron-Frobenius theory and Markov chains. November 19, 2013 Lecture overview Today we will look at: Permutation and graphs. Perron frobenius for non-negative. Stochastic, and their relation to theory. Hitting and hitting probabilities of chain.

More information

INTRODUCTION TO MARKOV CHAINS AND MARKOV CHAIN MIXING

INTRODUCTION TO MARKOV CHAINS AND MARKOV CHAIN MIXING INTRODUCTION TO MARKOV CHAINS AND MARKOV CHAIN MIXING ERIC SHANG Abstract. This paper provides an introduction to Markov chains and their basic classifications and interesting properties. After establishing

More information

STA 624 Practice Exam 2 Applied Stochastic Processes Spring, 2008

STA 624 Practice Exam 2 Applied Stochastic Processes Spring, 2008 Name STA 624 Practice Exam 2 Applied Stochastic Processes Spring, 2008 There are five questions on this test. DO use calculators if you need them. And then a miracle occurs is not a valid answer. There

More information

Chapter 5. Continuous-Time Markov Chains. Prof. Shun-Ren Yang Department of Computer Science, National Tsing Hua University, Taiwan

Chapter 5. Continuous-Time Markov Chains. Prof. Shun-Ren Yang Department of Computer Science, National Tsing Hua University, Taiwan Chapter 5. Continuous-Time Markov Chains Prof. Shun-Ren Yang Department of Computer Science, National Tsing Hua University, Taiwan Continuous-Time Markov Chains Consider a continuous-time stochastic process

More information

Markov Chains. As part of Interdisciplinary Mathematical Modeling, By Warren Weckesser Copyright c 2006.

Markov Chains. As part of Interdisciplinary Mathematical Modeling, By Warren Weckesser Copyright c 2006. Markov Chains As part of Interdisciplinary Mathematical Modeling, By Warren Weckesser Copyright c 2006 1 Introduction A (finite) Markov chain is a process with a finite number of states (or outcomes, or

More information

Lecture #5. Dependencies along the genome

Lecture #5. Dependencies along the genome Markov Chains Lecture #5 Background Readings: Durbin et. al. Section 3., Polanski&Kimmel Section 2.8. Prepared by Shlomo Moran, based on Danny Geiger s and Nir Friedman s. Dependencies along the genome

More information

CS 798: Homework Assignment 3 (Queueing Theory)

CS 798: Homework Assignment 3 (Queueing Theory) 1.0 Little s law Assigned: October 6, 009 Patients arriving to the emergency room at the Grand River Hospital have a mean waiting time of three hours. It has been found that, averaged over the period of

More information

Stochastic modelling of epidemic spread

Stochastic modelling of epidemic spread Stochastic modelling of epidemic spread Julien Arino Department of Mathematics University of Manitoba Winnipeg Julien Arino@umanitoba.ca 19 May 2012 1 Introduction 2 Stochastic processes 3 The SIS model

More information

Stochastic modelling of epidemic spread

Stochastic modelling of epidemic spread Stochastic modelling of epidemic spread Julien Arino Centre for Research on Inner City Health St Michael s Hospital Toronto On leave from Department of Mathematics University of Manitoba Julien Arino@umanitoba.ca

More information

An Introduction to Stochastic Modeling

An Introduction to Stochastic Modeling F An Introduction to Stochastic Modeling Fourth Edition Mark A. Pinsky Department of Mathematics Northwestern University Evanston, Illinois Samuel Karlin Department of Mathematics Stanford University Stanford,

More information

Math Homework 5 Solutions

Math Homework 5 Solutions Math 45 - Homework 5 Solutions. Exercise.3., textbook. The stochastic matrix for the gambler problem has the following form, where the states are ordered as (,, 4, 6, 8, ): P = The corresponding diagram

More information

Characterizing the BMAP/MAP/1 Departure Process via the ETAQA Truncation 1

Characterizing the BMAP/MAP/1 Departure Process via the ETAQA Truncation 1 Characterizing the BMAP/MAP/1 Departure Process via the ETAQA Truncation 1 Qi Zhang Armin Heindl Evgenia Smirni Department of Computer Science Computer Networks and Comm Systems Department of Computer

More information

On the Class of Quasi-Skip Free Processes: Stability & Explicit solutions when successively lumpable

On the Class of Quasi-Skip Free Processes: Stability & Explicit solutions when successively lumpable On the Class of Quasi-Skip Free Processes: Stability & Explicit solutions when successively lumpable DRAFT 2012-Nov-29 - comments welcome, do not cite or distribute without permission Michael N Katehakis

More information

Stochastic Processes

Stochastic Processes Stochastic Processes 8.445 MIT, fall 20 Mid Term Exam Solutions October 27, 20 Your Name: Alberto De Sole Exercise Max Grade Grade 5 5 2 5 5 3 5 5 4 5 5 5 5 5 6 5 5 Total 30 30 Problem :. True / False

More information

Structured Markov Chains

Structured Markov Chains Structured Markov Chains Ivo Adan and Johan van Leeuwaarden Where innovation starts Book on Analysis of structured Markov processes (arxiv:1709.09060) I Basic methods Basic Markov processes Advanced Markov

More information

Birth-death chain. X n. X k:n k,n 2 N k apple n. X k: L 2 N. x 1:n := x 1,...,x n. E n+1 ( x 1:n )=E n+1 ( x n ), 8x 1:n 2 X n.

Birth-death chain. X n. X k:n k,n 2 N k apple n. X k: L 2 N. x 1:n := x 1,...,x n. E n+1 ( x 1:n )=E n+1 ( x n ), 8x 1:n 2 X n. Birth-death chains Birth-death chain special type of Markov chain Finite state space X := {0,...,L}, with L 2 N X n X k:n k,n 2 N k apple n Random variable and a sequence of variables, with and A sequence

More information

Non-homogeneous random walks on a semi-infinite strip

Non-homogeneous random walks on a semi-infinite strip Non-homogeneous random walks on a semi-infinite strip Chak Hei Lo Joint work with Andrew R. Wade World Congress in Probability and Statistics 11th July, 2016 Outline Motivation: Lamperti s problem Our

More information

Statistics 992 Continuous-time Markov Chains Spring 2004

Statistics 992 Continuous-time Markov Chains Spring 2004 Summary Continuous-time finite-state-space Markov chains are stochastic processes that are widely used to model the process of nucleotide substitution. This chapter aims to present much of the mathematics

More information

Homework 3 posted, due Tuesday, November 29.

Homework 3 posted, due Tuesday, November 29. Classification of Birth-Death Chains Tuesday, November 08, 2011 2:02 PM Homework 3 posted, due Tuesday, November 29. Continuing with our classification of birth-death chains on nonnegative integers. Last

More information

Quantitative Model Checking (QMC) - SS12

Quantitative Model Checking (QMC) - SS12 Quantitative Model Checking (QMC) - SS12 Lecture 06 David Spieler Saarland University, Germany June 4, 2012 1 / 34 Deciding Bisimulations 2 / 34 Partition Refinement Algorithm Notation: A partition P over

More information

Chapter 16 focused on decision making in the face of uncertainty about one future

Chapter 16 focused on decision making in the face of uncertainty about one future 9 C H A P T E R Markov Chains Chapter 6 focused on decision making in the face of uncertainty about one future event (learning the true state of nature). However, some decisions need to take into account

More information

The SIS and SIR stochastic epidemic models revisited

The SIS and SIR stochastic epidemic models revisited The SIS and SIR stochastic epidemic models revisited Jesús Artalejo Faculty of Mathematics, University Complutense of Madrid Madrid, Spain jesus_artalejomat.ucm.es BCAM Talk, June 16, 2011 Talk Schedule

More information

Markov Chains Handout for Stat 110

Markov Chains Handout for Stat 110 Markov Chains Handout for Stat 0 Prof. Joe Blitzstein (Harvard Statistics Department) Introduction Markov chains were first introduced in 906 by Andrey Markov, with the goal of showing that the Law of

More information

MATH36001 Perron Frobenius Theory 2015

MATH36001 Perron Frobenius Theory 2015 MATH361 Perron Frobenius Theory 215 In addition to saying something useful, the Perron Frobenius theory is elegant. It is a testament to the fact that beautiful mathematics eventually tends to be useful,

More information

Necessary and sufficient conditions for strong R-positivity

Necessary and sufficient conditions for strong R-positivity Necessary and sufficient conditions for strong R-positivity Wednesday, November 29th, 2017 The Perron-Frobenius theorem Let A = (A(x, y)) x,y S be a nonnegative matrix indexed by a countable set S. We

More information

On Finding Optimal Policies for Markovian Decision Processes Using Simulation

On Finding Optimal Policies for Markovian Decision Processes Using Simulation On Finding Optimal Policies for Markovian Decision Processes Using Simulation Apostolos N. Burnetas Case Western Reserve University Michael N. Katehakis Rutgers University February 1995 Abstract A simulation

More information

6.207/14.15: Networks Lectures 4, 5 & 6: Linear Dynamics, Markov Chains, Centralities

6.207/14.15: Networks Lectures 4, 5 & 6: Linear Dynamics, Markov Chains, Centralities 6.207/14.15: Networks Lectures 4, 5 & 6: Linear Dynamics, Markov Chains, Centralities 1 Outline Outline Dynamical systems. Linear and Non-linear. Convergence. Linear algebra and Lyapunov functions. Markov

More information

The cost/reward formula has two specific widely used applications:

The cost/reward formula has two specific widely used applications: Applications of Absorption Probability and Accumulated Cost/Reward Formulas for FDMC Friday, October 21, 2011 2:28 PM No class next week. No office hours either. Next class will be 11/01. The cost/reward

More information

Lectures on Markov Chains

Lectures on Markov Chains Lectures on Markov Chains David M. McClendon Department of Mathematics Ferris State University 2016 edition 1 Contents Contents 2 1 Markov chains 4 1.1 The definition of a Markov chain.....................

More information

Irreducibility. Irreducible. every state can be reached from every other state For any i,j, exist an m 0, such that. Absorbing state: p jj =1

Irreducibility. Irreducible. every state can be reached from every other state For any i,j, exist an m 0, such that. Absorbing state: p jj =1 Irreducibility Irreducible every state can be reached from every other state For any i,j, exist an m 0, such that i,j are communicate, if the above condition is valid Irreducible: all states are communicate

More information

ECEN 689 Special Topics in Data Science for Communications Networks

ECEN 689 Special Topics in Data Science for Communications Networks ECEN 689 Special Topics in Data Science for Communications Networks Nick Duffield Department of Electrical & Computer Engineering Texas A&M University Lecture 8 Random Walks, Matrices and PageRank Graphs

More information

Stochastic Models: Markov Chains and their Generalizations

Stochastic Models: Markov Chains and their Generalizations Scuola di Dottorato in Scienza ed Alta Tecnologia Dottorato in Informatica Universita di Torino Stochastic Models: Markov Chains and their Generalizations Gianfranco Balbo e Andras Horvath Outline Introduction

More information

SIMILAR MARKOV CHAINS

SIMILAR MARKOV CHAINS SIMILAR MARKOV CHAINS by Phil Pollett The University of Queensland MAIN REFERENCES Convergence of Markov transition probabilities and their spectral properties 1. Vere-Jones, D. Geometric ergodicity in

More information

On a quadratic matrix equation associated with an M-matrix

On a quadratic matrix equation associated with an M-matrix Article Submitted to IMA Journal of Numerical Analysis On a quadratic matrix equation associated with an M-matrix Chun-Hua Guo Department of Mathematics and Statistics, University of Regina, Regina, SK

More information