Analysis of Markov Reward Models with Partial Reward Loss Based on a Time Reverse Approach

Similar documents
Analysis of Inhomogeneous Markov Reward Models

Analysis of inhomogeneous Markov reward models

Markovian techniques for performance analysis of computer and communication systems

MATH 307 Test 1 Study Guide

STOCHASTIC PROCESSES Basic notions

Alternative Characterization of Ergodicity for Doubly Stochastic Chains

PLEASE SCROLL DOWN FOR ARTICLE

RECURSION EQUATION FOR

IEOR 6711, HMWK 5, Professor Sigman

Markov decision processes and interval Markov chains: exploiting the connection

A FAST MATRIX-ANALYTIC APPROXIMATION FOR THE TWO CLASS GI/G/1 NON-PREEMPTIVE PRIORITY QUEUE

Definition and Properties

Language Acquisition and Parameters: Part II

Analysis of fluid queues in saturation with additive decomposition

Two-sided Bounds for the Convergence Rate of Markov Chains

Outlines. Discrete Time Markov Chain (DTMC) Continuous Time Markov Chain (CTMC)

Finite queues at the limit of saturation

Linear Algebra Solutions 1

(b) What is the variance of the time until the second customer arrives, starting empty, assuming that we measure time in minutes?

Analysis of generalized QBD queues with matrix-geometrically distributed batch arrivals and services

Stochastic modelling of epidemic spread

Markov chains. 1 Discrete time Markov chains. c A. J. Ganesh, University of Bristol, 2015

10. Linear Systems of ODEs, Matrix multiplication, superposition principle (parts of sections )

Lecture 20 : Markov Chains

Numerical Analysis of Large Markov. Miklos Telek. Department of Telecommunications, Sandor Racz. Department of Telecommunications and Telematics,

Acyclic discrete phase type distributions: properties and a parameter estimation algorithm

MAT 242 CHAPTER 4: SUBSPACES OF R n

Linear Ordinary Differential Equations

A Numerical Method for the Evaluation of the Distribution of Cumulative Reward Till Exit of a Subset of Transient States of a Markov Reward Model

Modelling in Systems Biology

Packet Loss Analysis of Load-Balancing Switch with ON/OFF Input Processes

Stochastic Shortest Path Problems

Markov Chains and Related Matters

Recap. Probability, stochastic processes, Markov chains. ELEC-C7210 Modeling and analysis of communication networks

EXTERNALLY AND INTERNALLY POSITIVE TIME-VARYING LINEAR SYSTEMS

Linear Algebra review Powers of a diagonalizable matrix Spectral decomposition

On Backward Product of Stochastic Matrices

. a m1 a mn. a 1 a 2 a = a n

Robotics & Automation. Lecture 06. Serial Kinematic Chain, Forward Kinematics. John T. Wen. September 11, 2008

Models of Molecular Evolution


The Scale Factor: A New Degree of Freedom in Phase Type Approximation

Chapter 3: Birth and Death processes

LINEAR SYSTEMS OF GENERALIZED ORDINARY DIFFERENTIAL EQUATIONS

A Model of Human Capital Accumulation and Occupational Choices. A simplified version of Keane and Wolpin (JPE, 1997)

CSL105: Discrete Mathematical Structures. Ragesh Jaiswal, CSE, IIT Delhi

Random Walks on Graphs. One Concrete Example of a random walk Motivation applications

Statistical analysis of peer-to-peer live streaming traffic

Fluid models in performance analysis

On Minimal Representations of Rational Arrival Processes

Curvature of Digital Curves

Direct Methods for Solving Linear Systems. Matrix Factorization

MAT SYS 5120 (Winter 2012) Assignment 5 (not to be submitted) There are 4 questions.

Deterministic Dynamic Programming

Uniformly Uniformly-ergodic Markov chains and BSDEs

6 Continuous-Time Birth and Death Chains

Uncertainty calculations for estimates of project durations

Impulsive stabilization of two kinds of second-order linear delay differential equations

Lecturer: Olga Galinina

Matrix Algebra & Elementary Matrices

Perron Frobenius Theory

Materials engineering Collage \\ Ceramic & construction materials department Numerical Analysis \\Third stage by \\ Dalya Hekmat

University of Oxford. Statistical Methods Autocorrelation. Identification and Estimation

Moment-based Availability Prediction for Bike-Sharing Systems

Statistics 992 Continuous-time Markov Chains Spring 2004

CDA6530: Performance Models of Computers and Networks. Chapter 3: Review of Practical Stochastic Processes

Data analysis and stochastic modeling

Lecture 5. If we interpret the index n 0 as time, then a Markov chain simply requires that the future depends only on the present and not on the past.

Stochastic modelling of epidemic spread

Stochastic processes. MAS275 Probability Modelling. Introduction and Markov chains. Continuous time. Markov property

Markov Processes. Stochastic process. Markov process

Part I Stochastic variables and Markov chains

= P{X 0. = i} (1) If the MC has stationary transition probabilities then, = i} = P{X n+1

Simulation of Contrast Agent Enhanced Ultrasound Imaging based on Field II

6 Linear Equation. 6.1 Equation with constant coefficients

CS 246 Review of Linear Algebra 01/17/19

T. Liggett Mathematics 171 Final Exam June 8, 2011

Stochastic Petri Net

TCOM 501: Networking Theory & Fundamentals. Lecture 6 February 19, 2003 Prof. Yannis A. Korilis

Upon successful completion of MATH 220, the student will be able to:

Round-off error propagation and non-determinism in parallel applications

Linear Algebra review Powers of a diagonalizable matrix Spectral decomposition

Irreducibility. Irreducible. every state can be reached from every other state For any i,j, exist an m 0, such that. Absorbing state: p jj =1

Discrete time Markov chains. Discrete Time Markov Chains, Definition and classification. Probability axioms and first results

Computing Battery Lifetime Distributions

Non-homogeneous random walks on a semi-infinite strip

Math 1553, Introduction to Linear Algebra

Lecture 5: Random Walks and Markov Chain

Solving Dynamic Equations: The State Transition Matrix

The story of the film so far... Mathematics for Informatics 4a. Continuous-time Markov processes. Counting processes

M. Matrices and Linear Algebra

On asymptotic behavior of a finite Markov chain

Example: physical systems. If the state space. Example: speech recognition. Context can be. Example: epidemics. Suppose each infected

MATH 2030: MATRICES ,, a m1 a m2 a mn If the columns of A are the vectors a 1, a 2,...,a n ; A is represented as A 1. .

General Lorentz Boost Transformations, Acting on Some Important Physical Quantities

On prediction and density estimation Peter McCullagh University of Chicago December 2004

The Boundary Problem: Markov Chain Solution

Kevin James. MTHSC 3110 Section 2.1 Matrix Operations

Markov Chains Absorption Hamid R. Rabiee

Outline. Linear maps. 1 Vector addition is commutative: summands can be swapped: 2 addition is associative: grouping of summands is irrelevant:

Transcription:

Analysis of Markov Reward Models with Partial Reward Loss Based on a Time Reverse Approach Gábor Horváth, Miklós Telek Technical University of Budapest, 1521 Budapest, Hungary {hgabor,telek}@webspn.hit.bme.hu M Telek, Markov Anniversary Meeting, June 2006. Analysis of Markov Reward Models with Partial Reward Loss - p. 1/15

Markov Reward models with reward loss The difficulty of time forward approach The time reverse analysis approach Properties of the obtained solution s M Telek, Markov Anniversary Meeting, June 2006. Analysis of Markov Reward Models with Partial Reward Loss - p. 2/15

Markov Reward models without reward loss Markov reward models (MRM) Markov Reward models without reward loss Markov Reward models with total reward loss Markov Reward models with partial reward loss a finite state CTMC, non negative reward rates (r i ), performance measures: reward accumulated up to time t, time to accumulate reward w. B(t) r i r k r j r k Z(t) k j i t t M Telek, Markov Anniversary Meeting, June 2006. Analysis of Markov Reward Models with Partial Reward Loss - p. 3/15

Markov Reward models with total reward loss Markov Reward models without reward loss Markov Reward models with total reward loss Markov Reward models with partial reward loss We consider first order MRM (deterministic dependence on Z(t)), without impulse reward, but with potential reward loss at state transition. B(t) r i r k r j Z(t) k j r k t i t M Telek, Markov Anniversary Meeting, June 2006. Analysis of Markov Reward Models with Partial Reward Loss - p. 4/15

Markov Reward models with partial reward los In case of partial reward loss: Markov Reward models without reward loss Markov Reward models with total reward loss Markov Reward models with partial reward loss α i remaining portion of reward when leaving state i, the lost reward is proportional to: total accumulated reward partial total loss, reward accumulated in the last state partial incremental loss. M Telek, Markov Anniversary Meeting, June 2006. Analysis of Markov Reward Models with Partial Reward Loss - p. 5/15

Markov Reward models with partial reward los Markov Reward models without reward loss Markov Reward models with total reward loss Markov Reward models with partial reward loss In case of partial reward loss: α i remaining portion of reward when leaving state i, the lost reward is proportional to: total accumulated reward partial total loss, reward accumulated in the last state partial incremental loss. B(t) Z(t) k r i r k B(T 1 )α i r j B(T 2 )α k r k B(t) B(T3 )α j r k r i r k α k r j α j α j [B(T3 ) B(T 2 )] r i α i B(T 2 ) t Z(t) t k r j r k j j i i T 1 T 2 T 3 t T 1 T 2 T 3 t M Telek, Markov Anniversary Meeting, June 2006. Analysis of Markov Reward Models with Partial Reward Loss - p. 5/15

Time forward approach Time forward approach Time reverse approach Possible interpretation: Reduced (r i α i ) reward accumulation up to the last state transition, and total (r i ) reward accumulation in the last state without reward loss. B(t) r i r i α i r k r k α k r j r j α j r k α j [B(T3 ) B(T 2 )] B(T 2 ) Z(t) k t j i T 1 T 2 T 3 t M Telek, Markov Anniversary Meeting, June 2006. Analysis of Markov Reward Models with Partial Reward Loss - p. 6/15

Time forward approach Time forward approach Time reverse approach Possible interpretation: Reduced (r i α i ) reward accumulation up to the last state transition, and total (r i ) reward accumulation in the last state without reward loss. B(t) r i r i α i r k r k α k r j r j α j r k α j [B(T3 ) B(T 2 )] B(T 2 ) Z(t) k t j i T 1 T 2 T 3 t Unfortunately, the last state transition before time T is not a stopping time. M Telek, Markov Anniversary Meeting, June 2006. Analysis of Markov Reward Models with Partial Reward Loss - p. 6/15

Time reverse approach Time forward approach Time reverse approach Behaviour of the time reverse process: Inhomogeneous CTMC with initial probability γ (0) = γ(t) and generator Q(τ) = { q ij (τ)}, where q ij (τ) = γ j (T τ) γ i (T τ) q ji if i j, γ k (T τ) γ i (T τ) q ki if i = j. k S,k i M Telek, Markov Anniversary Meeting, June 2006. Analysis of Markov Reward Models with Partial Reward Loss - p. 7/15

Time reverse approach Time forward approach Time reverse approach Behaviour of the time reverse process: Inhomogeneous CTMC with initial probability γ (0) = γ(t) and generator Q(τ) = { q ij (τ)}, where q ij (τ) = γ j (T τ) γ i (T τ) q ji if i j, γ k (T τ) γ i (T τ) q ki if i = j. k S,k i Total (r i ) reward accumulation in the first state, and reduced (r i α i ) reward accumulation in all consecutive states without reward loss. M Telek, Markov Anniversary Meeting, June 2006. Analysis of Markov Reward Models with Partial Reward Loss - p. 7/15

Time reverse approach Potential model description: Time reverse approach duplicate the state space to describe the total reward accumulation in the first state (r i ), and the reduced reward accumulation in all further states (r i α i ). M Telek, Markov Anniversary Meeting, June 2006. Analysis of Markov Reward Models with Partial Reward Loss - p. 8/15

Time reverse approach Potential model description: Time reverse approach duplicate the state space to describe the total reward accumulation in the first state (r i ), and the reduced reward accumulation in all further states (r i α i ). π (0) = [γ(t), 0], Q (τ) = Q D (τ) Q(τ) Q D (τ) 0 Q(τ), R = R 0 0 R α M Telek, Markov Anniversary Meeting, June 2006. Analysis of Markov Reward Models with Partial Reward Loss - p. 8/15

Inhomogeneous differential Introducing Y i (τ, w) = Pr( B(τ) w, Z (τ) = i) Inhomogeneous differential Homogeneous differential Block structure of the differential Moments of accumulated reward we can apply the analysis approach available for inhomogeneous MRMs. It is based on the solution of the inhomogeneous partial differential τ Y (τ, w) + Y (τ, w)r = Y (τ, w) Q(τ), w where Y (τ, w) = { Y i (τ, w)}. M Telek, Markov Anniversary Meeting, June 2006. Analysis of Markov Reward Models with Partial Reward Loss - p. 9/15

Inhomogeneous differential Introducing Y i (τ, w) = Pr( B(τ) w, Z (τ) = i) Inhomogeneous differential Homogeneous differential Block structure of the differential Moments of accumulated reward we can apply the analysis approach available for inhomogeneous MRMs. It is based on the solution of the inhomogeneous partial differential τ Y (τ, w) + Y (τ, w)r = Y (τ, w) Q(τ), w where Y (τ, w) = { Y i (τ, w)}. But a drawback of this approach is that it requires the computation of Q(τ). M Telek, Markov Anniversary Meeting, June 2006. Analysis of Markov Reward Models with Partial Reward Loss - p. 9/15

Homogeneous differential To overcome this drawback we introduce the conditional distribution of reward accumulated by the reverse process V i (τ, w) = Pr( B(τ) w Z (τ) = i) Inhomogeneous differential Homogeneous differential Block structure of the differential Moments of accumulated reward and the row vector V (τ, w) = { V i (τ, w)}. M Telek, Markov Anniversary Meeting, June 2006. Analysis of Markov Reward Models with Partial Reward Loss - p. 10/15

Homogeneous differential To overcome this drawback we introduce the conditional distribution of reward accumulated by the reverse process V i (τ, w) = Pr( B(τ) w Z (τ) = i) Inhomogeneous differential Homogeneous differential Block structure of the differential Moments of accumulated reward and the row vector V (τ, w) = { V i (τ, w)}. Using this performance measure we have to solve τ V (τ, w) + V (τ, w)r = V (τ, w)q T, w where Q T is the transpose of Q. M Telek, Markov Anniversary Meeting, June 2006. Analysis of Markov Reward Models with Partial Reward Loss - p. 10/15

Block structure of the differential Utilizing the special block structure of the Q (τ) and the R matrices (of size 2#S) we can obtain two homogeneous partial differential s of size #S: Inhomogeneous differential Homogeneous differential Block structure of the differential Moments of accumulated reward and τ τ X1(τ, w) + X1(τ, w)r = X1(τ, w)q D, w X2(τ, w)+ X2(τ, w)r α = X1(τ, w)(q Q D ) T + X2(τ, w)q T, w M Telek, Markov Anniversary Meeting, June 2006. Analysis of Markov Reward Models with Partial Reward Loss - p. 11/15

Moments of accumulated reward The analysis approach available for inhomogeneous MRMs allows to describe the moments of IMRMs with an inhomogeneous ordinary differential. Inhomogeneous differential Homogeneous differential Block structure of the differential Moments of accumulated reward Similar to the reward distribution case, this approach is also applicable for our model, but it requires the the computation of Q(τ). M Telek, Markov Anniversary Meeting, June 2006. Analysis of Markov Reward Models with Partial Reward Loss - p. 12/15

Moments of accumulated reward The analysis approach available for inhomogeneous MRMs allows to describe the moments of IMRMs with an inhomogeneous ordinary differential. Inhomogeneous differential Homogeneous differential Block structure of the differential Moments of accumulated reward Similar to the reward distribution case, this approach is also applicable for our model, but it requires the the computation of Q(τ). Using similar state dependent moment measures we obtain homogeneous ordinary differential s and d dτ M1 (n) (τ) = n M1 (n 1) (τ)r + M1 (n) (τ)q D, d dτ M2 (n) (τ) = n M2 (n 1) (τ)r α + M1 (n) (τ)(q Q D ) T + M2 (n) (τ)q M Telek, Markov Anniversary Meeting, June 2006. Analysis of Markov Reward Models with Partial Reward Loss - p. 12/15

Randomization based numerical method The ordinary differential with constant coefficients allows to compose a randomization based numerical method. Randomization based numerical method and where M1 (n) (τ) = τ n e R n E D (τ), M2 (n) (τ) = n!d n k=0 λτ (λτ)k e k! D (n) (k), e (I A k D ) n = 0 D (n) (k) = 0 k n, n 1 D (n 1) (k 1)S α + D (n) (k 1)A+ ) e S n A k 1 n D (A A D ) k > n, n 1 ( k 1 n M Telek, Markov Anniversary Meeting, June 2006. Analysis of Markov Reward Models with Partial Reward Loss - p. 13/15

Numerical Example α N =0.5 α N 1 =0.5 α N 2 =0.5 α 0 =1 r N =Nr r N 1 =(N 1)r r N 2 =(N 2)r r 0 =0 10 N N λ (N 1)λ N 1 N 2 λ 0 1 0.1 Numerical Example ρ σ σ σ M σ 0.01 0.001 r M =0 α M =1 Structure of the Markov chain 0.0001 1e-05 1st moment 2nd moment 3rd moment 4th moment 5th moment 0.5 1 1.5 2 2.5 3 3.5 4 4.5 5 Moments of the accumulated reward With parameters N = 500000, λ = 0.000004, σ = 1.5, ρ = 0.1, r = 0.000002, α = 0.5, M Telek, Markov Anniversary Meeting, June 2006. Analysis of Markov Reward Models with Partial Reward Loss - p. 14/15

The analysis of partial loss MRM is usually rather complex. We propose an analysis method with the following features: M Telek, Markov Anniversary Meeting, June 2006. Analysis of Markov Reward Models with Partial Reward Loss - p. 15/15

The analysis of partial loss MRM is usually rather complex. We propose an analysis method with the following features: non stopping time time reverse approach M Telek, Markov Anniversary Meeting, June 2006. Analysis of Markov Reward Models with Partial Reward Loss - p. 15/15

The analysis of partial loss MRM is usually rather complex. We propose an analysis method with the following features: non stopping time time reverse approach inhomogeneous differential proper performance measure, M Telek, Markov Anniversary Meeting, June 2006. Analysis of Markov Reward Models with Partial Reward Loss - p. 15/15

The analysis of partial loss MRM is usually rather complex. We propose an analysis method with the following features: non stopping time time reverse approach inhomogeneous differential proper performance measure, partial differential ordinary differential s, M Telek, Markov Anniversary Meeting, June 2006. Analysis of Markov Reward Models with Partial Reward Loss - p. 15/15

The analysis of partial loss MRM is usually rather complex. We propose an analysis method with the following features: non stopping time time reverse approach inhomogeneous differential proper performance measure, partial differential ordinary differential s, numerical stability, error control randomization based analysis. M Telek, Markov Anniversary Meeting, June 2006. Analysis of Markov Reward Models with Partial Reward Loss - p. 15/15

The analysis of partial loss MRM is usually rather complex. We propose an analysis method with the following features: non stopping time time reverse approach inhomogeneous differential proper performance measure, partial differential ordinary differential s, numerical stability, error control randomization based analysis. Thanks for your attention. M Telek, Markov Anniversary Meeting, June 2006. Analysis of Markov Reward Models with Partial Reward Loss - p. 15/15