THE VARIANCE CONSTANT FOR THE ACTUAL WAITING TIME OF THE PH/PH/1 QUEUE. By Mogens Bladt National University of Mexico

Similar documents
Central limit theorems for ergodic continuous-time Markov chains with applications to single birth processes

PROBABILISTIC METHODS FOR A LINEAR REACTION-HYPERBOLIC SYSTEM WITH CONSTANT COEFFICIENTS

NEW FRONTIERS IN APPLIED PROBABILITY

Geometric ρ-mixing property of the interarrival times of a stationary Markovian Arrival Process

Rare event simulation for the ruin problem with investments via importance sampling and duality

Analysis of an Infinite-Server Queue with Markovian Arrival Streams

Exact Simulation of the Stationary Distribution of M/G/c Queues

ABC methods for phase-type distributions with applications in insurance risk problems

HEAVY-TRAFFIC EXTREME-VALUE LIMITS FOR QUEUES

Recap. Probability, stochastic processes, Markov chains. ELEC-C7210 Modeling and analysis of communication networks

Stability and Rare Events in Stochastic Models Sergey Foss Heriot-Watt University, Edinburgh and Institute of Mathematics, Novosibirsk

RELATING TIME AND CUSTOMER AVERAGES FOR QUEUES USING FORWARD COUPLING FROM THE PAST

A NEW PROOF OF THE WIENER HOPF FACTORIZATION VIA BASU S THEOREM

On the Central Limit Theorem for an ergodic Markov chain

LIMITS FOR QUEUES AS THE WAITING ROOM GROWS. Bell Communications Research AT&T Bell Laboratories Red Bank, NJ Murray Hill, NJ 07974

2 light traffic derivatives for the GI/G/ queue. We shall see that our proof of analyticity is mainly based on some recursive formulas very similar to

Contents Preface The Exponential Distribution and the Poisson Process Introduction to Renewal Theory

Stochastic Volatility and Correction to the Heat Equation

Regenerative Processes. Maria Vlasiou. June 25, 2018

MATH 56A: STOCHASTIC PROCESSES CHAPTER 6

N.G.Bean, D.A.Green and P.G.Taylor. University of Adelaide. Adelaide. Abstract. process of an MMPP/M/1 queue is not a MAP unless the queue is a

Characterizations on Heavy-tailed Distributions by Means of Hazard Rate

Stochastic relations of random variables and processes

On the martingales obtained by an extension due to Saisho, Tanemura and Yor of Pitman s theorem

Supermodular ordering of Poisson arrays

On Reflecting Brownian Motion with Drift

Stability of the Defect Renewal Volterra Integral Equations

Stationary Probabilities of Markov Chains with Upper Hessenberg Transition Matrices

IEOR 6711, HMWK 5, Professor Sigman

Operations Research Letters. Instability of FIFO in a simple queueing system with arbitrarily low loads

The Azéma-Yor Embedding in Non-Singular Diffusions

An M/M/1 Queue in Random Environment with Disasters

Overload Analysis of the PH/PH/1/K Queue and the Queue of M/G/1/K Type with Very Large K

MS&E 321 Spring Stochastic Systems June 1, 2013 Prof. Peter W. Glynn Page 1 of 7

Some open problems related to stability. 1 Multi-server queue with First-Come-First-Served discipline

Departure Processes of a Tandem Network

MS&E 321 Spring Stochastic Systems June 1, 2013 Prof. Peter W. Glynn Page 1 of 10. x n+1 = f(x n ),

An EM algorithm for Batch Markovian Arrival Processes and its comparison to a simpler estimation procedure

One important issue in the study of queueing systems is to characterize departure processes. Study on departure processes was rst initiated by Burke (

A note on the GI/GI/ system with identical service and interarrival-time distributions

Advanced Queueing Theory

EXPLICIT DISTRIBUTIONAL RESULTS IN PATTERN FORMATION. By V. T. Stefanov 1 and A. G. Pakes University of Western Australia

Busy Period Length and Higher level First passage probabilities

On the static assignment to parallel servers

Data analysis and stochastic modeling

ANNALES DE L I. H. P., SECTION B

IDLE PERIODS FOR THE FINITE G/M/1 QUEUE AND THE DEFICIT AT RUIN FOR A CASH RISK MODEL WITH CONSTANT DIVIDEND BARRIER

Asymptotics for Polling Models with Limited Service Policies

Markov Chains. X(t) is a Markov Process if, for arbitrary times t 1 < t 2 <... < t k < t k+1. If X(t) is discrete-valued. If X(t) is continuous-valued

Adventures in Stochastic Processes

Class 11 Non-Parametric Models of a Service System; GI/GI/1, GI/GI/n: Exact & Approximate Analysis.

Kesten s power law for difference equations with coefficients induced by chains of infinite order

IEOR 8100: Topics in OR: Asymptotic Methods in Queueing Theory. Fall 2009, Professor Whitt. Class Lecture Notes: Wednesday, September 9.

An Introduction to Probability Theory and Its Applications

Stochastic process. X, a series of random variables indexed by t

Convergence rates in weighted L 1 spaces of kernel density estimators for linear processes

HITTING TIME IN AN ERLANG LOSS SYSTEM

M/M/3/3 AND M/M/4/4 RETRIAL QUEUES. Tuan Phung-Duc, Hiroyuki Masuyama, Shoji Kasahara and Yutaka Takahashi

Faithful couplings of Markov chains: now equals forever

A Queueing System with Queue Length Dependent Service Times, with Applications to Cell Discarding in ATM Networks

COMPOSITION SEMIGROUPS AND RANDOM STABILITY. By John Bunge Cornell University

Lecture 10: Semi-Markov Type Processes

Asymptotic Irrelevance of Initial Conditions for Skorohod Reflection Mapping on the Nonnegative Orthant

Stochastic models in product form: the (E)RCAT methodology

Existence, Uniqueness and Stability of Invariant Distributions in Continuous-Time Stochastic Models

Uniformly Uniformly-ergodic Markov chains and BSDEs

Convergence Rates for Renewal Sequences

THIELE CENTRE. The M/M/1 queue with inventory, lost sale and general lead times. Mohammad Saffari, Søren Asmussen and Rasoul Haji

Waiting-Time Distribution of a Discrete-time Multiserver Queue with Correlated. Arrivals and Deterministic Service Times: D-MAP/D/k System

Queueing Networks and Insensitivity

STABILIZATION OF AN OVERLOADED QUEUEING NETWORK USING MEASUREMENT-BASED ADMISSION CONTROL

Duality and Other Results for M/G/1 and GI/M/1 Queues, via a New Ballot Theorem

Pointwise convergence rates and central limit theorems for kernel density estimators in linear processes

1 Continuous-time chains, finite state space

Renewal theory and its applications

Logarithmic scaling of planar random walk s local times

PITMAN S 2M X THEOREM FOR SKIP-FREE RANDOM WALKS WITH MARKOVIAN INCREMENTS

Analysis of an M/G/1 queue with customer impatience and an adaptive arrival process

Distribution of the first ladder height of a stationary risk process perturbed by α-stable Lévy motion

Generalized Resolvents and Harris Recurrence of Markov Processes

Jae-Ho Lee. Markov Random Walks Driven by General Markov Chains and Their Applications to Semi-Markov Queues

Fluid Heuristics, Lyapunov Bounds and E cient Importance Sampling for a Heavy-tailed G/G/1 Queue

arxiv: v2 [math.pr] 24 Mar 2018

Asymptotic Coupling of an SPDE, with Applications to Many-Server Queues

Prime numbers and Gaussian random walks

Synchronized Queues with Deterministic Arrivals

Asymptotics and Simulation of Heavy-Tailed Processes

A tandem queueing model with coupled processors

WEIBULL RENEWAL PROCESSES

LECTURE #6 BIRTH-DEATH PROCESS

The Distributions of Stopping Times For Ordinary And Compound Poisson Processes With Non-Linear Boundaries: Applications to Sequential Estimation.

Finite-time Ruin Probability of Renewal Model with Risky Investment and Subexponential Claims

QUEUING SYSTEM. Yetunde Folajimi, PhD

CS 798: Homework Assignment 3 (Queueing Theory)

57:022 Principles of Design II Final Exam Solutions - Spring 1997

A Joining Shortest Queue with MAP Inputs

The Kemeny constant of a Markov chain

(b) What is the variance of the time until the second customer arrives, starting empty, assuming that we measure time in minutes?

Lectures on Stochastic Stability. Sergey FOSS. Heriot-Watt University. Lecture 4. Coupling and Harris Processes

A regeneration proof of the central limit theorem for uniformly ergodic Markov chains

Transcription:

The Annals of Applied Probability 1996, Vol. 6, No. 3, 766 777 THE VARIANCE CONSTANT FOR THE ACTUAL WAITING TIME OF THE PH/PH/1 QUEUE By Mogens Bladt National University of Mexico In this paper we consider the asymptotic time average variance of the actual waiting time of the PH/PH/1 queue. To this end Poisson s equation plays a major role, since the solution to Poisson s equation is shown to be intimately connected with the desired variance. We derive explicit solutions to Poisson s equation using phase-type methodology, and present some numerical examples. 1. Introduction. Let W n denote the actual waiting time in the PH/PH/1 queue, that is, the waiting time the nth arriving customer faces upon arrival until service is initiated. We are interested in the asymptotic behavior of n 1 1 f W n k k= To this end we apply Poisson s equation to show that this functional converges to the same limit as a related martingale; this martingale, in turn, converges to a normally distributed random variable. In particular, we are interested in calculating the variance of this limiting normal distribution, and we shall refer to this value as the (time average) variance constant. The work of this paper can be viewed as an extension of Glynn [8]. For discrete-time Markov chains, the relevant basic facts can be found in Neveu [13], Revuz [15], Nummelin [14] or Glynn [8]. For the continuous-time case and its application to calculating the variance constant of the virtual waiting time, see Bladt and Asmussen [7] for details. All results in this paper have been proved in [8] for the M/G/1 queue (and in [7] for the MAP/MMPH/1 virtual waiting time; MAP stands for Markovian arrival process and MMPH for service times that are Markov modulated phase-type, that is, phase-type distributed, where the parameters in the distribution depend on an underlying Markovian environment). In order to extend these results to the PH/PH/1 queue (interarrival times and service times are phase-type distributed), we need to put the theory into a slightly more general framework. The paper is organized as follows. In Section 2 we introduce the PH/PH/1 queue, and present a central limit theorem in order to obtain the asymptotic variance. In Section 3 explicit calculations are performed using ladder heights and renewal theory for phase-type distributions. In Section 4 we present two numerical examples. Received October 1994; revised January 1996. AMS 1991 subject classifications. Primary 6K25; secondary 6K5, 6G42. Key words and phrases. Actual waiting time, phase-type distribution, Poisson s equation, martingale, asymptotic time average variance, numerical methods. 766

VARIANCE CONSTANTS FOR WAITING TIMES 767 2. Poisson s equation and a central limit theorem for the PH/PH/1 queue. We use the following notation for phase-type distributions. A phasetype distribution has representation S if it has a density of the form f x = e Sx s x Here e Sx is a matrix exponential, a row vector and s a column vector that satisfies Se = s, where e is the column vector, the components of which are equal to 1. A phase-type distribution is the distribution of the life-time of a terminating Markov jump process with finitely many states, and in this context plays the role of the initial distribution, S the intensity matrix among the nonabsorbing states (subintensity matrix) and s the exit rate vector. For more details about phase-type distributions, see, for example, Asmussen [2], Bladt and Asmussen [6], Neuts [11] or Neuts [12]. The assumption of phase-type distributed service times and interarrival times is necessary in Section 4, where we rely on the probabilistic interpretation that a phase-type distribution is the life-time of a terminating Markov jump process. Extensions to matrix-exponential distributions are not immediate, and will require some new ideas as replacements for the probabilistic reasoning (e.g., transform methods). Consider the actual waiting time sequence W = W n n in the PH/PH/1 queue. Then W n+1 = W n + U n T n + where W =, U n denotes the i.i.d. sequence of service times with common phase-type distribution A and T n denotes the i.i.d. sequence of interarrival times with common phase-type distribution B. Let ρ = E U n /E T n be the traffic intensity. Then W n is recurrent if ρ 1, since will be visited infinitely often. Thus W n is a Harris chain, as it follows from being a regeneration set; see [1], page 15. Throughout this paper we will assume ρ < 1. Let µ be a measure on R + and let f R + R be a measurable function. Moreover, let P n denote the transition kernel of the Harris chain, P n x A = P W n A W = x, and define P = P 1. Then we define (1) µf = µp n = P n f = f y µ dy P n y µ dy f y P n dy It is known from the theory of Harris chains (see, e.g., [1]) that there exists a σ-finite stationary measure ν in the sense that ν = νp n for all n, and that ν can be represented by ( T 1 ) ν = E I W k k=

768 M. BLADT where E is the conditional expectation given W = and T = inf n 1 W n =. Similarly, P x and E x refer to the case where W = x. Moreover [see, e.g., [1], Theorem 2.3(b)], (2) ν = U + = k= G k + where U + is the renewal measure of the ascending ladder process that corresponds to the random walk S n, G + denotes the ascending ladder height distribution and G+ k is its k-fold convolution. In our case, where ρ < 1, the distribution of G + is defective. The mass of ν, defined as ν = ν dy, is ν = G + k 1 = 1 G + k= Introduce the distribution π = 1 G + ν as the normalized distribution of ν. We notice that π is in fact the distribution function of the maximum of the random walk S n. Indeed, G k + x 1 G + is the probability that the maximum is attained in exactly k ladder steps and is less than or equal to x, since 1 G + is the probability of no further ladder steps. Thus the result follows by summing over k. Poisson s equation is an equation of the form I P g = f We want to find a solution g to this equation. Thus we will look for a solution kernel Ɣ, which is a kernel such that g x = Ɣf x. By a kernel we understand a family Ɣ x x R + of σ-finite measures such that x Ɣ x A is a measurable function for any measurable set A. Poisson s equation can only have π-integrable solutions if f is such that πf = or, equivalently, νf =. Define T 1 Ɣ x = Ɣ x = E x k= I W k Defining Ɣf x = Ɣ x f similarly to (1), we can prove the following theorems in a similar way as in Glynn [8]. Theorem 2.1. Assume that f is π-integrable. If πf =, then Ɣf x solves Poisson s equation. We note that if g solves Poisson s equation, then so does g + c, where c is an arbitrary constant. The next theorem, the proof of which is similar to that in [8], states that the solutions of the form g + c are in fact the only solutions. Theorem 2.2. Let g be a π-integrable solution to Poisson s equation that is finite everywhere, g x < for all x. Then g = Ɣf + c for some constant c.

VARIANCE CONSTANTS FOR WAITING TIMES 769 In the proof of Theorem 2.1 it is assumed that Ɣ x g < for all x. To establish this rigorously, let f x y = P W 1 dy W = x be a transition density with respect to the Lebesgue measure. Assume that for all x z there exists a constant < c = c x z < such that (3) f x y cf z y for all y. This assumption is satisfied for the PH/PH/1 queue. Indeed, we have that f x u = c x δ u + h u x for u, where δ is the degenerate distribution at and h is the density of U n T n. Assume that representations of U n and T n are, respectively, S and T. Then for z we have that h z = e S t+z s e Tt t dt = e Sz S T 1 s t where and denote, respectively, the Kronecker product and the Kronecker sum; see Graham [9] for details. Similarly for z <. Thus h z decays as c 1 e αz as z for some positive constants c 1 and α. Thus it is clear that assumption (3) is satisfied. Using (3) and the assumptions in Theorem 2.2, we then prove that Ɣ x g < for all x in the following way. Assume without loss of generality that g. Then T 1 Ɣ x g = g x + E x = g x + c n=1 g W n f x y Ɣ y g dy f z y Ɣ y g dy so Ɣ x g < implies that f z y Ɣ y g dy < and hence Ɣ z g <, since Ɣ z g = g z + f z y Ɣ y g dy Thus, if there exists an x such that Ɣ x g <, then Ɣ x g < for all x, but this is ensured by π g = Ɣ g <. Next we link the solution of Poisson s equation to a central limit theorem, and prove that the variance constant can be found in terms of this solution. Define D k = g W k Pg W k 1, and consider the martingale M n = n D k + g W k=1 Since D k = g W k E Wk 1 g W k we see that D k is stationary when the initial distribution is π, and ergodicity of the sequence follows from Asmussen [1], page 182. Then, from Billingsley [5], page 26, we get the following result.

77 M. BLADT Theorem 2.3. Let g be a π-square-integrable solution to Poisson s equation I P g = f ˆ where f is π-integrable. Then, independently of the initial distribution, where σ 2 f = 2 π ˆ fg π ˆ f 2. 1 n M n EM N σ 2 f Note that the square-integrability condition of the martingale has been removed, the reason being as follows. The π-square-integrability of g implies that the limit theorem is valid for the initial distribution π. Since ρ < 1, W n is an aperiodic regenerative Harris chain (see, e.g., [1], page 182) and hence it admits a coupling to its stationary distribution π (see, e.g., Asmussen [1], Proposition 3.13). Thus, at least from a certain time S (which is finite a.s.), the square-integrability is ensured by πg 2 < for sufficiently large n. It is clear that any polynomial g will satisfy the conditions of the theorem for the PH/PH/1 queue (perform repeated partial integrations), which is sufficient for our purposes. Let us link the asymptotic behavior of the martingale M n with the asymptotic variance of n 1 i= f W i. First note that n M n+1 = f W i + g W n+1 i= Since W n converges weakly to a limit W, say, then g W n / n weakly as n, so the asymptotic behavior of n 1 i= f W i / n is identical to that of M n / n. 3. Explicit calculation of the variance constant. In this section we shall repeatedly make use of the following integration rules for matrix exponentials: for any matrix A we have that (4) A x e At dt = e Ax I where I is the identity matrix of the same dimension as A. Moreover, for any two matrices A and B such that any sum of an eigenvalue of A and an eigenvalue of B is negative, and any vectors a, b, and of appropriate dimensions, we have that (5) x ae At be Bt dt = a b A B 1( e A B x I ) where I is the identity matrix of the same dimension as A B. For the PH/PH/1 queue we are able to find an explicit formula for the stationary measure ν. (6) Proposition 3.1. For all y, ν dy = δ dy + I y + exp S + s + y s dy

VARIANCE CONSTANTS FOR WAITING TIMES 771 where + is the solution to the nonlinear matrix equation (7) and  B = ebx A dx. + =  S + s + Proof. By (2), it is enough to find U +. To this end we use a result for the GI/PH/1 queue in Asmussen [2] which states that, for a random walk S n = X 1 + + X n, S = with X i X = U T, U being phase-type with representation S and T having some arbitrary distribution A, the ascending ladder height distribution G + is phase-type with representation + S, and + is the solution of the nonlinear matrix equation + =  S + s + which is known to exist; see Asmussen [2] for details. Since U + is the renewal measure for the ascending ladder height process, for all y. Thus, for all y, U + dy = δ dy + + exp S + s + y s dy ν dy = δ dy + I y + exp S + s + y s dy Equation (7) can be solved iteratively by choosing + = and n+1 + =  S + s n + ; see [2] for details. Let U x = E x k= I S η k, where η k denotes the descending ladder epochs for S n. Proposition 3.2. Defining + similarly to + as the solution to the nonlinear equation + = ˆB T + t + and ˆB C = ecx B dx, we have that U x dy = δ x dy + I x y + exp T + t + x y t dy Proof. We can rewrite U x as ( ) ( (8) U x A = E x I S ηk A = E I S ηk k= k= ) x A Now apply the result of [2] again. The distribution of U n does not matter with respect to the following argument, and it may even help to consider it as general for a moment. Thus, considering U n as general, we consider a PH/G/1 queue. Since η k are descending ladder epochs for S n, they are also ascending ladder epochs for the random walk S n. (Note that there are no problems concerning weak versus strict ladder heights, since phase-type densities are positive everywhere and cannot have an atom at.) Now S n corresponds to the GI/PH/1 case, and we may then calculate U x using this result. For S n we know that the ascending ladder height distribution is phase-type with representation + T. Thus, by (8), U x is the renewal measure of the ascending ladder process for S n evaluated on the set x A, and hence U x dy = δ x dy + I x y + exp T + t + x y t dy

772 M. BLADT The formulas for U x and ν are now used to find an explicit form of the solution kernel for Poisson s equation. Define Proposition 3.3. Q = S + s + T + t + 1 s t The solution kernel Ɣ x dz has the following form: Ɣ x dz = δ x dz + I x z + exp T + t + x z t dz + I x z + exp S + s + z x s dz + + exp S + s + z + exp T + t + x Q dz I x z + + exp T + t + x z Q dz I x z + exp S + s + z x + Q dz where all vectors and matrices are described above. Proof. By (6) we see that ν A x = ν du A x = δ x A + Using Glynn [8], Proposition 4.1, we get (9) Ɣ x A = = A ν A y U x dy ( δ y A + I x u + exp S + s + u x s du A ) I y z + exp S + s + z y s dz ( δ x dy + I x y + exp T + t + x y t dy ) The integral can be calculated in a straightforward manner using the integration rules (4) and (5). Note that the condition for (5) to hold is satisfied for A = S + s + and B = T +t +, since B has exactly one eigenvalue that is and the rest are negative, whereas all eigenvalues of A are negative (B and A being, respectively, an intensity matrix and subintensity matrix). Let f k x = x k and put fˆ k = f πf k. The mass G + is given by such that G + = + e Sx s dx = + S 1 s π = 1 + + S 1 s ν Put r = 1 + + S 1 s and s i = + S + s + i s. We want to find the solutions g to the equations I P g = fˆ 1. First note that simple partial integration gives that πf 1 = rs 2, so, in particular, fˆ 1 y = y rs 2

VARIANCE CONSTANTS FOR WAITING TIMES 773 Next note that we cannot integrate the term exp T + t + z in the usual way since the matrix T +t + is singular. This is due to the fact that the traffic intensity ρ < 1, which implies that S n has negative drift. Therefore T +t + is the intensity matrix of a nondefective renewal process (pure renewal process). This also implies that + e = 1. So in order to integrate exp T + t + z, we use a trick involving generalized inverses. Let be the stationary distribution of the Markov jump process with intensity matrix T + t + ; that is, satisfies T + t + =. Put D = T + t + e 1 + e Then the indefinite integral exp T + t + y dy = xe + D exp T + t + x and, in particular, (1) x exp T + t + y dy = xe + D exp T + t + x I The following constants are needed to describe the solution to Poisson s equation given in Proposition 3.4: k 1 = 1 2 t + Q k 2 = 1 + + De t + Dt + S + s + 1 s + + De Q + + + D Q + + S + s + 1 + Q + πf 1 + Q t k 3 = + D 2 t + + S + s + 2 s + + + D 2 Q + S + s + 2 + Q πf 1 + πf 1 + S + s + 1 s πf 1 + S + s + 1 + Q Proposition 3.4. Let M x = exp T + t + x. In the PH/PH/1 queue Ɣ x is a σ-finite measure. Thus Ɣ defines a solution kernel, and a solution to Poisson s equation I P g = fˆ 1 is given by g x = Ɣfˆ 1 x = zɣ x dz πf 1 Ɣ x dz = k 1 x 2 + k 2 x + k 3 + + D 2 M x t + + S + s + 2 + M x Q + + D 2 M x Q + πf 1 + S + s + 1 + M x Q + πf 1 + + D M x I Q πf 1 + D M x I t Proof. That Ɣ x is a σ-finite measure follows by the fact that Ɣ x dz <. The rest of the proof follows by simple integration and partial integration using the integration rules (4), (5) and (1).

774 M. BLADT In order to find the expression for the variance constant using the formula 2π fˆ 1 g π ˆ, it is useful to notice the following result. f 2 1 Lemma 3.1. x k + exp S + s + x s dx = 1 k+1 k!s k+1 Proof. The result follows by integrating by parts k times using (4). The following constant is needed for calculating the variance constant in Proposition 3.5. Lemma 3.2. π ˆ f 2 1 = r3 s 2 2 2rs 3 r 3 s 2 2 s 1 2r 2 s 2 2 Proof. The result follows from partially integrating fˆ 2 1 x = x2 + r 2 s 2 2 2xrs 2 and using (4). The variance constant in Proposition 3.5 is given in terms of the following constants. Definition 3.1. = + S + s + 2 + 1 = + + D 2 2 = rs 2 + S + s + 1 + 3 = + + D = T + t + I m 1 = 6rk 1 s 4 + 2k 1 r 2 s 2 s 3 m 2 = 2k 2 rs 3 k 2 r 2 s 2 2 m 3 = k 3 r 2 s 2 + rk 3 s 2 + k 3 r 2 s 2 s 1 m 4 = r 2 s 2 + D 2 t + r 2 s 2 + D 2 + T + t + S + s + 1 t s + r + D 2 + T + t + S + s + 2 t s m 5 = r 2 s 2 Q + r + S + s + 2 Q s + r 2 s 2 + S + s + 1 Q s m 6 = r 2 s 2 1 Q + r 2 s 2 1 + S + s + 1 Q s + r 1 + S + s + 2 Q s m 7 = r 2 s 2 2 Q + r 2 + S + s + 2 Q s + r 2 s 2 2 + S + s + 1 Q s

VARIANCE CONSTANTS FOR WAITING TIMES 775 m 8 = r 3 s 2 2 3Q r 3 s 2 2 3 + S + s + 1 Q s + r 2 s 2 3 + S + s + 2 Q s + r 3 s 2 3 Q r 2 s 2 s 1Q + r 3 s 2 2 s 1 3 Q m 9 = r 2 s 2 + D + T + t + S + s + 2 t s r 3 s 2 2 +D + T + t + S + s + 1 t s + r 3 s 2 2 s 1 + Dt + r 2 s 2 2 +Dt Proposition 3.5. The variance constant for f x = x is given by 2 m 1 + m 2 + + m 9 π ˆ f 1 2 Proof. Again the proof is a simple exercise in applying the integration rules (4), (5) and (1). 4. Examples. The formula above was implemented in Pascal on a SUN workstation, and some examples were considered. It was necessary to restrict the matrix dimensions in the program to be less than 5. The highest order of matrices appearing in the formula for the variance constant is p 1 p 2 2, where p 1 is the dimension of the interarrival distribution and p 2 the dimension of the service time distribution. By the dimension of a phase-type distribution we mean the dimension of its corresponding vectors (e.g., the exit rate vector). This means that our examples are restricted to lower dimensions, though in the special case of exponential service times we would be able to consider an interarrival distribution with up to 5 phases. A free copy of the program for calculating the variance constant is available upon request from the author. A program for fitting phase-type distributions to densities or data can be obtained free of charge by e-mail at the following address: olleh@math.chalmers.se. For further information, please see Asmussen, Nerman and Olson [3] and Häggström, Asmussen and Nerman [1]. Example 4.1. Consider the M/M/1 queue. Then it is well known (see, e.g., [16]) that the variance constant is given by where ρ is the traffic intensity. ρ 2 + 5ρ 4ρ2 + ρ 3 1 ρ 4 Example 4.2 (Erlang distributions). We consider the E k /E n /1 queue for k = 1 2 3 4 5, n = 1 2 3 and for traffic intensities ρ = 3 6 9. We fix the mean of the service time distribution to 1 and let the mean of the interarrival distribution vary accordingly. The variance constants for the E k /E n /1 queue are given in Table 1.

776 M. BLADT Table 1 3 k 1 k 2 k 3 k 4 k 5 n = 1 3.957 1.313.782.573.464 n = 2 1.684.391.182.112.79 n = 3 1.18.224.92.5.32 6 k 1 k 2 k 3 k 4 k 5 n = 1 88.5 36.36 24.98 2.25 17.7 n = 2 36.48 1.51 5.891 4.184 3.328 n = 3 25.28 5.99 2.957 1.915 1.422 9 k 1 k 2 k 3 k 4 k 5 n = 1 35,9 15,265 1,758 8,881 7,867 n = 2 14,993 4,476 2,62 1,91 1,549 n = 3 1,476 2,575 1,323 889 681 Example 4.3 (M/G/1 queue). Consider the M/G/1 queue, where the service time distribution is a phase-type distribution with representation T, where = ( 9731 152 16 1 ) 28 648 28 532 89 27 T = 12 8 255 8 63 86 133 17 5 87 5 296 1 12 111 2 176 (this phase-type distribution was originally fitted to a log-normal distribution; see [4]). As we let the arrival intensity vary, we present the values for the asymptotic variance (as a function of the traffic intensity) in Table 2. Table 2 Traffic intensity Variance.1..2.8.3.62.4.32.5 1.338.6 5.74.7 18.56.8 66.95.9 233.

VARIANCE CONSTANTS FOR WAITING TIMES 777 Acknowledgments. I thank the referees and editor for their constructive comments and for pointing out an important reference concerning the martingale central limit theorem. REFERENCES [1] Asmussen, S. (1987). Applied Probability and Queues. Wiley, New York. [2] Asmussen, S. (1992). Phase-type representations in random walk and queueing problems. Ann. Probab. 2 772 789. [3] Asmussen, S., Nerman, O. and Olson, M. (1994). Fitting phase-type distributions via the EM-algorithm. Preprint, Dept. Math., Chalmers Univ. Technology, Gothenburg, Sweden. [4] Asmussen, S. and Rolski, T. (1991). Computational methods in risk theory: a matrixalgorithmic approach. Insurance Math. Econom. 1 259 274. [5] Billingsley, P. (1968). Convergence of Probability Measures. Wiley, New York. [6] Bladt, M. and Asmussen, S. (1996). Renewal theory and queueing algorithms for matrix-exponential distributions. In Matrix-Analytic Methods in Stochastic Models (S. Chakravarthy and A. S. Alta, eds.) 313 341. Dekker, New York. [7] Bladt, M. and Asmussen, S. (1994). Poisson s equation for queues driven by a Markovian point process. QUESTA 17 235 274. [8] Glynn, P. W. (1989). Poisson s equation for the recurrent M/G/1 queue. Adv. in Appl. Probab. 26 144 162. [9] Graham, A. (1981). Kronecker Products and Matrix Calculus with Applications. Ellis Horwood Series, London. [1] Häggström, O., Asmussen, S. and Nerman, O. (1992). EMPHT a program for fitting phasetype distributions. Preprint, Dept. Math., Chalmers Univ. Technology, Gothenburg, Sweden. [11] Neuts, M. F. (1978). Renewal processes of phase type. Naval Res. Logist. Quart. 25 445 454. [12] Neuts, M. F. (1981). Matrix-Geometric Solutions in Stochastic Models. Johns Hopkins Univ. Press. [13] Neveu, J. (1972). Potentiel Markovien récurrent des chaînes de Harris. Ann. Inst. Fourier (Grenoble) 22 85 13. [14] Nummelin, E. (1985). On the Poisson equation in the potential theory of a single kernel. Technical report, Univ. Helsinki. [15] Revuz, D. (1984). Markov Chains. North-Holland, New York. [16] Whitt, W. (1989). Planning queueing simulations. Management Sci. 35 1341 1366. Department of Statistics Instituto de Investigaciones en Matematicas Aplicadas y en Sistemas Universidad Nacional Autonoma de Mexico Apartado Postal 2-726 1 Mexico, D.F. Mexico E-mail: bladt@uxestad1.iimas.unam.mx