Figure 10.1: Recording when the event E occurs

Similar documents
Continuous Time Processes

Queueing Theory. VK Room: M Last updated: October 17, 2013.

Data analysis and stochastic modeling

Part I Stochastic variables and Markov chains

Performance Evaluation of Queuing Systems

Introduction to queuing theory

Lecture 7: Simulation of Markov Processes. Pasi Lassila Department of Communications and Networking

Non Markovian Queues (contd.)

Since D has an exponential distribution, E[D] = 0.09 years. Since {A(t) : t 0} is a Poisson process with rate λ = 10, 000, A(0.

IEOR 6711, HMWK 5, Professor Sigman

Networking = Plumbing. Queueing Analysis: I. Last Lecture. Lecture Outline. Jeremiah Deng. 29 July 2013

The story of the film so far... Mathematics for Informatics 4a. Continuous-time Markov processes. Counting processes

Kendall notation. PASTA theorem Basics of M/M/1 queue

Systems Simulation Chapter 6: Queuing Models

Part II: continuous time Markov chain (CTMC)

Markov processes and queueing networks

Continuous-time Markov Chains

Exponential Distribution and Poisson Process

Queueing Theory I Summary! Little s Law! Queueing System Notation! Stationary Analysis of Elementary Queueing Systems " M/M/1 " M/M/m " M/M/1/K "

Stochastic process. X, a series of random variables indexed by t

Lecture 4a: Continuous-Time Markov Chain Models

QUEUING SYSTEM. Yetunde Folajimi, PhD

All models are wrong / inaccurate, but some are useful. George Box (Wikipedia). wkc/course/part2.pdf

CDA5530: Performance Models of Computers and Networks. Chapter 3: Review of Practical

Class 11 Non-Parametric Models of a Service System; GI/GI/1, GI/GI/n: Exact & Approximate Analysis.

(b) What is the variance of the time until the second customer arrives, starting empty, assuming that we measure time in minutes?

CPSC 531: System Modeling and Simulation. Carey Williamson Department of Computer Science University of Calgary Fall 2017

Continuous-Time Markov Chain

Lecture 20: Reversible Processes and Queues

Chapter 2. Poisson Processes. Prof. Shun-Ren Yang Department of Computer Science, National Tsing Hua University, Taiwan

Exercises Stochastic Performance Modelling. Hamilton Institute, Summer 2010

CDA6530: Performance Models of Computers and Networks. Chapter 3: Review of Practical Stochastic Processes

Elementary queueing system

SOLUTIONS IEOR 3106: Second Midterm Exam, Chapters 5-6, November 8, 2012

LECTURE #6 BIRTH-DEATH PROCESS

Introduction to queuing theory

Queueing Systems: Lecture 3. Amedeo R. Odoni October 18, Announcements

The exponential distribution and the Poisson process

Chapter 1. Introduction. 1.1 Stochastic process

57:022 Principles of Design II Final Exam Solutions - Spring 1997

Queueing Review. Christos Alexopoulos and Dave Goldsman 10/25/17. (mostly from BCNN) Georgia Institute of Technology, Atlanta, GA, USA

λ, µ, ρ, A n, W n, L(t), L, L Q, w, w Q etc. These

Recap. Probability, stochastic processes, Markov chains. ELEC-C7210 Modeling and analysis of communication networks

Introduction to Markov Chains, Queuing Theory, and Network Performance

Introduction to Queuing Theory. Mathematical Modelling

Other properties of M M 1

Queueing Review. Christos Alexopoulos and Dave Goldsman 10/6/16. (mostly from BCNN) Georgia Institute of Technology, Atlanta, GA, USA

Queues and Queueing Networks

IEOR 6711: Stochastic Models I, Fall 2003, Professor Whitt. Solutions to Final Exam: Thursday, December 18.

Markov chains. 1 Discrete time Markov chains. c A. J. Ganesh, University of Bristol, 2015

Solutions to Homework Discrete Stochastic Processes MIT, Spring 2011

Stochastic Models in Computer Science A Tutorial

EXAM IN COURSE TMA4265 STOCHASTIC PROCESSES Wednesday 7. August, 2013 Time: 9:00 13:00

Slides 9: Queuing Models

2905 Queueing Theory and Simulation PART III: HIGHER DIMENSIONAL AND NON-MARKOVIAN QUEUES

Performance Modelling of Computer Systems

Computer Networks More general queuing systems

Chapter 5. Continuous-Time Markov Chains. Prof. Shun-Ren Yang Department of Computer Science, National Tsing Hua University, Taiwan

CDA5530: Performance Models of Computers and Networks. Chapter 4: Elementary Queuing Theory

Simple queueing models

Solution: The process is a compound Poisson Process with E[N (t)] = λt/p by Wald's equation.

Computer Systems Modelling

Chapter 6 Queueing Models. Banks, Carson, Nelson & Nicol Discrete-Event System Simulation

1 IEOR 4701: Continuous-Time Markov Chains

Computer Systems Modelling

The effect of probabilities of departure with time in a bank

CS418 Operating Systems

Statistics 150: Spring 2007

Name of the Student:

Markov Processes and Queues

M/G/1 and Priority Queueing

BIRTH DEATH PROCESSES AND QUEUEING SYSTEMS

NICTA Short Course. Network Analysis. Vijay Sivaraman. Day 1 Queueing Systems and Markov Chains. Network Analysis, 2008s2 1-1

Q = (c) Assuming that Ricoh has been working continuously for 7 days, what is the probability that it will remain working at least 8 more days?

THE QUEEN S UNIVERSITY OF BELFAST

MATH 56A: STOCHASTIC PROCESSES CHAPTER 6

PBW 654 Applied Statistics - I Urban Operations Research

GI/M/1 and GI/M/m queuing systems

Waiting Line Models: Queuing Theory Basics. Metodos Cuantitativos M. En C. Eduardo Bustos Farias 1

Queueing systems. Renato Lo Cigno. Simulation and Performance Evaluation Queueing systems - Renato Lo Cigno 1

Introduction to Queueing Theory

Exercises Solutions. Automation IEA, LTH. Chapter 2 Manufacturing and process systems. Chapter 5 Discrete manufacturing problems

MAT SYS 5120 (Winter 2012) Assignment 5 (not to be submitted) There are 4 questions.

Link Models for Circuit Switching

Renewal processes and Queues

2905 Queueing Theory and Simulation PART IV: SIMULATION

5/15/18. Operations Research: An Introduction Hamdy A. Taha. Copyright 2011, 2007 by Pearson Education, Inc. All rights reserved.

Probability and Statistics Concepts

Continuous time Markov chains

The Transition Probability Function P ij (t)

Introduction to Statistical Data Analysis Lecture 3: Probability Distributions

Review of Queuing Models

M/G/1 and M/G/1/K systems

Chapter 10. Queuing Systems. D (Queuing Theory) Queuing theory is the branch of operations research concerned with waiting lines.

Solutions For Stochastic Process Final Exam

Queuing Theory. Queuing Theory. Fatih Cavdur April 27, 2015

Chapter 8 Queuing Theory Roanna Gee. W = average number of time a customer spends in the system.

Chapter 2. Poisson Processes

Question Points Score Total: 70

ELEMENTS OF PROBABILITY THEORY

Transcription:

10 Poisson Processes Let T R be an interval. A family of random variables {X(t) ; t T} is called a continuous time stochastic process. We often consider T = [0, 1] and T = [0, ). As X(t) is a random variable for each t T, it has another variable ω Ω. When we need to explicitly refer to ω, we write X(t, ω) or X t (ω). For fixed ω Ω, the function t X(t, ω) is called a sample path of the stochastic process {X(t)}. It is the central idea of stochastic processes that a random evolution in the real world is expressed by a single sample path selected randomly from all the possible sample paths. The most fundamental continuous time stochastic processes are the Poisson process and the Brownian motion (Wiener process). In the recent study of mathematical physics and mathematical finance, a kind of composition of these two processes, called the Lévy process (or additive process), has received much attention. 10.1 Heuristic Introduction Let us imagine that the number of objects changes as time goes on. The number at time t is modelled by a random variable X t and we wish to construct a stochastic process {X t }. In this case X t takes values in {0, 1, 2,... }. In general, such a stochastic process is called a counting process. There are many different variations of randomness and so wide variations of counting processes. We below consider the simple situation as follows: We focus an event E which occurs repeatedly at random as time goes on. For example, (i) alert of receiving an e-mail; (ii) telephone call received by a call center; (iii) passengers making a queue at a bus stop; (iv) customers visiting a shop; (v) occurrence of defect of a machine; (vi) traffic accident at a corner; (vii) radiation from an atom. Let fix a time origin as t = 0. We count the number of occurrence of the event E during the time interval [0, t] and denote it by X t. Let t 1, t 2,... be the time when E occurs, see Fig. 10.1. 0 t t t tn Figure 10.1: Recording when the event E occurs There are two quantities which we measure. (i) The number of occurrence of E up to time t, say, X t. Then {X t ; t 0} becomes a counting process. (ii) The waiting time of the nth occurrence after the (n 1)th occurrence, say, T n. Here T 1 is defined to be the waiting time of the first occurrence of E after starting the observation. Then {T n ; n = 1, 2,... } is a sequence of random variables taking values in [0, ). 39

X t T n T T 0 t 0 t t t tn Figure 10.2: The counting process and waiting times We will introduce heuristically a stochastic process {X t } from the viewpoint of (i). It is convenient to start with discrete time approximation. Fix t > 0 and divide the time interval [0, t] into n small intervals. Let t = t n be the length of the small intervals and number from the time origin in order. t 1 2 n 0 t We assume the following conditions on the occurrence of the event E: (1) There exists a constant λ > 0 such that P(E occurs just once in a small time interval of length t) = λ t + o( t), P(E does not occur in a mall time interval of length t) = 1 λ t + o( t), P(E occurs more than once in a small time interval of length t) = o( t). (2) Occurrence of E in disjoint time intervals is independent. Some more accounts. Let us imagine the alert of receiving an e-mail. That P(E occurs more than once in a small time interval of length t) = o( t) means that two occurrences of the event E is always separated. That P(E occurs just once in a small time interval of length t) = λ t + o( t) means that when t is small the probability of occurrence of E in a time interval is proportional to the length of the time interval. We understand from (2) that occurrence of E is independent of the past occurrence. 40

Let Z i denote the number of occurrence of the event E in the ith time interval. Then Z 1, Z 2,..., Z n become a sequence of independent random variables with an identical distribution such that P(Z i = 0) = 1 λ t + o( t), P(Z i = 1) = λ t + o( t), P(Z i 2) = o( t). The number of occurrence of E during the time interval [0, t] is given by n Z i. The length t is introduced for a technical reason and is not essential in the probability model so letting t 0 or equivalently n, we define X t by X t = lim t 0 n Z i. (10.1) Although the limit does require mathematical justification, we obtain heuristically a continuous time stochastic process {X t }, which gives the number of occurrence of the event E up to time t. This is called a Poisson process with parameter λ > 0. A Poisson process belongs to the calss of continuous time Markov chains. Theorem 10.1.1 A Poisson process {X t ; t 0} satisfies the following properties: (1) (counting process) X t takes vales in {0, 1, 2,... }; (2) X 0 = 0; (3) (monotone increasing) X s X t for 0 s t; (4) (independent increment) if 0 t 1 < t 2 < < t k, then are independent; X t2 X t1, X t3 X t2,..., X tk X tk 1, (5) (stationarity) for 0 s < t and h 0, the distributions of X t+h X s+h and X t X s are identical; (6) X t obeys the Poisson distribution with parameter λt. (7) there exists a constant λ > 0 such that P(X h = 1) = λh + o(h), P(X h 2) = o(h). Proof (1) Since X t obeys the Poisson distribution with parameter λt, it takes values in non-negative integers almost surely. (2) Obvious by definition. (3) Let s = m t, t = n t, m < n. Then we have obviously X s = lim t 0 m Z i lim t 0 n Z i = X t. (4) Suppose t 1 = n 1 t,..., t k = n k t with n 1 < < n k. Then we have X t2 X t1 = lim t 0 n 2 Z i lim t 0 n 1 Z i = lim n 2 t 0 i=n 1 +1 In other words, X t2 X t1 is the sum of Z i s corresponding to the small time intervals contained in [t 2, t 1 ). Hence, X t2 X t1,..., X tk X tk 1 are the sums of Z i s and there is no common Z i appearing in the summands. Since {Z i } are independent, so are X t2 X t1,..., X tk X tk 1. Z i. 41

(5) Since X t+h X s+h and X t X s are defined from the sums of Z i s and the numbers of the terms coincide. Therefore the distributions are the same. (6) We note that n ( ) n P Z i = k = (λ t) k (1 λ t) n k + o( t). k In other words, X t obeys apporoximately B(n, λ t). In view of t = t/n we let n tend to the infinity and obtain (λt) k n(n 1)... (n k + 1) ( P(X t = k) = lim t 0 k! n k 1 λt ) n k = (λt)k n k! e λt. This proves the assertion. (7) Since X h obeys the Poisson distribution with parameter λh, we have P(X h = 0) = e λh = 1 λh + = 1 λh + o(h), P(X h = 1) = λhe λh = λh(1 λh +... ) = λh + o(h). Therefore we have P(X h 2) = 1 P(X h = 0) P(X h = 1) = o(h). Remark 10.1.2 The essence of the above argument in (7) is the Poisson s law of small numbers which says that the binomial distribution B(n, p) is approximated by Poisson distribution with parameter λ = np when n is large and p is small. The following table shows the distributions of B(100, 0.02) and the Poisson distribution with parameter λ = 2. k 0 1 2 3 4 5 6 Binomial 0.1326 0.2707 0.2734 0.1823 0.0902 0.0353 0.0114 Poisson 0.1353 0.2707 0.2707 0.1804 0.0902 0.0361 0.0120 Example 10.1.3 The average number of customers visiting a certain service gate is two per minute. Using the Poisson model, calculate the following probabilities. (1) The probability that no customer visits during the first two minutes after the gate opens. (2) The probability that no customer visits during a time interval of two minutes. (3) The probability that no customer visits during the first two minutes after the gate opens and that two customers visit during the next one minute. Let X t be the number of visitors up to time t. By assumption {X t } is a Poisson process with parameter λ = 2. (1) We need to calculate P(X 2 = 0). Since X 2 obeys the Poisson distribution with parameter 2λ = 4, we have P(X 2 = 0) = 40 0! e 4 0.018. (2) Suppose that the time interval starts at t 0. Then the probability under discussion is given by P(X t0 +2 X t0 = 0). By stationarity we have P(X t0 +2 X t0 = 0) = P(X 2 X 0 = 0) = P(X 2 = 0), which coincides with (1). (3) We need calculate the probability P(X 2 = 0, X 3 X 2 = 2). Since X 2 and X 3 X 2 are independent, P(X 2 = 0, X 3 X 2 = 2) = P(X 2 = 0)P(X 3 X 2 = 2). 42

By stationarity we have = P(X 2 = 0)P(X 1 = 2) = 40 0! e 4 22 2! e 2 0.00496. Problem 19 Let {X t } be a Poisson process. Show that ( ) n ( s ) k ( P(X s = k X t = n) = 1 s ) n k, k = 0, 1,..., n, k t t for 0 < s < t. Then give an intuitive explanation of the above formula. Problem 20 How often do you receive a message through internet? Check your iphone (or other similar devises) and collect the data focussing on how many messages are received during one hour. Probably you must check the records during at least 1 week (168 hours). (1) Make up a table that shows the relative frequency of the number of received messages per one hour. (2) Calculate the mean value and variance of the number of messages received during one hour. (If they are close each other, your data should well fit to the Poisson distribution.) (3) Suppose that your data obeys Poisson distribution. Using the Poisson model, find the probability that 0 messages arrive during 1 hour and 2 messages arrive during the next 2 hours. 10.2 Waiting Time Let {X t ; t 0} be a Poisson process with parameter λ. By definition X 0 = 0 and X t increases by one as time passes. Recall that the Poisson process counts the number of events occurring up to time t. First we set T 1 = inf{t 0 ; X t 1}. (10.2) This is the waiting time for the first occurrence of the event E. Let T 2 be the waiting time for the second occurrence of the event E after the first occurrence, i.e., In a similar fashion, we set T 2 = inf{t 0 ; X t 2} T 1. T n = inf{t 0 ; X t n} T n 1, n = 2, 3,.... (10.3) Theorem 10.2.1 Let {X t } be a Poisson process with parameter λ. Define the waiting time T n by (10.2) and (10.3). Then, {T n ; n = 1, 2,... } becomes a sequence of iid random variables, of which distribution is the exponential distribution with parameter λ. In particular, the waiting time for occurence of an event in the Poisson process obeys the exponential distribution with parameter λ. Proof Set t = n t and consider the approximation by refinement of the time interval. Recall that to each small time interval of length t a random variable Z i is associated. Then we know that ( P(T 1 > t) = lim P(Z 1 = = Z n = 0) = lim (1 λ t) n = lim 1 λt ) n = e λt. t 0 t 0 t 0 n Therefore, P(T 1 t) = 1 e λt = t 0 λe λs ds, which shows that T 1 obeys the exponential distribution with parameter λ. The distributions of T 2, T 3,... are similar. 43

Remark 10.2.2 Let {X t } be a Poisson process with parameter λ. We know that E(X 1 ) = λ, which means the average number of occurrence of the event during the unit time interval. Hence, it is expected that the average waiting time between two occurrences is 1/λ. Theorem 10.2.1 says that the waiting time obeys the exponential distribution with parameter λ so its mean value is 1/λ. Thus, our rough consideration gives the correct answer. Problem 21 Let {X t } be a Poisson process with parameter λ. The waiting time for n occurrence of the events is defined by S n = T 1 + T 2 + + T n, where T n is given in Theorem 10.2.1. Calculate P(S 2 t) and find the probability density function of S 2. [In general, S n obeys a gamma distribution.] 10.3 The Rigorous Definition of Poisson Processes The definition of a Poisson process in (10.1) is intuitive and instructive for modeling random phenomena. However, strictly speaking, the argument is not sufficient to define a stochastic process {X t }. For example, the probability space (Ω, F, P) on which {X t } is defined is not at all clear. We need to start with the waiting time {T n }. First we prepare a sequence of iid random variables {T n ; n = 1, 2,... }, of which the distribution is the exponential distribution with parameter λ > 0. Here the probability space (Ω, F, P) is clearly defined. Next we set and for t 0, S 0 = 0, S n = T 1 + + T n, n = 1, 2,..., X t = max{n 0 ; S n t}. It is obvious that for each t 0, X t is a random variable defined on the probability space (Ω, F, P). In other words, {X t ; t 0} becomes a continuous time stochastic process. This is called Poisson process with parameter λ by definition. Starting with the above definition one can prove the properties in mentioned Theorem 10.1.1. 10.4 Generating Function Let {X t } be a Poisson process with parameter λ > 0. For simplicity, we set By observing the transition we have Letting t 0, we have The last equation is easily solved: p n (t) = P(X t = n), t 0, n = 0, 1, 2,.... p n (t + t) = λ t p n 1 (t) + (1 λ t)p n (t), n 1, p 0 (t + t) = (1 λ t)p 0 (t). p n(t) = λp n 1 (t) λp n (t), n 1, p 0 (t) = λp 0(t). p 0 (t) = e λt, where the initial condition p 0 (0) = P(X 0 = 0) = 0 is taken into account. Define the generating function by f (x, t) = p n (t)x n. 44

Then we have f t = p n(t)x n = λp 0 (t) + { λpn 1 (t)x n λp n (t)x n} n=1 = λ p n (t)x n + λx p n (t)x n = ( λ + λx) f (x, t). Moreover, Thus, from which we have f (0, t) = p 0 (t) = e λt. f (x, t) = e ( λ+λx)t = e λt λt (λt)n p n (t) = e. n! (λt) n n! x n, 45

11 Queueing Theory 11.1 Modeling Queues In our daily life, we observe often waiting lines or queues of customers for services. Agner Krarup Erlang (1878 1929, Danish engineer at the Copenhagen Telephone Exchange) published in 1909 the paper entitled: The Theory of Probabilities and Telephone Conversations, which opened the door to the research field of queueing theory. Such a queue is modeled in terms of a system consisting of servers and a waiting room. Customers arriving at the system are served at once if there is an idle server. Otherwise, the customer waits for a vacant server in a waiting room. After being served, the customer leaves the system. system servers waiting room departure arrival customers In most of the qeueing models, a customer arrives at random and the service time is also random. So we are interested in relevant statistics such as (1) sojourn time (time of a customer staying in the system) (2) waiting time (= sojourn time - service time) (3) the number of customers in the system Apparently, many different conditions may be introduced for the queueing system. In 1953, David G. Kendall introduced the so-called Kendall s notation A/B/c/K/m/Z for describing the characteristics of a queuing model, where A: arrival process, B: service time distribution, c: number of servers, K: number of places in the system (or in the waiting room), m: calling population, Z: queue s discipline or priority order, e.g., FIFO (First In First Out) The first model analyzed by Erlang in 1909 was the M/D/1 queue in Kendall s notation, where M means that arrivals occur according to a Poisson process, and D stands for deterministic (i.e., service time is not random but constant). Most of queueing models are classified into four categories by the behavior of customers as follows: 46

(I) Delay models: customers wait in line until they can be served. Example: M/M/c queue, where (i) customers arrive according to a Poisson process with rate λ; (ii) there are c servers and there is an infinite waiting space; (iii) each customer requires an exponential service time with mean 1/µ; (iv) customers who upon arrival find all servers busy wait in line to be served. (II) Loss models: customers leave the system when they find all servers busy upon arrival. Example: Erlang s loss model M/G/c/c, where (i) customers arrive according to a Poisson process with rate λ; (ii) there are c servers and the capacity of the system is limited to c customers, i.e., there is no waiting space; (iii) each customer requires a generally distributed service time; (iv) customers who upon arrival find all servers busy are rejected forever. (III) Retrial models: customers who do not find an idle server upon arrival leave the system only temporarily, and try to reenter some random time later. Example: the Palm/Erlang-A queue, where (i) customers arrive according to a Poisson process with rate λ; (ii) there are c servers and there is an infinite waiting space; (iii) each customer requires an exponential service time with mean 1/µ; (iv) customers who upon arrival find all servers busy wait in line to be served; (v) customers wait in line only an exponentially distributed time with mean 1/θ (patience time). (IV) Abandonment models: customers waiting in line will leave the system before being served after their patience time has expired. References 1. J. W. Cohen: The Single Server Queue (2nd rev. ed.), Elsevier, 2012. 2.,, 2014. 3.,, 2006. 11.2 M/M/1 Queue This is the most fundamental model, which satisfies the following conditions: (i) arrivals occur according to a Poisson process with parameter λ; (ii) service times obey an exponential distribution with parameter µ; (iii) arrivals of customers and service times are independent; (iii) the system contains a single server; (iv) the size of waiting room is infinite; (v) (FIFO = First In First Out) customers are served from the front of the queue, i.e., according to a first-come, first-served discipline. 47

Thus there are two parameters characterizing an M/M/1 queue, that is, the parameter λ > 0 for the Poisson arrival and the one µ > 0 for the exponential service. In other words, a customer arrives at the system with average time interval 1/λ and the average service time is 1/µ. In the queuing theory λ is called the mean arrival rate and µ the mean service rate. Let X(t) be the number of customers in the system at time t. It is the proved that {X(t) ; t 0} becomes a continuous time Markov chain on {0, 1, 2, 3,...}. In fact, the letter M stands for Markov or memoryless. Our main objective is p n (t) = P(X(t) = n X(0) = 0), i.e., the probabbility of finding n customers in the system at time t > 0 subject to the initial condition X(0) = 0. Let us consider the change of the system during the small time interval [t, t + t]. It is assumed that during the small time interval t only one event happens, namely, a new customer arrives, a customer under service leaves the system, or nothing changes. The probabilities of these events are given by λ t, µ t, 1 λ t µ t. λ t n+1 n+1 µ t n n n n µ t n 1 n 1 λ t t t + t t t + t Therefore, P(X(t) = n X(0) = 0) fulfills the following equation: P(X(t + t) = n X(0) = 0) = P(X(t + t) = n X(t) = n 1)P(X(t) = n 1 X(0) = 0) + P(X(t + t) = n X(t) = n)p(x(t) = n X(0) = 0) + P(X(t + t) = n X(t) = n + 1)P(X(t) = n + 1 X(0) = 0) = λ tp(x(t) = n 1 X(0) = 0) + (1 λ t µ t)p(x(t) = n X(0) = 0) + µ tp(x(t) = n + 1 X(0) = 0), P(X(t + t) = 0 X(0) = 0) = (1 λ t)p(x(t) = 0 X(0) = 0) + µ tp(x(t) = 1 X(0) = 0). Hence for p n (t) = P(X(t) = n X(0) = 0) we have p n(t) = λp n 1 (t) (λ + µ)p n (t) + µp n+1 (t), n = 1, 2,..., p 0 (t) = λp 0(t) + µp 1 (t). (11.1) The initial condition is as follows: p 0 (0) = 1, p n (0) = 0 for n 1. (11.2) Solving the linear system (??) with the initial condition (??) is not difficult with the help of linear algebra and spectral theory. However, the explicit solution is not so simple and is omitted. We only mention that most important characteristics are obtained from the explicit p n (t). Here we focus on the equilibrium solution (limit transition probability), i.e., p n = lim t p n (t) 48

whenever the limit exists. Since in the equilibrium the derivative of the left hand side of (??) is 0, we have λp n 1 (λ + µ)p n + µp n+1 = 0 n = 1, 2,..., λp 0 + µp 1 = 0. (11.3) A general solution to (??) is easily derived: ( n λ C 1 + C 2, λ µ, p n = µ) C 1 + C 2 n, λ = µ. Since p n gives a probability distribution, we have p n 0 and we have p n = p n = 1. This occurs only when λ < µ and ( 1 λ ) ( ) n λ, n = 0, 1, 2,.... µ µ This is the geometric distribution with parameter λ/µ. In queuing theory, the ratio of the mean arrival rate λ and the mean service rate µ is called the utilization: ρ = λ µ. Utilization stands for how busy the system is. It was shown above that the number of customers in the system after long time obeys the geometric distribution with parameter ρ. If ρ < 1, the system functions well. Otherwise, the queue will continue to grow as time goes on. After long time, i.e., in the equilibrium the number of customers in the system obeys the geometric distribution: (1 ρ)ρ n, n = 0, 1, 2,.... In particular, the probability that the server is free is 1 ρ and the probability that the server is busy and the customer need to wait is ρ. This is the origin of the term utilization. Note also that the mean number of the customers in the system is given by np n = ρ 1 ρ = λ µ λ. Example 11.2.1 There is an ATM, where each customer arrives with average time interval 5 minutes and spends 3 minutes in average for the service. Using an M/M/1 queue, we know some statistical characteristics. We set λ = 1 5, µ = 1 3, ρ = λ µ = 3 5. Then the probability that the ATM is free is p 0 = 1 ρ = 2. The probability that the ATM is busy but there 5 is no waiting customer is p 1 = 2 5 3 5 = 6 25. Hence the probability that the ATM is busy and there is some waiting customers is 1 p 0 p 1 = 1 2 5 6 25 = 9 25 = 0.36. So, roughly speaking, a customer needs to make a queue once per three visits. 49

Problem 22 (M/M/1/1 queue) There is a single server and no waiting space. Customers arrive according to the Poisson process with parameter λ, and their service time obeys the exponential distribution with parameter µ. Let Q(t) be the number of customers in the system at time t. In fact, 1, server is busy, Q(t) = 0, server is idle, (1) Find p 0 (t) = P(Q(t) = 0 Q(0) = 0), p 1 (t) = P(Q(t) = 1 Q(0) = 0) by solving a linear system satisfied by those p 0 (t) and p 1 (t). (2) Using the results in (1), calculate p 0 = lim t p 0 (t), p 1 = lim t p 1 (t), (3) Find the mean number of customers in the system in the long time limit: 11.3 Birth-and-Death Processes lim E[Q(t) Q(0) = 0]. t The Markov process X(t) appearing in the M/M/1 queuing model is studied more generally within the framework of birth-and-death process. 50