A monotonic property of the optimal admission control to an M/M/1 queue under periodic observations with average cost criterion

Similar documents
OPTIMALITY OF RANDOMIZED TRUNK RESERVATION FOR A PROBLEM WITH MULTIPLE CONSTRAINTS

ADMISSION CONTROL IN THE PRESENCE OF PRIORITIES: A SAMPLE PATH APPROACH

Individual, Class-based, and Social Optimal Admission Policies in Two-Priority Queues

The M/G/1 FIFO queue with several customer classes

Queueing Theory I Summary! Little s Law! Queueing System Notation! Stationary Analysis of Elementary Queueing Systems " M/M/1 " M/M/m " M/M/1/K "

Dynamic Control of a Tandem Queueing System with Abandonments

Total Expected Discounted Reward MDPs: Existence of Optimal Policies

Continuous-Time Markov Decision Processes. Discounted and Average Optimality Conditions. Xianping Guo Zhongshan University.

LIMITS FOR QUEUES AS THE WAITING ROOM GROWS. Bell Communications Research AT&T Bell Laboratories Red Bank, NJ Murray Hill, NJ 07974

On the static assignment to parallel servers

A Note on Powers in Finite Fields

Strategic Dynamic Jockeying Between Two Parallel Queues

Online algorithms for parallel job scheduling and strip packing Hurink, J.L.; Paulus, J.J.

1 Markov decision processes

REMARKS ON THE EXISTENCE OF SOLUTIONS IN MARKOV DECISION PROCESSES. Emmanuel Fernández-Gaucherand, Aristotle Arapostathis, and Steven I.

STABILIZATION OF AN OVERLOADED QUEUEING NETWORK USING MEASUREMENT-BASED ADMISSION CONTROL

Operations Research Letters. Instability of FIFO in a simple queueing system with arbitrarily low loads

Analytical Studies of the Influence of Regional Groundwater Flow by on the Performance of Borehole Heat Exchangers

A numerical study of vortex-induced vibrations (viv) in an elastic cantilever

Tilburg University. Strongly Regular Graphs with Maximal Energy Haemers, W. H. Publication date: Link to publication

On Finding Optimal Policies for Markovian Decision Processes Using Simulation

Graphs with given diameter maximizing the spectral radius van Dam, Edwin

Tutorial: Optimal Control of Queueing Networks

ON ESSENTIAL INFORMATION IN SEQUENTIAL DECISION PROCESSES

COMPARISON OF THERMAL HAND MODELS AND MEASUREMENT SYSTEMS

Chapter 2 SOME ANALYTICAL TOOLS USED IN THE THESIS

Performance Evaluation of Queuing Systems

Optimal Sleep Scheduling for a Wireless Sensor Network Node

Resistivity survey at Stora Uppåkra, Sweden

Stationary Probabilities of Markov Chains with Upper Hessenberg Transition Matrices

EQUILIBRIUM STRATEGIES IN AN M/M/1 QUEUE WITH SETUP TIMES AND A SINGLE VACATION POLICY

Technical Appendix for: When Promotions Meet Operations: Cross-Selling and Its Effect on Call-Center Performance

NEW FRONTIERS IN APPLIED PROBABILITY

Minimizing response times and queue lengths in systems of parallel queues

On the Unique D1 Equilibrium in the Stackelberg Model with Asymmetric Information Janssen, M.C.W.; Maasland, E.

Section Notes 9. Midterm 2 Review. Applied Math / Engineering Sciences 121. Week of December 3, 2018

Lecture notes for Analysis of Algorithms : Markov decision processes

De Prisco, Renato; Comunian, Michele; Grespan, Franscesco; Piscent, Andrea; Karlsson, Anders

Technical Appendix for: When Promotions Meet Operations: Cross-Selling and Its Effect on Call-Center Performance

Tilburg University. Two-Dimensional Minimax Latin Hypercube Designs van Dam, Edwin. Document version: Publisher's PDF, also known as Version of record

Multi Stage Queuing Model in Level Dependent Quasi Birth Death Process

A Queueing System with Queue Length Dependent Service Times, with Applications to Cell Discarding in ATM Networks

The dipole moment of a wall-charged void in a bulk dielectric

Inventory Ordering Control for a Retrial Service Facility System Semi- MDP

Citation for pulished version (APA): Henning, M. A., & Yeo, A. (2016). Transversals in 4-uniform hypergraphs. Journal of Combinatorics, 23(3).

UNCORRECTED PROOFS. P{X(t + s) = j X(t) = i, X(u) = x(u), 0 u < t} = P{X(t + s) = j X(t) = i}.

A tandem queue under an economical viewpoint

Some notes on Markov Decision Theory

Maximizing throughput in zero-buffer tandem lines with dedicated and flexible servers

HITTING TIME IN AN ERLANG LOSS SYSTEM

Tilburg University. Two-dimensional maximin Latin hypercube designs van Dam, Edwin. Published in: Discrete Applied Mathematics

Aalborg Universitet. The number of independent sets in unicyclic graphs Pedersen, Anders Sune; Vestergaard, Preben Dahl. Publication date: 2005

Transmuted distributions and extrema of random number of variables

Infinite-Horizon Average Reward Markov Decision Processes

Admission policies for a two class loss system with random rewards

Lecture 20: Reversible Processes and Queues

Strongly 2-connected orientations of graphs

Room Allocation Optimisation at the Technical University of Denmark

Markov Chains. X(t) is a Markov Process if, for arbitrary times t 1 < t 2 <... < t k < t k+1. If X(t) is discrete-valued. If X(t) is continuous-valued

Near-Optimal Control of Queueing Systems via Approximate One-Step Policy Improvement

Aalborg Universitet. All P3-equipackable graphs Randerath, Bert; Vestergaard, Preben Dahl. Publication date: 2008

Non Markovian Queues (contd.)

On the Einstein-Stern model of rotational heat capacities

MAT SYS 5120 (Winter 2012) Assignment 5 (not to be submitted) There are 4 questions.

A zero-free interval for chromatic polynomials of graphs with 3-leaf spanning trees

Series Expansions in Queues with Server

MDP Preliminaries. Nan Jiang. February 10, 2019

Value Iteration and Action ɛ-approximation of Optimal Policies in Discounted Markov Decision Processes

Optimal Control of an Inventory System with Joint Production and Pricing Decisions

Operations Research, Vol. 30, No. 2. (Mar. - Apr., 1982), pp

Optimality of Myopic Sensing in Multi-Channel Opportunistic Access

Time Reversibility and Burke s Theorem

Aalborg Universitet. Bounds on information combining for parity-check equations Land, Ingmar Rüdiger; Hoeher, A.; Huber, Johannes

Robustness of policies in Constrained Markov Decision Processes

Dynamic control of a tandem system with abandonments

Stability of the two queue system

Aalborg Universitet. Reparametrizations with given stop data Raussen, Martin Hubert. Publication date: 2008

THIELE CENTRE. The M/M/1 queue with inventory, lost sale and general lead times. Mohammad Saffari, Søren Asmussen and Rasoul Haji

Exact Simulation of the Stationary Distribution of M/G/c Queues

Höst, Stefan; Johannesson, Rolf; Zigangirov, Kamil; Zyablov, Viktor V.

Optimal Control of a Production-Inventory System with both Backorders and Lost Sales

Weighted Sums of Orthogonal Polynomials Related to Birth-Death Processes with Killing

EFFECTS OF SYSTEM PARAMETERS ON THE OPTIMAL POLICY STRUCTURE IN A CLASS OF QUEUEING CONTROL PROBLEMS

Learning Algorithms for Minimizing Queue Length Regret

Benchmarking of optimization methods for topology optimization problems

IEEE TRANSACTIONS ON AUTOMATIC CONTROL, VOL. 43, NO. 3, MARCH

DELAY, MEMORY, AND MESSAGING TRADEOFFS IN DISTRIBUTED SERVICE SYSTEMS

Notation. state space

STABILITY OF MULTICLASS QUEUEING NETWORKS UNDER LONGEST-QUEUE AND LONGEST-DOMINATING-QUEUE SCHEDULING

Stochastic Optimization with Inequality Constraints Using Simultaneous Perturbations and Penalty Functions

Dynamic Call Center Routing Policies Using Call Waiting and Agent Idle Times Online Supplement

Generating the optimal magnetic field for magnetic refrigeration

Solving convex optimization with side constraints in a multi-class queue by adaptive c rule

Published in: Proceedings of the 21st Symposium on Mathematical Theory of Networks and Systems

A product form solution to a system with multi-type jobs and multi-type servers Visschers, J.W.C.H.; Adan, I.J.B.F.; Weiss, G.

Dynamic Call Center Routing Policies Using Call Waiting and Agent Idle Times Online Supplement

Session-Based Queueing Systems

MULTIPLE CHOICE QUESTIONS DECISION SCIENCE

Modelling Salmonella transfer during grinding of meat

On the Hausdorff Dimension of Fat Generalised Hyperbolic Attractors

Transcription:

A monotonic property of the optimal admission control to an M/M/1 queue under periodic observations with average cost criterion Cao, Jianhua; Nyberg, Christian Published in: Seventeenth Nordic Teletraffic Seminar, NTS 17, Fornebu, Norway, 25-27 August 2004 Published: 2004-01-01 Link to publication Citation for published version (APA): Cao, J., & Nyberg, C. (2004). A monotonic property of the optimal admission control to an M/M/1 queue under periodic observations with average cost criterion. In Seventeenth Nordic Teletraffic Seminar, NTS 17, Fornebu, Norway, 25-27 August 2004 Fornebu : Telenor. General rights Copyright and moral rights for the publications made accessible in the public portal are retained by the authors and/or other copyright owners and it is a condition of accessing publications that users recognise and abide by the legal requirements associated with these rights. Users may download and print one copy of any publication from the public portal for the purpose of private study or research. You may not further distribute the material or use it for any profit-making activity or commercial gain You may freely distribute the URL identifying the publication in the public portal Take down policy If you believe that this document breaches copyright please contact us providing details, and we will remove access to the work immediately and investigate your claim. L UNDUNI VERS I TY PO Box117 22100L und +46462220000

Download date: 29. Sep. 2018

This is an author produced version of a paper presented at the 17th Nordic Teletraffic Seminar (NTS 17), Fornebu, Norway, 25-27 August, 2004. This paper may not include the final publisher proof-corrections or pagination. Citation for the published paper: J. Cao and C. Nyberg, 2004, "A Monotonic Property of the Optimal Admission Control to an M/M/1 Queue under Periodic Observations with Average Cost Criterion", Seventeenth Nordic Teletraffic Seminar, NTS 17, Fornebu, Norway, 25-27 August 2004. ISBN: 82-423-0595-1. Publisher: Fornebu : Telenor.

A Monotonic Property of the Optimal Admission Control to an M/M/1 Queue under Periodic Observations with Average Cost Criterion Jianhua Cao and Christian Nyberg Department of Communication Systems Lund University, Sweden {jcao, cn}@telecom.lth.se Abstract We consider the problem of admission control to an M/M/1 queue under periodic observations with average cost criterion. The admission controller receives the system state information every τ:th second and can accordingly adjust the acceptance probability for customers who arrive before the next state information update instance. For a period of τ seconds, the cost is a linear function of the time average of customer populations and the total number of served customers in that period. The objective is to find a stationary deterministic control policy that minimizes the long run average cost. The problem is formulated as a discrete time Markov decision process whose states are fully observable. By taking the control period τ to 0 or to, the model in question generalizes two classical queueing control problems: the open and the closed loop admission control to an M/M/1 queue. We show that the optimal policy is to admit customers with a non-increasing probability with respect to the observed number of customers in the system. Numerical examples are also given. 1 Introduction We consider the problem of admission control to an M/M/1 queue under periodic observations with average cost criterion. The admission controller receives the system state information every τ:the second and can accordingly adjust the control policy which is the probability that an arrival will be accepted into the system. For a period of τ seconds, the cost for making a particular decision is a linear function of the time average of customer populations waiting in queue and the total number of served customers in that period. The objective is to find a stationary deterministic control policy that minimizes long run average cost. Our objective function definition which rewards departures and penalizes a long queue resembles to that of Naor (1969) and many others, see Stidham Jr. (1985); Stidham Jr. (1988, 2002) and references therein. Along this line of objective function definition, the optimal admission control in the cases of complete observable and nonobservable queueing system are well understood see, e.g., Hassin and Haviv (2003). In

a brief summary, when the queue length is completely observable, the optimal policy is of threshold type and when the queue length is not observable, the optimal policy accepts customers with a fixed probability. Many models of queueing system admission control, including ours, are based on Markov decision theory, see(stidham Jr. and Weber, 1993, Section 4) and references therein. One ramification to the existing admission control problems is reducing the amount of information used for decision. Much work has been done along this direction. Fukuda (1986) also considered a queueing system under periodic observations. There are two types of customers in his system with the same service time distribution and different arrivals processes. The control policy considered is call gapping control where the low priority customers are blocked for the next period whenever the observed queueing system state exceeds a predetermined limit. Kuri and Kumar (1995) studied the discrete time admission control and routing to a queueing system consisting of two parallel queues with delayed queue length information. Their objective is to minimize the total expected discounted cost in which a fixed reward is received for each admitted customers and linear holding cost is incurred for customers waiting in queue. The optimal policy in case of one period information delay is of threshold type. Lin and Ross (2003) considered a multiple-server loss model where the admission controller is informed when an admitted customer finds all servers are busy but not informed when customers depart. In this system a cost is incurred if a new arrival is blocked and an even larger cost induced if an admitted customer is blocked by servers. They proved that the threshold type policy that blocks for a certain amount of time after an admitted arrival is optimal in case of single servers. In this paper we establish that the average cost optimal policy is nonincreasing. This is an important property that can be exploited to accelerate numerical computations. This paper is organized as follows: In Section 2 we define the Markov decision process model for our problem. We show that the average optimal solution is nonincreasing in Section 3. Numerical examples are give in Section 4. Finally, we conclude with some remarks. 2 The Markov Decision Process Model 2.1 Classes of Control Policies and Induced Markov Chains We control the arrival rate to an M/M/1 queue using percentage blocking. For convenience, we assume that the service rate is 1 in the rest of this paper. The arrival rate to the system is λ which can be less or greater than 1. There is an admission controller that rejects arrivals in order to maintain a reasonable response time for admitted customers. The controller is characterized by the admission probability a, 0 a 1. We are allowed to adjust the admission probability every τ:th seconds when the updated system state information becomes available. Fig. 1 illustrates the model. Let S := {0, 1, 2, } be the set of system states. In our problem, the system state is the number of customers in the queue. A stationary deterministic control policy is an infinite sequence (π (0), π (1), ) =: π, with 0 π (i) 1, i S. It specifies that whenever the observed number of customers in the system is i S, the admission

(1 a) λ λ a λ µ a The controller receives the system state information from the queue and adjusts acceptance rate a at those time instances t t+ τ t+2 τ Figure 1: Arrival rate control to an M/M/1 queue with periodic observations. probability should be adjusted to π (i) and remain the same for the next τ seconds. Let Π be the set of stationary deterministic control policies. A more generic control class is, e.g., {π k (i, t) [0, 1] t [0, τ), k N, i S}, where π k (i, t) is the admission probability at time kτ + t when the observed number of customers at time kτ is i. We shall, however, content ourselves with the class of stationary deterministic control policies in this paper. Let t 0, t 1, t 2, be the time instances when the admission probability is adjusted and t 0 = 0,t k+1 t k = τ, k N. Let X(t) S be the system state at time t. Clearly {X (t k )} k=0 can be regarded as a discrete time Markov chain. For convenience denote X k := X (t k ). The transition probability between states i and j is denoted as P (j i, π (i)). It is actually the probability that at time τ there are j customers in an M/M/1 queue with arrival rate π (i) λ when there are i customers in the system initially. Once a control policy π Π and the initial state are given, the associated Markov chain {X k } is fully specified. Notice that the transition probability P (, a) is continuous in a. 2.2 One Step Cost Given a fixed arrival rate, one would like to accept as few customers as possible in order maintain a short mean response time. But, on the other hand, to have a high throughput, the arrival rate ought to be close to the service rate. Hence one has to balance low response time and high throughput. One way to resolve this dilemma is to formulate an constrained optimization problem where the throughput is the objective to be maximized while the response time is subject to a given constraint. Another way is to consider an unconstrained optimization problem whose objective is to minimize

a cost which is a function of response time and throughput. This paper discusses the second approach. To make the definition of cost more precise, we introduce some auxiliary notation. Let N (t, i, a), N : [0, τ) S [0, 1] [0, ), be the expected number of customers in an M/M/1 queue at time t when initially there are i customers and the admission probability is a. The time average number of customers in the system between time 0 and τ is 1 τ τ 0 N (t, i, a) dt. The average number of severed customers(not including the rejected customers) between time 0 and τ is 1 ( ) i + aλτ N (τ, i, a). τ Let C be the cost to maintain one customer in the system per time unit and R be the reward for a departure. To avoid the trivial case, we assume that 0 < C < R. The one-step cost, i.e. the time average of net cost incurred in a period of τ seconds between two adjacent control parameter adjustment instances, is denoted as c (i, a), and is defined as follows, c (i, a) := C τ τ 0 N (t, i, a) dt R τ ( i + aλτ N (τ, i, a) ). (1) It is easy to convince oneself that the step cost c (i, a) is not necessarily monotonic in i or a. 2.3 Average Cost and β-discount Cost The expected long-run average cost or average cost incurred by a policy π Π is defined as [ ] N 1 J (i, π) := lim sup Ei π 1 c (X k, π (X k )), N N where i is the initial number of customers in the system. The optimal average cost for any initial state i S is defined as k=0 J (i) := inf {J (i, π)}. π Π A policy π is average cost optimal, or average optimal, if J (i, π) = J (i) for all i S. While the average cost is of primal interest of this paper, β-discounted cost is helpful in the proof of the existence of the average optimal policy. Given some β (0, 1), the β-discount cost incurred by a policy π Π is defined as [ ] J β (i, π) := Ei π β k c (X k, π (X k )). k=0

For any initial state i S, the optimal β-discount cost is defined as J β (i) := inf π Π {J β (i, π)} A policy π is β-discount cost optimal, or β-discount optimal, if J β (i, π) = J β (i) for all i S. 2.4 The Existence of Average Cost Optimal Policy There are at least three well developed methods to prove the existence of the average optimal policy when the state space is enumerable infinite and step cost is unbounded see Arapostathis et al. (1993, Section 5.2 and Section 5.3) for review. The first one is Hordijk s Lyapunov stability condition (Hordijk, 1974). Sennott s three necessary conditions (Sennott, 1989, 1998, 2000) can be counted as the second. The third one is Borkar s convex analytic approach (Borkar, 1988, 2000). Sennott s conditions are particularly well suited for queueing problems (Sennott, 1998) where the step cost often, if not always, grows as more customers accumulate in the queue. Our proof of the existence of the average optimal policy for our particular model is based on Sennott s conditions. A complete proof of the existence of average optimal policy for a Markov decision process satisfies Sennott s three conditions relies on the vanishing discounted approach. The general idea is that one treats the average cost case as the limit of the discounted cost problem. The vanishing discounted approach was also used by Blackwell (1962) to prove the existence of average optimal policy for a model with finite state space and action set, and Derman (1966) for countable state space, finite action set and bounded costs. The existence β-discount optimal policy of our problem is easy to check, see Arapostathis et al. (1993, Lemma 2.1). We shall verify Sennott s three conditions in our particular problem in Lemma 1 which guarantees existence of average optimal policy using vanishing discount approach see e.g. Arapostathis et al. (1993, Theorem 5.9) Lemma 1. (a) For every i S and every β (0, 1), Jβ (i) <. (b) There exists a nonnegative integer L such that h β (i) := J β (i) J β (0) L (c) There exists a function M : S R such that h β (i) M (i) for all i S and β (0, 1) and for every i S and a [0, 1] such that j P (j i, a) M (j) <. Proof. (a). Clearly there exists a stationary deterministic policy, e.g π := (0, 0, ) such that the induced MC is ergodic and J (i, π) = 0 for all i S. By the Tauberian Theorem, we have lim inf β 1 (1 β) Jβ (i) lim sup (1 β) Jβ (i) J (i, π) β 1 Hence Jβ (i) is finite for all i S and β (0, 1). (b) Consider a queuing system with two M/M/1 queues, one queue with i customers initially but no arrivals, another queue starts empty and the β-discount optimal policy

is employed. The discounted cost for the whole system is therefore Jβ (0) + J β (i, 0), where 0 = (0, 0, ) is a policy that rejects all arrivals. Clearly Jβ (0) + J β (i, 0) Jβ (i), hence we have J β (i) J β (0) J β (i, 0) Let n := R/C. For all i S, J β (i, 0) J β (n, 0) nr. Therefore h β (i) R/C R. (c) Let Π be the class of policies inducing irreducible, ergodic MC and c i,0 (π) the expected total cost of a first passage from i to 0. Clearly Π is not empty. Let M (0) = 0 and for i 1, M (i) = inf π Π c i,0 (π). We then have Jβ (i) M (i) + J β (0) and for all i S and a [0, 1], and some π Π, P (j i, a) c i,0 (π) = c i,0 (π) <. j S P (j i, a) M (j) j 3 The Monotonic Property of the Average Cost Optimal Policy Intuitively when more customers are observed in one control instance, less should be admitted into the system in the next control period of τ seconds. Hence it is reasonable to conjecture that the average optimal policy ϕ is nonincreasing in i, i.e. ϕ (i) ϕ (i + 1), i S. This fact can be easily verified in the cases of closed loop and open loop admission control. In this section we shall prove that this is true in general when 0 τ <. Let N (t, i, a) be the number of customers in a queue (with arrival rate aλ and service rate 1) at time t with initial condition N (0, i, a) = i. Notice that given t, i, a, N (t, i, a) is a random variable. A function f : N R is discrete convex if f (i + 2) f (i + 1) f (i + 1) f (i) for all i N. Let F be the space of discrete convex functions that map from S to R. The following lemma follows. Lemma 2. Given t > 0, for all f F, E [f (N (t, i, a))] is supermodular in (i, a). Proof. We have to show that for all f F, i S, 0 b < a 1, E [f (N (t, i + 1, a))] + E [f (N (t, i, b))] E [f (N (t, i, a))] + E [f (N (t, i + 1, b))] The following stochastic order relations can be verified by the normalization technique: N (t, i + 1, a) st N (t, i, a), N (t, i + 1, a) st N (t, i + 1, b), N (t, i + 1, b) st N (t, i, b), N (t, i + 1, a) N (t, i, a) st N (t, i + 1, b) N (t, i, a). Since f is discrete convex, we have f (N (t, i + 1, a)) f (N (t, i, a)) st f (N (t, i + 1, b)) f (N (t, i, a)) which leads to the claim after taking expectation on both sides. Lemma 3. Given t > 0, for all f F. i S, 0 b < a 1,

[ ( ( E [f (N (t, i, a))] + E [f (N (t, i + 2, b))] 2E f N t, i + 1, a + b ))] 2 Proof. Since f is discrete convex, it is enough to show that ( N (t, i, a) + N (t, i + 2, b) st 2N t, i + 1, a + b ). (2) 2 Consider a system consisting of two identical but independent M/M/1 queues. The total arrival rate is (a + b) λ and initially there are 2i + 2 customers in total in two queues. Suppose we are allowed to distribute 2i + 2 customers into to two queues at time 0 and split the incoming traffic by random routing. Total number of customers in the whole system at time t can be minimized (in stochastic order sense) by placing i + 1 customers in each queue and splitting traffic evenly. Thus (2) holds. Lemma 4. Define g : F S [0, 1] R as follows g(f) (i, a) := c (i, a)+e [f (N (τ, i, a))], then g (f) (i, a)is supermodular in (i, a) and min a g (f) (i, a) is convex in i. Proof. (Supermodularity) By Lemma 2, E [N (t, i, a)] is supermodular in (i, a). Recall the definition of the step cost c (i, a) = C τ τ 0 E [N (t, i, a)] dt R τ (i + aλτ E [N (τ, i, a)]). Hence c (i, a) is supermodular in (i, a) too. Also from Lemma 2, E [f (N (τ, i, a))] is supermodular. Thus g (f) (i, a) is supermodular in (i, a). (Convexity) We have to show that i S min a g (f) (i + 2, a) + min g (f) (i, a) 2 min g (f) (i + 1, a). a a We shall show that there exist some w [0, 1] such that min g (f) (i + 2, a) + min g (f) (i, a) 2g (f) (i, w) (3) a a Let u := arg min a g (f) (i, a), v := arg min a g (f) (i + 2, a). By the supermodularity of g, v u. If u = v, then let w = u = v, the inequality above is evident. If v < u then let w = u+v 2. By Lemma 3, and E [N (t, i + 2, v)] + E [N (t, i, u)] 2E [N (t, i + 1, w)] E [f (N (t, i + 2, v))] + E [f (N (t, i, u))] 2E [f (N (t, i + 1, w))]. Therefore (3) is valid. Theorem 1. If ϕ Π is average optimal, then ϕ (i) ϕ (i + 1) for all i S.

Proof. Consider the value iteration. The value function V k : S R at the k:th iteration is defined as follows V k+1 (i) = min a c (i, a) + V k (j) P (j i, a) j with V 0 (j) = 0, j S. The average optimal policy is defined as follows ϕ (i) = lim arg min c (i, a) + k a j V k (j) P (j i, a), i S By induction, we have that for all k, V k (i) is convex in i and c (i, a)+ j V k (j) P (j i, a) is supermodular in (i, a). Hence ϕ (i) is nonincreasing in i. 4 Numerical Examples In our model, the state space is countable infinite. For numerical calculations, however, the state space must be truncated (cutting off the tail). Rigorous analysis of the effects of state space truncation for dynamic programming in general can be found in Fox (1971); Whitt (1978, 1979). The approximating sequence method described in Sennott (1998, Chapter 8) is also very illuminating. In our particular case, we note that the difference in the optimal cost between the original model and the state space truncated model decreases as the probability of unexpected blocking diminishes. Once the state space is truncated, both the value iteration algorithm and the policy iteration algorithm (Puterman, 1994, Section 8.5 and 8.6) can be used to calculate the optimal policy and the optimal average cost. The policy iteration algorithm which is used in our numerical examples is described briefly as follows. Step 1 Initialization: An arbitrary stationary policy π is chosen. Step 2 Value determination: For the current policy π, compute the unique solution {J, h i } for the following system of linear equations h i = c (i, π (i)) J + j S h j P (j i, π (i)), i S h 0 = 0 where S is the truncated state space. Step 3 Policy improvement: The new policy π is obtained as follows π (i) = arg min 0 a 1 c (i, a) J + j S h j P (j i, a) i S. Step 4 Convergence test. Let d (π, π) be a distance measure of two policies such as follows d (π, π) = π i S (i) π (i). If the new policy π is within the give distance ( ) ɛ of the old policy π, i.e. d π, π < ɛ, the algorithm stops, otherwise goto step 2 with π replaced by π.

-5 C=50-10 -15 J -20-25 λ=0.7 λ=1.0 λ=1.5 λ=2.0 an upper bound a lower bound when λ<2.0-30 -35 0 1 2 3 4 5 6 7 8 9 10 τ Figure 2: Average optimal cost versus the control interval τ when C = 50. Note that due the state space truncation, the expression for the average number of customers in the system N (i, t, a) and the transition probability P (j i, a) must be swapped from the M/M/1 model to the M/M/1/K model where K is the truncated queue size. Using the same techniques in proving Lemma 4 and Theorem 1, one can show that the function to be minimized in the policy improvement step is convex and supermodular if the policy chosen in the initialization step is nonincreasing. As a result many convex programming techniques can be used to find the minimum [ in step ] 3 and further, the search region for π (i) can be narrowed from [0, 1] to 0, π (i 1). In the following discussions, we set the reward coefficient to a constant R = 100. In Fig. 2, we show how the optimal average cost changes as the control interval τ increases for different combinations of cost coefficient C = 50 and arrival rates λ = 0.7, 1.0, 1.5, 2.0. Notice that when the offered traffic intensity and the cost of waiting are high, e.g. C = 50, λ = 2.0, the optimal average cost for models with short control intervals is sensitive to τ. We also plot the theoretical lower and the upper bound of the optimal average cost in these figures. We have shown in Section 3 that the average optimal policy is nonincreasing. In Fig. 3, we give examples of the average optimal control policies in cases of τ = 0.1, 5.0, 10.0 when λ = 1.0 and C = 30. The optimal control policy in the case of open loop control τ is also shown. If the closed loop control is used (i.e. τ = 0.0) when λ = 1.0, C = 30, the average optimal policy is of threshold type and the optimal threshold is 1.

1 τ=0.1 τ=5.0 τ=10.0 τ 0.8 0.6 a i 0.4 0.2 0 0 2 4 6 8 10 i Figure 3: The average optimal control policy when λ = 1.0 and C = 30. 5 Conclusions We have presented an admission control model for an M/M/1 queue under periodic observations with average cost criterion. This model degenerates into two well known queueing control problems when the observation interval becomes zero and infinite. The corresponding discrete time Markov decision process is obtained via embedding the state transition instances at the state information update epochs. The state transition probability is related to the transient solution of state probability distribution of an M/M/1 queue. The step cost is linearly proportional to the time average number of customers in the queue and the time average number of departures per time unit between two adjacent control instances. With help of value iteration algorithm and induction, we proved that the average optimal policy is nonincreasing. Several numerical examples are also provided in a separate section. Many claims/arguments in the paper are not specific to the M/M/1 queue, e.g. Lemma 1 used in the proof of the existence of average optimal policy; Lemma 2 and 3 in the proof of the monotonic property of optimal policy. However one must proceed with caution when the approach in defining the discrete time Markov decision process for the M/M/1 model in this paper is extended to an M/G/1 or an G/M/1 queue. References Arapostathis, A., Borkar, V., Fernández-Gaucherand, E., Ghosh, M. K., and Marcus, S. I. (1993). Discrete-time controlled Markov processes with average cost criterion: a survey. SIAM J. Control and Optimization, 31(2):282 344.

Blackwell, D. (1962). Discrete dynamic programming. The Annals of Mathematical Statistics, 33(2):719 726. Borkar, V. S. (1988). Stochastic Differential Systems, Stochastic Control Theory and Applications, volume 10 of The IMA Volumes in Mathematics and Its Applications, chapter Control of Markov chains with long-run average cost criterion, pages 57 77. Springer-Verlag. Borkar, V. S. (2000). Markov Decision Processes: Model, Methods and Open Problems, chapter Convex analytic methods in Markov decision processes. Kluwer 2000. Derman, C. (1966). Denumerable state Markov decision processes, average cost criterion. Ann. Math. Statist., 37:1545 1553. Fox, B. L. (1971). Finite-State Approximations to Denumerable-State Dynamic Programs. J. Math. Anal. Appl, 34(665 670). Fukuda, A. (1986). Input regulation control based on periodical monitoring using call gapping control. Electronics Comm. Japan, Part 1, 69(11):84 92. Hassin, R. and Haviv, M. (2003). To Queue or Not to Queue: Equilibrium Behavior in Queueing Systems. Kluwer Academic Publishers. Hordijk, A. (1974). Dynamic Programming and Markov Potential Theory. Technical report, Math. Centre Tract No. 51, Mathematisch Centrum, Amsterdam. Kuri, J. and Kumar, A. (1995). Optimal control of arrivals to queues with delayed queue length information. IEEE Tran. Automatic Control, 40(8):1444 1450. Lin, K. Y. and Ross, S. M. (2003). Admission control with incomplete information of a queueing system. Operations Research, 51(4):645 654. Naor, P. (1969). The regulation of queue size by levying tolls. Econometrica, 37(1). Puterman, M. L. (1994). Markov Decision Processes, Discrete Stochastic Dynamic Programming. John Wiley & Sons, Inc. Sennott, L. I. (1989). Average cost optimal stationary policies in infinite state Markov decision processes with unbounded costs. Operations Research, 37:626 633. Sennott, L. I. (1998). Stochastic Dynamic Programming and the Control of Queueing Systems. John Wiley & Sons. Sennott, L. I. (2000). Markov Decision Processes: Model, Methods and Open Problems, chapter Average reward optimization theory for denumerable state spaces. Kluwer 2000. Stidham Jr., S. (1985). Optimal control of admission to a queueing system. IEEE Tran. on Automatic Control, 30(8).

Stidham Jr., S. (1988). Stochastic differential systems: stochastic control theory and applications, chapter Scheduling, Routing, and Flow Control in Stochastic Networks, pages 529 561. Springer-Verlag. Stidham Jr., S. (2002). Analysis, design and control of queueing systems. Operations Research, 50(1):197 216. Stidham Jr., S. and Weber, R. (1993). A survey of Markov decision models for control of networks of queues. Queueing Systems, 13:291 314. Whitt, W. (1978). Approximations of dynamic programs I. Mathematics of Operations Research, 3(3):231 243. Whitt, W. (1979). Approximations of dynamic programs, II. Mathematics of Operations Research, 4(2):179 185.