Point Process Control

Similar documents
Dynamic Pricing for Non-Perishable Products with Demand Learning

Optional Stopping Theorem Let X be a martingale and T be a stopping time such

Liquidation in Limit Order Books. LOBs with Controlled Intensity

Order book modeling and market making under uncertainty.

An essay on the general theory of stochastic processes

1 Delayed Renewal Processes: Exploiting Laplace Transforms

Jump Processes. Richard F. Bass

Filtrations, Markov Processes and Martingales. Lectures on Lévy Processes and Stochastic Calculus, Braunschweig, Lecture 3: The Lévy-Itô Decomposition

Multiple-Item Dynamic Pricing under a Common Pricing Constraint

Chapter 1. Poisson processes. 1.1 Definitions

Figure 10.1: Recording when the event E occurs

CIMPA SCHOOL, 2007 Jump Processes and Applications to Finance Monique Jeanblanc

Functional Limit theorems for the quadratic variation of a continuous time random walk and for certain stochastic integrals

Optimal Control of an Inventory System with Joint Production and Pricing Decisions

Exponential Distribution and Poisson Process

Manual for SOA Exam MLC.

ELEMENTS OF PROBABILITY THEORY

Let (Ω, F) be a measureable space. A filtration in discrete time is a sequence of. F s F t

Deterministic Dynamic Programming

Modern Discrete Probability Branching processes

Chapter 2. Poisson Processes. Prof. Shun-Ren Yang Department of Computer Science, National Tsing Hua University, Taiwan

1. Stochastic Process

Modelling the risk process

STAT 331. Martingale Central Limit Theorem and Related Results

Solutions For Stochastic Process Final Exam

Data analysis and stochastic modeling

1 Basics of probability theory

Lecture 6 Random walks - advanced methods

µ X (A) = P ( X 1 (A) )

On the martingales obtained by an extension due to Saisho, Tanemura and Yor of Pitman s theorem

(b) What is the variance of the time until the second customer arrives, starting empty, assuming that we measure time in minutes?

Theory and Applications of Stochastic Systems Lecture Exponential Martingale for Random Walk

Simplex Algorithm for Countable-state Discounted Markov Decision Processes

Tyler Hofmeister. University of Calgary Mathematical and Computational Finance Laboratory

arxiv:math/ v4 [math.pr] 12 Apr 2007

Limit theorems for Hawkes processes

On the Approximate Linear Programming Approach for Network Revenue Management Problems

The exponential distribution and the Poisson process

Fundamental Inequalities, Convergence and the Optional Stopping Theorem for Continuous-Time Martingales

arxiv: v3 [math.oc] 11 Dec 2018

Stochastic integration. P.J.C. Spreij

L2 gains and system approximation quality 1

Chapter 2 SOME ANALYTICAL TOOLS USED IN THE THESIS

Renewal theory and its applications

HJB equations. Seminar in Stochastic Modelling in Economics and Finance January 10, 2011

PROBABILITY: LIMIT THEOREMS II, SPRING HOMEWORK PROBLEMS

Class 11 Non-Parametric Models of a Service System; GI/GI/1, GI/GI/n: Exact & Approximate Analysis.

Stocking Retail Assortments Under Dynamic Consumer Substitution

Some Aspects of Universal Portfolio

Markov Chains. X(t) is a Markov Process if, for arbitrary times t 1 < t 2 <... < t k < t k+1. If X(t) is discrete-valued. If X(t) is continuous-valued

Lecture 4a: Continuous-Time Markov Chain Models

Optimal Control. Macroeconomics II SMU. Ömer Özak (SMU) Economic Growth Macroeconomics II 1 / 112

Markov Chain BSDEs and risk averse networks

Poisson Processes. Stochastic Processes. Feb UC3M

A Randomized Linear Program for the Network Revenue Management Problem with Customer Choice Behavior. (Research Paper)

Optimal stopping for non-linear expectations Part I

Stochastic Processes II/ Wahrscheinlichkeitstheorie III. Lecture Notes

A Revisit to the Markup Practice of Dynamic Pricing

Lecture 21 Representations of Martingales

Other properties of M M 1

2905 Queueing Theory and Simulation PART III: HIGHER DIMENSIONAL AND NON-MARKOVIAN QUEUES

Almost sure convergence to zero in stochastic growth models

Lecture 2. We now introduce some fundamental tools in martingale theory, which are useful in controlling the fluctuation of martingales.

1 Some basic renewal theory: The Renewal Reward Theorem

Risk Measures in non-dominated Models

Birgit Rudloff Operations Research and Financial Engineering, Princeton University

PROBABILITY: LIMIT THEOREMS II, SPRING HOMEWORK PROBLEMS

An Uncertain Control Model with Application to. Production-Inventory System

Likelihood Function for Multivariate Hawkes Processes

MATH 56A: STOCHASTIC PROCESSES CHAPTER 6

A tractable consideration set structure for network revenue management

Distributed Optimization. Song Chong EE, KAIST

March 1, Florida State University. Concentration Inequalities: Martingale. Approach and Entropy Method. Lizhe Sun and Boning Yang.

Jump-type Levy Processes

Risk-Minimality and Orthogonality of Martingales

Exponential martingales: uniform integrability results and applications to point processes

OPTIMALITY OF RANDOMIZED TRUNK RESERVATION FOR A PROBLEM WITH MULTIPLE CONSTRAINTS

STAT331 Lebesgue-Stieltjes Integrals, Martingales, Counting Processes

Dynamic Pricing Strategies for Multi-Product Revenue Management Problems

Properties of an infinite dimensional EDS system : the Muller s ratchet

Oblivious Equilibrium: A Mean Field Approximation for Large-Scale Dynamic Games

Thermodynamic limit and phase transitions in non-cooperative games: some mean-field examples

Lecture 17 Brownian motion as a Markov process

Gallego et al. [Gallego, G., G. Iyengar, R. Phillips, A. Dubey Managing flexible products on a network.

STAT331. Combining Martingales, Stochastic Integrals, and Applications to Logrank Test & Cox s Model

MASSACHUSETTS INSTITUTE OF TECHNOLOGY 6.265/15.070J Fall 2013 Lecture 7 9/25/2013

Introduction Optimality and Asset Pricing

I forgot to mention last time: in the Ito formula for two standard processes, putting

Dynkin (λ-) and π-systems; monotone classes of sets, and of functions with some examples of application (mainly of a probabilistic flavor)

n E(X t T n = lim X s Tn = X s

CONTROL SYSTEMS, ROBOTICS AND AUTOMATION Vol. XI Stochastic Stability - H.J. Kushner

Poisson random measure: motivation

Selected Exercises on Expectations and Some Probability Inequalities

Exercises Measure Theoretic Probability

The Stochastic Knapsack Revisited: Switch-Over Policies and Dynamic Pricing

Optimal Execution Tracking a Benchmark

Mathematical Methods for Neurosciences. ENS - Master MVA Paris 6 - Master Maths-Bio ( )

Lecture Notes on Risk Theory

Poisson Processes. Particles arriving over time at a particle detector. Several ways to describe most common model.

Random Times and Their Properties

Transcription:

Point Process Control The following note is based on Chapters I, II and VII in Brémaud s book Point Processes and Queues (1981). 1 Basic Definitions Consider some probability space (Ω, F, P). A real-valued mapping X : Ω R is a random variable if for every C B(R) the pre-image X 1 (C) F. A filtration (or history) of a measurable space (Ω, F) is a collection (F t ) t of sub-σ-fields of F such that for all s t F s F t. We denote by F = ( ) t F t := σ t F t. A family (X t ) t of real-valued random variables is called a stochastic process. generated by X t is F X t := σ (X s : s [, t]). The filtration For a fixed ω Ω, the function X t (ω) is called a path of the stochastic process. We say that the stochastic process is adapted to the filtration F t if Ft X F t for all t. We say that X t is F t -progressive if for all t the mapping (t, ω) X t (ω) from [, t] Ω R is B([, t]) F t -measurable. Let F t be a filtration. We define the F t -predictable σ-field P(F t ) as follows: P(F t ) := σ ((s, t] A : s [, t] and A F s ). A stochastic process X t is F t -predictable if X t is P(F t )-measurable. Proposition 1 A real-valued stochastic process X t adapted to F t and left-continuous is F t -predictable Given a filtration F t, a process X t is called a F t -martingale over [, c] if the following three conditions are satisfied 1. X t is adapted to F t. 2. E[ X t ] < for all t [, c].

3. E[X t F s ] = X s a.s., for all s t c. If the equality in (3) is replaced by ( ) then X t is called a submartingale (supermartingale). Exercise 1: Let X t be a real-valued process with independent increment, that is, for all s t, X t X s is independent of F X s. Suppose that X t is integrable and E[ X t ] =. Show that X t is a F X t - martingale. If in addition X 2 t is integrable then X 2 t is a F X t -submartingale and X 2 t E[X 2 t ] is a F t -martingale. 2 Counting Processes Definition 1 A sequence of random variables {T n n, T n F and : n } is called a point process if for each ω Ω T (ω) = and T n (ω) < T n+1 (ω) whenever T n <. We will only consider nonexplosive point process, that is, process for which lim n T n = P-a.s. Associated to a nonexplosive point process {T n }, we define the corresponding counting process {N t : t } as follows N t = n if t [T n, T n+1 ). Thus, N t is a right-continuous step function starting at (see figure). Since we are considering N t 5 4 3 2 1 T T 1 T 2 T 3 T 4 T 5 t nonexplosive point processes, N t < for all t P-a.s. In addition, if E[N t ] is finite for all t then the point process is said to be integrable. Example 1: Simple Renewal Process 2

Consider a sequence {X n : n 1} of iid nonnegative random variables. We define the point process recursively as follows: T = and T n = T n 1 + X n, n 1. A sufficient condition for the process T n to be nonexplosive is E[X] >. One of the most famous renewal process is the Poisson process. In this case X, the inter-arrival interval, has an exponential distribution with time homogenous rate λ. Example 2: Queueing Process A simple queueing process Q t is a nonnegative integer-valued process o of the form Q t = Q + A t D t, where A t (arrival process) and D t (departure process) are two nonexplosive point processes without common jumps. Note that by definition D t Q + A t. Exercise 2: Let N t (1) and N t (2) be two nonexplosive point processes without common jumps and let Q be a nonnegative integer-valued random variable. Define X t = Q + N t (1) N t (2) and m t = min s t {Xs }. Show that Q t = X t m t is a simple queueing process with arrival process A t = N t (1), departure process D t = t 1(Q s > ) dn s (2), and m t = t 1(Q s = ) dn s (2). Poisson processes are commonly used in practice to represent point process. One possible explanation for this popularity is its inherent mathematical tractability. However, a well-known result in renewal theory says that the sum of a large number of independent renewal processes converges as the number of summands goes to infinity to a Poisson process. So, the Poisson process can be in fact a good approximation to many applications in practice. To make the Poisson process even more realistic we would like to have more flexibility on the arrival rate λ. We can achieve this generalization as follows. Definition 2 (Doubly Stochastic or Conditional Poisson Process) Let N t be a point process adapted to a filtration F t, and let λ t be a nonnegative measurable process. Suppose that and that λ t is F measurable for all t, t λs ds < P-a.s., for all t. If for all s [, t] the increment N t N s is independent of F s given F and P[N t N s = k F s ] = 1 k! exp ( t s ) ( t ) k λ u du λ u du s then N t is called a (P F t )-doubly stochastic Poisson process with stochastic intensity λ t. 3

The process λ t is referred as the intensity of the process. A special case of a doubly stochastic Poisson process occurs when λ t = λ F. Another example is the case where λ t = f(t, Y t ) for a measurable nonnegative function f and a process Y t such that F Y F. Exercise 3: Show that for a doubly stochastic Poisson process N t is such that E[ t λ s ds] < for all t then is a F t -martingale. M t = N t t λ s ds Based on this observation we have the following important result. Proposition 2 If N t is an integrable doubly stochastic Poisson process with F t -intensity λ t, then for all nonnegative F t -predictable processes C t [ ] [ E C s dn s = E ] C s λ s ds where t C s dn s := n 1 C T n 1(T n t). It turns out that the converse is also true. This was first proved by Watanabe in a less general setting. Proposition 3 (Watanabe (1964)) Let N t be a point process adapted to the filtration F t, and let λ(t) be a locally integrable nonnegative function. Suppose that N t t λ(s) ds is an F t-martingale. Then N t is Poisson process with intensity λ(t), that is, for all s t, N t N s is a Poisson random variable with parameter t s λ u du independent of F s. Motivated by this result we define the notion of stochastic intensity for an arbitrary point process as follows. Definition 3 (Stochastic Intensity) Let N t be a point process adapted to some filtration F t, and let λ t be a nonnegative F t -progressive process such that for all t t λ s ds < P a.s. If for all nonnegative F t predictable processes C t, the equality [ ] [ ] E C s dn s = E C s λ s ds is verified, then we say that N t admits the F t -intensity λ t. 4

Exercise 4: Let N t be a point process with the F t -intensity λ t. Show that if λ t id G t -progressive for some filtration G t such that F N t G t F t t then λ t is also the G t -intensity N t. Similarly to the Poisson process, we can connect point processes with stochastic intensities to martingales. Proposition 4 (Integration Theorem) If N t admits the F t -intensity λ t (where t λ s ds < a.s.) then N t is nonexplosive and 1. M t = N t t λ s ds is an F t -local martingale. 2. if X t is F t -predictable process such that E[ t X s λ s ds] < then t X s dm s is an F t - martingale. 3. if X t is F t -predictable process such that t X s λ s ds < a.s. then t X s dm s is an F t -local martingale. 3 Optimal Intensity Control In this section we study the problem of controlling a point process. In particular, we focus on the case where the controller can affect the intensity of the point process. This type of control differs from impulsive control where the controller has the ability to add or erase some of the point in the sequence. We consider a point process N t that we wish to control. The control u belongs to a set U of admissible controls. We will assume that U consists on the set of real-valued processes defined on (Ω, F) adapted to Ft N in addition for each t [, T ] we assume that u t U t. In addition, for each u U the point process N t admits a (P u, F t )-intensity λ t (u). Here, F t is some filtration associated to N t. The performance measure that we will consider is given by [ T ] J(u) = E u C s (u) ds + φ T (u). (1) The expectation in J(u) above is taken with respect to P u. The function C t (u) is an F t -progressive process and φ T (u) is a F T -measurable random variable. We will consider a problem with complete information so that F t Ft N. In addition, we assume local dynamics u t = u(t, N t ) is Ft N predictable λ t (u) = λ(t, N t, u t ) C t (u) = C(t, N t, u t ) φ T = φ T (T, N T ). 5

Exercise 5: Consider the cost function J(u) = E <T n T k Tn (u). Where k t (u) is a nonnegative F t -measurable process. Show that this cost function can be written in the from given by equation (1). 3.1 Dynamic Programming for Intensity Control Theorem 1 (Hamilton-Jacobi Sufficient Conditions) Suppose there exists for each n N + a differentiable bounded F t -progressive mapping V (t, ω, n) such that all ω Omega and all n N + V (t, ω, n) t + inf {λ(t, ω, n, v) [V (t, ω, n + 1) V (t, ω, n)] + C(t, ω, n, v)} = v U t (2) V (T, ω, n) = φ(t, ω, n). (3) and suppose there exists for each n N + an F N t -predictable process u (t, ω, n) such that u (t, ω, n) achieves the minimum in equation (2). Then, u is the optimal control. Exercise 6: Proof the theorem. This Theorem lacks of practical applicability because of Value Function is in general path dependent. The analysis can be greatly simplified if we assume that the problem is Markovian. Corollary 1 (Markovian Control) Suppose that λ(t, ω, n, v), C(t, ω, n, v), and φ(t, ω, n) do not dependent on ω and that there is a function V (t, n) such that V (t, n) t + inf {λ(t, n, v) [V (t, n + 1) V (t, n)] + C(t, n, v)} = v U t (4) V (T, n) = φ(t, n). (5) suppose that the minimum is achieved by a measurable function U (t, n). Then, u is the optimal control. 4 Applications to Revenue Management In this section we present an application of point process optimal control based on the work by Gallego and van Ryzin (1994). Optimal Dynamic Pricing of Inventories with Stochastic Demand over Finite Horizons, Mgmnt. Sci. 4, 999-12. 6

4.1 Model Description and HJB Equation Consider a seller that owns I units of a product that wants to sell over a fixed time period T. Demand for the product is characterized by a point process N t. Given a price policy p, N t admits the intensity λ(p t ). The seller s problem is to select a price strategy {p t : t [, T ]} (a predictable process) that maximizes the expected revenue over the selling horizon. That is, [ T ] max J p (t, I) := E p p s dn s subject to T dn s I P p a.s. In order to ensure that the problem is feasible we will assume that there exist a price p such that λ(p ) = a.s. In the case, we define the set of admissible pricing policies A, as the set of predictable p(t, I N t ) policies such that p(t, ) = p. The seller s problem becomes to maximize over p A the expected revenue J p (t, I). Using corollary 1, we can write the optimality condition for this problem as follows: V (t, n) + sup {λ(p) [V (t, n) V (t, n 1)] p λ(p)} = t p V (T, n) =. Let us make the following transformation of time t T t, that is, t measures the remaining selling time. Also, instead of looking at the price p as the decision variable we use λ as the control and p(λ) as the inverse demand function. The revenue rate r(λ) := λ p(λ) is assumed to be regular, i.e., continuous, bounded, concave, has a bounded maximizer λ = min{ λ = argmax{r(λ)}} and such that lim λ r(λ) =. Under this new definitions the HJB equation becomes V (t, n) = sup {r(λ) λ [V (t, n) V (t, n 1)]} = t λ V (, n) =. Proposition 5 If λ(p) is a regular demand function then there exists a unique solution to the HJB equation. Further, the optimal intensities satisfies λ (t, n) λ for all n for all t T. Closed-form solution to the HJB equation are generally intractable however, the optimality condition can be exploited to get some qualitative results about the optimal solution. Theorem 2 The optimal value function V (t, n) is strictly increasing and strictly concave in both n and t. Furthermore, there exists an optimal intensity λ (n, t) that is strictly increasing in n and strictly decreasing in t. The following figure plots a sample path of price and inventory under an optimal policy. 7

22 2 18 16 14 Price 12 1 8 6 Inventory 4 2 1 2 3 4 5 6 7 8 9 1 Time Figure 1: Path of an optimal price policy and its inventory level. Demand is a time homogeneous Poisson process with intensity λ(p) = exp(.1p), the initial inventory is C = 2, and the selling horizon is H = 1. The dashed line corresponds to the minimum price p min = 1. 4.2 Bounds and Heuristics The fact that the HJB is intractable in most cases creates the need for alternative solution methods. One possibility, that we consider here, is the use of the certainty equivalent version of the problem. That is, the deterministic control problem resulting from changing all uncertainty by its expected value. In this case, the deterministic version of the problem is given by subject to V D (T, n) = max T λ s ds x. λ T r(λ s ) ds The solution to this time-homogeneous problem can be found easily. Let λ (T, n) = n T, that is, the, run-out rate. Then, it is straightforward to show that the optimal deterministic rate is λ D = min{λ, λ (T, n)} and the optimal expect revenue is V D (T, n) = T min{r(λ ), r(λ (T, n))}. At least two things make the deterministic solution interesting. Theorem 3 If λ(p) is a regular demand function then for all n and t V (t, n) V D (t, n). 8

Thus, the deterministic value function provides an upper bound on the optimal expected revenue. In addition, if we fixed the price at p D and we denote by V F P (t, n) the expected revenue collected from this fixed price strategy then we have the following important result. Theorem 4 V F P (t, n) V (t, n) 1 1 2 min{n, λ t}. Therefore, the fixed price policy is asymptotically optimal as the number of product n or the selling horizon t become large. 9