On Ergodic Impulse Control with Constraint

Size: px
Start display at page:

Download "On Ergodic Impulse Control with Constraint"

Transcription

1 On Ergodic Impulse Control with Constraint Maurice Robin Based on joint papers with J.L. Menaldi University Paris-Sanclay 9119 Saint-Aubin, France ( IMA, Minneapolis, MN May 7 11, 218 MR (UP-S) Ergodic Impulse Control with Constraint 1 / 19

2 Content - Statement of the problem - HJB equation - Solution of the HJB equation - Existence of an optimal control - Extensions - References MR (UP-S) Ergodic Impulse Control with Constraint 2 / 19

3 Introduction Statement of the Problem Statement of the Problem (as in JL Menaldi s talk except for the cost and ergodic assumptions) - The uncontrolled state is described by a Markov-Feller process x t (values in E metric compact) - impulse control ν = (θ i, ξ i ) i 1, θ i increasing sequence of stopping times, ξ i E valued random variable - constraint on impulse controls: θ i > and θ i is a jump time of the signal process y t y τn =, y t = t τ n for τ n t τ n+1, n 1, T n = τ n+1 τ n, conditionally to x t as IID random variables with intensity λ(x, y) - ξ i Γ(x θi ), Γ(x) closed set of E and ξ Γ(x), Γ(ξ) Γ(x) MR (UP-S) Ergodic Impulse Control with Constraint 3 / 19

4 Introduction Statement of the Problem (2) Statement of the Problem (2) - running cost f (x, y) and impulse cost c(x, ξ), both positive bounded and continuous, c(x, ξ) c > and c(x, ξ) + c(ξ, ξ ) c(x, ξ ) - Mg(x) inf ξ Γ(x) {c(x, ξ) + g(x)} is assumed to be continuous if g is continuous and there exists a measurable selector ˆξ(x, g). V will denote the set of admissible controls V = {(θ i, ξ i ), i 1, θ 1 >, y θi = } V the set of admissible controls satisfying the constraint, but θ 1 = is allowed MR (UP-S) Ergodic Impulse Control with Constraint 4 / 19

5 Introduction Statement of the Problem (3) Statement of the Problem (3) - Controlled process The controlled process for a control ν is defined on the product space Ω, Ω = D(R + ; E R + ) by a probability P ν xy, and (x ν t, y ν t ) = (x i t, y i t ) for θ i 1 t < θ i evolves as the uncontrolled process between impulses instants. MR (UP-S) Ergodic Impulse Control with Constraint 5 / 19

6 Introduction Statement of the Problem (4) Statement of the Problem (4) - The average cost to be minimized: J(x, y, ν) = lim inf T 1 { T T Eν xy f (xs ν, ys ν )ds + i } 1 θi T c(x i 1 θ i, ξ i ) µ(x, y) = inf{j(x, y, ν) : ν V} We will use an auxiliary problem J(x, 1 { y, ν) = lim inf n E ν E ν τ n xy f (xs ν, ys ν )ds + xyτ n i } 1 θi T c(x i 1 θ i, ξ i ) µ (x, y) = inf{j(x, y, ν) : ν V } MR (UP-S) Ergodic Impulse Control with Constraint 6 / 19

7 Introduction Additional Assumptions Additional Assumptions λ(x, y) is bounded and continuous and Ergodicity assumption: < a 1 E x (τ 1 ) a 2 P(x, B) = E x 1 B (x τ1 ) B B(E) satisfies: there exists a positive measure m on E s.t. < m(e) 1 and P(x, B) m(b) B B(E) Example: reflected diffusions and reflected diffusions with jumps for which the transition density satisfies p(x, t, x ) k(ε) on E [ε, [ E MR (UP-S) Ergodic Impulse Control with Constraint 7 / 19

8 Solution HJB Equation HJB Equation Heuristic argument with the discounted problem { { u α (x, ) = min Mu α τ 1 } } (x, ), E x e αt f dt + e ατ 1 u α (x τ1, ) Let m α = inf u α(x, ), w α = uα m α, { { w α (x, ) = min Mw α τ 1 } } (x, ), E x e αt (f αm α )dt+e ατ 1 w α (x τ1, ) Assuming w α w a function, and αm α µ a constant, { { τ 1 } } w (x, ) = min Mw (x, ), E x (f µ )dt + w (x τ1, ) One can also use heuristic argument on { u T (t, x, y) = inf E ν T t x f dt + i } 1 θi T tc(x i 1 θ i, ξ i ). MR (UP-S) Ergodic Impulse Control with Constraint 8 / 19

9 Solution HJB Equation HJB Equation (2) - For the auxiliary problem: Find (w, µ ), µ constant, such that { { τ 1 } } w (x, ) = min Mw (x, ), E x [f µ ]ds + w (x τ1, ) then w (x, y), for y >, is given by w (x, y) = E xy { τ 1 } [f µ ]ds + w (x τ1, ) - For the initial problem: (w, µ ) gives w(x, y) as w(x, y) = E xy { τ 1 } [f µ ]ds + w (x τ1, ) MR (UP-S) Ergodic Impulse Control with Constraint 9 / 19

10 Solution Solution (µ, w ) Solution (µ, w ) - A discrete time HJB equation for ( µ, w (x, ) ) : define l(x) = E x { τ 1 } f (x s, y s )ds, Pg(x) = E x g(x τ1 ) τ(x) = E x τ 1, w (x) w (x, ), then { w (x) = min inf ξ Γ(x) { c(x, ξ) + w (ξ) }, l(x) µ τ(x) + Pw (x)} is equivalent to the previous HJB equation Proposition There exists a solution (µ, w ) in R + C(E) of the HJB equation. Remark: If w is solution, w + constant is also solution. The uniqueness of µ will come from the stochastic interpretation. MR (UP-S) Ergodic Impulse Control with Constraint 1 / 19

11 Solution Solution (µ, w ) Arguments The proof uses the following equivalent form of the HJB equation w (x) = inf ξ Γ(x) {x} { l(ξ) + 1ξ x c(x, ξ) µ τ(ξ) + Pw (ξ) } = Rw and the fact that P(x, β) τ(x)γ(β) for a positive measure γ on E satisfying γ(e) > 1 β τ(x), < β < 1. Then R is a contraction on C(E). w is the unique fixed point and µ = E w (x)γ(dx) MR (UP-S) Ergodic Impulse Control with Constraint 11 / 19

12 Solution Existence of an Optimal Control Existence of an Optimal Control Additional assumptions: the (uncontrolled) Markov process (x t, y t ) has a unique invariant measure ζ and there exists a continuous function h(x, y) s.t. { τ } E xy h(x τ, y τ ) = h(x, y) E xy [f (x t, y t ) f ]dt, for any finite stopping time τ, with f = f (x, y)dζ. E R + Remark: if f (x, y) = f (x), then it is sufficient to assume that the Poisson equation for x t, i.e. A x h = f (x) f has a continuous solution. MR (UP-S) Ergodic Impulse Control with Constraint 12 / 19

13 Solution Existence of an Optimal Control (2) Existence of an Optimal Control (2) Theorem With the additional assumption, we have µ = inf ( J(x,, ν), ν V ) and there exists an optimal control ˆν of the auxiliary problem case 1. µ = f : then it is optimal to do nothing case 2. µ < f : one can rewrite the HJB equation w(x) = min { ψ(x), l(x) + P w(x) } with w = w h(x, ), ψ = Mw h(x, ), l(x) = ( f µ )E x τ 1. This is the HJB equation of a discrete optimal stopping problem which has an optimal control ˆη = inf { n : w(x n ) = ψ(x n ) }, i.e., ˆη = inf { n : w (x n ) = Mw (x n ) } where x n is the Markov chain x τn. From this, we deduce an optimal control with θ 1 = τˆη, and θ i = θˆηi with ˆη i = inf { n ˆη i : w (x n ) = Mw (x n ) }. MR (UP-S) Ergodic Impulse Control with Constraint 13 / 19

14 Solution Existence of an Optimal Control (3) Existence of an Optimal Control (3) Corollary µ = inf { J(x, y, ν), v V } and the optimal control ˆν obtained by translation by τ 1 of the control ˆν. The final result is given by Theorem µ = inf{j(x, y, ν) : ν V} = J(x, y, ˆν) MR (UP-S) Ergodic Impulse Control with Constraint 14 / 19

15 Solution Existence of an Optimal Control (4) Existence of an Optimal Control (4) A first step is to prove: Proposition (µ, w ) being the solution previously obtained and recalling that (µ, w) is solution of w(x, y) = E x { τ 1 } [f µ ]ds + w (x τ1, ) A xy w(x, y) + λ(x, y)[w(x, ) Mw(x, )] + = f µ. where A xy is the (weak) infinitesimal generator of the uncontrolled process A xy ϕ = A x ϕ + ϕ y + λ(x, y)[ ϕ(x, ) ϕ(x, y) ]. MR (UP-S) Ergodic Impulse Control with Constraint 15 / 19

16 Solution Existence of an Optimal Control (5) Existence of an Optimal Control (5) To prove the proposition, one first shows which gives w (x) = min{w(x, ), Mw(x, )} { τ 1 w(x, y) = E xy [f µ ]dt + w(x τ1, ) [w(x τ1, ) Mw] +} from which we deduce that equation. Next, the proposition allows us to show that M T = T [f (x t, y t ) µ ]ds + w(x T, y T ) is a submartingale. This gives µ J(x, y, ν), ν V, and, from the first expression of w(x, y), on obtains µ = J(x, y, ν). MR (UP-S) Ergodic Impulse Control with Constraint 16 / 19

17 Solution Existence of an Optimal Control (5) Extensions E locally compact. If we assume Φ(t)C (E) C (E) all other assumptions being unchanged, one can still obtain the results on the HJB equation. However, the additional assumption is no longer sufficient to get the result on the existence of an optimal cost. adding h(x, ) bounded would be sufficient (e.g., if A xy h = f f has a bounded solution), but more general assumptions would require further work. When λ is independent of x and f independent of y, one can obtain an optimal control without an additional assumption. MR (UP-S) Ergodic Impulse Control with Constraint 17 / 19

18 Solution Existence of an Optimal Control (5) Extensions (2) P(x, B) m(b) is also restrictive for E locally compact Replacing by a localized condition like P(x, B) α1 K (x)m(b), allows to obtains the results on the HJB equations and, if h(x, ) is bounded, the existence of an optimal control. MR (UP-S) Ergodic Impulse Control with Constraint 18 / 19

19 Solution References (among others) References (among others) Besides the classical works on impulse control, A. Bensoussan and J.L. Lions (book 1982), etc,..., For discrete time control: A. Bensoussan (book, 211), O. Hernandez-Lerma and J. Lasserre (books 1996, 1999),... For average control, ergodicity: Bensoussan (book 1988), M. Kurano (paper 1989), F. Luque-Vasquez and O. Hernandez-Lerma (paper 1999), L. Stettner (paper 1986), D. Gatarek and L. Stettner (paper 199), A. Arapostathis et al. (survey paper 1993), M.G. Garroni and J.L. Menaldi (book 22), Several papers with J.L. Menaldi (1997, 213,... ), Papers to appear with J.L. Menaldi, H. Jasso-Fuentes, T. Prieto-Rumeau. MR (UP-S) Ergodic Impulse Control with Constraint 19 / 19

On Stopping Times and Impulse Control with Constraint

On Stopping Times and Impulse Control with Constraint On Stopping Times and Impulse Control with Constraint Jose Luis Menaldi Based on joint papers with M. Robin (216, 217) Department of Mathematics Wayne State University Detroit, Michigan 4822, USA (e-mail:

More information

On Some Ergodic Impulse Control Problems with Constraint

On Some Ergodic Impulse Control Problems with Constraint Wayne State University Mathematics Faculty Research Publications Mathematics 7-19-218 On Some Ergodic Impulse Control Problems with Constraint J. L. Menaldi Wayne State University, menaldi@wayne.edu Maurice

More information

Zero-sum semi-markov games in Borel spaces: discounted and average payoff

Zero-sum semi-markov games in Borel spaces: discounted and average payoff Zero-sum semi-markov games in Borel spaces: discounted and average payoff Fernando Luque-Vásquez Departamento de Matemáticas Universidad de Sonora México May 2002 Abstract We study two-person zero-sum

More information

SUPPLEMENT TO CONTROLLED EQUILIBRIUM SELECTION IN STOCHASTICALLY PERTURBED DYNAMICS

SUPPLEMENT TO CONTROLLED EQUILIBRIUM SELECTION IN STOCHASTICALLY PERTURBED DYNAMICS SUPPLEMENT TO CONTROLLED EQUILIBRIUM SELECTION IN STOCHASTICALLY PERTURBED DYNAMICS By Ari Arapostathis, Anup Biswas, and Vivek S. Borkar The University of Texas at Austin, Indian Institute of Science

More information

Multi-dimensional Stochastic Singular Control Via Dynkin Game and Dirichlet Form

Multi-dimensional Stochastic Singular Control Via Dynkin Game and Dirichlet Form Multi-dimensional Stochastic Singular Control Via Dynkin Game and Dirichlet Form Yipeng Yang * Under the supervision of Dr. Michael Taksar Department of Mathematics University of Missouri-Columbia Oct

More information

Continuous dependence estimates for the ergodic problem with an application to homogenization

Continuous dependence estimates for the ergodic problem with an application to homogenization Continuous dependence estimates for the ergodic problem with an application to homogenization Claudio Marchi Bayreuth, September 12 th, 2013 C. Marchi (Università di Padova) Continuous dependence Bayreuth,

More information

Feller Processes and Semigroups

Feller Processes and Semigroups Stat25B: Probability Theory (Spring 23) Lecture: 27 Feller Processes and Semigroups Lecturer: Rui Dong Scribe: Rui Dong ruidong@stat.berkeley.edu For convenience, we can have a look at the list of materials

More information

Laplace s Equation. Chapter Mean Value Formulas

Laplace s Equation. Chapter Mean Value Formulas Chapter 1 Laplace s Equation Let be an open set in R n. A function u C 2 () is called harmonic in if it satisfies Laplace s equation n (1.1) u := D ii u = 0 in. i=1 A function u C 2 () is called subharmonic

More information

Markov processes and queueing networks

Markov processes and queueing networks Inria September 22, 2015 Outline Poisson processes Markov jump processes Some queueing networks The Poisson distribution (Siméon-Denis Poisson, 1781-1840) { } e λ λ n n! As prevalent as Gaussian distribution

More information

Consistency of the maximum likelihood estimator for general hidden Markov models

Consistency of the maximum likelihood estimator for general hidden Markov models Consistency of the maximum likelihood estimator for general hidden Markov models Jimmy Olsson Centre for Mathematical Sciences Lund University Nordstat 2012 Umeå, Sweden Collaborators Hidden Markov models

More information

1 Continuous-time chains, finite state space

1 Continuous-time chains, finite state space Université Paris Diderot 208 Markov chains Exercises 3 Continuous-time chains, finite state space Exercise Consider a continuous-time taking values in {, 2, 3}, with generator 2 2. 2 2 0. Draw the diagramm

More information

The tree-valued Fleming-Viot process with mutation and selection

The tree-valued Fleming-Viot process with mutation and selection The tree-valued Fleming-Viot process with mutation and selection Peter Pfaffelhuber University of Freiburg Joint work with Andrej Depperschmidt and Andreas Greven Population genetic models Populations

More information

Variational approach to mean field games with density constraints

Variational approach to mean field games with density constraints 1 / 18 Variational approach to mean field games with density constraints Alpár Richárd Mészáros LMO, Université Paris-Sud (based on ongoing joint works with F. Santambrogio, P. Cardaliaguet and F. J. Silva)

More information

ON THE POLICY IMPROVEMENT ALGORITHM IN CONTINUOUS TIME

ON THE POLICY IMPROVEMENT ALGORITHM IN CONTINUOUS TIME ON THE POLICY IMPROVEMENT ALGORITHM IN CONTINUOUS TIME SAUL D. JACKA AND ALEKSANDAR MIJATOVIĆ Abstract. We develop a general approach to the Policy Improvement Algorithm (PIA) for stochastic control problems

More information

1. Stochastic Processes and filtrations

1. Stochastic Processes and filtrations 1. Stochastic Processes and 1. Stoch. pr., A stochastic process (X t ) t T is a collection of random variables on (Ω, F) with values in a measurable space (S, S), i.e., for all t, In our case X t : Ω S

More information

REMARKS ON THE EXISTENCE OF SOLUTIONS IN MARKOV DECISION PROCESSES. Emmanuel Fernández-Gaucherand, Aristotle Arapostathis, and Steven I.

REMARKS ON THE EXISTENCE OF SOLUTIONS IN MARKOV DECISION PROCESSES. Emmanuel Fernández-Gaucherand, Aristotle Arapostathis, and Steven I. REMARKS ON THE EXISTENCE OF SOLUTIONS TO THE AVERAGE COST OPTIMALITY EQUATION IN MARKOV DECISION PROCESSES Emmanuel Fernández-Gaucherand, Aristotle Arapostathis, and Steven I. Marcus Department of Electrical

More information

Average-cost temporal difference learning and adaptive control variates

Average-cost temporal difference learning and adaptive control variates Average-cost temporal difference learning and adaptive control variates Sean Meyn Department of ECE and the Coordinated Science Laboratory Joint work with S. Mannor, McGill V. Tadic, Sheffield S. Henderson,

More information

The coupling method - Simons Counting Complexity Bootcamp, 2016

The coupling method - Simons Counting Complexity Bootcamp, 2016 The coupling method - Simons Counting Complexity Bootcamp, 2016 Nayantara Bhatnagar (University of Delaware) Ivona Bezáková (Rochester Institute of Technology) January 26, 2016 Techniques for bounding

More information

Mean Field Games on networks

Mean Field Games on networks Mean Field Games on networks Claudio Marchi Università di Padova joint works with: S. Cacace (Rome) and F. Camilli (Rome) C. Marchi (Univ. of Padova) Mean Field Games on networks Roma, June 14 th, 2017

More information

SUPPLEMENT TO PAPER CONVERGENCE OF ADAPTIVE AND INTERACTING MARKOV CHAIN MONTE CARLO ALGORITHMS

SUPPLEMENT TO PAPER CONVERGENCE OF ADAPTIVE AND INTERACTING MARKOV CHAIN MONTE CARLO ALGORITHMS Submitted to the Annals of Statistics SUPPLEMENT TO PAPER CONERGENCE OF ADAPTIE AND INTERACTING MARKO CHAIN MONTE CARLO ALGORITHMS By G Fort,, E Moulines and P Priouret LTCI, CNRS - TELECOM ParisTech,

More information

LECTURES ON MEAN FIELD GAMES: II. CALCULUS OVER WASSERSTEIN SPACE, CONTROL OF MCKEAN-VLASOV DYNAMICS, AND THE MASTER EQUATION

LECTURES ON MEAN FIELD GAMES: II. CALCULUS OVER WASSERSTEIN SPACE, CONTROL OF MCKEAN-VLASOV DYNAMICS, AND THE MASTER EQUATION LECTURES ON MEAN FIELD GAMES: II. CALCULUS OVER WASSERSTEIN SPACE, CONTROL OF MCKEAN-VLASOV DYNAMICS, AND THE MASTER EQUATION René Carmona Department of Operations Research & Financial Engineering PACM

More information

u xx + u yy = 0. (5.1)

u xx + u yy = 0. (5.1) Chapter 5 Laplace Equation The following equation is called Laplace equation in two independent variables x, y: The non-homogeneous problem u xx + u yy =. (5.1) u xx + u yy = F, (5.) where F is a function

More information

Deforming conformal metrics with negative Bakry-Émery Ricci Tensor on manifolds with boundary

Deforming conformal metrics with negative Bakry-Émery Ricci Tensor on manifolds with boundary Deforming conformal metrics with negative Bakry-Émery Ricci Tensor on manifolds with boundary Weimin Sheng (Joint with Li-Xia Yuan) Zhejiang University IMS, NUS, 8-12 Dec 2014 1 / 50 Outline 1 Prescribing

More information

Convergence of Feller Processes

Convergence of Feller Processes Chapter 15 Convergence of Feller Processes This chapter looks at the convergence of sequences of Feller processes to a iting process. Section 15.1 lays some ground work concerning weak convergence of processes

More information

Ergodic Theorems. Samy Tindel. Purdue University. Probability Theory 2 - MA 539. Taken from Probability: Theory and examples by R.

Ergodic Theorems. Samy Tindel. Purdue University. Probability Theory 2 - MA 539. Taken from Probability: Theory and examples by R. Ergodic Theorems Samy Tindel Purdue University Probability Theory 2 - MA 539 Taken from Probability: Theory and examples by R. Durrett Samy T. Ergodic theorems Probability Theory 1 / 92 Outline 1 Definitions

More information

Mean-field dual of cooperative reproduction

Mean-field dual of cooperative reproduction The mean-field dual of systems with cooperative reproduction joint with Tibor Mach (Prague) A. Sturm (Göttingen) Friday, July 6th, 2018 Poisson construction of Markov processes Let (X t ) t 0 be a continuous-time

More information

Impulse Control of Stochastic Navier-Stokes Equations

Impulse Control of Stochastic Navier-Stokes Equations Wayne State University Mathematics Faculty Research Publications Mathematics 1-1-23 Impulse Control of Stochastic Navier-Stokes Equations J. L. Menaldi Wayne State University, menaldi@wayne.edu S. S. Sritharan

More information

Threshold behavior and non-quasiconvergent solutions with localized initial data for bistable reaction-diffusion equations

Threshold behavior and non-quasiconvergent solutions with localized initial data for bistable reaction-diffusion equations Threshold behavior and non-quasiconvergent solutions with localized initial data for bistable reaction-diffusion equations P. Poláčik School of Mathematics, University of Minnesota Minneapolis, MN 55455

More information

LONG-TERM OPTIMAL REAL INVESTMENT STRATEGIES IN THE PRESENCE OF ADJUSTMENT COSTS

LONG-TERM OPTIMAL REAL INVESTMENT STRATEGIES IN THE PRESENCE OF ADJUSTMENT COSTS LONG-TERM OPTIMAL REAL INVESTMENT STRATEGIES IN THE PRESENCE OF ADJUSTMENT COSTS ARNE LØKKA AND MIHAIL ZERVOS Abstract. We consider the problem of determining in a dynamical way the optimal capacity level

More information

Microlocal analysis and inverse problems Lecture 4 : Uniqueness results in admissible geometries

Microlocal analysis and inverse problems Lecture 4 : Uniqueness results in admissible geometries Microlocal analysis and inverse problems Lecture 4 : Uniqueness results in admissible geometries David Dos Santos Ferreira LAGA Université de Paris 13 Wednesday May 18 Instituto de Ciencias Matemáticas,

More information

Noncooperative continuous-time Markov games

Noncooperative continuous-time Markov games Morfismos, Vol. 9, No. 1, 2005, pp. 39 54 Noncooperative continuous-time Markov games Héctor Jasso-Fuentes Abstract This work concerns noncooperative continuous-time Markov games with Polish state and

More information

Value Iteration and Action ɛ-approximation of Optimal Policies in Discounted Markov Decision Processes

Value Iteration and Action ɛ-approximation of Optimal Policies in Discounted Markov Decision Processes Value Iteration and Action ɛ-approximation of Optimal Policies in Discounted Markov Decision Processes RAÚL MONTES-DE-OCA Departamento de Matemáticas Universidad Autónoma Metropolitana-Iztapalapa San Rafael

More information

CONNECTIONS BETWEEN BOUNDED-VARIATION CONTROL AND DYNKIN GAMES

CONNECTIONS BETWEEN BOUNDED-VARIATION CONTROL AND DYNKIN GAMES CONNECTIONS BETWEEN BOUNDED-VARIATION CONTROL AND DYNKIN GAMES IOANNIS KARATZAS Departments of Mathematics and Statistics, 619 Mathematics Building Columbia University, New York, NY 127 ik@math.columbia.edu

More information

Singular Perturbations of Stochastic Control Problems with Unbounded Fast Variables

Singular Perturbations of Stochastic Control Problems with Unbounded Fast Variables Singular Perturbations of Stochastic Control Problems with Unbounded Fast Variables Joao Meireles joint work with Martino Bardi and Guy Barles University of Padua, Italy Workshop "New Perspectives in Optimal

More information

Point Process Control

Point Process Control Point Process Control The following note is based on Chapters I, II and VII in Brémaud s book Point Processes and Queues (1981). 1 Basic Definitions Consider some probability space (Ω, F, P). A real-valued

More information

{σ x >t}p x. (σ x >t)=e at.

{σ x >t}p x. (σ x >t)=e at. 3.11. EXERCISES 121 3.11 Exercises Exercise 3.1 Consider the Ornstein Uhlenbeck process in example 3.1.7(B). Show that the defined process is a Markov process which converges in distribution to an N(0,σ

More information

THE MULTIPLICATIVE ERGODIC THEOREM OF OSELEDETS

THE MULTIPLICATIVE ERGODIC THEOREM OF OSELEDETS THE MULTIPLICATIVE ERGODIC THEOREM OF OSELEDETS. STATEMENT Let (X, µ, A) be a probability space, and let T : X X be an ergodic measure-preserving transformation. Given a measurable map A : X GL(d, R),

More information

Polynomial Preserving Jump-Diffusions on the Unit Interval

Polynomial Preserving Jump-Diffusions on the Unit Interval Polynomial Preserving Jump-Diffusions on the Unit Interval Martin Larsson Department of Mathematics, ETH Zürich joint work with Christa Cuchiero and Sara Svaluto-Ferro 7th General AMaMeF and Swissquote

More information

Random and Deterministic perturbations of dynamical systems. Leonid Koralov

Random and Deterministic perturbations of dynamical systems. Leonid Koralov Random and Deterministic perturbations of dynamical systems Leonid Koralov - M. Freidlin, L. Koralov Metastability for Nonlinear Random Perturbations of Dynamical Systems, Stochastic Processes and Applications

More information

On the Bellman equation for control problems with exit times and unbounded cost functionals 1

On the Bellman equation for control problems with exit times and unbounded cost functionals 1 On the Bellman equation for control problems with exit times and unbounded cost functionals 1 Michael Malisoff Department of Mathematics, Hill Center-Busch Campus Rutgers University, 11 Frelinghuysen Road

More information

Information geometry for bivariate distribution control

Information geometry for bivariate distribution control Information geometry for bivariate distribution control C.T.J.Dodson + Hong Wang Mathematics + Control Systems Centre, University of Manchester Institute of Science and Technology Optimal control of stochastic

More information

INVARIANT MANIFOLDS WITH BOUNDARY FOR JUMP-DIFFUSIONS

INVARIANT MANIFOLDS WITH BOUNDARY FOR JUMP-DIFFUSIONS INVARIANT MANIFOLDS WITH BOUNDARY FOR JUMP-DIFFUSIONS DAMIR FILIPOVIĆ, STFAN TAPP, AND JOSF TICHMANN Abstract. We provide necessary and sufficient conditions for stochastic invariance of finite dimensional

More information

Reflected Brownian Motion

Reflected Brownian Motion Chapter 6 Reflected Brownian Motion Often we encounter Diffusions in regions with boundary. If the process can reach the boundary from the interior in finite time with positive probability we need to decide

More information

Markov Chain Approximation of Pure Jump Processes. Nikola Sandrić (joint work with Ante Mimica and René L. Schilling) University of Zagreb

Markov Chain Approximation of Pure Jump Processes. Nikola Sandrić (joint work with Ante Mimica and René L. Schilling) University of Zagreb Markov Chain Approximation of Pure Jump Processes Nikola Sandrić (joint work with Ante Mimica and René L. Schilling) University of Zagreb 8 th International Conference on Lévy processes Angers, July 25-29,

More information

Asymptotics of minimax stochastic programs

Asymptotics of minimax stochastic programs Asymptotics of minimax stochastic programs Alexander Shapiro Abstract. We discuss in this paper asymptotics of the sample average approximation (SAA) of the optimal value of a minimax stochastic programming

More information

Two viewpoints on measure valued processes

Two viewpoints on measure valued processes Two viewpoints on measure valued processes Olivier Hénard Université Paris-Est, Cermics Contents 1 The classical framework : from no particle to one particle 2 The lookdown framework : many particles.

More information

Stochastic Processes

Stochastic Processes Stochastic Processes 8.445 MIT, fall 20 Mid Term Exam Solutions October 27, 20 Your Name: Alberto De Sole Exercise Max Grade Grade 5 5 2 5 5 3 5 5 4 5 5 5 5 5 6 5 5 Total 30 30 Problem :. True / False

More information

Optimality Inequalities for Average Cost Markov Decision Processes and the Optimality of (s, S) Policies

Optimality Inequalities for Average Cost Markov Decision Processes and the Optimality of (s, S) Policies Optimality Inequalities for Average Cost Markov Decision Processes and the Optimality of (s, S) Policies Eugene A. Feinberg Department of Applied Mathematics and Statistics State University of New York

More information

PROBABILITY: LIMIT THEOREMS II, SPRING HOMEWORK PROBLEMS

PROBABILITY: LIMIT THEOREMS II, SPRING HOMEWORK PROBLEMS PROBABILITY: LIMIT THEOREMS II, SPRING 218. HOMEWORK PROBLEMS PROF. YURI BAKHTIN Instructions. You are allowed to work on solutions in groups, but you are required to write up solutions on your own. Please

More information

process on the hierarchical group

process on the hierarchical group Intertwining of Markov processes and the contact process on the hierarchical group April 27, 2010 Outline Intertwining of Markov processes Outline Intertwining of Markov processes First passage times of

More information

TD Learning. Sean Meyn. Department of Electrical and Computer Engineering University of Illinois & the Coordinated Science Laboratory

TD Learning. Sean Meyn. Department of Electrical and Computer Engineering University of Illinois & the Coordinated Science Laboratory TD Learning Sean Meyn Department of Electrical and Computer Engineering University of Illinois & the Coordinated Science Laboratory Joint work with I.-K. Cho, Illinois S. Mannor, McGill V. Tadic, Sheffield

More information

Strong approximation for additive functionals of geometrically ergodic Markov chains

Strong approximation for additive functionals of geometrically ergodic Markov chains Strong approximation for additive functionals of geometrically ergodic Markov chains Florence Merlevède Joint work with E. Rio Université Paris-Est-Marne-La-Vallée (UPEM) Cincinnati Symposium on Probability

More information

USING FUNCTIONAL ANALYSIS AND SOBOLEV SPACES TO SOLVE POISSON S EQUATION

USING FUNCTIONAL ANALYSIS AND SOBOLEV SPACES TO SOLVE POISSON S EQUATION USING FUNCTIONAL ANALYSIS AND SOBOLEV SPACES TO SOLVE POISSON S EQUATION YI WANG Abstract. We study Banach and Hilbert spaces with an eye towards defining weak solutions to elliptic PDE. Using Lax-Milgram

More information

Stochastic Dynamic Programming: The One Sector Growth Model

Stochastic Dynamic Programming: The One Sector Growth Model Stochastic Dynamic Programming: The One Sector Growth Model Esteban Rossi-Hansberg Princeton University March 26, 2012 Esteban Rossi-Hansberg () Stochastic Dynamic Programming March 26, 2012 1 / 31 References

More information

MASSACHUSETTS INSTITUTE OF TECHNOLOGY 6.265/15.070J Fall 2013 Lecture 22 12/09/2013. Skorokhod Mapping Theorem. Reflected Brownian Motion

MASSACHUSETTS INSTITUTE OF TECHNOLOGY 6.265/15.070J Fall 2013 Lecture 22 12/09/2013. Skorokhod Mapping Theorem. Reflected Brownian Motion MASSACHUSETTS INSTITUTE OF TECHNOLOGY 6.265/15.7J Fall 213 Lecture 22 12/9/213 Skorokhod Mapping Theorem. Reflected Brownian Motion Content. 1. G/G/1 queueing system 2. One dimensional reflection mapping

More information

Translation Invariant Exclusion Processes (Book in Progress)

Translation Invariant Exclusion Processes (Book in Progress) Translation Invariant Exclusion Processes (Book in Progress) c 2003 Timo Seppäläinen Department of Mathematics University of Wisconsin Madison, WI 53706-1388 December 11, 2008 Contents PART I Preliminaries

More information

Lecture 5. If we interpret the index n 0 as time, then a Markov chain simply requires that the future depends only on the present and not on the past.

Lecture 5. If we interpret the index n 0 as time, then a Markov chain simply requires that the future depends only on the present and not on the past. 1 Markov chain: definition Lecture 5 Definition 1.1 Markov chain] A sequence of random variables (X n ) n 0 taking values in a measurable state space (S, S) is called a (discrete time) Markov chain, if

More information

Optimality Inequalities for Average Cost MDPs and their Inventory Control Applications

Optimality Inequalities for Average Cost MDPs and their Inventory Control Applications 43rd IEEE Conference on Decision and Control December 14-17, 2004 Atlantis, Paradise Island, Bahamas FrA08.6 Optimality Inequalities for Average Cost MDPs and their Inventory Control Applications Eugene

More information

Stochastic Volatility and Correction to the Heat Equation

Stochastic Volatility and Correction to the Heat Equation Stochastic Volatility and Correction to the Heat Equation Jean-Pierre Fouque, George Papanicolaou and Ronnie Sircar Abstract. From a probabilist s point of view the Twentieth Century has been a century

More information

On the convergence of sequences of random variables: A primer

On the convergence of sequences of random variables: A primer BCAM May 2012 1 On the convergence of sequences of random variables: A primer Armand M. Makowski ECE & ISR/HyNet University of Maryland at College Park armand@isr.umd.edu BCAM May 2012 2 A sequence a :

More information

Branching Brownian motion seen from the tip

Branching Brownian motion seen from the tip Branching Brownian motion seen from the tip J. Berestycki 1 1 Laboratoire de Probabilité et Modèles Aléatoires, UPMC, Paris 09/02/2011 Joint work with Elie Aidekon, Eric Brunet and Zhan Shi J Berestycki

More information

Weak convergence and large deviation theory

Weak convergence and large deviation theory First Prev Next Go To Go Back Full Screen Close Quit 1 Weak convergence and large deviation theory Large deviation principle Convergence in distribution The Bryc-Varadhan theorem Tightness and Prohorov

More information

On the submartingale / supermartingale property of diffusions in natural scale

On the submartingale / supermartingale property of diffusions in natural scale On the submartingale / supermartingale property of diffusions in natural scale Alexander Gushchin Mikhail Urusov Mihail Zervos November 13, 214 Abstract Kotani 5 has characterised the martingale property

More information

STEADY STATE AND SCALING LIMIT FOR A TRAFFIC CONGESTION MODEL

STEADY STATE AND SCALING LIMIT FOR A TRAFFIC CONGESTION MODEL STEADY STATE AND SCALING LIMIT FOR A TRAFFIC CONGESTION MODEL ILIE GRIGORESCU AND MIN KANG Abstract. In a general model used in internet congestion control, the time dependent data flow vector x(t) > undergoes

More information

Weak solutions of mean-field stochastic differential equations

Weak solutions of mean-field stochastic differential equations Weak solutions of mean-field stochastic differential equations Juan Li School of Mathematics and Statistics, Shandong University (Weihai), Weihai 26429, China. Email: juanli@sdu.edu.cn Based on joint works

More information

EXCHANGEABLE MARKOV PROCESSES ON GRAPHS: FELLER CASE

EXCHANGEABLE MARKOV PROCESSES ON GRAPHS: FELLER CASE EXCHANGEABLE MARKOV PROCESSES ON GRAPHS: FELLER CASE HARRY CRANE Abstract. The transition law of every exchangeable Feller process on the space of countable graphs is determined by a σ-finite measure on

More information

for all f satisfying E[ f(x) ] <.

for all f satisfying E[ f(x) ] <. . Let (Ω, F, P ) be a probability space and D be a sub-σ-algebra of F. An (H, H)-valued random variable X is independent of D if and only if P ({X Γ} D) = P {X Γ}P (D) for all Γ H and D D. Prove that if

More information

Controllability of linear PDEs (I): The wave equation

Controllability of linear PDEs (I): The wave equation Controllability of linear PDEs (I): The wave equation M. González-Burgos IMUS, Universidad de Sevilla Doc Course, Course 2, Sevilla, 2018 Contents 1 Introduction. Statement of the problem 2 Distributed

More information

A probabilistic approach to second order variational inequalities with bilateral constraints

A probabilistic approach to second order variational inequalities with bilateral constraints Proc. Indian Acad. Sci. (Math. Sci.) Vol. 113, No. 4, November 23, pp. 431 442. Printed in India A probabilistic approach to second order variational inequalities with bilateral constraints MRINAL K GHOSH,

More information

PCMI LECTURE NOTES ON PROPERTY (T ), EXPANDER GRAPHS AND APPROXIMATE GROUPS (PRELIMINARY VERSION)

PCMI LECTURE NOTES ON PROPERTY (T ), EXPANDER GRAPHS AND APPROXIMATE GROUPS (PRELIMINARY VERSION) PCMI LECTURE NOTES ON PROPERTY (T ), EXPANDER GRAPHS AND APPROXIMATE GROUPS (PRELIMINARY VERSION) EMMANUEL BREUILLARD 1. Lecture 1, Spectral gaps for infinite groups and non-amenability The final aim of

More information

Practical conditions on Markov chains for weak convergence of tail empirical processes

Practical conditions on Markov chains for weak convergence of tail empirical processes Practical conditions on Markov chains for weak convergence of tail empirical processes Olivier Wintenberger University of Copenhagen and Paris VI Joint work with Rafa l Kulik and Philippe Soulier Toronto,

More information

Optimality of Walrand-Varaiya Type Policies and. Approximation Results for Zero-Delay Coding of. Markov Sources. Richard G. Wood

Optimality of Walrand-Varaiya Type Policies and. Approximation Results for Zero-Delay Coding of. Markov Sources. Richard G. Wood Optimality of Walrand-Varaiya Type Policies and Approximation Results for Zero-Delay Coding of Markov Sources by Richard G. Wood A thesis submitted to the Department of Mathematics & Statistics in conformity

More information

On semilinear elliptic equations with measure data

On semilinear elliptic equations with measure data On semilinear elliptic equations with measure data Andrzej Rozkosz (joint work with T. Klimsiak) Nicolaus Copernicus University (Toruń, Poland) Controlled Deterministic and Stochastic Systems Iasi, July

More information

MS&E 321 Spring Stochastic Systems June 1, 2013 Prof. Peter W. Glynn Page 1 of 10

MS&E 321 Spring Stochastic Systems June 1, 2013 Prof. Peter W. Glynn Page 1 of 10 MS&E 321 Spring 12-13 Stochastic Systems June 1, 2013 Prof. Peter W. Glynn Page 1 of 10 Section 3: Regenerative Processes Contents 3.1 Regeneration: The Basic Idea............................... 1 3.2

More information

Discrete time processes

Discrete time processes Discrete time processes Predictions are difficult. Especially about the future Mark Twain. Florian Herzog 2013 Modeling observed data When we model observed (realized) data, we encounter usually the following

More information

Zdzis law Brzeźniak and Tomasz Zastawniak

Zdzis law Brzeźniak and Tomasz Zastawniak Basic Stochastic Processes by Zdzis law Brzeźniak and Tomasz Zastawniak Springer-Verlag, London 1999 Corrections in the 2nd printing Version: 21 May 2005 Page and line numbers refer to the 2nd printing

More information

Martingale Problems. Abhay G. Bhatt Theoretical Statistics and Mathematics Unit Indian Statistical Institute, Delhi

Martingale Problems. Abhay G. Bhatt Theoretical Statistics and Mathematics Unit Indian Statistical Institute, Delhi s Abhay G. Bhatt Theoretical Statistics and Mathematics Unit Indian Statistical Institute, Delhi Lectures on Probability and Stochastic Processes III Indian Statistical Institute, Kolkata 20 24 November

More information

Poisson random measure: motivation

Poisson random measure: motivation : motivation The Lévy measure provides the expected number of jumps by time unit, i.e. in a time interval of the form: [t, t + 1], and of a certain size Example: ν([1, )) is the expected number of jumps

More information

Representation of Markov chains. Stochastic stability AIMS 2014

Representation of Markov chains. Stochastic stability AIMS 2014 Markov chain model Joint with J. Jost and M. Kell Max Planck Institute for Mathematics in the Sciences Leipzig 0th July 204 AIMS 204 We consider f : M! M to be C r for r 0 and a small perturbation parameter

More information

The Contraction Method on C([0, 1]) and Donsker s Theorem

The Contraction Method on C([0, 1]) and Donsker s Theorem The Contraction Method on C([0, 1]) and Donsker s Theorem Henning Sulzbach J. W. Goethe-Universität Frankfurt a. M. YEP VII Probability, random trees and algorithms Eindhoven, March 12, 2010 joint work

More information

Walsh Diffusions. Andrey Sarantsev. March 27, University of California, Santa Barbara. Andrey Sarantsev University of Washington, Seattle 1 / 1

Walsh Diffusions. Andrey Sarantsev. March 27, University of California, Santa Barbara. Andrey Sarantsev University of Washington, Seattle 1 / 1 Walsh Diffusions Andrey Sarantsev University of California, Santa Barbara March 27, 2017 Andrey Sarantsev University of Washington, Seattle 1 / 1 Walsh Brownian Motion on R d Spinning measure µ: probability

More information

On the stochastic nonlinear Schrödinger equation

On the stochastic nonlinear Schrödinger equation On the stochastic nonlinear Schrödinger equation Annie Millet collaboration with Z. Brzezniak SAMM, Paris 1 and PMA Workshop Women in Applied Mathematics, Heraklion - May 3 211 Outline 1 The NL Shrödinger

More information

On the posterior structure of NRMI

On the posterior structure of NRMI On the posterior structure of NRMI Igor Prünster University of Turin, Collegio Carlo Alberto and ICER Joint work with L.F. James and A. Lijoi Isaac Newton Institute, BNR Programme, 8th August 2007 Outline

More information

On the Convergence of Optimal Actions for Markov Decision Processes and the Optimality of (s, S) Inventory Policies

On the Convergence of Optimal Actions for Markov Decision Processes and the Optimality of (s, S) Inventory Policies On the Convergence of Optimal Actions for Markov Decision Processes and the Optimality of (s, S) Inventory Policies Eugene A. Feinberg Department of Applied Mathematics and Statistics Stony Brook University

More information

Harmonic Functions and Brownian motion

Harmonic Functions and Brownian motion Harmonic Functions and Brownian motion Steven P. Lalley April 25, 211 1 Dynkin s Formula Denote by W t = (W 1 t, W 2 t,..., W d t ) a standard d dimensional Wiener process on (Ω, F, P ), and let F = (F

More information

The range of tree-indexed random walk

The range of tree-indexed random walk The range of tree-indexed random walk Jean-François Le Gall, Shen Lin Institut universitaire de France et Université Paris-Sud Orsay Erdös Centennial Conference July 2013 Jean-François Le Gall (Université

More information

MATH 220: MIDTERM OCTOBER 29, 2015

MATH 220: MIDTERM OCTOBER 29, 2015 MATH 22: MIDTERM OCTOBER 29, 25 This is a closed book, closed notes, no electronic devices exam. There are 5 problems. Solve Problems -3 and one of Problems 4 and 5. Write your solutions to problems and

More information

PROBABILITY: LIMIT THEOREMS II, SPRING HOMEWORK PROBLEMS

PROBABILITY: LIMIT THEOREMS II, SPRING HOMEWORK PROBLEMS PROBABILITY: LIMIT THEOREMS II, SPRING 15. HOMEWORK PROBLEMS PROF. YURI BAKHTIN Instructions. You are allowed to work on solutions in groups, but you are required to write up solutions on your own. Please

More information

Iowa State University. Instructor: Alex Roitershtein Summer Homework #5. Solutions

Iowa State University. Instructor: Alex Roitershtein Summer Homework #5. Solutions Math 50 Iowa State University Introduction to Real Analysis Department of Mathematics Instructor: Alex Roitershtein Summer 205 Homework #5 Solutions. Let α and c be real numbers, c > 0, and f is defined

More information

Dynamic and Stochastic Brenier Transport via Hopf-Lax formulae on Was

Dynamic and Stochastic Brenier Transport via Hopf-Lax formulae on Was Dynamic and Stochastic Brenier Transport via Hopf-Lax formulae on Wasserstein Space With many discussions with Yann Brenier and Wilfrid Gangbo Brenierfest, IHP, January 9-13, 2017 ain points of the

More information

Introduction to self-similar growth-fragmentations

Introduction to self-similar growth-fragmentations Introduction to self-similar growth-fragmentations Quan Shi CIMAT, 11-15 December, 2017 Quan Shi Growth-Fragmentations CIMAT, 11-15 December, 2017 1 / 34 Literature Jean Bertoin, Compensated fragmentation

More information

Large deviations and averaging for systems of slow fast stochastic reaction diffusion equations.

Large deviations and averaging for systems of slow fast stochastic reaction diffusion equations. Large deviations and averaging for systems of slow fast stochastic reaction diffusion equations. Wenqing Hu. 1 (Joint work with Michael Salins 2, Konstantinos Spiliopoulos 3.) 1. Department of Mathematics

More information

Chapter 4. Multiple Integrals. 1. Integrals on Rectangles

Chapter 4. Multiple Integrals. 1. Integrals on Rectangles Chapter 4. Multiple Integrals Let be a rectangle in I 2 given by 1. Integrals on ectangles = [a 1, b 1 [a 2, b 2 := {(x, y) I 2 : a 1 x b 1, a 2 y b 2 }. Let P 1 be a partition of [a 1, b 1 : P 1 : a 1

More information

Exercises with solutions (Set D)

Exercises with solutions (Set D) Exercises with solutions Set D. A fair die is rolled at the same time as a fair coin is tossed. Let A be the number on the upper surface of the die and let B describe the outcome of the coin toss, where

More information

Rare Events in Random Walks and Queueing Networks in the Presence of Heavy-Tailed Distributions

Rare Events in Random Walks and Queueing Networks in the Presence of Heavy-Tailed Distributions Rare Events in Random Walks and Queueing Networks in the Presence of Heavy-Tailed Distributions Sergey Foss Heriot-Watt University, Edinburgh and Institute of Mathematics, Novosibirsk The University of

More information

On Finite-Time Ruin Probabilities in a Risk Model Under Quota Share Reinsurance

On Finite-Time Ruin Probabilities in a Risk Model Under Quota Share Reinsurance Applied Mathematical Sciences, Vol. 11, 217, no. 53, 269-2629 HIKARI Ltd, www.m-hikari.com https://doi.org/1.12988/ams.217.7824 On Finite-Time Ruin Probabilities in a Risk Model Under Quota Share Reinsurance

More information

On the Convergence of Optimal Actions for Markov Decision Processes and the Optimality of (s, S) Inventory Policies

On the Convergence of Optimal Actions for Markov Decision Processes and the Optimality of (s, S) Inventory Policies On the Convergence of Optimal Actions for Markov Decision Processes and the Optimality of (s, S) Inventory Policies Eugene A. Feinberg, 1 Mark E. Lewis 2 1 Department of Applied Mathematics and Statistics,

More information

Local Uniqueness and Refined Spike Profiles of Ground States for Two-Dimensional Attractive Bose-Einstein Condensates

Local Uniqueness and Refined Spike Profiles of Ground States for Two-Dimensional Attractive Bose-Einstein Condensates Local Uniqueness and Refined Spie Profiles of Ground States for Two-Dimensional Attractive Bose-Einstein Condensates Yujin Guo, Changshou Lin, and Juncheng Wei May 1, 17 Abstract We consider ground states

More information

Zero-Sum Average Semi-Markov Games: Fixed Point Solutions of the Shapley Equation

Zero-Sum Average Semi-Markov Games: Fixed Point Solutions of the Shapley Equation Zero-Sum Average Semi-Markov Games: Fixed Point Solutions of the Shapley Equation Oscar Vega-Amaya Departamento de Matemáticas Universidad de Sonora May 2002 Abstract This paper deals with zero-sum average

More information

Chapter 2 Application to Markov Processes

Chapter 2 Application to Markov Processes Chapter 2 Application to Markov Processes 2.1 Problem Let X t be a standard Markov process with the state space. The time interval [, is denoted by T.Leta be a fixed state and σ a the hitting time for

More information