Deterministic Minimax Impulse Control

Similar documents
Differential games withcontinuous, switching and impulse controls

Prof. Erhan Bayraktar (University of Michigan)

Differential Games II. Marc Quincampoix Université de Bretagne Occidentale ( Brest-France) SADCO, London, September 2011

Chain differentials with an application to the mathematical fear operator

On the Bellman equation for control problems with exit times and unbounded cost functionals 1

Nonlinear Control Systems

On the Multi-Dimensional Controller and Stopper Games

An introduction to Mathematical Theory of Control

A deterministic approach to the Skorokhod problem

Min-Max Certainty Equivalence Principle and Differential Games

Sébastien Chaumont a a Institut Élie Cartan, Université Henri Poincaré Nancy I, B. P. 239, Vandoeuvre-lès-Nancy Cedex, France. 1.

Hausdorff Continuous Viscosity Solutions of Hamilton-Jacobi Equations

Numerical Methods for Optimal Control Problems. Part I: Hamilton-Jacobi-Bellman Equations and Pontryagin Minimum Principle

Multi-dimensional Stochastic Singular Control Via Dynkin Game and Dirichlet Form

UNCERTAINTY FUNCTIONAL DIFFERENTIAL EQUATIONS FOR FINANCE

HOMEWORK ASSIGNMENT 6

Optimal Control and Viscosity Solutions of Hamilton-Jacobi-Bellman Equations

The Skorokhod problem in a time-dependent interval

INFINITE TIME HORIZON OPTIMAL CONTROL OF THE SEMILINEAR HEAT EQUATION

Example 1. Hamilton-Jacobi equation. In particular, the eikonal equation. for some n( x) > 0 in Ω. Here 1 / 2

On differential games with long-time-average cost

An introduction to Birkhoff normal form

Uniqueness of unbounded viscosity solutions for impulse control problem

AN OVERVIEW OF STATIC HAMILTON-JACOBI EQUATIONS. 1. Introduction

Viscosity Solutions of the Bellman Equation for Perturbed Optimal Control Problems with Exit Times 0

EXISTENCE THEOREMS FOR FUNCTIONAL DIFFERENTIAL EQUATIONS IN BANACH SPACES. 1. Introduction

Limit value of dynamic zero-sum games with vanishing stage duration

Passivity-based Stabilization of Non-Compact Sets

CLASSIFICATIONS OF THE FLOWS OF LINEAR ODE

A Monotone Scheme for Hamilton-Jacobi Equations via the Nonstandard Finite Difference Method

2 Statement of the problem and assumptions

arxiv: v3 [math.ds] 22 Feb 2012

HJ equations. Reachability analysis. Optimal control problems

Robustness for a Liouville type theorem in exterior domains

Computation of an Over-Approximation of the Backward Reachable Set using Subsystem Level Set Functions. Stanford University, Stanford, CA 94305

A NOTE ON ALMOST PERIODIC VARIATIONAL EQUATIONS

Institut für Mathematik

Asymptotic Perron Method for Stochastic Games and Control

The Arzelà-Ascoli Theorem

Linear Quadratic Zero-Sum Two-Person Differential Games Pierre Bernhard June 15, 2013

MINIMAL TIME PROBLEM WITH IMPULSIVE CONTROLS

The main motivation of this paper comes from the following, rather surprising, result of Ecker and Huisken [13]: for any initial data u 0 W 1,

Noncooperative continuous-time Markov games

SPACES ENDOWED WITH A GRAPH AND APPLICATIONS. Mina Dinarvand. 1. Introduction

Threshold behavior and non-quasiconvergent solutions with localized initial data for bistable reaction-diffusion equations

Exam February h

Closed-Loop Impulse Control of Oscillating Systems

Research Article Almost Periodic Viscosity Solutions of Nonlinear Parabolic Equations

Solution of Stochastic Optimal Control Problems and Financial Applications

ON WEAKLY NONLINEAR BACKWARD PARABOLIC PROBLEM

Viscosity Solutions of Fully Nonlinear Second Order Parabolic Equations with L 1 Dependence in Time and Neumann Boundary Conditions

Stability of Feedback Solutions for Infinite Horizon Noncooperative Differential Games

Observer design for a general class of triangular systems

Convergence of a first order scheme for a non local eikonal equation

Functional Analysis HW #3

Local semiconvexity of Kantorovich potentials on non-compact manifolds

Identification of Parameters in Neutral Functional Differential Equations with State-Dependent Delays

1 Lyapunov theory of stability

Non-Zero-Sum Stochastic Differential Games of Controls and St

Optimal stopping for non-linear expectations Part I

Further Results on the Bellman Equation for Optimal Control Problems with Exit Times and Nonnegative Instantaneous Costs

The Minimum Speed for a Blocking Problem on the Half Plane

u( x) = g( y) ds y ( 1 ) U solves u = 0 in U; u = 0 on U. ( 3)

Risk-Sensitive Control with HARA Utility

STOCHASTIC PERRON S METHOD AND VERIFICATION WITHOUT SMOOTHNESS USING VISCOSITY COMPARISON: OBSTACLE PROBLEMS AND DYNKIN GAMES

Approximation of Minimal Functions by Extreme Functions

Notes on uniform convergence

Linear Quadratic Zero-Sum Two-Person Differential Games

The principle of least action and two-point boundary value problems in orbital mechanics

HAMILTON-JACOBI EQUATIONS : APPROXIMATIONS, NUMERICAL ANALYSIS AND APPLICATIONS. CIME Courses-Cetraro August 29-September COURSES

AC&ST AUTOMATIC CONTROL AND SYSTEM THEORY SYSTEMS AND MODELS. Claudio Melchiorri

Propagating terraces and the dynamics of front-like solutions of reaction-diffusion equations on R

THE SKOROKHOD OBLIQUE REFLECTION PROBLEM IN A CONVEX POLYHEDRON

Deterministic Dynamic Programming

Proof. We indicate by α, β (finite or not) the end-points of I and call

On the Well-Posedness of the Cauchy Problem for a Neutral Differential Equation with Distributed Prehistory

Piecewise Smooth Solutions to the Burgers-Hilbert Equation

New Discretizations of Turbulent Flow Problems

GLOBAL EXISTENCE RESULTS AND UNIQUENESS FOR DISLOCATION EQUATIONS

GENERAL EXISTENCE OF SOLUTIONS TO DYNAMIC PROGRAMMING PRINCIPLE. 1. Introduction

Converse Lyapunov theorem and Input-to-State Stability

L2 gains and system approximation quality 1

Linear-Quadratic Stochastic Differential Games with General Noise Processes

Lecture notes on Ordinary Differential Equations. S. Sivaji Ganesh Department of Mathematics Indian Institute of Technology Bombay

A maximum principle for optimal control system with endpoint constraints

ON THE PATHWISE UNIQUENESS OF SOLUTIONS OF STOCHASTIC DIFFERENTIAL EQUATIONS

Parameter Dependent Quasi-Linear Parabolic Equations

Viscosity Solutions for Dummies (including Economists)

On Stopping Times and Impulse Control with Constraint

Limit value of dynamic zero-sum games with vanishing stage duration

Homogenization of first order equations with (u/ε)-periodic Hamiltonians. Part I: local equations

THE CANTOR GAME: WINNING STRATEGIES AND DETERMINACY. by arxiv: v1 [math.ca] 29 Jan 2017 MAGNUS D. LADUE

Weak solutions of mean-field stochastic differential equations

Optimal stopping time formulation of adaptive image filtering

Master Thesis. Nguyen Tien Thinh. Homogenization and Viscosity solution

SOME PROPERTIES ON THE CLOSED SUBSETS IN BANACH SPACES

A MAXIMUM PRINCIPLE FOR SEMICONTINUOUS FUNCTIONS APPLICABLE TO INTEGRO-PARTIAL DIFFERENTIAL EQUATIONS

Nonlinear L 2 -gain analysis via a cascade

A TWO PARAMETERS AMBROSETTI PRODI PROBLEM*

EXISTENCE OF SOLUTIONS TO ASYMPTOTICALLY PERIODIC SCHRÖDINGER EQUATIONS

Transcription:

Deterministic Minimax Impulse Control N. El Farouq, Guy Barles, and P. Bernhard March 02, 2009, revised september, 15, and october, 21, 2009 Abstract We prove the uniqueness of the viscosity solution of an Isaacs quasivariational inequality arising in an impulse control minimax problem, motivated by an application in mathematical finance. Keywords: impulse control, robust control, differential games, quasi-variational inequality, viscosity solution. Mathematical subject classification: 35D40, 49L25, 49N70, 91A23, 91G80 1 Introduction The concept of viscosity solution has become central to control theory ever since the seminal paper of Crandall & Lions [6 following the volume [11. For firstorder Hamilton-Jacobi Equations and applications to optimal control and differential games, achievements due to this concept can be found noticeably in [2 and [1. Its application to differential games was pioneered by W. Fleming, before its formalization by P.L. Lions et al., in [10, and in modern form by Lions & Souganidis [12, Evans & Souganidis [9, and Souganidis [14. It is a relatively simple matter to show that the value function of control problems is a viscosity solution of a relevant Hamilton-Jacobi Equation. The strength of the concept, though, lies in the uniqueness. Early uniqueness proofs were based upon the comparison theorem of Crandall & Lions [6 and its variations. The proof we develop here for the QVI of an impulse minimax problem is based upon the Lipschitz character of the dependence of the solution of a variational inequality upon its data. Previous results in the same direction were obtained by Yong [15 in the finite horizon case but allowing less general jumps. We want also to mention, in the University Blaise Pascal, Clermont-Ferrand, France(Elfarouq@i3s.unice.fr). University François Rabelais, Tours, France(Barles@lmpt.univ-tours.fr). INRIA-Sophia Antipolis-Méditerranée, France(Pierre.Bernhard@sophia.inria.fr). 1

infinite horizon case, the works of Dharmatti and Shaiju [13, 7 and Dharmatti and Ramaswamy [8 involving also hybrid controls. See also [3. 2 The problem 2.1 System 2.1.1 Dynamics Let a two-player differential game system be defined by the solution of following dynamical equations ẏ(t) = f(t, y(t), τ(t)), y(t 0 ) = x R n, y(t + k ) = y(t k ) + g(t k, y(t k ), ξ k), t k t 0, ξ k 0. Here and below, the time variable belongs to [t 0, T where T > t 0 0 are given. The state at time t, y(t) lies in R n. The system is driven by two controls, a continuous control τ(t) K R l, where K is a compact set, and an impulsive control defined by a finite sequence of impulse times t k and the controls ξ k R m controlling the jumps in y(t k ). Let ψ = ({t k }, {ξ k }), where k N. For any initial condition (t 0, x) and controls τ( ) and ψ generating a trajectory y( ) of this system, let a pay-off J be defined as J(t 0, x, ψ, τ(.)) = T t 0 L(t, y(t), τ(t)) dt + k C(ξ k ) + G(y(T )). (1) 2.1.2 Regularity assumptions In all the paper, we assume the following 1. f(t, y, τ) and L(t, y, τ) are continuous with respect to t uniformly in y and τ, Lipschitz continuous with respect to y uniformly in t and τ with constant c f and c L respectively, and continuous with respect to τ. 2. g(t, y, ξ) is Lipschitz continuous with respect to t, uniformly in y and ξ, with constant c gg and it is Lipschitz continuous with respect to y uniformly in t and ξ, with constant c g. 3. L, f, and G are bounded. 4. C is continuous and inf ξ C(ξ) = γ > 0. 2

5. G is Lipschitz continuous with constant c G. It follows, inter alia, that there exists a unique solution y( ) and a unique J for any measurable τ( ) and any sequence ψ, and thus for any nonanticipative strategy Φ as defined hereafter. 2.2 Strategies 2.2.1 Multiple jumps It may be to the best advantage of the minimizer to make a jump at some time t, immediately followed, at the same time, by another jump, and so on. As any such jump entails a cost not less than γ, a near optimal strategy will never attempt to make an infinite number of jumps. As a matter of fact, we shall show that it follows from the boundedness of L and J that the number of jumps may be restricted, with no loss of generality, to be less than K = 2[(T t 0 ) L + G + 1/2/γ. To allow for the possibility of several successive but simultaneous jumps, we proceed as follows. Let Ξ = K p=1 (R m ) p. We extend g and C from R m to Ξ in the natural way: let ξ (R m ) p be a multiple jump of multiplicity p. Let (ξ 1, ξ 2,..., ξ p ) be its components. For a given y R n, let z 0 = y, and for l {1,..., p}, z l = z l 1 + g(t, z l 1, ξ l ). Then we set g(t, y, ξ) = z p y C(ξ) = p C(ξ l ). There is no point in considering the possibility of several successive multiple jumps at the same jump time, since a sequence of simultaneous multiple jumps is a multiple jump. It easily follows that g so extended is still Lipschitz continuous in t and in y, uniformly in the other two variables, with new coefficients c gg and c g respectively. From now on, when we refer to jumps, it will always be multiple jumps, unless specifically referred to as simple jumps. Simple jumps are the same thing as a multiple jump of multiplicity 1. And of course, in a control ψ = ({t k }, {ξ k }) the ξ k are to be considered as multiple jumps. But the t k s are always assumed to be distinct. We also state the following definition: l=1 3

Definition 2.1 For any bounded function V : [t 0, T R n R, let the operator M be given by It enjoys the following properties: MV (t, x) = inf [V (t, x + g(t, x, ξ)) + C(ξ) ξ Ξ Lemma 2.1 The operator M is continuous for the topology. Moreover, if the function V is bounded and Lipschitz continuous, so is the function MV. Proof The first statement is elementary. Clearly, if, for all (t, x), V (t, x) V, then, clearly MV (t, x) V + γ. To prove the regularity of MV, we need to make the following remark, which as akin to the lemma 3.5 below: the range of ξ in the inf ξ operator defining MV may be restricted to jumps of multiplicity no larger than 1 + 2 V /γ. As a matter of fact, the simple jumps 2 to K will cost at least (K 1)γ and cannot decrease V by more than 2 V. Therefore, if K 1 > 2 V /γ, for any jump of multiplicity K, say ξ = (ξ 1,..., ξ K ), V (t, x + g(t, x, ξ 1 )) + C(ξ 1 ) < V (t, x + g(t, x, ξ)) + C(ξ). Over this retricted range Ξ, (t, x) g(t, x, ξ) is Lipschitz continuous uniformly in ξ: there exists a constant C g > 0: for all (t 1, x 1 ), (t 2, x 2 ), ξ Ξ, g(t 1, x 1, ξ) g(t 2, x 2, ξ) C g ( t 1 t 2 + x 1 x 2 ). Assume V is Lipschitz continuous, more precisely that (t 1, x 1 ), (t 2, x 2 ), V (t 1, x 1 ) V (t 2, x 2 ) λ( t 1 t 2 + x 1 x 2 ). We need now evaluate = MV (t 1, x 1 ) MV (t 2, x 2 ). Choose an arbitrary ε > 0 and a ξ ε such that Then V (t 2, x 2 + g(t 2, x 2, ξ ε )) + C(ξ ε ) MV (t 2, x 2 ) + ε. MV (t 1, x 1 ) V (t 2, x 2 + g(t 2, x 2, ξ ε )) C(ξ ε ) + ε V (t 1, x 1 + g(t 1, x 1, ξ ε )) + C(ξ ε ) V (t 2, x 2 + g(t 2, x 2, ξ ε )) C(ξ ε ) + ε λ(1 + C g )[ t 1 t 2 + x 1 x 2 + ε. Since, on the one hand ε was arbitrary, and on the other hand, (t 1, x 1 ) and (t 2, x 2 ) play symmetrical roles, we may conclude that MV (t 1, x 1 ) MV (t 2, x 2 ) λ(1 + C g )[ t 1 t 2 + x 1 x 2. 4

2.2.2 Admissible strategies We want to investigate the problem of minimizing sup τ(.) Ω J through the impulse control. We mean to allow closed loop strategies for the minimizing control. We remark that, being only interested in the inf sup problem, and not a possible saddle point, there is no loss of generality in restricting τ( ) to open loop controls τ( ) Ω = {measurable functions [t 0, T K} We shall sometimes write τ Ω instead of τ( ) Ω. Let Ψ be the set of all finite sequences ψ. We now define the admissible closed loop strategies Φ for the minimizing impulse control ψ, as nonanticipative strategies. We shall let Π be the set of all such nonanticipative strategies. Definition 2.2 A map Φ : Ω Ψ is called a nonanticipative strategy if for any two controls τ 1 ( ) and τ 2 ( ), and any t [t 0, T, the condition on their restrictions to [t 0, t[: τ 1 [t0,t[ = τ 2 [t0,t[ implies Φ(τ 1 ) [t0,t = Φ(τ 2 ) [t0,t. 3 The value function We define the value function of the problem W : [0, T R n R as W (t 0, x) = inf sup Φ Π τ(.) Ω 3.1 Dynamic Programming Principle J(t 0, x, Φ(τ( )), τ(.)). (2) The solution of the minimax impulse control problem at hand depends on the following basic properties Proposition 3.1 (Isaacs Tenet of transition) For all t t [t 0, T [, for all x R n, t W (t, x) = inf sup L(s, y(s), τ(s))ds + C(ξ k ) + W (t, y(t )) Φ Π τ Ω t t k <t Proof Assume first that for some x, t and t > t, t W (t, x) > inf sup L(s, y(s), τ(s)) ds + C(ξ k ) + W (t, y(t )), Φ Π τ Ω t t k <t 5

and let the difference be 3ε. Choose an admissible strategy Φ ε 1 that approximates the infimum in the r.h.s. up to ε. Then, for any τ( ) defined over [t, t, t t L(s, y(s), τ(s)) ds + t k <t C(ξ k ) + W (t, y(t )) W (t, x) 2ε. Whatever y(t ), from t on, choose a non anticipative strategy Φ ε 2 of the game over [t, T that approximates the Value W (t, y(t )) again up to ε. The concatenation Φ ε of Φ ε 1 and Φε 2 is a nonanticipative strategy of the game over [t, T. It yields, for all τ( ), J(t, x, Φ ε, τ( )) W (t, x) ε, a contradiction. Assume to the contrary that t W (t, x) < inf sup L(s, y(s), τ(s)) ds + C(ξ k ) + W (t, y(t )), Φ Π τ Ω t t k <t and let the difference be 2ε. Choose an admissible strategy Φ ε that approximates W (t, x) up to ε, and denote (t ε k, ξε k ) the jumps it produces. Then, [ t sup L(s, y(s), τ(s)) ds + C(ξk ε ) + J(t, y(t ), Φ ε, τ) τ( ) t t ε k [ <t t inf sup L(s, y(s), τ(s)) ds + C(ξ k ) + W (t, y(t )) ε. Φ τ( ) t t k <t The above inequality can also be written in terms of the restrictions τ and τ of τ( ) to [t, t and [t, T respectively, as [ t sup L(s, y(s), τ (s)) ds + C(ξk ε ) + sup J(t, y(t ), Φ ε, τ ) τ t t ε k <t τ [ t inf sup L(s, y(s), τ (s)) ds + (3) C(ξ k ) + W (t, y(t )) ε. Φ τ t t k <t Observe that the knowledge of τ is useless to evaluate J(t, y(t ), Φ, τ ) once y(t ) is known. Therefore, the restriction of Φ ε to [t, T cannot do better than a nonanticipative strategy of the game over [t, T. As a result, W (t, y(t )) sup τ J(t, y(t ), Φ ε, τ ). But placing that in the l.h.s. of inequality (3) leads to a contradiction. 6

Proposition 3.2 For all t [t 0, T and x R n, W (t, x) MW (t, x). Proof Among the admissible Φ s there are those that place a jump at time t. Using the same reasoning as above, minimizing over the jump parameter ξ at time t, one gets the required inequality. Proposition 3.3 For any admissible strategy Φ, let t 1 Φ(τ( )) be the first impulse time after t. Then, for all x R n, [ t1 W (t, x) = inf sup L(s, y(s), τ(s)) ds + MW (t 1, y(t 1 )) Φ τ t Proof Proceed as in proposition 3.1 with t replaced by t 1 as specified by the strategy Φ ε chosen, and notice that the inf over [t 1, T involves a inf ξ1. 3.2 Regularity of the value function In this section, we prove the following theorem Theorem 3.4 Under the standing assumptions (paragraph 2.1.2) the value function W is bounded, Lipschitz continuous in x uniformly in t, and Lipschitz continuous in t, uniformly in x. The rest of the section is devoted to the proof of this theorem. 3.2.1 Boundedness Let us first show that thanks to hypothesis 3, W as defined by (2) is bounded. On the one hand, a particular strategy Φ is the one where we have no impulse time. In this case k C(ξ k) = 0. Since L and G are bounded, it is easy to see that for all t [t 0, T and x R n, W (t, x) (T t 0 ) L + G. On the other hand, since the costs C(ξ k ) are positive numbers and again since L and G are bounded, then again it is easy to see that for all t [t 0, T and x R n, W (t, x) (T t 0 ) L G. We can then say that proving the first result. It follows the important property: W (T t 0 ) L + G 7

Lemma 3.5 There exists K N depending only on the data of the problem such that, in the definition of W, the inf Φ Π can be replaced by inf Φ Π K where Π K is the set of all nonanticipative strategies who s sums of the multiplicity of their jumps is less or equal to K. Proof of the lemma Let us choose a strategy Φ such that [ T K W (t, x) sup L(s, y(s), τ) ds + C( ξ k ) + G(y(T )) 1 τ Since then t k=1 [ T (T t) L G sup L(s, y(s), τ) ds + G(y(T )) τ t K C( ξ k ) (T t) L + G + W + 1 k=1 with γ = inf ξ C(ξ), we obtain then 2((T t 0 ) L + G ) + 1 K 1 γ (2((T t 0) L + G ) + 1). This proves that, for the strategies which are close to optimality, the number of impulses, simple or multiple, is bounded, with a bound depending only on the data of the problem. 3.2.2 Lipschitz continuity in x Let us now prove that the value function W is Lipschitz continuous in x, uniformly in t. To that aim, we estimate the difference = W (t, x 1 ) W (t, x 2 ) Fix an arbitrary ε > 0. Let Φ 2 be a strategy such that Then sup J(t, x 2, Φ 2, τ) inf sup J(t, x 2, Φ, τ) + ε = W (t, x 2 ) + ε. τ Φ τ W (t, x 1 ) sup τ J(t, x 2, Φ 2, τ) + ε sup τ J(t, x 1, Φ 2, τ) sup τ J(t, x 2, Φ 2, τ) + ε. 8

Now, pick τ 1 such that Then J(t, x 1, Φ 2, τ 1 ) sup J(t, x 1, Φ 2, τ) ε. τ J(t, x 1, Φ 2, τ 1 ) sup τ J(t, x 2, Φ 2, τ) + 2ε J(t, x 1, Φ 2, τ 1 ) J(t, x 2, Φ 2, τ 1 ) + 2ε. Consider now the control ψ 2 = Φ 2 (τ 1 ), composed of jump instants t 1, t 2...t K in the interval [t, T, with jumps ξ 1, ξ 2...ξ K, and let y 1 ( ) and y 2 ( ) be the trajectories generated by (Φ 2, τ 1 ), or equivalently (ψ 2, τ 1 ), from y i (t) = x i, i = 1, 2. By Gronwall s lemma, we can say for all s [t, t 1, y 1 (s) y 2 (s) exp(c f (s t)) x 1 x 2. Looking more carefully at the first jump and using the Lipschitz continuity of g, we have y 1 (t + 1 ) y 2(t + 1 ) = (y 1 (t 1 )+g(t 1, y 1 (t 1 ), ξ 1)) (y 2 (t 1 )+g(t 1, y 2 (t 1 ), ξ 1)) (1 + c g ) y 1 (t 1 ) y 2(t 1 ) exp(c f (t 1 t))(1 + c g ) x 1 x 2. Repeating inductively the same argument, we have, for an impulse time t k, k {1, 2...K}, Finally y 1 (t + k ) y 2(t + k ) exp(c f (t k t))(1 + c g ) k x 1 x 2. for all s [t, T, y 1 (s) y 2 (s)) exp(c f (s t))(1 + c g ) K x 1 x 2, where we recall that K is the number of impulses in ψ 2. Then, J(t, x 1, Φ 2, τ 1 ) J(t, x 2, Φ 2, τ 1 ) T c L y 1 (s) y 2 (s) ds + c G y 1 (T ) y 2 (T ) t [ T (1 + c g ) K x 1 x 2 c L exp(c f (s t)) ds + c G exp(c f (T t)) t [ 1 (1 + c g ) K x 1 x 2 (exp(c f (T t 0 )) 1) + c G exp(c f (T t 0 )), c f 9

and we can say now W (t, x 1 ) W (t, x 2 ) C 1 x 1 x 2 + 2ε and, since ε was arbitrary, W (t, x 1 ) W (t, x 2 ) C 1 x 1 x 2. where the constant C 1 = (1 + c g ) K [ 1 (exp(c f (T t 0 )) 1) + c G exp(c f (T t 0 )), c f with K the maximum number of impulses according to the lemma. The roles of x 1, x 2 being symmetrical, we finally have W (t, x 1 ) W (t, x 2 ) C 1 x 1 x 2. 3.2.3 Lipschitz continuity in t Let us examine now the difference W (s 1, x) W (s 2, x). Assume first that s 2 > s 1. By the dynamic programming principle, W (s 1, x) [ W (s 2, x) = s2 inf sup L(s, y(s), τ(s)) ds + C(ξ k ) + W (s 2, y(s 2 )) Φ τ s 1 t k <s 2 W (s 2, y(s 1 )) We choose the strategy Φ where there is no impulse time between s 1 and s 2, then W (s 1, x) W (s 2, x) [ s2 sup L(s, y(s), τ(s)) ds + W (s 2, y(s 2 )) W (s 2, y(s 1 )) τ s 1 L s 2 s 1 + C 1 f s 2 s 1 c 2 s 2 s 1 with c 2 = L + C 1 f Assume now that s 1 > s 2. We denote by y 2 ( ) a trajectory generated by the controls considered from y(s 2 ) = x. By the dynamic programming principle, we have W (s 1, x) W (s 2, x) = s1 W (s 1, x) inf sup L(s, y 2 (s), τ(s)) ds + C(ξ k ) + W (s 1, y 2 (s 1 )). Φ τ s 2 k t k <s 1 10

Choose an arbirary ε > 0, and a strategy Φ ε, such that the inf Φ above is reached within ε. We have W (s 1, x) W (s 2, x) s1 W (s 1, x) sup L(s, y 2 (s), τ(s)) ds + C(ξ k ) + W (s 1, y 2 (s 1 )) + ε s 2 k t k <s 1 W (s 1, x) τ s1 s 2 L(s, y 2 (s), τ(s)) ds k t k <s 1 C(ξ k ) W (s 1, y 2 (s 1 )) + ε for any given τ( ). Let us, from now on fix a τ( ), and let Φ ε (τ( )) exhibit K jumps (ξ 1,..., ξ K ) in the interval [s 2, s 1 [. And let y 2 ( ) be the trajectory generated by these controls from y 2 (s 2 ) = x. We now bound W (s 1, x) using this sequence of jumps and Proposition 3.2. First, consider the effect of the multiple jump ξ = (ξ 1,..., ξ K ) at time s 1 on x. Let therefore z K = x + g(s 1, x, ξ), constructed as in the definition of multiple jumps, via z 0 = x and z k = z k 1 + g(s 1, z k 1, ξ k ). 1 Using Proposition 3.2, we have K W (s 1, x) W (s 1, z K ) + C(ξ k ) Therefore W (s 1, x) W (s 2, x) W (s 1, z K ) s1 s 2 k=1 L(s, y 2 (s), τ(s)) ds W (s 1, y 2 (s 1 )) + ε L s 1 s 2 + W (s 1, z K ) W (s 1, y 2 (s 1 )) + ε. We can write, for any τ( ), W (s 1, z K ) W (s 1, y 2 (s 1 )) = W (s 1, z K ) W (s 1, y 2 (t + K )) + W (s 1, y 2 (t + K )) W (s 1, y 2 (s 1 )) C 1 z K y 2 (t + K ) + C 1 f s 1 s 2. Now, for k {1,.., K}, we examine the difference k = y 2 (t + k ) z k. k+1 = y 2 (t k+1 ) + g(t k+1, y 2 (t k+1 ), ξ k+1) z k g(s 1, z k, ξ k+1 ) (1 + c g ) k + (1 + c g ) f t k+1 t k + c gg t k+1 s 1. 1 as opposed to the construction in subsection 2.2.1, here each ξ k may be a multiple jump itself. 11

We introduce δ k = (1 + c g ) f t k+1 t k + c gg t k+1 s 1, then which can be written as k k+1 (1 + c g ) k+1 k (1 + c g ) k + δ k (1 + c g ) k+1 (4) u k+1 u k + δ k (1 + c g ) k+1 with u k =. The strategy considered here is with K impulse times where (1 + c g ) k K is bounded. So, k=k 1 δ k u K u 1 + (1 + c g ) k+1 On the one hand, for any k {1,..K} k=1 δ k ((1 + c g ) f + c gg ) s 1 s 2 = δ = C 2 s 1 s 2. On the other hand, u 1 = 1 1 + c g, with Finally, and hence Therefore, where 1 = y 2 (t 1 ) + g(t 1, y 2 (t 1 ), ξ 1) y 2 (s 2 ) g(s 1, y 2 (s 2 ), ξ 1 ) δ We let ε 0 to obtain K 1 1 u K δ (1 + c g ) k+1 Kδ Kδ, k=0 K (1 + c g ) K Kδ (1 + c g ) K Kδ. W (s 1, x) W (s 2, x) C 3 s 1 s 2 + ε C 3 = L + C 1 [ f + C 2 (1 + c g ) K K. W (s 1, x) W (s 2, x) C 3 s 1 s 2. Since the inequality is symmetric in s 1 and s 2, it follows that W (s 1, x) W (s 2, x) C 3 s 1 s 2. We then proved that W is Lipschitz continuous with respect to time t, uniformly in x. 12

3.3 Terminal value Because of the possible jumps at the terminal time T, it is easy to see that, in general, W (t, x) does not tend to G(x) as t tends to T. Extend the set of multiple jumps to include jumps of zero multiplicity, meaning no jump. Call this extended set Ξ 0, extend trivially the operator M to a function independant from t, and let G(x) = inf ξ Ξ 0 [G(x + g(t, x, ξ)) + C(ξ) = min{g(x), MG(T, x)}. (5) We know from lemma 2.1 that G(x) is Lipschitz continuous. We claim Lemma 3.6 W (t, x) G(x) as t T. Proof Fix (t, x) and a strategy Φ. As in the previous proof, for each τ( ), gather all jumps of ψ = Φ(τ), if any, in a multiple jump ξ at the time T, and let z = x + g(t, x, ξ). The same argument as previously shows that there exists a constant C T such that, for all Φ and τ or J(t, x, Φ, τ) [G(z) + C( ξ) C T (T t). J(t, x, Φ, τ) = G(z) + C( ξ) + O(T t). The right hand side above only depends on ξ, not on τ( ) itself. It follows that inf sup J(t, x, Φ, τ) = inf (G(z) + C(ξ)) + O(T t) = G(x) + O(T t). φ ξ Ξ 0 τ The result follows letting t T. Remark 3.1 Because of the uniform convergence of W (t, x) to G(x) as t T, using the lemma 2.1 we have MW (t, x) M G(T, x) as t T, uniformly in R n, and using proposition 3.2, G(x) M G(T, x). 4 Isaacs quasi-variational inequality In this section we proove that the value function W is a viscosity solution of the Hamilton-Jacobi-Isaacs quasi-variational inequality, that we replace by an equivalent QVI easier to investigate. 13

4.1 Natural Isaacs Quasi-Variational Inequality In the domain [t 0, T [ R n, we consider the QVI { [ max min W τ K t W f(t, x, τ) L(t, x, τ), x } W (t, x) MW (t, x) = 0. (6) with the terminal condition: W (T, x) = G(x) in R n, where G is given by (5). Notice that it follows from hypothesis 1 that the term in square brackets in equation (6) above is continuous with respect to τ so that the minimum in τ over the compact K exists. Theorem 4.1 The function: (t, x) W (t, x) is a viscosity solution of the quasivariational inequality (6). Proof The proof will be in two parts (i) Let φ C 1 ([t 0, T [ R n ) and let ( t, x) be a local maximum for W φ. We have to prove that ( t, x) satisfies { [ max min φ τ t φ x f( t, x, τ) L( t, x, τ), } W ( t, x) inf [W ( t, x + g( t, x, ξ)) + C(ξ) 0. ξ (ii) Let φ C 1 ([t 0, T [ R n ) and let (t, x) be a local minimum for W φ. We have to prove that (t, x) satisfies { [ max min φ τ t φ f(t, x, τ) L(t, x, τ), x } W (t, x) inf [W (t, x + g(t, x, ξ)) + C(ξ) 0. ξ Let us prove (i). By Proposition 3.2, we have W ( t, x) inf ξ [W ( t, x + g( t, x, ξ)) + C(ξ) 0. Next for t > t, t close to t, we choose a strategy Φ where we have no jumps between t and t. By Proposition 3.1, we have [ t W ( t, x) sup τ( ) t L(s, y(s), τ(s)) ds + W (t, y(t )) 14.

By hypothesis, there exists a neighborhood N of ( t, x) such that W φ is maximum at ( t, x) in that neighborhood. For small enough t t > 0, since f is bounded, so is ẏ and then we can conclude that (t, y(t )) N. Hence we get thus This yields If 0 inf τ( ) 0 sup τ( ) [ t W (t, y(t )) W ( t, x) φ(t, y(t )) φ( t, x), t [ t L(s, y(s), τ(s)) ds + φ(t, y(t )) φ( t, x). t ( φ s φ y ) f(s, y(s), τ(s)) L(s, y(s), τ(s)) ds. (7) [ min φ τ t ( t, x) φ y f( t, x, τ) L( t, x, τ) > 0, then, this inequality is also true in a neighborhood of ( t, x), for any choice of τ K. Then t t ( φ s φ y ) f(s, y(s), τ(s)) L(s, y(s), τ(s)) ds > 0, which is in contradiction with inequality (7). Then, we conclude [ min φ τ t ( t, x) φ y f( t, x, τ) L( t, x, τ) 0. Let us now prove (ii). Let (t, x) be a local minimum for W φ. We have, by Proposition 3.2, If W (t, x) inf ξ W (t, x) inf ξ then we are done. Otherwise W (t, x) inf ξ But we have, by Proposition 3.3, W (t, x) = inf Φ sup τ [ t1 t [W (t, x + g(t, x, ξ)) + C(ξ) 0. [W (t, x + g(t, x, ξ)) + C(ξ) = 0, [W (t, x + g(t, x, ξ)) + C(ξ) < 0. (8) L(s, y(s), τ(s)) ds + MW (t 1, y(t 1 )), 15

t 1 being the first impulse time after t. For any ε > 0, there exists a strategy Φ ε where t ε 1 is the first impulse time after t, such that [ t ε 1 W (t, x) sup L(s, y(s), τ(s)) ds + MW (t ε 1, y(t ε 1)) ε. τ t We prove in the sequel that there exists δ > 0 such that t ε 1 t+δ for all ε. Assume to the contrary that there exists a sequence t εn 1 that converges towards t when ε n converges towards 0. Let η > 0 be any fixed real number. Since L is bounded, for n large enough, it holds that, for any τ( ), t εn 1 L(s, y(s), τ(s)) ds L t εn 1 t η. t Hence W (t, x) MW (t εn 1, y(tεn 1 )) ε n η. Now we choose ξ n such that MW (t εn 1, y(tεn 1 )) W (tεn 1, y(tεn 1 ) + g(tn 1, y(t εn 1 ), ξ n)) + C(ξ n ) η. Using the continuity of W and of g, it follows that, again for large enough n, W (t εn 1, y(tεn 1 ) + g(tn 1, y(t εn 1 ), ξ n)) W (t, x + g(t, x, ξ n )) η. Finally, we also have W (t, x + g(t, x, ξ n )) + C(ξ n ) MW (t, x). Combining the last four inequalities, we get, in the limit in n W (t, x) MW (t, x) 3η. This being true for any positive η, we conclude W (t, x) MW (t, x). This inequality is in contradiction with inequality (8). This means that t ε 1 t + δ. Now we choose t t, t + δ[. By Proposition 3.1, we have [ t W (t, x) = inf L(s, y(s), τ(s)) ds + W (t, y(t )) Φ τ Ω t [ t = sup L(s, y(s), τ(s)) ds + W (t, y(t )) τ Ω t t t L(s, y(s), τ) ds + W (t, y(t )), 16

for any constant τ K in the last line above. Since (t, x) is a local minimum of W φ, then for t close enough to t, for the same reasons than in (i), (t, y(t )) is in the neighborhood of (t, x), then we have that Then 0 0 t t t t L(s, y(s), τ) ds + φ(t, y(t )) φ(t, y), ( L(s, y(s), τ) + φ s + φ y Dividing by t t and letting t tend to t, we obtain ) f(s, y(s), τ) ds. φ t φ f(t, x, τ) L(t, x, τ) 0. y and then [ min φ τ t φ y f(t, x, τ) L(t, x, τ) 0. 4.2 An Equivalent Quasi-Variational Inequality In this section, we consider the new function Γ given by the classical change of variable Γ(t, x) = exp(t)w (t, x), for any t [t 0, T and x R n. Of course, the function Γ is bounded and uniformly continuous with respect to its arguments. A second property is given by the Proposition 4.2 W is a viscosity solution of (6) if and only if Γ is a viscosity solution to the following quasi-variational inequality in [t 0, T [ R n, { [ 0 = max min Γ Γ + Γ(t, x) f(t, x, τ) exp(t)l(t, x, τ), τ t } x (9) Γ(t, x) N Γ(t, x) where N Γ(t, x) = inf(γ(t, x + g(t, x, ξ)) + exp(t)c(ξ)). The terminal condition ξ for Γ is: Γ(T, x) = exp(t ) G(x) in R n. Proof It suffices to prove that if W is a viscosity solution to (6), then Γ is a viscosity solution to (9). The relation between W and Γ is symmetric and the proof the other way around would be the same. The proof will be in two parts 17

(i) Let φ C 1 ([t 0, T [ R n ) and let ( t, x) be a local maximum for Γ φ. We have to prove that ( t, x) satisfies { [ max min φ τ t ( t, x) + Γ( t, x) φ x ( t, x)f( t, x, τ) e t L( t, x, τ), } Γ( t, x) N Γ( t, x) 0 (ii) Let φ C 1 ([t 0, T [ R n ) and let (t, x) be a local minimum for Γ φ. We have to prove that (t, x) satisfies { [ max min φ φ (t, x) + Γ(t, x) τ t x (t, x)f(t, x, τ) et L(t, x, τ), } Γ(t, x) N Γ(t, x) 0 Let us prove (i). For (t, x) in the neighborhood of ( t, x), we have then M 1 = (Γ φ)( t, x) (Γ φ)(t, x) = exp(t)w (t, x) φ(t, x) 0 (W Φ)(t, x) where Φ(t, x) = exp( t)(φ(t, x) + M 1 ). Remark that ( t, x) is also a local maximum to W Φ. But since W is a viscosity solution to (6), then { [ max min exp( t)(φ( t, x) + M 1 ) exp( t) φ τ t ( t, x) exp( t) φ x ( t, x)f( t, x, τ) L( t, x, τ), } exp( t)γ( t, x) inf (exp( t)γ(t, x + g(t, x, ξ)) + C(ξ)) 0 ξ and then max { [ min Γ( t, x) φ τ t ( t, x) φ x ( t, x)f( t, x, τ) exp( t)l( t, x, τ), } Γ( t, x) inf ξ Γ(t, x + g(t, x, ξ)) + exp( t)c(ξ)) This means that Γ is a viscosity subsolution to (9). The idea of the proof is the same for point (ii). We then proved that Γ is a viscosity solution to (9). 0 18

5 Uniqueness We are now going to use the equivalent QVI to prove uniqueness of the viscosity solution. To do so, we introduce a variational inequality which we first study. 5.1 General Obstacle Problem In this section, we are interested by the following variational inequality where the second term under the max in the quasi-variational inequality (9) (Γ N Γ)(t, x) is replaced by the more general term (Γ P )(t, x) where P is in BUC([t 0, T [ R n ): for all (t, x) [t 0, T [ R n, 0 = max { [ min Γ τ t with the terminal condition: Γ + Γ(t, x) } x (Γ P )(t, x), f(t, x, τ) exp(t)l(t, x, τ), (10) x R n, Γ(T, x) = exp(t ) G(x). (11) For compatibility reasons, in order to have a solution for such obstacle problems, one has to assume that the function P satisfies P (T, x) exp(t ) G(x); by Remark 3.1, this is indeed the case for P := N Γ(t, x). In all the obstacle problems we consider below, we assume implicitly that such condition is fulfilled. Let Γ 1 and Γ 2 be respectively viscosity solutions of (10) in BUC, where L is equal to L 1, P is equal to P 1 and G is equal to G 1, respectively L is equal to L 2, P is equal to P 2 and G is equal to G 2, that is, { [ max min Γ 1 τ t + Γ 1(t, x) Γ 1 x f(t, x, τ) exp(t)l 1(t, x, τ), } (Γ 1 P 1 )(t, x) = 0, with the terminal condition: for all x R n, Γ 1 (T, x) = exp(t )G 1 (x), and { [ max min Γ 2 τ t + Γ 2(t, x) Γ 2 x f(t, x, τ) exp(t)l 2(t, x, τ), } (Γ 2 P 2 )(t, x) = 0, with the terminal condition: for all x R n, Γ 2 (T, x) = exp(t )G 2 (x). 19

Theorem 5.1 Assume that both Γ 1 and Γ 2 are bounded and uniformly continuous with respect to their arguments. If f, L 2, and G satisfy the assumptions 1, 3, and 5 then, (Γ 1 Γ 2 ) + } max {e T (L 1 L 2 ) +, (P 1 P 2 ) +, e T (G 1 G 2 ) +. (12) The end of this section is devoted to the proof of that theorem. We build the following test-function: let t, s [t 0, T and x, z R n, φ α,β,γ (t, x, s, z) = Γ 1 (t, x) Γ 2 (s, z) α( x 2 + z 2 ) x z 2 t s 2 β 2 γ 2. Let ( t, x, s, z) be a maximum point of φ α,β,γ which exists, since this is a continuous function going to infinity when x or z does, and s and t range over a compact set. We need the following lemma for the proof of Theorem (5.1). Lemma 5.2 Assume that M = sup [t0,t R n(γ 1 Γ 2 )(t, x) > 0, let ( t, x, s, z) be a maximal point of φ α,β,γ, and M α,β,γ = φ α,β,γ ( t, x, s, z). Then, for any ε > 0, there exists α 0, β 0 and γ 0 such that, for any α α 0, β β 0, γ γ 0, α( x 2 + z 2 ) + x z 2 β 2 + t s 2 γ 2 ε. (13) Proof By definition, ( t, x, s, z) satisfies, for all t, s [t 0, T, x, z R n Γ 1 ( t, x) Γ 2 ( s, z) α( x 2 + z 2 x z 2 ) β 2 t s 2 γ 2 Γ 1 (t, x) Γ 2 (s, z) α( x 2 + z 2 x z 2 t s 2 ) β 2 γ 2. Let us take t = s and x = z, then M α,β,γ Γ 1 (t, x) Γ 2 (t, x) 2α x 2. Let (t, x ) be a point where the sup(γ 1 Γ 2 ) is reached within δ, then Γ 1 (t, x ) Γ 2 (t, x ) M δ, where δ > 0 can be chosen arbitrarily small. We choose it and α such that M δ 2α x 2 > 0, which is always possible because x depends only on δ. We have in particular that M α,β,γ Γ 1 (t, x ) Γ 2 (t, x ) 2α x 2 M δ 2α x 2 > 0. (14) 20

Let r 2 = Γ 1 + Γ 2, then 0 < M δ 2α x 2 M α,β,γ r 2 α( x 2 + z 2 ) Thus, x z 2 β 2 t s 2 γ 2 x z rβ and t s rγ. (15) The function Γ 2 is assumed uniformly continuous in t and also in x. Let us therefore introduce the following functions: m(u) = n(v) = sup Γ 2 (t, x) Γ 2 (t, z), t, x z u Γ 2 (t, x) Γ 2 (s, x). sup t s v,x (16) Clearly, m and n decrease to 0 with their arguments. With these notations, and using (15), we have Γ 1 ( t, x) Γ 2 ( s, z) = Γ 1 ( t, x) Γ 2 ( t, x) + Γ 2 ( t, x) Γ 2 ( t, z) + Γ 2 ( t, z) Γ 2 ( s, z) M + m(rβ) + n(rγ). Place this in the definition of M α,β,γ and use (14): M δ 2α x 2 M α,β,γ or equivalently M + m(rβ) + n(rγ) α( x 2 + z 2 ) α( x 2 + z 2 ) + x z 2 β 2 t s 2 γ 2 x z 2 β 2 + t s 2 γ 2 δ + 2α x 2 + m(rβ) + n(rγ). Pick ε < 5M/3, take δ = ε/5, α 0 = ε/(5 x 2 ) if x 0, α 0 = 1 if x = 0, m(rβ 0 ) = n(rγ 0 ) = ε/5 to get (13). Let us now give the proof of Theorem (5.1). Let ( t, x, s, z) be a maximum point of φ α,β,γ. Case t and s different from T Assume first that both t and s are different from T. Then, for all x R n, t [t 0, T, Γ 1 ( t, x) Γ 2 ( s, z) α( x 2 + z 2 x z 2 ) β 2 t s 2 γ 2 Γ 1 (t, x) Γ 2 ( s, z) α( x 2 + z 2 x z 2 t s 2 ) β 2 γ 2. 21

Let φ 1 C 1 ([t 0, T [ R n ) be defined as φ 1 (t, x) = Γ 2 ( s, z) + α( x 2 + z 2 ) + x z 2 β 2 + t s 2 γ 2. This last inequality means that ( t, x) is a maximal point of Γ 1 (t, x) φ 1 (t, x). We also have, for all z R n, s [t 0, T Γ 1 ( t, x) Γ 2 ( s, z) α( x 2 + z 2 x z 2 ) β 2 t s 2 γ 2 Γ 1 ( t, x) Γ 2 (s, z) α( x 2 + z 2 x z 2 ) β 2 t s 2 γ 2. Let also φ 2 C 1 ([t 0, T [ R n ) be defined as φ 2 (s, z) = Γ 1 ( t, x) α( x 2 + z 2 ) x z 2 β 2 t s 2 γ 2. This inequality means that ( s, z) is a minimal point of Γ 2 (s, z) φ 2 (s, z). Then, { [ max min φ 1 τ t ( t, x) + Γ 1 ( t, x) φ 1 x ( t, x)f( t, x, τ) e t L 1 ( t, x, τ), } (17) Γ 1 ( t, x) P 1 ( t, x) 0, and { [ max min φ 2 τ s ( s, z) + Γ 2( s, z) φ 2 z ( s, z)f( s, z, τ) e s L 2 ( s, z, τ), } Γ 2 ( s, z) P 2 ( s, z) 0. (18) Inequality (17) can be decomposed into two inequalities (19), where we have set p β = 2( x z)/β 2, and (20): [ 2( t s) min τ γ 2 + Γ 1 ( t, x) p β + 2α x, f( t, x, τ) e t L 1 ( t, x, τ) 0, (19) Γ 1 ( t, x) P 1 ( t, x) 0, (20) Inequality (18) is equivalent to either inequality (21), or inequality (22): [ 2( t s) min τ γ 2 + Γ 2 ( s, z) p β 2α z, f( s, z, τ) e s L 2 ( s, z, τ) 0, (21) 22

Γ 2 ( s, z) P 2 ( s, z) 0. (22) In (21), min τ may be replaced by for all τ. Whenever it holds, [ Γ 1 ( t, x) Γ 2 ( s, z) sup p β, f( t, x, τ) f( s, z, τ) τ +2α x, f( t, x, τ) + 2α z, f( s, z, τ) +e t L 1 ( t, x, τ) e s L 2 ( s, z, τ) where sup[a + B + D + E + F + G τ A = p β, f( t, x, τ) f( t, z, τ), B = p β, f( t, z, τ) f( s, z, τ), D = 2α f ( x + z ), E = e t (L 1 ( t, x, τ) L 2 ( t, x, τ)), F = e t (L 2 ( t, x, τ) L 2 ( s, z, τ)), G = (e t e s )L 2 ( s, z, τ). It follows from the proof of Lemma 5.2 that, for α α 0, t s γ ε and x z β ε. L 2 is continuous in t, uniformly in (x, τ), and Lipschitz continuous in x with constant c L uniform in (t, τ). Choose γ 1 γ 0 such that, for t s γ 1 ε, e T L 2 (t, x, τ) L 2 (s, x, τ) ε for all α α 0, β β 0, and γ γ 1, and β 1 = min{β 0, e T ε/c L }. This insures that, for all α α 0, β β 1, γ γ 1, F 2ε. Finally, pick γ 2 = min{γ 1, e T ε/ L 2 }, so that for γ γ 2, we furthermore insure that G ε. We have A 2c f x z 2 /β 2 2c f ε for all α α 0, β β 0 and γ γ 0 and a fortiori if β β 1 and γ γ 2. Let us turn to B. Clearly, B p β f( t, z, τ) f( s, z, τ). Fix β = β 1. Then, for small α and γ, p β 2 ε/β 1. Since f is continuous in t, uniformly in (x, τ), using the bound t s γ ε, it is possible to find γ 3 γ 2 such that, for γ γ 3, f( t, z, τ) f( s, z, τ) εβ 1, hence B 2ε. Finally, it follows also from the lemma 5.2 that, for α, β, and γ as in the lemma, α x 2 ε, i.e. α x αε, and similarily for α z. Hence, keeping β = β 1, γ = γ 3, we may choose α 1 = min{α 0, ε f 2 } to insure that D 4ε. 23

Hence in the case under investigation, there exists a triple (α 1, β 2, γ 3 ) such that for this triple, Γ 1 ( t, x) Γ 2 ( s, z) e t sup[l 1 ( t, x, τ) L 2 ( t, x, τ) + aε τ e T (L 1 L 2 ) + + aε. (23) where a := 9 + 2c f does not depend on ε nor (α, β, γ). Let us assume now that (22) happened. Substracting (22) from (20), we obtain Γ 1 ( t, x) Γ 2 ( s, z) P 1 ( t, x) P 2 ( s, z). We use the bounds on t s and x z arising from the lemma and the uniform continuity of P 1 (or of P 2 ) to conclude that (α, β, γ) can be chosen small enough to insure Γ 1 ( t, x) Γ 2 ( s, z) (P 1 P 2 ) + + aε. In every cases, if t and s T, we have either (23) or the above inequality, thus, there always exist (α, β, γ) such that } Γ 1 ( t, x) Γ 2 ( s, z) max {e T (L 1 L 2 ) +, (P 1 P 2 ) + + aε. (24) Case t or s equal to T Now, we shall examine the case where t or s are equal to T. Then, (10) is not available. Assume that t = T. Provided that α α 0, β β 0, and γ γ 0, we may rewrite the definition (16) using the bounds (15) on t s and x z as Γ 2 (T, x) Γ 2 ( s, z) m(rβ) + n(rγ). From the boundary condition (11), we get Hence we obtain Γ 1 (T, x) = e T G 1 ( x) and Γ 2 (T, x) = e T G 2 ( x). Γ 1 (T, x) Γ 2 ( s, z) exp(t )(G 1 ( x) G 2 ( x)) + m(rβ) + n(rγ). The same reasoning holds for the case where s = T. Therefore, whenever t or s = T, we may again choose α α 0, β β 0, and γ γ 0 such that Γ 1 ( t, x) Γ 2 ( s, z) e T (G 1 G 2 ) + + aε. (25) And clearly, this also holds if both s = t = T, since then, using the continuity of G 2 and the bound (15), again if α α 0, β β 0, and γ γ 0, Γ 1 (T, x) Γ 2 (T, z) = e T (G 1 ( x) G 2 ( z)) e T (G 1 ( x) G 2 ( x)) + e T c G rβ, which is less than e T (G 1 G 2 ) + +aε provided that β min{β 0, aε/(e T c G r)}. 24

Synthesis Using (25) if either t = T or s = T (or both), and (24) otherwise, we may conclude that there always exist (α, β, γ) such that Γ 1 ( t, x) Γ 2 ( s, z) } max {e T (L 1 L 2 ) +, (P 1 P 2 ) +, e T (G 1 G 2 ) + + aε. (26) Finally, it follows from (14) that, for all (α, β, γ) used to get (26), M ε M δ 2α x 2 M α,β,γ Γ 1 ( t, x) Γ 2 ( s, z). Hence, for all (t, x) [t 0, T R n, using (26) we have Γ 1 (t, x) Γ 2 (t, x) M } max {e T (L 1 L 2 ) +, (P 1 P 2 ) +, e T (G 1 G 2 ) + + (a + 1)ε. And as ε was arbitrary, it follows (12), proving theorem 5.1. 5.2 Uniqueness of the Viscosity Solution of the New Quasi-variational Inequality In this section, we give the uniqueness result for the viscosity solution of the Hamilton-Jacobi-Isaacs quasi-variational inequality (9). Theorem 5.3 The quasi-variational inequality (9) has a unique bounded uniformly continuous viscosity solution. Proof Assume that (9) has two bounded uniformly continuous viscosity solutions Γ 1 and Γ 2. Let us first remark that for 0 < µ < 1, Θ = µγ 1 is a viscosity solution to the following quasi-variational inequality { [ 0 = max min Θ Θ + Θ(t, x) τ t x f(t, x, τ) et µl(t, x, τ), } (27) Θ(t, x) KΘ(t, x), where KΘ(t, x) = inf ξ (Θ(t, x + g(t, x, ξ)) + µe t C(ξ)), with the terminal condition: Θ(T, x) = µ exp(t ) G(x) in R n. We then apply the result (12) obtained in the previous section, with Θ = µγ 1 is the viscosity solution to the variational inequality (27) where the obstacle P 1 = KΘ and Γ 2 is a viscosity solution of (9) with the obstacle P 2 = N Γ 2. So, we have (Θ Γ 2 ) + max ( e T ((µ 1)L) +, KΘ N Γ 2 ) +, e T ((µ 1) G) + ) (28) 25

We have that, for any (t, x) [t 0, T R n (KΘ N Γ 2 )(t, x) sup(µγ 1 (t, x + g(t, x, ξ)) Γ 2 (t, x + g(t, x, ξ)) + ξ sup((µ 1) exp(t)c(ξ)) ξ We recall that C(ξ) inf ξ Hence C(ξ) = γ > 0: since µ < 1, this yields sup((µ 1) exp(t)c(ξ)) = (µ 1) exp(t)γ < γ < 0. ξ (KΘ N Γ 2 )(t, x) sup(µγ 1 (t, x + g(t, x, ξ)) Γ 2 (t, x + g(t, x, ξ))) + γ ξ and then (KΘ N Γ 2 )) + < (µγ 1 Γ 2 ) +. Equation (28) and the last inequality together imply (µγ 1 Γ 2 ) + max (e T ((µ 1)L) +, e T ((µ 1) G) ) +. Now, let µ 1, then since L and G are bounded, we finally obtain (Γ 1 Γ 2 ) + 0. This clearly implies that, for any (t, x) [t 0, T R n, Γ 1 (t, x) Γ 2 (t, x) 0. If Γ 1, Γ 2 are solutions of (9), we can exchange their role and obtain as well Γ 2 (t, x) Γ 1 (t, x), for all (t, x) [t 0, T R n. Finally Γ 1 (t, x) = Γ 2 (t, x) for any (t, x) [t 0, T R n. 6 Conclusion We can now give the result of this paper. Corollary 6.1 Under the assumptions of paragraph 2.1.2, the value function W is the unique bounded and uniformly continuous viscosity solution of the quasivariational inequality (6). 26

As an example of a use of this resut, one may consider the option pricing problem of references [4, 5. If the piecewise linear transaction costs are replaced by a more realistic piecewise affine cost, i.e. a fixed cost is charged for any transaction in addition to a variable part, then the problem at hand is exactly that considered here. This was actually the motivation for the present analysis. The problem with no fixed cost, investigated by other means in these references, leads to a more difficult problem in terms of uniqueness of the viscosity solution, since it corresponds to the case γ = 0 in this paper. As far as we know, the uniqueness of the bounded uniformly continuous viscosity solution in that case is still an open problem. References [1 Bardi M, Capuzzo-Dolcetta I (1997) Optimal and viscosity solutions of Hamilton-Jacobi-Bellman equations. Birkhaüser. Boston, Basel, Berlin [2 Barles G (1994) Solutions de viscosité des équations de Hamilton-Jacobi. Springer-Verlag, Mathématiques & Applications. Berlin, Heidelberg, New York [3 Barles G (1985) Deterministic impulse control problems. SIAM Journal on Control and Optimization 23:419 432 [4 Bernhard P (2005) A robust control approach to option pricing including transaction costs. Annals of the ISDG 7:391 416. Birkhaüser [5 Bernhard P, El Farouq N, Thiery S (2006) An Impulsive Differential Game Arising in Finance with Interesting Singularities. Annals of the ISDG 8:335 363. Birkhaüser [6 Crandall MG, Lions PL (1983) Viscosity solutions of Hamilton Jacobi equations. Transactions of the American Mathematical Society 177:1 42 [7 Dharmatti S, Shaiju AJ (2007) Infinite dimensional differential games with hybrid controls. Proceedings of Indian Academy of Sciences, Mathematics 117:233 257 [8 Dharmatti S, Ramaswamy M (2006) Zero-sum differential games involving hybrid controls. Journal of Optimization Theory and Applications 128:75 102 [9 Evans LC, Souganidis PE (1984) Differential games and representation formulas for the solution of Hamilton-Jacobi-Isaacs equations. Indiana University Journal of Mathematics 33:773 797 27

[10 Fleming WH (1964) The convergence problem for differential games, 2. Annals of Mathematical Study 52:195 210 [11 Lions PL (1982) Generalized solutions of Hamilton-Jacobi equations. Pitman, Boston [12 Lions PL, Souganidis PE (1985) Differential games, optimal control and directional derivatives of viscosity solutions of Bellman s and Isaacs Equations. SIAM Journal on Control and Optimization 23:566 583 [13 Shaiju AJ and S. Dharmatti S (2005) Differential games with continuous, switching and impulse controls. Nonlinear Analysis 63:23 41 [14 Souganidis PE (1985) Max-min representations and product formulas for the viscosity solutions of Hamilton-Jacobi equations with applications to differential games. Nonlinear Analysis, Theory, Methods and Applications 9:217 257 [15 Yong JM (1994) Zero-sum differential games involving impulse controls. Applied Mathematics and Optimization 29:243 261 28