Deterministic Minimax Impulse Control

Size: px
Start display at page:

Download "Deterministic Minimax Impulse Control"

Transcription

1 Deterministic Minimax Impulse Control N. El Farouq, Guy Barles, and P. Bernhard March 02, 2009, revised september, 15, and october, 21, 2009 Abstract We prove the uniqueness of the viscosity solution of an Isaacs quasivariational inequality arising in an impulse control minimax problem, motivated by an application in mathematical finance. Keywords: impulse control, robust control, differential games, quasi-variational inequality, viscosity solution. Mathematical subject classification: 35D40, 49L25, 49N70, 91A23, 91G80 1 Introduction The concept of viscosity solution has become central to control theory ever since the seminal paper of Crandall & Lions [6 following the volume [11. For firstorder Hamilton-Jacobi Equations and applications to optimal control and differential games, achievements due to this concept can be found noticeably in [2 and [1. Its application to differential games was pioneered by W. Fleming, before its formalization by P.L. Lions et al., in [10, and in modern form by Lions & Souganidis [12, Evans & Souganidis [9, and Souganidis [14. It is a relatively simple matter to show that the value function of control problems is a viscosity solution of a relevant Hamilton-Jacobi Equation. The strength of the concept, though, lies in the uniqueness. Early uniqueness proofs were based upon the comparison theorem of Crandall & Lions [6 and its variations. The proof we develop here for the QVI of an impulse minimax problem is based upon the Lipschitz character of the dependence of the solution of a variational inequality upon its data. Previous results in the same direction were obtained by Yong [15 in the finite horizon case but allowing less general jumps. We want also to mention, in the University Blaise Pascal, Clermont-Ferrand, France(Elfarouq@i3s.unice.fr). University François Rabelais, Tours, France(Barles@lmpt.univ-tours.fr). INRIA-Sophia Antipolis-Méditerranée, France(Pierre.Bernhard@sophia.inria.fr). 1

2 infinite horizon case, the works of Dharmatti and Shaiju [13, 7 and Dharmatti and Ramaswamy [8 involving also hybrid controls. See also [3. 2 The problem 2.1 System Dynamics Let a two-player differential game system be defined by the solution of following dynamical equations ẏ(t) = f(t, y(t), τ(t)), y(t 0 ) = x R n, y(t + k ) = y(t k ) + g(t k, y(t k ), ξ k), t k t 0, ξ k 0. Here and below, the time variable belongs to [t 0, T where T > t 0 0 are given. The state at time t, y(t) lies in R n. The system is driven by two controls, a continuous control τ(t) K R l, where K is a compact set, and an impulsive control defined by a finite sequence of impulse times t k and the controls ξ k R m controlling the jumps in y(t k ). Let ψ = ({t k }, {ξ k }), where k N. For any initial condition (t 0, x) and controls τ( ) and ψ generating a trajectory y( ) of this system, let a pay-off J be defined as J(t 0, x, ψ, τ(.)) = T t 0 L(t, y(t), τ(t)) dt + k C(ξ k ) + G(y(T )). (1) Regularity assumptions In all the paper, we assume the following 1. f(t, y, τ) and L(t, y, τ) are continuous with respect to t uniformly in y and τ, Lipschitz continuous with respect to y uniformly in t and τ with constant c f and c L respectively, and continuous with respect to τ. 2. g(t, y, ξ) is Lipschitz continuous with respect to t, uniformly in y and ξ, with constant c gg and it is Lipschitz continuous with respect to y uniformly in t and ξ, with constant c g. 3. L, f, and G are bounded. 4. C is continuous and inf ξ C(ξ) = γ > 0. 2

3 5. G is Lipschitz continuous with constant c G. It follows, inter alia, that there exists a unique solution y( ) and a unique J for any measurable τ( ) and any sequence ψ, and thus for any nonanticipative strategy Φ as defined hereafter. 2.2 Strategies Multiple jumps It may be to the best advantage of the minimizer to make a jump at some time t, immediately followed, at the same time, by another jump, and so on. As any such jump entails a cost not less than γ, a near optimal strategy will never attempt to make an infinite number of jumps. As a matter of fact, we shall show that it follows from the boundedness of L and J that the number of jumps may be restricted, with no loss of generality, to be less than K = 2[(T t 0 ) L + G + 1/2/γ. To allow for the possibility of several successive but simultaneous jumps, we proceed as follows. Let Ξ = K p=1 (R m ) p. We extend g and C from R m to Ξ in the natural way: let ξ (R m ) p be a multiple jump of multiplicity p. Let (ξ 1, ξ 2,..., ξ p ) be its components. For a given y R n, let z 0 = y, and for l {1,..., p}, z l = z l 1 + g(t, z l 1, ξ l ). Then we set g(t, y, ξ) = z p y C(ξ) = p C(ξ l ). There is no point in considering the possibility of several successive multiple jumps at the same jump time, since a sequence of simultaneous multiple jumps is a multiple jump. It easily follows that g so extended is still Lipschitz continuous in t and in y, uniformly in the other two variables, with new coefficients c gg and c g respectively. From now on, when we refer to jumps, it will always be multiple jumps, unless specifically referred to as simple jumps. Simple jumps are the same thing as a multiple jump of multiplicity 1. And of course, in a control ψ = ({t k }, {ξ k }) the ξ k are to be considered as multiple jumps. But the t k s are always assumed to be distinct. We also state the following definition: l=1 3

4 Definition 2.1 For any bounded function V : [t 0, T R n R, let the operator M be given by It enjoys the following properties: MV (t, x) = inf [V (t, x + g(t, x, ξ)) + C(ξ) ξ Ξ Lemma 2.1 The operator M is continuous for the topology. Moreover, if the function V is bounded and Lipschitz continuous, so is the function MV. Proof The first statement is elementary. Clearly, if, for all (t, x), V (t, x) V, then, clearly MV (t, x) V + γ. To prove the regularity of MV, we need to make the following remark, which as akin to the lemma 3.5 below: the range of ξ in the inf ξ operator defining MV may be restricted to jumps of multiplicity no larger than V /γ. As a matter of fact, the simple jumps 2 to K will cost at least (K 1)γ and cannot decrease V by more than 2 V. Therefore, if K 1 > 2 V /γ, for any jump of multiplicity K, say ξ = (ξ 1,..., ξ K ), V (t, x + g(t, x, ξ 1 )) + C(ξ 1 ) < V (t, x + g(t, x, ξ)) + C(ξ). Over this retricted range Ξ, (t, x) g(t, x, ξ) is Lipschitz continuous uniformly in ξ: there exists a constant C g > 0: for all (t 1, x 1 ), (t 2, x 2 ), ξ Ξ, g(t 1, x 1, ξ) g(t 2, x 2, ξ) C g ( t 1 t 2 + x 1 x 2 ). Assume V is Lipschitz continuous, more precisely that (t 1, x 1 ), (t 2, x 2 ), V (t 1, x 1 ) V (t 2, x 2 ) λ( t 1 t 2 + x 1 x 2 ). We need now evaluate = MV (t 1, x 1 ) MV (t 2, x 2 ). Choose an arbitrary ε > 0 and a ξ ε such that Then V (t 2, x 2 + g(t 2, x 2, ξ ε )) + C(ξ ε ) MV (t 2, x 2 ) + ε. MV (t 1, x 1 ) V (t 2, x 2 + g(t 2, x 2, ξ ε )) C(ξ ε ) + ε V (t 1, x 1 + g(t 1, x 1, ξ ε )) + C(ξ ε ) V (t 2, x 2 + g(t 2, x 2, ξ ε )) C(ξ ε ) + ε λ(1 + C g )[ t 1 t 2 + x 1 x 2 + ε. Since, on the one hand ε was arbitrary, and on the other hand, (t 1, x 1 ) and (t 2, x 2 ) play symmetrical roles, we may conclude that MV (t 1, x 1 ) MV (t 2, x 2 ) λ(1 + C g )[ t 1 t 2 + x 1 x 2. 4

5 2.2.2 Admissible strategies We want to investigate the problem of minimizing sup τ(.) Ω J through the impulse control. We mean to allow closed loop strategies for the minimizing control. We remark that, being only interested in the inf sup problem, and not a possible saddle point, there is no loss of generality in restricting τ( ) to open loop controls τ( ) Ω = {measurable functions [t 0, T K} We shall sometimes write τ Ω instead of τ( ) Ω. Let Ψ be the set of all finite sequences ψ. We now define the admissible closed loop strategies Φ for the minimizing impulse control ψ, as nonanticipative strategies. We shall let Π be the set of all such nonanticipative strategies. Definition 2.2 A map Φ : Ω Ψ is called a nonanticipative strategy if for any two controls τ 1 ( ) and τ 2 ( ), and any t [t 0, T, the condition on their restrictions to [t 0, t[: τ 1 [t0,t[ = τ 2 [t0,t[ implies Φ(τ 1 ) [t0,t = Φ(τ 2 ) [t0,t. 3 The value function We define the value function of the problem W : [0, T R n R as W (t 0, x) = inf sup Φ Π τ(.) Ω 3.1 Dynamic Programming Principle J(t 0, x, Φ(τ( )), τ(.)). (2) The solution of the minimax impulse control problem at hand depends on the following basic properties Proposition 3.1 (Isaacs Tenet of transition) For all t t [t 0, T [, for all x R n, t W (t, x) = inf sup L(s, y(s), τ(s))ds + C(ξ k ) + W (t, y(t )) Φ Π τ Ω t t k <t Proof Assume first that for some x, t and t > t, t W (t, x) > inf sup L(s, y(s), τ(s)) ds + C(ξ k ) + W (t, y(t )), Φ Π τ Ω t t k <t 5

6 and let the difference be 3ε. Choose an admissible strategy Φ ε 1 that approximates the infimum in the r.h.s. up to ε. Then, for any τ( ) defined over [t, t, t t L(s, y(s), τ(s)) ds + t k <t C(ξ k ) + W (t, y(t )) W (t, x) 2ε. Whatever y(t ), from t on, choose a non anticipative strategy Φ ε 2 of the game over [t, T that approximates the Value W (t, y(t )) again up to ε. The concatenation Φ ε of Φ ε 1 and Φε 2 is a nonanticipative strategy of the game over [t, T. It yields, for all τ( ), J(t, x, Φ ε, τ( )) W (t, x) ε, a contradiction. Assume to the contrary that t W (t, x) < inf sup L(s, y(s), τ(s)) ds + C(ξ k ) + W (t, y(t )), Φ Π τ Ω t t k <t and let the difference be 2ε. Choose an admissible strategy Φ ε that approximates W (t, x) up to ε, and denote (t ε k, ξε k ) the jumps it produces. Then, [ t sup L(s, y(s), τ(s)) ds + C(ξk ε ) + J(t, y(t ), Φ ε, τ) τ( ) t t ε k [ <t t inf sup L(s, y(s), τ(s)) ds + C(ξ k ) + W (t, y(t )) ε. Φ τ( ) t t k <t The above inequality can also be written in terms of the restrictions τ and τ of τ( ) to [t, t and [t, T respectively, as [ t sup L(s, y(s), τ (s)) ds + C(ξk ε ) + sup J(t, y(t ), Φ ε, τ ) τ t t ε k <t τ [ t inf sup L(s, y(s), τ (s)) ds + (3) C(ξ k ) + W (t, y(t )) ε. Φ τ t t k <t Observe that the knowledge of τ is useless to evaluate J(t, y(t ), Φ, τ ) once y(t ) is known. Therefore, the restriction of Φ ε to [t, T cannot do better than a nonanticipative strategy of the game over [t, T. As a result, W (t, y(t )) sup τ J(t, y(t ), Φ ε, τ ). But placing that in the l.h.s. of inequality (3) leads to a contradiction. 6

7 Proposition 3.2 For all t [t 0, T and x R n, W (t, x) MW (t, x). Proof Among the admissible Φ s there are those that place a jump at time t. Using the same reasoning as above, minimizing over the jump parameter ξ at time t, one gets the required inequality. Proposition 3.3 For any admissible strategy Φ, let t 1 Φ(τ( )) be the first impulse time after t. Then, for all x R n, [ t1 W (t, x) = inf sup L(s, y(s), τ(s)) ds + MW (t 1, y(t 1 )) Φ τ t Proof Proceed as in proposition 3.1 with t replaced by t 1 as specified by the strategy Φ ε chosen, and notice that the inf over [t 1, T involves a inf ξ Regularity of the value function In this section, we prove the following theorem Theorem 3.4 Under the standing assumptions (paragraph 2.1.2) the value function W is bounded, Lipschitz continuous in x uniformly in t, and Lipschitz continuous in t, uniformly in x. The rest of the section is devoted to the proof of this theorem Boundedness Let us first show that thanks to hypothesis 3, W as defined by (2) is bounded. On the one hand, a particular strategy Φ is the one where we have no impulse time. In this case k C(ξ k) = 0. Since L and G are bounded, it is easy to see that for all t [t 0, T and x R n, W (t, x) (T t 0 ) L + G. On the other hand, since the costs C(ξ k ) are positive numbers and again since L and G are bounded, then again it is easy to see that for all t [t 0, T and x R n, W (t, x) (T t 0 ) L G. We can then say that proving the first result. It follows the important property: W (T t 0 ) L + G 7

8 Lemma 3.5 There exists K N depending only on the data of the problem such that, in the definition of W, the inf Φ Π can be replaced by inf Φ Π K where Π K is the set of all nonanticipative strategies who s sums of the multiplicity of their jumps is less or equal to K. Proof of the lemma Let us choose a strategy Φ such that [ T K W (t, x) sup L(s, y(s), τ) ds + C( ξ k ) + G(y(T )) 1 τ Since then t k=1 [ T (T t) L G sup L(s, y(s), τ) ds + G(y(T )) τ t K C( ξ k ) (T t) L + G + W + 1 k=1 with γ = inf ξ C(ξ), we obtain then 2((T t 0 ) L + G ) + 1 K 1 γ (2((T t 0) L + G ) + 1). This proves that, for the strategies which are close to optimality, the number of impulses, simple or multiple, is bounded, with a bound depending only on the data of the problem Lipschitz continuity in x Let us now prove that the value function W is Lipschitz continuous in x, uniformly in t. To that aim, we estimate the difference = W (t, x 1 ) W (t, x 2 ) Fix an arbitrary ε > 0. Let Φ 2 be a strategy such that Then sup J(t, x 2, Φ 2, τ) inf sup J(t, x 2, Φ, τ) + ε = W (t, x 2 ) + ε. τ Φ τ W (t, x 1 ) sup τ J(t, x 2, Φ 2, τ) + ε sup τ J(t, x 1, Φ 2, τ) sup τ J(t, x 2, Φ 2, τ) + ε. 8

9 Now, pick τ 1 such that Then J(t, x 1, Φ 2, τ 1 ) sup J(t, x 1, Φ 2, τ) ε. τ J(t, x 1, Φ 2, τ 1 ) sup τ J(t, x 2, Φ 2, τ) + 2ε J(t, x 1, Φ 2, τ 1 ) J(t, x 2, Φ 2, τ 1 ) + 2ε. Consider now the control ψ 2 = Φ 2 (τ 1 ), composed of jump instants t 1, t 2...t K in the interval [t, T, with jumps ξ 1, ξ 2...ξ K, and let y 1 ( ) and y 2 ( ) be the trajectories generated by (Φ 2, τ 1 ), or equivalently (ψ 2, τ 1 ), from y i (t) = x i, i = 1, 2. By Gronwall s lemma, we can say for all s [t, t 1, y 1 (s) y 2 (s) exp(c f (s t)) x 1 x 2. Looking more carefully at the first jump and using the Lipschitz continuity of g, we have y 1 (t + 1 ) y 2(t + 1 ) = (y 1 (t 1 )+g(t 1, y 1 (t 1 ), ξ 1)) (y 2 (t 1 )+g(t 1, y 2 (t 1 ), ξ 1)) (1 + c g ) y 1 (t 1 ) y 2(t 1 ) exp(c f (t 1 t))(1 + c g ) x 1 x 2. Repeating inductively the same argument, we have, for an impulse time t k, k {1, 2...K}, Finally y 1 (t + k ) y 2(t + k ) exp(c f (t k t))(1 + c g ) k x 1 x 2. for all s [t, T, y 1 (s) y 2 (s)) exp(c f (s t))(1 + c g ) K x 1 x 2, where we recall that K is the number of impulses in ψ 2. Then, J(t, x 1, Φ 2, τ 1 ) J(t, x 2, Φ 2, τ 1 ) T c L y 1 (s) y 2 (s) ds + c G y 1 (T ) y 2 (T ) t [ T (1 + c g ) K x 1 x 2 c L exp(c f (s t)) ds + c G exp(c f (T t)) t [ 1 (1 + c g ) K x 1 x 2 (exp(c f (T t 0 )) 1) + c G exp(c f (T t 0 )), c f 9

10 and we can say now W (t, x 1 ) W (t, x 2 ) C 1 x 1 x 2 + 2ε and, since ε was arbitrary, W (t, x 1 ) W (t, x 2 ) C 1 x 1 x 2. where the constant C 1 = (1 + c g ) K [ 1 (exp(c f (T t 0 )) 1) + c G exp(c f (T t 0 )), c f with K the maximum number of impulses according to the lemma. The roles of x 1, x 2 being symmetrical, we finally have W (t, x 1 ) W (t, x 2 ) C 1 x 1 x Lipschitz continuity in t Let us examine now the difference W (s 1, x) W (s 2, x). Assume first that s 2 > s 1. By the dynamic programming principle, W (s 1, x) [ W (s 2, x) = s2 inf sup L(s, y(s), τ(s)) ds + C(ξ k ) + W (s 2, y(s 2 )) Φ τ s 1 t k <s 2 W (s 2, y(s 1 )) We choose the strategy Φ where there is no impulse time between s 1 and s 2, then W (s 1, x) W (s 2, x) [ s2 sup L(s, y(s), τ(s)) ds + W (s 2, y(s 2 )) W (s 2, y(s 1 )) τ s 1 L s 2 s 1 + C 1 f s 2 s 1 c 2 s 2 s 1 with c 2 = L + C 1 f Assume now that s 1 > s 2. We denote by y 2 ( ) a trajectory generated by the controls considered from y(s 2 ) = x. By the dynamic programming principle, we have W (s 1, x) W (s 2, x) = s1 W (s 1, x) inf sup L(s, y 2 (s), τ(s)) ds + C(ξ k ) + W (s 1, y 2 (s 1 )). Φ τ s 2 k t k <s 1 10

11 Choose an arbirary ε > 0, and a strategy Φ ε, such that the inf Φ above is reached within ε. We have W (s 1, x) W (s 2, x) s1 W (s 1, x) sup L(s, y 2 (s), τ(s)) ds + C(ξ k ) + W (s 1, y 2 (s 1 )) + ε s 2 k t k <s 1 W (s 1, x) τ s1 s 2 L(s, y 2 (s), τ(s)) ds k t k <s 1 C(ξ k ) W (s 1, y 2 (s 1 )) + ε for any given τ( ). Let us, from now on fix a τ( ), and let Φ ε (τ( )) exhibit K jumps (ξ 1,..., ξ K ) in the interval [s 2, s 1 [. And let y 2 ( ) be the trajectory generated by these controls from y 2 (s 2 ) = x. We now bound W (s 1, x) using this sequence of jumps and Proposition 3.2. First, consider the effect of the multiple jump ξ = (ξ 1,..., ξ K ) at time s 1 on x. Let therefore z K = x + g(s 1, x, ξ), constructed as in the definition of multiple jumps, via z 0 = x and z k = z k 1 + g(s 1, z k 1, ξ k ). 1 Using Proposition 3.2, we have K W (s 1, x) W (s 1, z K ) + C(ξ k ) Therefore W (s 1, x) W (s 2, x) W (s 1, z K ) s1 s 2 k=1 L(s, y 2 (s), τ(s)) ds W (s 1, y 2 (s 1 )) + ε L s 1 s 2 + W (s 1, z K ) W (s 1, y 2 (s 1 )) + ε. We can write, for any τ( ), W (s 1, z K ) W (s 1, y 2 (s 1 )) = W (s 1, z K ) W (s 1, y 2 (t + K )) + W (s 1, y 2 (t + K )) W (s 1, y 2 (s 1 )) C 1 z K y 2 (t + K ) + C 1 f s 1 s 2. Now, for k {1,.., K}, we examine the difference k = y 2 (t + k ) z k. k+1 = y 2 (t k+1 ) + g(t k+1, y 2 (t k+1 ), ξ k+1) z k g(s 1, z k, ξ k+1 ) (1 + c g ) k + (1 + c g ) f t k+1 t k + c gg t k+1 s 1. 1 as opposed to the construction in subsection 2.2.1, here each ξ k may be a multiple jump itself. 11

12 We introduce δ k = (1 + c g ) f t k+1 t k + c gg t k+1 s 1, then which can be written as k k+1 (1 + c g ) k+1 k (1 + c g ) k + δ k (1 + c g ) k+1 (4) u k+1 u k + δ k (1 + c g ) k+1 with u k =. The strategy considered here is with K impulse times where (1 + c g ) k K is bounded. So, k=k 1 δ k u K u 1 + (1 + c g ) k+1 On the one hand, for any k {1,..K} k=1 δ k ((1 + c g ) f + c gg ) s 1 s 2 = δ = C 2 s 1 s 2. On the other hand, u 1 = c g, with Finally, and hence Therefore, where 1 = y 2 (t 1 ) + g(t 1, y 2 (t 1 ), ξ 1) y 2 (s 2 ) g(s 1, y 2 (s 2 ), ξ 1 ) δ We let ε 0 to obtain K 1 1 u K δ (1 + c g ) k+1 Kδ Kδ, k=0 K (1 + c g ) K Kδ (1 + c g ) K Kδ. W (s 1, x) W (s 2, x) C 3 s 1 s 2 + ε C 3 = L + C 1 [ f + C 2 (1 + c g ) K K. W (s 1, x) W (s 2, x) C 3 s 1 s 2. Since the inequality is symmetric in s 1 and s 2, it follows that W (s 1, x) W (s 2, x) C 3 s 1 s 2. We then proved that W is Lipschitz continuous with respect to time t, uniformly in x. 12

13 3.3 Terminal value Because of the possible jumps at the terminal time T, it is easy to see that, in general, W (t, x) does not tend to G(x) as t tends to T. Extend the set of multiple jumps to include jumps of zero multiplicity, meaning no jump. Call this extended set Ξ 0, extend trivially the operator M to a function independant from t, and let G(x) = inf ξ Ξ 0 [G(x + g(t, x, ξ)) + C(ξ) = min{g(x), MG(T, x)}. (5) We know from lemma 2.1 that G(x) is Lipschitz continuous. We claim Lemma 3.6 W (t, x) G(x) as t T. Proof Fix (t, x) and a strategy Φ. As in the previous proof, for each τ( ), gather all jumps of ψ = Φ(τ), if any, in a multiple jump ξ at the time T, and let z = x + g(t, x, ξ). The same argument as previously shows that there exists a constant C T such that, for all Φ and τ or J(t, x, Φ, τ) [G(z) + C( ξ) C T (T t). J(t, x, Φ, τ) = G(z) + C( ξ) + O(T t). The right hand side above only depends on ξ, not on τ( ) itself. It follows that inf sup J(t, x, Φ, τ) = inf (G(z) + C(ξ)) + O(T t) = G(x) + O(T t). φ ξ Ξ 0 τ The result follows letting t T. Remark 3.1 Because of the uniform convergence of W (t, x) to G(x) as t T, using the lemma 2.1 we have MW (t, x) M G(T, x) as t T, uniformly in R n, and using proposition 3.2, G(x) M G(T, x). 4 Isaacs quasi-variational inequality In this section we proove that the value function W is a viscosity solution of the Hamilton-Jacobi-Isaacs quasi-variational inequality, that we replace by an equivalent QVI easier to investigate. 13

14 4.1 Natural Isaacs Quasi-Variational Inequality In the domain [t 0, T [ R n, we consider the QVI { [ max min W τ K t W f(t, x, τ) L(t, x, τ), x } W (t, x) MW (t, x) = 0. (6) with the terminal condition: W (T, x) = G(x) in R n, where G is given by (5). Notice that it follows from hypothesis 1 that the term in square brackets in equation (6) above is continuous with respect to τ so that the minimum in τ over the compact K exists. Theorem 4.1 The function: (t, x) W (t, x) is a viscosity solution of the quasivariational inequality (6). Proof The proof will be in two parts (i) Let φ C 1 ([t 0, T [ R n ) and let ( t, x) be a local maximum for W φ. We have to prove that ( t, x) satisfies { [ max min φ τ t φ x f( t, x, τ) L( t, x, τ), } W ( t, x) inf [W ( t, x + g( t, x, ξ)) + C(ξ) 0. ξ (ii) Let φ C 1 ([t 0, T [ R n ) and let (t, x) be a local minimum for W φ. We have to prove that (t, x) satisfies { [ max min φ τ t φ f(t, x, τ) L(t, x, τ), x } W (t, x) inf [W (t, x + g(t, x, ξ)) + C(ξ) 0. ξ Let us prove (i). By Proposition 3.2, we have W ( t, x) inf ξ [W ( t, x + g( t, x, ξ)) + C(ξ) 0. Next for t > t, t close to t, we choose a strategy Φ where we have no jumps between t and t. By Proposition 3.1, we have [ t W ( t, x) sup τ( ) t L(s, y(s), τ(s)) ds + W (t, y(t )) 14.

15 By hypothesis, there exists a neighborhood N of ( t, x) such that W φ is maximum at ( t, x) in that neighborhood. For small enough t t > 0, since f is bounded, so is ẏ and then we can conclude that (t, y(t )) N. Hence we get thus This yields If 0 inf τ( ) 0 sup τ( ) [ t W (t, y(t )) W ( t, x) φ(t, y(t )) φ( t, x), t [ t L(s, y(s), τ(s)) ds + φ(t, y(t )) φ( t, x). t ( φ s φ y ) f(s, y(s), τ(s)) L(s, y(s), τ(s)) ds. (7) [ min φ τ t ( t, x) φ y f( t, x, τ) L( t, x, τ) > 0, then, this inequality is also true in a neighborhood of ( t, x), for any choice of τ K. Then t t ( φ s φ y ) f(s, y(s), τ(s)) L(s, y(s), τ(s)) ds > 0, which is in contradiction with inequality (7). Then, we conclude [ min φ τ t ( t, x) φ y f( t, x, τ) L( t, x, τ) 0. Let us now prove (ii). Let (t, x) be a local minimum for W φ. We have, by Proposition 3.2, If W (t, x) inf ξ W (t, x) inf ξ then we are done. Otherwise W (t, x) inf ξ But we have, by Proposition 3.3, W (t, x) = inf Φ sup τ [ t1 t [W (t, x + g(t, x, ξ)) + C(ξ) 0. [W (t, x + g(t, x, ξ)) + C(ξ) = 0, [W (t, x + g(t, x, ξ)) + C(ξ) < 0. (8) L(s, y(s), τ(s)) ds + MW (t 1, y(t 1 )), 15

16 t 1 being the first impulse time after t. For any ε > 0, there exists a strategy Φ ε where t ε 1 is the first impulse time after t, such that [ t ε 1 W (t, x) sup L(s, y(s), τ(s)) ds + MW (t ε 1, y(t ε 1)) ε. τ t We prove in the sequel that there exists δ > 0 such that t ε 1 t+δ for all ε. Assume to the contrary that there exists a sequence t εn 1 that converges towards t when ε n converges towards 0. Let η > 0 be any fixed real number. Since L is bounded, for n large enough, it holds that, for any τ( ), t εn 1 L(s, y(s), τ(s)) ds L t εn 1 t η. t Hence W (t, x) MW (t εn 1, y(tεn 1 )) ε n η. Now we choose ξ n such that MW (t εn 1, y(tεn 1 )) W (tεn 1, y(tεn 1 ) + g(tn 1, y(t εn 1 ), ξ n)) + C(ξ n ) η. Using the continuity of W and of g, it follows that, again for large enough n, W (t εn 1, y(tεn 1 ) + g(tn 1, y(t εn 1 ), ξ n)) W (t, x + g(t, x, ξ n )) η. Finally, we also have W (t, x + g(t, x, ξ n )) + C(ξ n ) MW (t, x). Combining the last four inequalities, we get, in the limit in n W (t, x) MW (t, x) 3η. This being true for any positive η, we conclude W (t, x) MW (t, x). This inequality is in contradiction with inequality (8). This means that t ε 1 t + δ. Now we choose t t, t + δ[. By Proposition 3.1, we have [ t W (t, x) = inf L(s, y(s), τ(s)) ds + W (t, y(t )) Φ τ Ω t [ t = sup L(s, y(s), τ(s)) ds + W (t, y(t )) τ Ω t t t L(s, y(s), τ) ds + W (t, y(t )), 16

17 for any constant τ K in the last line above. Since (t, x) is a local minimum of W φ, then for t close enough to t, for the same reasons than in (i), (t, y(t )) is in the neighborhood of (t, x), then we have that Then 0 0 t t t t L(s, y(s), τ) ds + φ(t, y(t )) φ(t, y), ( L(s, y(s), τ) + φ s + φ y Dividing by t t and letting t tend to t, we obtain ) f(s, y(s), τ) ds. φ t φ f(t, x, τ) L(t, x, τ) 0. y and then [ min φ τ t φ y f(t, x, τ) L(t, x, τ) An Equivalent Quasi-Variational Inequality In this section, we consider the new function Γ given by the classical change of variable Γ(t, x) = exp(t)w (t, x), for any t [t 0, T and x R n. Of course, the function Γ is bounded and uniformly continuous with respect to its arguments. A second property is given by the Proposition 4.2 W is a viscosity solution of (6) if and only if Γ is a viscosity solution to the following quasi-variational inequality in [t 0, T [ R n, { [ 0 = max min Γ Γ + Γ(t, x) f(t, x, τ) exp(t)l(t, x, τ), τ t } x (9) Γ(t, x) N Γ(t, x) where N Γ(t, x) = inf(γ(t, x + g(t, x, ξ)) + exp(t)c(ξ)). The terminal condition ξ for Γ is: Γ(T, x) = exp(t ) G(x) in R n. Proof It suffices to prove that if W is a viscosity solution to (6), then Γ is a viscosity solution to (9). The relation between W and Γ is symmetric and the proof the other way around would be the same. The proof will be in two parts 17

18 (i) Let φ C 1 ([t 0, T [ R n ) and let ( t, x) be a local maximum for Γ φ. We have to prove that ( t, x) satisfies { [ max min φ τ t ( t, x) + Γ( t, x) φ x ( t, x)f( t, x, τ) e t L( t, x, τ), } Γ( t, x) N Γ( t, x) 0 (ii) Let φ C 1 ([t 0, T [ R n ) and let (t, x) be a local minimum for Γ φ. We have to prove that (t, x) satisfies { [ max min φ φ (t, x) + Γ(t, x) τ t x (t, x)f(t, x, τ) et L(t, x, τ), } Γ(t, x) N Γ(t, x) 0 Let us prove (i). For (t, x) in the neighborhood of ( t, x), we have then M 1 = (Γ φ)( t, x) (Γ φ)(t, x) = exp(t)w (t, x) φ(t, x) 0 (W Φ)(t, x) where Φ(t, x) = exp( t)(φ(t, x) + M 1 ). Remark that ( t, x) is also a local maximum to W Φ. But since W is a viscosity solution to (6), then { [ max min exp( t)(φ( t, x) + M 1 ) exp( t) φ τ t ( t, x) exp( t) φ x ( t, x)f( t, x, τ) L( t, x, τ), } exp( t)γ( t, x) inf (exp( t)γ(t, x + g(t, x, ξ)) + C(ξ)) 0 ξ and then max { [ min Γ( t, x) φ τ t ( t, x) φ x ( t, x)f( t, x, τ) exp( t)l( t, x, τ), } Γ( t, x) inf ξ Γ(t, x + g(t, x, ξ)) + exp( t)c(ξ)) This means that Γ is a viscosity subsolution to (9). The idea of the proof is the same for point (ii). We then proved that Γ is a viscosity solution to (9). 0 18

19 5 Uniqueness We are now going to use the equivalent QVI to prove uniqueness of the viscosity solution. To do so, we introduce a variational inequality which we first study. 5.1 General Obstacle Problem In this section, we are interested by the following variational inequality where the second term under the max in the quasi-variational inequality (9) (Γ N Γ)(t, x) is replaced by the more general term (Γ P )(t, x) where P is in BUC([t 0, T [ R n ): for all (t, x) [t 0, T [ R n, 0 = max { [ min Γ τ t with the terminal condition: Γ + Γ(t, x) } x (Γ P )(t, x), f(t, x, τ) exp(t)l(t, x, τ), (10) x R n, Γ(T, x) = exp(t ) G(x). (11) For compatibility reasons, in order to have a solution for such obstacle problems, one has to assume that the function P satisfies P (T, x) exp(t ) G(x); by Remark 3.1, this is indeed the case for P := N Γ(t, x). In all the obstacle problems we consider below, we assume implicitly that such condition is fulfilled. Let Γ 1 and Γ 2 be respectively viscosity solutions of (10) in BUC, where L is equal to L 1, P is equal to P 1 and G is equal to G 1, respectively L is equal to L 2, P is equal to P 2 and G is equal to G 2, that is, { [ max min Γ 1 τ t + Γ 1(t, x) Γ 1 x f(t, x, τ) exp(t)l 1(t, x, τ), } (Γ 1 P 1 )(t, x) = 0, with the terminal condition: for all x R n, Γ 1 (T, x) = exp(t )G 1 (x), and { [ max min Γ 2 τ t + Γ 2(t, x) Γ 2 x f(t, x, τ) exp(t)l 2(t, x, τ), } (Γ 2 P 2 )(t, x) = 0, with the terminal condition: for all x R n, Γ 2 (T, x) = exp(t )G 2 (x). 19

20 Theorem 5.1 Assume that both Γ 1 and Γ 2 are bounded and uniformly continuous with respect to their arguments. If f, L 2, and G satisfy the assumptions 1, 3, and 5 then, (Γ 1 Γ 2 ) + } max {e T (L 1 L 2 ) +, (P 1 P 2 ) +, e T (G 1 G 2 ) +. (12) The end of this section is devoted to the proof of that theorem. We build the following test-function: let t, s [t 0, T and x, z R n, φ α,β,γ (t, x, s, z) = Γ 1 (t, x) Γ 2 (s, z) α( x 2 + z 2 ) x z 2 t s 2 β 2 γ 2. Let ( t, x, s, z) be a maximum point of φ α,β,γ which exists, since this is a continuous function going to infinity when x or z does, and s and t range over a compact set. We need the following lemma for the proof of Theorem (5.1). Lemma 5.2 Assume that M = sup [t0,t R n(γ 1 Γ 2 )(t, x) > 0, let ( t, x, s, z) be a maximal point of φ α,β,γ, and M α,β,γ = φ α,β,γ ( t, x, s, z). Then, for any ε > 0, there exists α 0, β 0 and γ 0 such that, for any α α 0, β β 0, γ γ 0, α( x 2 + z 2 ) + x z 2 β 2 + t s 2 γ 2 ε. (13) Proof By definition, ( t, x, s, z) satisfies, for all t, s [t 0, T, x, z R n Γ 1 ( t, x) Γ 2 ( s, z) α( x 2 + z 2 x z 2 ) β 2 t s 2 γ 2 Γ 1 (t, x) Γ 2 (s, z) α( x 2 + z 2 x z 2 t s 2 ) β 2 γ 2. Let us take t = s and x = z, then M α,β,γ Γ 1 (t, x) Γ 2 (t, x) 2α x 2. Let (t, x ) be a point where the sup(γ 1 Γ 2 ) is reached within δ, then Γ 1 (t, x ) Γ 2 (t, x ) M δ, where δ > 0 can be chosen arbitrarily small. We choose it and α such that M δ 2α x 2 > 0, which is always possible because x depends only on δ. We have in particular that M α,β,γ Γ 1 (t, x ) Γ 2 (t, x ) 2α x 2 M δ 2α x 2 > 0. (14) 20

21 Let r 2 = Γ 1 + Γ 2, then 0 < M δ 2α x 2 M α,β,γ r 2 α( x 2 + z 2 ) Thus, x z 2 β 2 t s 2 γ 2 x z rβ and t s rγ. (15) The function Γ 2 is assumed uniformly continuous in t and also in x. Let us therefore introduce the following functions: m(u) = n(v) = sup Γ 2 (t, x) Γ 2 (t, z), t, x z u Γ 2 (t, x) Γ 2 (s, x). sup t s v,x (16) Clearly, m and n decrease to 0 with their arguments. With these notations, and using (15), we have Γ 1 ( t, x) Γ 2 ( s, z) = Γ 1 ( t, x) Γ 2 ( t, x) + Γ 2 ( t, x) Γ 2 ( t, z) + Γ 2 ( t, z) Γ 2 ( s, z) M + m(rβ) + n(rγ). Place this in the definition of M α,β,γ and use (14): M δ 2α x 2 M α,β,γ or equivalently M + m(rβ) + n(rγ) α( x 2 + z 2 ) α( x 2 + z 2 ) + x z 2 β 2 t s 2 γ 2 x z 2 β 2 + t s 2 γ 2 δ + 2α x 2 + m(rβ) + n(rγ). Pick ε < 5M/3, take δ = ε/5, α 0 = ε/(5 x 2 ) if x 0, α 0 = 1 if x = 0, m(rβ 0 ) = n(rγ 0 ) = ε/5 to get (13). Let us now give the proof of Theorem (5.1). Let ( t, x, s, z) be a maximum point of φ α,β,γ. Case t and s different from T Assume first that both t and s are different from T. Then, for all x R n, t [t 0, T, Γ 1 ( t, x) Γ 2 ( s, z) α( x 2 + z 2 x z 2 ) β 2 t s 2 γ 2 Γ 1 (t, x) Γ 2 ( s, z) α( x 2 + z 2 x z 2 t s 2 ) β 2 γ 2. 21

22 Let φ 1 C 1 ([t 0, T [ R n ) be defined as φ 1 (t, x) = Γ 2 ( s, z) + α( x 2 + z 2 ) + x z 2 β 2 + t s 2 γ 2. This last inequality means that ( t, x) is a maximal point of Γ 1 (t, x) φ 1 (t, x). We also have, for all z R n, s [t 0, T Γ 1 ( t, x) Γ 2 ( s, z) α( x 2 + z 2 x z 2 ) β 2 t s 2 γ 2 Γ 1 ( t, x) Γ 2 (s, z) α( x 2 + z 2 x z 2 ) β 2 t s 2 γ 2. Let also φ 2 C 1 ([t 0, T [ R n ) be defined as φ 2 (s, z) = Γ 1 ( t, x) α( x 2 + z 2 ) x z 2 β 2 t s 2 γ 2. This inequality means that ( s, z) is a minimal point of Γ 2 (s, z) φ 2 (s, z). Then, { [ max min φ 1 τ t ( t, x) + Γ 1 ( t, x) φ 1 x ( t, x)f( t, x, τ) e t L 1 ( t, x, τ), } (17) Γ 1 ( t, x) P 1 ( t, x) 0, and { [ max min φ 2 τ s ( s, z) + Γ 2( s, z) φ 2 z ( s, z)f( s, z, τ) e s L 2 ( s, z, τ), } Γ 2 ( s, z) P 2 ( s, z) 0. (18) Inequality (17) can be decomposed into two inequalities (19), where we have set p β = 2( x z)/β 2, and (20): [ 2( t s) min τ γ 2 + Γ 1 ( t, x) p β + 2α x, f( t, x, τ) e t L 1 ( t, x, τ) 0, (19) Γ 1 ( t, x) P 1 ( t, x) 0, (20) Inequality (18) is equivalent to either inequality (21), or inequality (22): [ 2( t s) min τ γ 2 + Γ 2 ( s, z) p β 2α z, f( s, z, τ) e s L 2 ( s, z, τ) 0, (21) 22

23 Γ 2 ( s, z) P 2 ( s, z) 0. (22) In (21), min τ may be replaced by for all τ. Whenever it holds, [ Γ 1 ( t, x) Γ 2 ( s, z) sup p β, f( t, x, τ) f( s, z, τ) τ +2α x, f( t, x, τ) + 2α z, f( s, z, τ) +e t L 1 ( t, x, τ) e s L 2 ( s, z, τ) where sup[a + B + D + E + F + G τ A = p β, f( t, x, τ) f( t, z, τ), B = p β, f( t, z, τ) f( s, z, τ), D = 2α f ( x + z ), E = e t (L 1 ( t, x, τ) L 2 ( t, x, τ)), F = e t (L 2 ( t, x, τ) L 2 ( s, z, τ)), G = (e t e s )L 2 ( s, z, τ). It follows from the proof of Lemma 5.2 that, for α α 0, t s γ ε and x z β ε. L 2 is continuous in t, uniformly in (x, τ), and Lipschitz continuous in x with constant c L uniform in (t, τ). Choose γ 1 γ 0 such that, for t s γ 1 ε, e T L 2 (t, x, τ) L 2 (s, x, τ) ε for all α α 0, β β 0, and γ γ 1, and β 1 = min{β 0, e T ε/c L }. This insures that, for all α α 0, β β 1, γ γ 1, F 2ε. Finally, pick γ 2 = min{γ 1, e T ε/ L 2 }, so that for γ γ 2, we furthermore insure that G ε. We have A 2c f x z 2 /β 2 2c f ε for all α α 0, β β 0 and γ γ 0 and a fortiori if β β 1 and γ γ 2. Let us turn to B. Clearly, B p β f( t, z, τ) f( s, z, τ). Fix β = β 1. Then, for small α and γ, p β 2 ε/β 1. Since f is continuous in t, uniformly in (x, τ), using the bound t s γ ε, it is possible to find γ 3 γ 2 such that, for γ γ 3, f( t, z, τ) f( s, z, τ) εβ 1, hence B 2ε. Finally, it follows also from the lemma 5.2 that, for α, β, and γ as in the lemma, α x 2 ε, i.e. α x αε, and similarily for α z. Hence, keeping β = β 1, γ = γ 3, we may choose α 1 = min{α 0, ε f 2 } to insure that D 4ε. 23

24 Hence in the case under investigation, there exists a triple (α 1, β 2, γ 3 ) such that for this triple, Γ 1 ( t, x) Γ 2 ( s, z) e t sup[l 1 ( t, x, τ) L 2 ( t, x, τ) + aε τ e T (L 1 L 2 ) + + aε. (23) where a := 9 + 2c f does not depend on ε nor (α, β, γ). Let us assume now that (22) happened. Substracting (22) from (20), we obtain Γ 1 ( t, x) Γ 2 ( s, z) P 1 ( t, x) P 2 ( s, z). We use the bounds on t s and x z arising from the lemma and the uniform continuity of P 1 (or of P 2 ) to conclude that (α, β, γ) can be chosen small enough to insure Γ 1 ( t, x) Γ 2 ( s, z) (P 1 P 2 ) + + aε. In every cases, if t and s T, we have either (23) or the above inequality, thus, there always exist (α, β, γ) such that } Γ 1 ( t, x) Γ 2 ( s, z) max {e T (L 1 L 2 ) +, (P 1 P 2 ) + + aε. (24) Case t or s equal to T Now, we shall examine the case where t or s are equal to T. Then, (10) is not available. Assume that t = T. Provided that α α 0, β β 0, and γ γ 0, we may rewrite the definition (16) using the bounds (15) on t s and x z as Γ 2 (T, x) Γ 2 ( s, z) m(rβ) + n(rγ). From the boundary condition (11), we get Hence we obtain Γ 1 (T, x) = e T G 1 ( x) and Γ 2 (T, x) = e T G 2 ( x). Γ 1 (T, x) Γ 2 ( s, z) exp(t )(G 1 ( x) G 2 ( x)) + m(rβ) + n(rγ). The same reasoning holds for the case where s = T. Therefore, whenever t or s = T, we may again choose α α 0, β β 0, and γ γ 0 such that Γ 1 ( t, x) Γ 2 ( s, z) e T (G 1 G 2 ) + + aε. (25) And clearly, this also holds if both s = t = T, since then, using the continuity of G 2 and the bound (15), again if α α 0, β β 0, and γ γ 0, Γ 1 (T, x) Γ 2 (T, z) = e T (G 1 ( x) G 2 ( z)) e T (G 1 ( x) G 2 ( x)) + e T c G rβ, which is less than e T (G 1 G 2 ) + +aε provided that β min{β 0, aε/(e T c G r)}. 24

25 Synthesis Using (25) if either t = T or s = T (or both), and (24) otherwise, we may conclude that there always exist (α, β, γ) such that Γ 1 ( t, x) Γ 2 ( s, z) } max {e T (L 1 L 2 ) +, (P 1 P 2 ) +, e T (G 1 G 2 ) + + aε. (26) Finally, it follows from (14) that, for all (α, β, γ) used to get (26), M ε M δ 2α x 2 M α,β,γ Γ 1 ( t, x) Γ 2 ( s, z). Hence, for all (t, x) [t 0, T R n, using (26) we have Γ 1 (t, x) Γ 2 (t, x) M } max {e T (L 1 L 2 ) +, (P 1 P 2 ) +, e T (G 1 G 2 ) + + (a + 1)ε. And as ε was arbitrary, it follows (12), proving theorem Uniqueness of the Viscosity Solution of the New Quasi-variational Inequality In this section, we give the uniqueness result for the viscosity solution of the Hamilton-Jacobi-Isaacs quasi-variational inequality (9). Theorem 5.3 The quasi-variational inequality (9) has a unique bounded uniformly continuous viscosity solution. Proof Assume that (9) has two bounded uniformly continuous viscosity solutions Γ 1 and Γ 2. Let us first remark that for 0 < µ < 1, Θ = µγ 1 is a viscosity solution to the following quasi-variational inequality { [ 0 = max min Θ Θ + Θ(t, x) τ t x f(t, x, τ) et µl(t, x, τ), } (27) Θ(t, x) KΘ(t, x), where KΘ(t, x) = inf ξ (Θ(t, x + g(t, x, ξ)) + µe t C(ξ)), with the terminal condition: Θ(T, x) = µ exp(t ) G(x) in R n. We then apply the result (12) obtained in the previous section, with Θ = µγ 1 is the viscosity solution to the variational inequality (27) where the obstacle P 1 = KΘ and Γ 2 is a viscosity solution of (9) with the obstacle P 2 = N Γ 2. So, we have (Θ Γ 2 ) + max ( e T ((µ 1)L) +, KΘ N Γ 2 ) +, e T ((µ 1) G) + ) (28) 25

26 We have that, for any (t, x) [t 0, T R n (KΘ N Γ 2 )(t, x) sup(µγ 1 (t, x + g(t, x, ξ)) Γ 2 (t, x + g(t, x, ξ)) + ξ sup((µ 1) exp(t)c(ξ)) ξ We recall that C(ξ) inf ξ Hence C(ξ) = γ > 0: since µ < 1, this yields sup((µ 1) exp(t)c(ξ)) = (µ 1) exp(t)γ < γ < 0. ξ (KΘ N Γ 2 )(t, x) sup(µγ 1 (t, x + g(t, x, ξ)) Γ 2 (t, x + g(t, x, ξ))) + γ ξ and then (KΘ N Γ 2 )) + < (µγ 1 Γ 2 ) +. Equation (28) and the last inequality together imply (µγ 1 Γ 2 ) + max (e T ((µ 1)L) +, e T ((µ 1) G) ) +. Now, let µ 1, then since L and G are bounded, we finally obtain (Γ 1 Γ 2 ) + 0. This clearly implies that, for any (t, x) [t 0, T R n, Γ 1 (t, x) Γ 2 (t, x) 0. If Γ 1, Γ 2 are solutions of (9), we can exchange their role and obtain as well Γ 2 (t, x) Γ 1 (t, x), for all (t, x) [t 0, T R n. Finally Γ 1 (t, x) = Γ 2 (t, x) for any (t, x) [t 0, T R n. 6 Conclusion We can now give the result of this paper. Corollary 6.1 Under the assumptions of paragraph 2.1.2, the value function W is the unique bounded and uniformly continuous viscosity solution of the quasivariational inequality (6). 26

27 As an example of a use of this resut, one may consider the option pricing problem of references [4, 5. If the piecewise linear transaction costs are replaced by a more realistic piecewise affine cost, i.e. a fixed cost is charged for any transaction in addition to a variable part, then the problem at hand is exactly that considered here. This was actually the motivation for the present analysis. The problem with no fixed cost, investigated by other means in these references, leads to a more difficult problem in terms of uniqueness of the viscosity solution, since it corresponds to the case γ = 0 in this paper. As far as we know, the uniqueness of the bounded uniformly continuous viscosity solution in that case is still an open problem. References [1 Bardi M, Capuzzo-Dolcetta I (1997) Optimal and viscosity solutions of Hamilton-Jacobi-Bellman equations. Birkhaüser. Boston, Basel, Berlin [2 Barles G (1994) Solutions de viscosité des équations de Hamilton-Jacobi. Springer-Verlag, Mathématiques & Applications. Berlin, Heidelberg, New York [3 Barles G (1985) Deterministic impulse control problems. SIAM Journal on Control and Optimization 23: [4 Bernhard P (2005) A robust control approach to option pricing including transaction costs. Annals of the ISDG 7: Birkhaüser [5 Bernhard P, El Farouq N, Thiery S (2006) An Impulsive Differential Game Arising in Finance with Interesting Singularities. Annals of the ISDG 8: Birkhaüser [6 Crandall MG, Lions PL (1983) Viscosity solutions of Hamilton Jacobi equations. Transactions of the American Mathematical Society 177:1 42 [7 Dharmatti S, Shaiju AJ (2007) Infinite dimensional differential games with hybrid controls. Proceedings of Indian Academy of Sciences, Mathematics 117: [8 Dharmatti S, Ramaswamy M (2006) Zero-sum differential games involving hybrid controls. Journal of Optimization Theory and Applications 128: [9 Evans LC, Souganidis PE (1984) Differential games and representation formulas for the solution of Hamilton-Jacobi-Isaacs equations. Indiana University Journal of Mathematics 33:

28 [10 Fleming WH (1964) The convergence problem for differential games, 2. Annals of Mathematical Study 52: [11 Lions PL (1982) Generalized solutions of Hamilton-Jacobi equations. Pitman, Boston [12 Lions PL, Souganidis PE (1985) Differential games, optimal control and directional derivatives of viscosity solutions of Bellman s and Isaacs Equations. SIAM Journal on Control and Optimization 23: [13 Shaiju AJ and S. Dharmatti S (2005) Differential games with continuous, switching and impulse controls. Nonlinear Analysis 63:23 41 [14 Souganidis PE (1985) Max-min representations and product formulas for the viscosity solutions of Hamilton-Jacobi equations with applications to differential games. Nonlinear Analysis, Theory, Methods and Applications 9: [15 Yong JM (1994) Zero-sum differential games involving impulse controls. Applied Mathematics and Optimization 29:

Differential games withcontinuous, switching and impulse controls

Differential games withcontinuous, switching and impulse controls Differential games withcontinuous, switching and impulse controls A.J. Shaiju a,1, Sheetal Dharmatti b,,2 a TIFR Centre, IISc Campus, Bangalore 5612, India b Department of Mathematics, Indian Institute

More information

Prof. Erhan Bayraktar (University of Michigan)

Prof. Erhan Bayraktar (University of Michigan) September 17, 2012 KAP 414 2:15 PM- 3:15 PM Prof. (University of Michigan) Abstract: We consider a zero-sum stochastic differential controller-and-stopper game in which the state process is a controlled

More information

Differential Games II. Marc Quincampoix Université de Bretagne Occidentale ( Brest-France) SADCO, London, September 2011

Differential Games II. Marc Quincampoix Université de Bretagne Occidentale ( Brest-France) SADCO, London, September 2011 Differential Games II Marc Quincampoix Université de Bretagne Occidentale ( Brest-France) SADCO, London, September 2011 Contents 1. I Introduction: A Pursuit Game and Isaacs Theory 2. II Strategies 3.

More information

Chain differentials with an application to the mathematical fear operator

Chain differentials with an application to the mathematical fear operator Chain differentials with an application to the mathematical fear operator Pierre Bernhard I3S, University of Nice Sophia Antipolis and CNRS, ESSI, B.P. 145, 06903 Sophia Antipolis cedex, France January

More information

On the Bellman equation for control problems with exit times and unbounded cost functionals 1

On the Bellman equation for control problems with exit times and unbounded cost functionals 1 On the Bellman equation for control problems with exit times and unbounded cost functionals 1 Michael Malisoff Department of Mathematics, Hill Center-Busch Campus Rutgers University, 11 Frelinghuysen Road

More information

Nonlinear Control Systems

Nonlinear Control Systems Nonlinear Control Systems António Pedro Aguiar pedro@isr.ist.utl.pt 3. Fundamental properties IST-DEEC PhD Course http://users.isr.ist.utl.pt/%7epedro/ncs2012/ 2012 1 Example Consider the system ẋ = f

More information

On the Multi-Dimensional Controller and Stopper Games

On the Multi-Dimensional Controller and Stopper Games On the Multi-Dimensional Controller and Stopper Games Joint work with Yu-Jui Huang University of Michigan, Ann Arbor June 7, 2012 Outline Introduction 1 Introduction 2 3 4 5 Consider a zero-sum controller-and-stopper

More information

An introduction to Mathematical Theory of Control

An introduction to Mathematical Theory of Control An introduction to Mathematical Theory of Control Vasile Staicu University of Aveiro UNICA, May 2018 Vasile Staicu (University of Aveiro) An introduction to Mathematical Theory of Control UNICA, May 2018

More information

A deterministic approach to the Skorokhod problem

A deterministic approach to the Skorokhod problem Control and Cybernetics vol. 35 (26) No. 4 A deterministic approach to the Skorokhod problem by Piernicola Bettiol SISSA/ISAS via Beirut, 2-4 - 3413 Trieste, Italy e-mail: bettiol@sissa.it Abstract: We

More information

Min-Max Certainty Equivalence Principle and Differential Games

Min-Max Certainty Equivalence Principle and Differential Games Min-Max Certainty Equivalence Principle and Differential Games Pierre Bernhard and Alain Rapaport INRIA Sophia-Antipolis August 1994 Abstract This paper presents a version of the Certainty Equivalence

More information

Sébastien Chaumont a a Institut Élie Cartan, Université Henri Poincaré Nancy I, B. P. 239, Vandoeuvre-lès-Nancy Cedex, France. 1.

Sébastien Chaumont a a Institut Élie Cartan, Université Henri Poincaré Nancy I, B. P. 239, Vandoeuvre-lès-Nancy Cedex, France. 1. A strong comparison result for viscosity solutions to Hamilton-Jacobi-Bellman equations with Dirichlet condition on a non-smooth boundary and application to parabolic problems Sébastien Chaumont a a Institut

More information

Hausdorff Continuous Viscosity Solutions of Hamilton-Jacobi Equations

Hausdorff Continuous Viscosity Solutions of Hamilton-Jacobi Equations Hausdorff Continuous Viscosity Solutions of Hamilton-Jacobi Equations R Anguelov 1,2, S Markov 2,, F Minani 3 1 Department of Mathematics and Applied Mathematics, University of Pretoria 2 Institute of

More information

Numerical Methods for Optimal Control Problems. Part I: Hamilton-Jacobi-Bellman Equations and Pontryagin Minimum Principle

Numerical Methods for Optimal Control Problems. Part I: Hamilton-Jacobi-Bellman Equations and Pontryagin Minimum Principle Numerical Methods for Optimal Control Problems. Part I: Hamilton-Jacobi-Bellman Equations and Pontryagin Minimum Principle Ph.D. course in OPTIMAL CONTROL Emiliano Cristiani (IAC CNR) e.cristiani@iac.cnr.it

More information

Multi-dimensional Stochastic Singular Control Via Dynkin Game and Dirichlet Form

Multi-dimensional Stochastic Singular Control Via Dynkin Game and Dirichlet Form Multi-dimensional Stochastic Singular Control Via Dynkin Game and Dirichlet Form Yipeng Yang * Under the supervision of Dr. Michael Taksar Department of Mathematics University of Missouri-Columbia Oct

More information

UNCERTAINTY FUNCTIONAL DIFFERENTIAL EQUATIONS FOR FINANCE

UNCERTAINTY FUNCTIONAL DIFFERENTIAL EQUATIONS FOR FINANCE Surveys in Mathematics and its Applications ISSN 1842-6298 (electronic), 1843-7265 (print) Volume 5 (2010), 275 284 UNCERTAINTY FUNCTIONAL DIFFERENTIAL EQUATIONS FOR FINANCE Iuliana Carmen Bărbăcioru Abstract.

More information

HOMEWORK ASSIGNMENT 6

HOMEWORK ASSIGNMENT 6 HOMEWORK ASSIGNMENT 6 DUE 15 MARCH, 2016 1) Suppose f, g : A R are uniformly continuous on A. Show that f + g is uniformly continuous on A. Solution First we note: In order to show that f + g is uniformly

More information

Optimal Control and Viscosity Solutions of Hamilton-Jacobi-Bellman Equations

Optimal Control and Viscosity Solutions of Hamilton-Jacobi-Bellman Equations Martino Bardi Italo Capuzzo-Dolcetta Optimal Control and Viscosity Solutions of Hamilton-Jacobi-Bellman Equations Birkhauser Boston Basel Berlin Contents Preface Basic notations xi xv Chapter I. Outline

More information

The Skorokhod problem in a time-dependent interval

The Skorokhod problem in a time-dependent interval The Skorokhod problem in a time-dependent interval Krzysztof Burdzy, Weining Kang and Kavita Ramanan University of Washington and Carnegie Mellon University Abstract: We consider the Skorokhod problem

More information

INFINITE TIME HORIZON OPTIMAL CONTROL OF THE SEMILINEAR HEAT EQUATION

INFINITE TIME HORIZON OPTIMAL CONTROL OF THE SEMILINEAR HEAT EQUATION Nonlinear Funct. Anal. & Appl., Vol. 7, No. (22), pp. 69 83 INFINITE TIME HORIZON OPTIMAL CONTROL OF THE SEMILINEAR HEAT EQUATION Mihai Sîrbu Abstract. We consider here the infinite horizon control problem

More information

Example 1. Hamilton-Jacobi equation. In particular, the eikonal equation. for some n( x) > 0 in Ω. Here 1 / 2

Example 1. Hamilton-Jacobi equation. In particular, the eikonal equation. for some n( x) > 0 in Ω. Here 1 / 2 Oct. 1 0 Viscosity S olutions In this lecture we take a glimpse of the viscosity solution theory for linear and nonlinear PDEs. From our experience we know that even for linear equations, the existence

More information

On differential games with long-time-average cost

On differential games with long-time-average cost On differential games with long-time-average cost Martino Bardi Dipartimento di Matematica Pura ed Applicata Università di Padova via Belzoni 7, 35131 Padova, Italy bardi@math.unipd.it Abstract The paper

More information

An introduction to Birkhoff normal form

An introduction to Birkhoff normal form An introduction to Birkhoff normal form Dario Bambusi Dipartimento di Matematica, Universitá di Milano via Saldini 50, 0133 Milano (Italy) 19.11.14 1 Introduction The aim of this note is to present an

More information

Uniqueness of unbounded viscosity solutions for impulse control problem

Uniqueness of unbounded viscosity solutions for impulse control problem J. Math. Anal. Appl. 315 26 686 71 www.elsevier.com/locate/jmaa Uniqueness of unbounded viscosity solutions for impulse control problem Mythily Ramaswamy a,, Sheetal Dharmatti b a IISc-TIFR Mathematics

More information

AN OVERVIEW OF STATIC HAMILTON-JACOBI EQUATIONS. 1. Introduction

AN OVERVIEW OF STATIC HAMILTON-JACOBI EQUATIONS. 1. Introduction AN OVERVIEW OF STATIC HAMILTON-JACOBI EQUATIONS JAMES C HATELEY Abstract. There is a voluminous amount of literature on Hamilton-Jacobi equations. This paper reviews some of the existence and uniqueness

More information

Viscosity Solutions of the Bellman Equation for Perturbed Optimal Control Problems with Exit Times 0

Viscosity Solutions of the Bellman Equation for Perturbed Optimal Control Problems with Exit Times 0 Viscosity Solutions of the Bellman Equation for Perturbed Optimal Control Problems with Exit Times Michael Malisoff Department of Mathematics Louisiana State University Baton Rouge, LA 783-4918 USA malisoff@mathlsuedu

More information

EXISTENCE THEOREMS FOR FUNCTIONAL DIFFERENTIAL EQUATIONS IN BANACH SPACES. 1. Introduction

EXISTENCE THEOREMS FOR FUNCTIONAL DIFFERENTIAL EQUATIONS IN BANACH SPACES. 1. Introduction Acta Math. Univ. Comenianae Vol. LXXVIII, 2(29), pp. 287 32 287 EXISTENCE THEOREMS FOR FUNCTIONAL DIFFERENTIAL EQUATIONS IN BANACH SPACES A. SGHIR Abstract. This paper concernes with the study of existence

More information

Limit value of dynamic zero-sum games with vanishing stage duration

Limit value of dynamic zero-sum games with vanishing stage duration Limit value of dynamic zero-sum games with vanishing stage duration Sylvain Sorin IMJ-PRG Université P. et M. Curie - Paris 6 sylvain.sorin@imj-prg.fr Stochastic Methods in Game Theory National University

More information

Passivity-based Stabilization of Non-Compact Sets

Passivity-based Stabilization of Non-Compact Sets Passivity-based Stabilization of Non-Compact Sets Mohamed I. El-Hawwary and Manfredi Maggiore Abstract We investigate the stabilization of closed sets for passive nonlinear systems which are contained

More information

CLASSIFICATIONS OF THE FLOWS OF LINEAR ODE

CLASSIFICATIONS OF THE FLOWS OF LINEAR ODE CLASSIFICATIONS OF THE FLOWS OF LINEAR ODE PETER ROBICHEAUX Abstract. The goal of this paper is to examine characterizations of linear differential equations. We define the flow of an equation and examine

More information

A Monotone Scheme for Hamilton-Jacobi Equations via the Nonstandard Finite Difference Method

A Monotone Scheme for Hamilton-Jacobi Equations via the Nonstandard Finite Difference Method A Monotone Scheme for Hamilton-Jacobi Equations via the Nonstandard Finite Difference Method Roumen Anguelov, Jean M-S Lubuma and Froduald Minani Department of Mathematics and Applied Mathematics University

More information

2 Statement of the problem and assumptions

2 Statement of the problem and assumptions Mathematical Notes, 25, vol. 78, no. 4, pp. 466 48. Existence Theorem for Optimal Control Problems on an Infinite Time Interval A.V. Dmitruk and N.V. Kuz kina We consider an optimal control problem on

More information

arxiv: v3 [math.ds] 22 Feb 2012

arxiv: v3 [math.ds] 22 Feb 2012 Stability of interconnected impulsive systems with and without time-delays using Lyapunov methods arxiv:1011.2865v3 [math.ds] 22 Feb 2012 Sergey Dashkovskiy a, Michael Kosmykov b, Andrii Mironchenko b,

More information

HJ equations. Reachability analysis. Optimal control problems

HJ equations. Reachability analysis. Optimal control problems HJ equations. Reachability analysis. Optimal control problems Hasnaa Zidani 1 1 ENSTA Paris-Tech & INRIA-Saclay Graz, 8-11 September 2014 H. Zidani (ENSTA & Inria) HJ equations. Reachability analysis -

More information

Robustness for a Liouville type theorem in exterior domains

Robustness for a Liouville type theorem in exterior domains Robustness for a Liouville type theorem in exterior domains Juliette Bouhours 1 arxiv:1207.0329v3 [math.ap] 24 Oct 2014 1 UPMC Univ Paris 06, UMR 7598, Laboratoire Jacques-Louis Lions, F-75005, Paris,

More information

Computation of an Over-Approximation of the Backward Reachable Set using Subsystem Level Set Functions. Stanford University, Stanford, CA 94305

Computation of an Over-Approximation of the Backward Reachable Set using Subsystem Level Set Functions. Stanford University, Stanford, CA 94305 To appear in Dynamics of Continuous, Discrete and Impulsive Systems http:monotone.uwaterloo.ca/ journal Computation of an Over-Approximation of the Backward Reachable Set using Subsystem Level Set Functions

More information

A NOTE ON ALMOST PERIODIC VARIATIONAL EQUATIONS

A NOTE ON ALMOST PERIODIC VARIATIONAL EQUATIONS A NOTE ON ALMOST PERIODIC VARIATIONAL EQUATIONS PETER GIESL AND MARTIN RASMUSSEN Abstract. The variational equation of a nonautonomous differential equation ẋ = F t, x) along a solution µ is given by ẋ

More information

Institut für Mathematik

Institut für Mathematik U n i v e r s i t ä t A u g s b u r g Institut für Mathematik Martin Rasmussen, Peter Giesl A Note on Almost Periodic Variational Equations Preprint Nr. 13/28 14. März 28 Institut für Mathematik, Universitätsstraße,

More information

Asymptotic Perron Method for Stochastic Games and Control

Asymptotic Perron Method for Stochastic Games and Control Asymptotic Perron Method for Stochastic Games and Control Mihai Sîrbu, The University of Texas at Austin Methods of Mathematical Finance a conference in honor of Steve Shreve s 65th birthday Carnegie Mellon

More information

The Arzelà-Ascoli Theorem

The Arzelà-Ascoli Theorem John Nachbar Washington University March 27, 2016 The Arzelà-Ascoli Theorem The Arzelà-Ascoli Theorem gives sufficient conditions for compactness in certain function spaces. Among other things, it helps

More information

Linear Quadratic Zero-Sum Two-Person Differential Games Pierre Bernhard June 15, 2013

Linear Quadratic Zero-Sum Two-Person Differential Games Pierre Bernhard June 15, 2013 Linear Quadratic Zero-Sum Two-Person Differential Games Pierre Bernhard June 15, 2013 Abstract As in optimal control theory, linear quadratic (LQ) differential games (DG) can be solved, even in high dimension,

More information

MINIMAL TIME PROBLEM WITH IMPULSIVE CONTROLS

MINIMAL TIME PROBLEM WITH IMPULSIVE CONTROLS MINIMAL TIME PROBLEM WITH IMPULSIVE CONTROLS KARL KUNISCH AND ZHIPING RAO Abstract. Time optimal control problems for systems with impulsive controls are investigated. Sufficient conditions for the existence

More information

The main motivation of this paper comes from the following, rather surprising, result of Ecker and Huisken [13]: for any initial data u 0 W 1,

The main motivation of this paper comes from the following, rather surprising, result of Ecker and Huisken [13]: for any initial data u 0 W 1, Quasilinear parabolic equations, unbounded solutions and geometrical equations II. Uniqueness without growth conditions and applications to the mean curvature flow in IR 2 Guy Barles, Samuel Biton and

More information

Noncooperative continuous-time Markov games

Noncooperative continuous-time Markov games Morfismos, Vol. 9, No. 1, 2005, pp. 39 54 Noncooperative continuous-time Markov games Héctor Jasso-Fuentes Abstract This work concerns noncooperative continuous-time Markov games with Polish state and

More information

SPACES ENDOWED WITH A GRAPH AND APPLICATIONS. Mina Dinarvand. 1. Introduction

SPACES ENDOWED WITH A GRAPH AND APPLICATIONS. Mina Dinarvand. 1. Introduction MATEMATIČKI VESNIK MATEMATIQKI VESNIK 69, 1 (2017), 23 38 March 2017 research paper originalni nauqni rad FIXED POINT RESULTS FOR (ϕ, ψ)-contractions IN METRIC SPACES ENDOWED WITH A GRAPH AND APPLICATIONS

More information

Threshold behavior and non-quasiconvergent solutions with localized initial data for bistable reaction-diffusion equations

Threshold behavior and non-quasiconvergent solutions with localized initial data for bistable reaction-diffusion equations Threshold behavior and non-quasiconvergent solutions with localized initial data for bistable reaction-diffusion equations P. Poláčik School of Mathematics, University of Minnesota Minneapolis, MN 55455

More information

Exam February h

Exam February h Master 2 Mathématiques et Applications PUF Ho Chi Minh Ville 2009/10 Viscosity solutions, HJ Equations and Control O.Ley (INSA de Rennes) Exam February 2010 3h Written-by-hands documents are allowed. Printed

More information

Closed-Loop Impulse Control of Oscillating Systems

Closed-Loop Impulse Control of Oscillating Systems Closed-Loop Impulse Control of Oscillating Systems A. N. Daryin and A. B. Kurzhanski Moscow State (Lomonosov) University Faculty of Computational Mathematics and Cybernetics Periodic Control Systems, 2007

More information

Research Article Almost Periodic Viscosity Solutions of Nonlinear Parabolic Equations

Research Article Almost Periodic Viscosity Solutions of Nonlinear Parabolic Equations Hindawi Publishing Corporation Boundary Value Problems Volume 29, Article ID 873526, 15 pages doi:1.1155/29/873526 Research Article Almost Periodic Viscosity Solutions of Nonlinear Parabolic Equations

More information

Solution of Stochastic Optimal Control Problems and Financial Applications

Solution of Stochastic Optimal Control Problems and Financial Applications Journal of Mathematical Extension Vol. 11, No. 4, (2017), 27-44 ISSN: 1735-8299 URL: http://www.ijmex.com Solution of Stochastic Optimal Control Problems and Financial Applications 2 Mat B. Kafash 1 Faculty

More information

ON WEAKLY NONLINEAR BACKWARD PARABOLIC PROBLEM

ON WEAKLY NONLINEAR BACKWARD PARABOLIC PROBLEM ON WEAKLY NONLINEAR BACKWARD PARABOLIC PROBLEM OLEG ZUBELEVICH DEPARTMENT OF MATHEMATICS THE BUDGET AND TREASURY ACADEMY OF THE MINISTRY OF FINANCE OF THE RUSSIAN FEDERATION 7, ZLATOUSTINSKY MALIY PER.,

More information

Viscosity Solutions of Fully Nonlinear Second Order Parabolic Equations with L 1 Dependence in Time and Neumann Boundary Conditions

Viscosity Solutions of Fully Nonlinear Second Order Parabolic Equations with L 1 Dependence in Time and Neumann Boundary Conditions Viscosity Solutions of Fully Nonlinear Second Order Parabolic Equations with L 1 Dependence in Time and Neumann Boundary Conditions Mariane BOURGOING Laboratoire de Mathématiques et Physique Théorique

More information

Stability of Feedback Solutions for Infinite Horizon Noncooperative Differential Games

Stability of Feedback Solutions for Infinite Horizon Noncooperative Differential Games Stability of Feedback Solutions for Infinite Horizon Noncooperative Differential Games Alberto Bressan ) and Khai T. Nguyen ) *) Department of Mathematics, Penn State University **) Department of Mathematics,

More information

Observer design for a general class of triangular systems

Observer design for a general class of triangular systems 1st International Symposium on Mathematical Theory of Networks and Systems July 7-11, 014. Observer design for a general class of triangular systems Dimitris Boskos 1 John Tsinias Abstract The paper deals

More information

Convergence of a first order scheme for a non local eikonal equation

Convergence of a first order scheme for a non local eikonal equation Convergence of a first order scheme for a non local eikonal equation O. Alvarez, E. Carlini, R. Monneau, E. Rouy March 22, 2005 Abstract We prove the convergence of a first order finite difference scheme

More information

Functional Analysis HW #3

Functional Analysis HW #3 Functional Analysis HW #3 Sangchul Lee October 26, 2015 1 Solutions Exercise 2.1. Let D = { f C([0, 1]) : f C([0, 1])} and define f d = f + f. Show that D is a Banach algebra and that the Gelfand transform

More information

Local semiconvexity of Kantorovich potentials on non-compact manifolds

Local semiconvexity of Kantorovich potentials on non-compact manifolds Local semiconvexity of Kantorovich potentials on non-compact manifolds Alessio Figalli, Nicola Gigli Abstract We prove that any Kantorovich potential for the cost function c = d / on a Riemannian manifold

More information

Identification of Parameters in Neutral Functional Differential Equations with State-Dependent Delays

Identification of Parameters in Neutral Functional Differential Equations with State-Dependent Delays To appear in the proceedings of 44th IEEE Conference on Decision and Control and European Control Conference ECC 5, Seville, Spain. -5 December 5. Identification of Parameters in Neutral Functional Differential

More information

1 Lyapunov theory of stability

1 Lyapunov theory of stability M.Kawski, APM 581 Diff Equns Intro to Lyapunov theory. November 15, 29 1 1 Lyapunov theory of stability Introduction. Lyapunov s second (or direct) method provides tools for studying (asymptotic) stability

More information

Non-Zero-Sum Stochastic Differential Games of Controls and St

Non-Zero-Sum Stochastic Differential Games of Controls and St Non-Zero-Sum Stochastic Differential Games of Controls and Stoppings October 1, 2009 Based on two preprints: to a Non-Zero-Sum Stochastic Differential Game of Controls and Stoppings I. Karatzas, Q. Li,

More information

Optimal stopping for non-linear expectations Part I

Optimal stopping for non-linear expectations Part I Stochastic Processes and their Applications 121 (2011) 185 211 www.elsevier.com/locate/spa Optimal stopping for non-linear expectations Part I Erhan Bayraktar, Song Yao Department of Mathematics, University

More information

Further Results on the Bellman Equation for Optimal Control Problems with Exit Times and Nonnegative Instantaneous Costs

Further Results on the Bellman Equation for Optimal Control Problems with Exit Times and Nonnegative Instantaneous Costs Further Results on the Bellman Equation for Optimal Control Problems with Exit Times and Nonnegative Instantaneous Costs Michael Malisoff Systems Science and Mathematics Dept. Washington University in

More information

The Minimum Speed for a Blocking Problem on the Half Plane

The Minimum Speed for a Blocking Problem on the Half Plane The Minimum Speed for a Blocking Problem on the Half Plane Alberto Bressan and Tao Wang Department of Mathematics, Penn State University University Park, Pa 16802, USA e-mails: bressan@mathpsuedu, wang

More information

u( x) = g( y) ds y ( 1 ) U solves u = 0 in U; u = 0 on U. ( 3)

u( x) = g( y) ds y ( 1 ) U solves u = 0 in U; u = 0 on U. ( 3) M ath 5 2 7 Fall 2 0 0 9 L ecture 4 ( S ep. 6, 2 0 0 9 ) Properties and Estimates of Laplace s and Poisson s Equations In our last lecture we derived the formulas for the solutions of Poisson s equation

More information

Risk-Sensitive Control with HARA Utility

Risk-Sensitive Control with HARA Utility IEEE TRANSACTIONS ON AUTOMATIC CONTROL, VOL. 46, NO. 4, APRIL 2001 563 Risk-Sensitive Control with HARA Utility Andrew E. B. Lim Xun Yu Zhou, Senior Member, IEEE Abstract In this paper, a control methodology

More information

STOCHASTIC PERRON S METHOD AND VERIFICATION WITHOUT SMOOTHNESS USING VISCOSITY COMPARISON: OBSTACLE PROBLEMS AND DYNKIN GAMES

STOCHASTIC PERRON S METHOD AND VERIFICATION WITHOUT SMOOTHNESS USING VISCOSITY COMPARISON: OBSTACLE PROBLEMS AND DYNKIN GAMES STOCHASTIC PERRON S METHOD AND VERIFICATION WITHOUT SMOOTHNESS USING VISCOSITY COMPARISON: OBSTACLE PROBLEMS AND DYNKIN GAMES ERHAN BAYRAKTAR AND MIHAI SÎRBU Abstract. We adapt the Stochastic Perron s

More information

Approximation of Minimal Functions by Extreme Functions

Approximation of Minimal Functions by Extreme Functions Approximation of Minimal Functions by Extreme Functions Teresa M. Lebair and Amitabh Basu August 14, 2017 Abstract In a recent paper, Basu, Hildebrand, and Molinaro established that the set of continuous

More information

Notes on uniform convergence

Notes on uniform convergence Notes on uniform convergence Erik Wahlén erik.wahlen@math.lu.se January 17, 2012 1 Numerical sequences We begin by recalling some properties of numerical sequences. By a numerical sequence we simply mean

More information

Linear Quadratic Zero-Sum Two-Person Differential Games

Linear Quadratic Zero-Sum Two-Person Differential Games Linear Quadratic Zero-Sum Two-Person Differential Games Pierre Bernhard To cite this version: Pierre Bernhard. Linear Quadratic Zero-Sum Two-Person Differential Games. Encyclopaedia of Systems and Control,

More information

The principle of least action and two-point boundary value problems in orbital mechanics

The principle of least action and two-point boundary value problems in orbital mechanics The principle of least action and two-point boundary value problems in orbital mechanics Seung Hak Han and William M McEneaney Abstract We consider a two-point boundary value problem (TPBVP) in orbital

More information

HAMILTON-JACOBI EQUATIONS : APPROXIMATIONS, NUMERICAL ANALYSIS AND APPLICATIONS. CIME Courses-Cetraro August 29-September COURSES

HAMILTON-JACOBI EQUATIONS : APPROXIMATIONS, NUMERICAL ANALYSIS AND APPLICATIONS. CIME Courses-Cetraro August 29-September COURSES HAMILTON-JACOBI EQUATIONS : APPROXIMATIONS, NUMERICAL ANALYSIS AND APPLICATIONS CIME Courses-Cetraro August 29-September 3 2011 COURSES (1) Models of mean field, Hamilton-Jacobi-Bellman Equations and numerical

More information

AC&ST AUTOMATIC CONTROL AND SYSTEM THEORY SYSTEMS AND MODELS. Claudio Melchiorri

AC&ST AUTOMATIC CONTROL AND SYSTEM THEORY SYSTEMS AND MODELS. Claudio Melchiorri C. Melchiorri (DEI) Automatic Control & System Theory 1 AUTOMATIC CONTROL AND SYSTEM THEORY SYSTEMS AND MODELS Claudio Melchiorri Dipartimento di Ingegneria dell Energia Elettrica e dell Informazione (DEI)

More information

Propagating terraces and the dynamics of front-like solutions of reaction-diffusion equations on R

Propagating terraces and the dynamics of front-like solutions of reaction-diffusion equations on R Propagating terraces and the dynamics of front-like solutions of reaction-diffusion equations on R P. Poláčik School of Mathematics, University of Minnesota Minneapolis, MN 55455 Abstract We consider semilinear

More information

THE SKOROKHOD OBLIQUE REFLECTION PROBLEM IN A CONVEX POLYHEDRON

THE SKOROKHOD OBLIQUE REFLECTION PROBLEM IN A CONVEX POLYHEDRON GEORGIAN MATHEMATICAL JOURNAL: Vol. 3, No. 2, 1996, 153-176 THE SKOROKHOD OBLIQUE REFLECTION PROBLEM IN A CONVEX POLYHEDRON M. SHASHIASHVILI Abstract. The Skorokhod oblique reflection problem is studied

More information

Deterministic Dynamic Programming

Deterministic Dynamic Programming Deterministic Dynamic Programming 1 Value Function Consider the following optimal control problem in Mayer s form: V (t 0, x 0 ) = inf u U J(t 1, x(t 1 )) (1) subject to ẋ(t) = f(t, x(t), u(t)), x(t 0

More information

Proof. We indicate by α, β (finite or not) the end-points of I and call

Proof. We indicate by α, β (finite or not) the end-points of I and call C.6 Continuous functions Pag. 111 Proof of Corollary 4.25 Corollary 4.25 Let f be continuous on the interval I and suppose it admits non-zero its (finite or infinite) that are different in sign for x tending

More information

On the Well-Posedness of the Cauchy Problem for a Neutral Differential Equation with Distributed Prehistory

On the Well-Posedness of the Cauchy Problem for a Neutral Differential Equation with Distributed Prehistory Bulletin of TICMI Vol. 21, No. 1, 2017, 3 8 On the Well-Posedness of the Cauchy Problem for a Neutral Differential Equation with Distributed Prehistory Tamaz Tadumadze I. Javakhishvili Tbilisi State University

More information

Piecewise Smooth Solutions to the Burgers-Hilbert Equation

Piecewise Smooth Solutions to the Burgers-Hilbert Equation Piecewise Smooth Solutions to the Burgers-Hilbert Equation Alberto Bressan and Tianyou Zhang Department of Mathematics, Penn State University, University Park, Pa 68, USA e-mails: bressan@mathpsuedu, zhang

More information

New Discretizations of Turbulent Flow Problems

New Discretizations of Turbulent Flow Problems New Discretizations of Turbulent Flow Problems Carolina Cardoso Manica and Songul Kaya Merdan Abstract A suitable discretization for the Zeroth Order Model in Large Eddy Simulation of turbulent flows is

More information

GLOBAL EXISTENCE RESULTS AND UNIQUENESS FOR DISLOCATION EQUATIONS

GLOBAL EXISTENCE RESULTS AND UNIQUENESS FOR DISLOCATION EQUATIONS GLOBAL EXISTENCE RESULTS AND UNIQUENESS FOR DISLOCATION EQUATIONS GUY BARLES, PIERRE CARDALIAGUET, OLIVIER LEY & RÉGIS MONNEAU Abstract. We are interested in nonlocal Eikonal Equations arising in the study

More information

GENERAL EXISTENCE OF SOLUTIONS TO DYNAMIC PROGRAMMING PRINCIPLE. 1. Introduction

GENERAL EXISTENCE OF SOLUTIONS TO DYNAMIC PROGRAMMING PRINCIPLE. 1. Introduction GENERAL EXISTENCE OF SOLUTIONS TO DYNAMIC PROGRAMMING PRINCIPLE QING LIU AND ARMIN SCHIKORRA Abstract. We provide an alternative approach to the existence of solutions to dynamic programming equations

More information

Converse Lyapunov theorem and Input-to-State Stability

Converse Lyapunov theorem and Input-to-State Stability Converse Lyapunov theorem and Input-to-State Stability April 6, 2014 1 Converse Lyapunov theorem In the previous lecture, we have discussed few examples of nonlinear control systems and stability concepts

More information

L2 gains and system approximation quality 1

L2 gains and system approximation quality 1 Massachusetts Institute of Technology Department of Electrical Engineering and Computer Science 6.242, Fall 24: MODEL REDUCTION L2 gains and system approximation quality 1 This lecture discusses the utility

More information

Linear-Quadratic Stochastic Differential Games with General Noise Processes

Linear-Quadratic Stochastic Differential Games with General Noise Processes Linear-Quadratic Stochastic Differential Games with General Noise Processes Tyrone E. Duncan Abstract In this paper a noncooperative, two person, zero sum, stochastic differential game is formulated and

More information

Lecture notes on Ordinary Differential Equations. S. Sivaji Ganesh Department of Mathematics Indian Institute of Technology Bombay

Lecture notes on Ordinary Differential Equations. S. Sivaji Ganesh Department of Mathematics Indian Institute of Technology Bombay Lecture notes on Ordinary Differential Equations S. Ganesh Department of Mathematics Indian Institute of Technology Bombay May 20, 2016 ii IIT Bombay Contents I Ordinary Differential Equations 1 1 Initial

More information

A maximum principle for optimal control system with endpoint constraints

A maximum principle for optimal control system with endpoint constraints Wang and Liu Journal of Inequalities and Applications 212, 212: http://www.journalofinequalitiesandapplications.com/content/212/231/ R E S E A R C H Open Access A maimum principle for optimal control system

More information

ON THE PATHWISE UNIQUENESS OF SOLUTIONS OF STOCHASTIC DIFFERENTIAL EQUATIONS

ON THE PATHWISE UNIQUENESS OF SOLUTIONS OF STOCHASTIC DIFFERENTIAL EQUATIONS PORTUGALIAE MATHEMATICA Vol. 55 Fasc. 4 1998 ON THE PATHWISE UNIQUENESS OF SOLUTIONS OF STOCHASTIC DIFFERENTIAL EQUATIONS C. Sonoc Abstract: A sufficient condition for uniqueness of solutions of ordinary

More information

Parameter Dependent Quasi-Linear Parabolic Equations

Parameter Dependent Quasi-Linear Parabolic Equations CADERNOS DE MATEMÁTICA 4, 39 33 October (23) ARTIGO NÚMERO SMA#79 Parameter Dependent Quasi-Linear Parabolic Equations Cláudia Buttarello Gentile Departamento de Matemática, Universidade Federal de São

More information

Viscosity Solutions for Dummies (including Economists)

Viscosity Solutions for Dummies (including Economists) Viscosity Solutions for Dummies (including Economists) Online Appendix to Income and Wealth Distribution in Macroeconomics: A Continuous-Time Approach written by Benjamin Moll August 13, 2017 1 Viscosity

More information

On Stopping Times and Impulse Control with Constraint

On Stopping Times and Impulse Control with Constraint On Stopping Times and Impulse Control with Constraint Jose Luis Menaldi Based on joint papers with M. Robin (216, 217) Department of Mathematics Wayne State University Detroit, Michigan 4822, USA (e-mail:

More information

Limit value of dynamic zero-sum games with vanishing stage duration

Limit value of dynamic zero-sum games with vanishing stage duration Limit value of dynamic zero-sum games with vanishing stage duration Sylvain Sorin IMJ-PRG Université P. et M. Curie - Paris 6 sylvain.sorin@imj-prg.fr Workshop on Analysis and Applications of Stochastic

More information

Homogenization of first order equations with (u/ε)-periodic Hamiltonians. Part I: local equations

Homogenization of first order equations with (u/ε)-periodic Hamiltonians. Part I: local equations Homogenization of first order equations with u/-periodic Hamiltonians. Part I: local equations Cyril Imbert, Régis Monneau February 23, 2006 Abstract. In this paper, we present a result of homogenization

More information

THE CANTOR GAME: WINNING STRATEGIES AND DETERMINACY. by arxiv: v1 [math.ca] 29 Jan 2017 MAGNUS D. LADUE

THE CANTOR GAME: WINNING STRATEGIES AND DETERMINACY. by arxiv: v1 [math.ca] 29 Jan 2017 MAGNUS D. LADUE THE CANTOR GAME: WINNING STRATEGIES AND DETERMINACY by arxiv:170109087v1 [mathca] 9 Jan 017 MAGNUS D LADUE 0 Abstract In [1] Grossman Turett define the Cantor game In [] Matt Baker proves several results

More information

Weak solutions of mean-field stochastic differential equations

Weak solutions of mean-field stochastic differential equations Weak solutions of mean-field stochastic differential equations Juan Li School of Mathematics and Statistics, Shandong University (Weihai), Weihai 26429, China. Email: juanli@sdu.edu.cn Based on joint works

More information

Optimal stopping time formulation of adaptive image filtering

Optimal stopping time formulation of adaptive image filtering Optimal stopping time formulation of adaptive image filtering I. Capuzzo Dolcetta, R. Ferretti 19.04.2000 Abstract This paper presents an approach to image filtering based on an optimal stopping time problem

More information

Master Thesis. Nguyen Tien Thinh. Homogenization and Viscosity solution

Master Thesis. Nguyen Tien Thinh. Homogenization and Viscosity solution Master Thesis Nguyen Tien Thinh Homogenization and Viscosity solution Advisor: Guy Barles Defense: Friday June 21 th, 2013 ii Preface Firstly, I am grateful to Prof. Guy Barles for helping me studying

More information

SOME PROPERTIES ON THE CLOSED SUBSETS IN BANACH SPACES

SOME PROPERTIES ON THE CLOSED SUBSETS IN BANACH SPACES ARCHIVUM MATHEMATICUM (BRNO) Tomus 42 (2006), 167 174 SOME PROPERTIES ON THE CLOSED SUBSETS IN BANACH SPACES ABDELHAKIM MAADEN AND ABDELKADER STOUTI Abstract. It is shown that under natural assumptions,

More information

A MAXIMUM PRINCIPLE FOR SEMICONTINUOUS FUNCTIONS APPLICABLE TO INTEGRO-PARTIAL DIFFERENTIAL EQUATIONS

A MAXIMUM PRINCIPLE FOR SEMICONTINUOUS FUNCTIONS APPLICABLE TO INTEGRO-PARTIAL DIFFERENTIAL EQUATIONS Dept. of Math. University of Oslo Pure Mathematics ISBN 82 553 1382 6 No. 18 ISSN 0806 2439 May 2003 A MAXIMUM PRINCIPLE FOR SEMICONTINUOUS FUNCTIONS APPLICABLE TO INTEGRO-PARTIAL DIFFERENTIAL EQUATIONS

More information

Nonlinear L 2 -gain analysis via a cascade

Nonlinear L 2 -gain analysis via a cascade 9th IEEE Conference on Decision and Control December -7, Hilton Atlanta Hotel, Atlanta, GA, USA Nonlinear L -gain analysis via a cascade Peter M Dower, Huan Zhang and Christopher M Kellett Abstract A nonlinear

More information

A TWO PARAMETERS AMBROSETTI PRODI PROBLEM*

A TWO PARAMETERS AMBROSETTI PRODI PROBLEM* PORTUGALIAE MATHEMATICA Vol. 53 Fasc. 3 1996 A TWO PARAMETERS AMBROSETTI PRODI PROBLEM* C. De Coster** and P. Habets 1 Introduction The study of the Ambrosetti Prodi problem has started with the paper

More information

EXISTENCE OF SOLUTIONS TO ASYMPTOTICALLY PERIODIC SCHRÖDINGER EQUATIONS

EXISTENCE OF SOLUTIONS TO ASYMPTOTICALLY PERIODIC SCHRÖDINGER EQUATIONS Electronic Journal of Differential Equations, Vol. 017 (017), No. 15, pp. 1 7. ISSN: 107-6691. URL: http://ejde.math.txstate.edu or http://ejde.math.unt.edu EXISTENCE OF SOLUTIONS TO ASYMPTOTICALLY PERIODIC

More information