Deterministic Minimax Impulse Control
|
|
- Barry Nash
- 5 years ago
- Views:
Transcription
1 Deterministic Minimax Impulse Control N. El Farouq, Guy Barles, and P. Bernhard March 02, 2009, revised september, 15, and october, 21, 2009 Abstract We prove the uniqueness of the viscosity solution of an Isaacs quasivariational inequality arising in an impulse control minimax problem, motivated by an application in mathematical finance. Keywords: impulse control, robust control, differential games, quasi-variational inequality, viscosity solution. Mathematical subject classification: 35D40, 49L25, 49N70, 91A23, 91G80 1 Introduction The concept of viscosity solution has become central to control theory ever since the seminal paper of Crandall & Lions [6 following the volume [11. For firstorder Hamilton-Jacobi Equations and applications to optimal control and differential games, achievements due to this concept can be found noticeably in [2 and [1. Its application to differential games was pioneered by W. Fleming, before its formalization by P.L. Lions et al., in [10, and in modern form by Lions & Souganidis [12, Evans & Souganidis [9, and Souganidis [14. It is a relatively simple matter to show that the value function of control problems is a viscosity solution of a relevant Hamilton-Jacobi Equation. The strength of the concept, though, lies in the uniqueness. Early uniqueness proofs were based upon the comparison theorem of Crandall & Lions [6 and its variations. The proof we develop here for the QVI of an impulse minimax problem is based upon the Lipschitz character of the dependence of the solution of a variational inequality upon its data. Previous results in the same direction were obtained by Yong [15 in the finite horizon case but allowing less general jumps. We want also to mention, in the University Blaise Pascal, Clermont-Ferrand, France(Elfarouq@i3s.unice.fr). University François Rabelais, Tours, France(Barles@lmpt.univ-tours.fr). INRIA-Sophia Antipolis-Méditerranée, France(Pierre.Bernhard@sophia.inria.fr). 1
2 infinite horizon case, the works of Dharmatti and Shaiju [13, 7 and Dharmatti and Ramaswamy [8 involving also hybrid controls. See also [3. 2 The problem 2.1 System Dynamics Let a two-player differential game system be defined by the solution of following dynamical equations ẏ(t) = f(t, y(t), τ(t)), y(t 0 ) = x R n, y(t + k ) = y(t k ) + g(t k, y(t k ), ξ k), t k t 0, ξ k 0. Here and below, the time variable belongs to [t 0, T where T > t 0 0 are given. The state at time t, y(t) lies in R n. The system is driven by two controls, a continuous control τ(t) K R l, where K is a compact set, and an impulsive control defined by a finite sequence of impulse times t k and the controls ξ k R m controlling the jumps in y(t k ). Let ψ = ({t k }, {ξ k }), where k N. For any initial condition (t 0, x) and controls τ( ) and ψ generating a trajectory y( ) of this system, let a pay-off J be defined as J(t 0, x, ψ, τ(.)) = T t 0 L(t, y(t), τ(t)) dt + k C(ξ k ) + G(y(T )). (1) Regularity assumptions In all the paper, we assume the following 1. f(t, y, τ) and L(t, y, τ) are continuous with respect to t uniformly in y and τ, Lipschitz continuous with respect to y uniformly in t and τ with constant c f and c L respectively, and continuous with respect to τ. 2. g(t, y, ξ) is Lipschitz continuous with respect to t, uniformly in y and ξ, with constant c gg and it is Lipschitz continuous with respect to y uniformly in t and ξ, with constant c g. 3. L, f, and G are bounded. 4. C is continuous and inf ξ C(ξ) = γ > 0. 2
3 5. G is Lipschitz continuous with constant c G. It follows, inter alia, that there exists a unique solution y( ) and a unique J for any measurable τ( ) and any sequence ψ, and thus for any nonanticipative strategy Φ as defined hereafter. 2.2 Strategies Multiple jumps It may be to the best advantage of the minimizer to make a jump at some time t, immediately followed, at the same time, by another jump, and so on. As any such jump entails a cost not less than γ, a near optimal strategy will never attempt to make an infinite number of jumps. As a matter of fact, we shall show that it follows from the boundedness of L and J that the number of jumps may be restricted, with no loss of generality, to be less than K = 2[(T t 0 ) L + G + 1/2/γ. To allow for the possibility of several successive but simultaneous jumps, we proceed as follows. Let Ξ = K p=1 (R m ) p. We extend g and C from R m to Ξ in the natural way: let ξ (R m ) p be a multiple jump of multiplicity p. Let (ξ 1, ξ 2,..., ξ p ) be its components. For a given y R n, let z 0 = y, and for l {1,..., p}, z l = z l 1 + g(t, z l 1, ξ l ). Then we set g(t, y, ξ) = z p y C(ξ) = p C(ξ l ). There is no point in considering the possibility of several successive multiple jumps at the same jump time, since a sequence of simultaneous multiple jumps is a multiple jump. It easily follows that g so extended is still Lipschitz continuous in t and in y, uniformly in the other two variables, with new coefficients c gg and c g respectively. From now on, when we refer to jumps, it will always be multiple jumps, unless specifically referred to as simple jumps. Simple jumps are the same thing as a multiple jump of multiplicity 1. And of course, in a control ψ = ({t k }, {ξ k }) the ξ k are to be considered as multiple jumps. But the t k s are always assumed to be distinct. We also state the following definition: l=1 3
4 Definition 2.1 For any bounded function V : [t 0, T R n R, let the operator M be given by It enjoys the following properties: MV (t, x) = inf [V (t, x + g(t, x, ξ)) + C(ξ) ξ Ξ Lemma 2.1 The operator M is continuous for the topology. Moreover, if the function V is bounded and Lipschitz continuous, so is the function MV. Proof The first statement is elementary. Clearly, if, for all (t, x), V (t, x) V, then, clearly MV (t, x) V + γ. To prove the regularity of MV, we need to make the following remark, which as akin to the lemma 3.5 below: the range of ξ in the inf ξ operator defining MV may be restricted to jumps of multiplicity no larger than V /γ. As a matter of fact, the simple jumps 2 to K will cost at least (K 1)γ and cannot decrease V by more than 2 V. Therefore, if K 1 > 2 V /γ, for any jump of multiplicity K, say ξ = (ξ 1,..., ξ K ), V (t, x + g(t, x, ξ 1 )) + C(ξ 1 ) < V (t, x + g(t, x, ξ)) + C(ξ). Over this retricted range Ξ, (t, x) g(t, x, ξ) is Lipschitz continuous uniformly in ξ: there exists a constant C g > 0: for all (t 1, x 1 ), (t 2, x 2 ), ξ Ξ, g(t 1, x 1, ξ) g(t 2, x 2, ξ) C g ( t 1 t 2 + x 1 x 2 ). Assume V is Lipschitz continuous, more precisely that (t 1, x 1 ), (t 2, x 2 ), V (t 1, x 1 ) V (t 2, x 2 ) λ( t 1 t 2 + x 1 x 2 ). We need now evaluate = MV (t 1, x 1 ) MV (t 2, x 2 ). Choose an arbitrary ε > 0 and a ξ ε such that Then V (t 2, x 2 + g(t 2, x 2, ξ ε )) + C(ξ ε ) MV (t 2, x 2 ) + ε. MV (t 1, x 1 ) V (t 2, x 2 + g(t 2, x 2, ξ ε )) C(ξ ε ) + ε V (t 1, x 1 + g(t 1, x 1, ξ ε )) + C(ξ ε ) V (t 2, x 2 + g(t 2, x 2, ξ ε )) C(ξ ε ) + ε λ(1 + C g )[ t 1 t 2 + x 1 x 2 + ε. Since, on the one hand ε was arbitrary, and on the other hand, (t 1, x 1 ) and (t 2, x 2 ) play symmetrical roles, we may conclude that MV (t 1, x 1 ) MV (t 2, x 2 ) λ(1 + C g )[ t 1 t 2 + x 1 x 2. 4
5 2.2.2 Admissible strategies We want to investigate the problem of minimizing sup τ(.) Ω J through the impulse control. We mean to allow closed loop strategies for the minimizing control. We remark that, being only interested in the inf sup problem, and not a possible saddle point, there is no loss of generality in restricting τ( ) to open loop controls τ( ) Ω = {measurable functions [t 0, T K} We shall sometimes write τ Ω instead of τ( ) Ω. Let Ψ be the set of all finite sequences ψ. We now define the admissible closed loop strategies Φ for the minimizing impulse control ψ, as nonanticipative strategies. We shall let Π be the set of all such nonanticipative strategies. Definition 2.2 A map Φ : Ω Ψ is called a nonanticipative strategy if for any two controls τ 1 ( ) and τ 2 ( ), and any t [t 0, T, the condition on their restrictions to [t 0, t[: τ 1 [t0,t[ = τ 2 [t0,t[ implies Φ(τ 1 ) [t0,t = Φ(τ 2 ) [t0,t. 3 The value function We define the value function of the problem W : [0, T R n R as W (t 0, x) = inf sup Φ Π τ(.) Ω 3.1 Dynamic Programming Principle J(t 0, x, Φ(τ( )), τ(.)). (2) The solution of the minimax impulse control problem at hand depends on the following basic properties Proposition 3.1 (Isaacs Tenet of transition) For all t t [t 0, T [, for all x R n, t W (t, x) = inf sup L(s, y(s), τ(s))ds + C(ξ k ) + W (t, y(t )) Φ Π τ Ω t t k <t Proof Assume first that for some x, t and t > t, t W (t, x) > inf sup L(s, y(s), τ(s)) ds + C(ξ k ) + W (t, y(t )), Φ Π τ Ω t t k <t 5
6 and let the difference be 3ε. Choose an admissible strategy Φ ε 1 that approximates the infimum in the r.h.s. up to ε. Then, for any τ( ) defined over [t, t, t t L(s, y(s), τ(s)) ds + t k <t C(ξ k ) + W (t, y(t )) W (t, x) 2ε. Whatever y(t ), from t on, choose a non anticipative strategy Φ ε 2 of the game over [t, T that approximates the Value W (t, y(t )) again up to ε. The concatenation Φ ε of Φ ε 1 and Φε 2 is a nonanticipative strategy of the game over [t, T. It yields, for all τ( ), J(t, x, Φ ε, τ( )) W (t, x) ε, a contradiction. Assume to the contrary that t W (t, x) < inf sup L(s, y(s), τ(s)) ds + C(ξ k ) + W (t, y(t )), Φ Π τ Ω t t k <t and let the difference be 2ε. Choose an admissible strategy Φ ε that approximates W (t, x) up to ε, and denote (t ε k, ξε k ) the jumps it produces. Then, [ t sup L(s, y(s), τ(s)) ds + C(ξk ε ) + J(t, y(t ), Φ ε, τ) τ( ) t t ε k [ <t t inf sup L(s, y(s), τ(s)) ds + C(ξ k ) + W (t, y(t )) ε. Φ τ( ) t t k <t The above inequality can also be written in terms of the restrictions τ and τ of τ( ) to [t, t and [t, T respectively, as [ t sup L(s, y(s), τ (s)) ds + C(ξk ε ) + sup J(t, y(t ), Φ ε, τ ) τ t t ε k <t τ [ t inf sup L(s, y(s), τ (s)) ds + (3) C(ξ k ) + W (t, y(t )) ε. Φ τ t t k <t Observe that the knowledge of τ is useless to evaluate J(t, y(t ), Φ, τ ) once y(t ) is known. Therefore, the restriction of Φ ε to [t, T cannot do better than a nonanticipative strategy of the game over [t, T. As a result, W (t, y(t )) sup τ J(t, y(t ), Φ ε, τ ). But placing that in the l.h.s. of inequality (3) leads to a contradiction. 6
7 Proposition 3.2 For all t [t 0, T and x R n, W (t, x) MW (t, x). Proof Among the admissible Φ s there are those that place a jump at time t. Using the same reasoning as above, minimizing over the jump parameter ξ at time t, one gets the required inequality. Proposition 3.3 For any admissible strategy Φ, let t 1 Φ(τ( )) be the first impulse time after t. Then, for all x R n, [ t1 W (t, x) = inf sup L(s, y(s), τ(s)) ds + MW (t 1, y(t 1 )) Φ τ t Proof Proceed as in proposition 3.1 with t replaced by t 1 as specified by the strategy Φ ε chosen, and notice that the inf over [t 1, T involves a inf ξ Regularity of the value function In this section, we prove the following theorem Theorem 3.4 Under the standing assumptions (paragraph 2.1.2) the value function W is bounded, Lipschitz continuous in x uniformly in t, and Lipschitz continuous in t, uniformly in x. The rest of the section is devoted to the proof of this theorem Boundedness Let us first show that thanks to hypothesis 3, W as defined by (2) is bounded. On the one hand, a particular strategy Φ is the one where we have no impulse time. In this case k C(ξ k) = 0. Since L and G are bounded, it is easy to see that for all t [t 0, T and x R n, W (t, x) (T t 0 ) L + G. On the other hand, since the costs C(ξ k ) are positive numbers and again since L and G are bounded, then again it is easy to see that for all t [t 0, T and x R n, W (t, x) (T t 0 ) L G. We can then say that proving the first result. It follows the important property: W (T t 0 ) L + G 7
8 Lemma 3.5 There exists K N depending only on the data of the problem such that, in the definition of W, the inf Φ Π can be replaced by inf Φ Π K where Π K is the set of all nonanticipative strategies who s sums of the multiplicity of their jumps is less or equal to K. Proof of the lemma Let us choose a strategy Φ such that [ T K W (t, x) sup L(s, y(s), τ) ds + C( ξ k ) + G(y(T )) 1 τ Since then t k=1 [ T (T t) L G sup L(s, y(s), τ) ds + G(y(T )) τ t K C( ξ k ) (T t) L + G + W + 1 k=1 with γ = inf ξ C(ξ), we obtain then 2((T t 0 ) L + G ) + 1 K 1 γ (2((T t 0) L + G ) + 1). This proves that, for the strategies which are close to optimality, the number of impulses, simple or multiple, is bounded, with a bound depending only on the data of the problem Lipschitz continuity in x Let us now prove that the value function W is Lipschitz continuous in x, uniformly in t. To that aim, we estimate the difference = W (t, x 1 ) W (t, x 2 ) Fix an arbitrary ε > 0. Let Φ 2 be a strategy such that Then sup J(t, x 2, Φ 2, τ) inf sup J(t, x 2, Φ, τ) + ε = W (t, x 2 ) + ε. τ Φ τ W (t, x 1 ) sup τ J(t, x 2, Φ 2, τ) + ε sup τ J(t, x 1, Φ 2, τ) sup τ J(t, x 2, Φ 2, τ) + ε. 8
9 Now, pick τ 1 such that Then J(t, x 1, Φ 2, τ 1 ) sup J(t, x 1, Φ 2, τ) ε. τ J(t, x 1, Φ 2, τ 1 ) sup τ J(t, x 2, Φ 2, τ) + 2ε J(t, x 1, Φ 2, τ 1 ) J(t, x 2, Φ 2, τ 1 ) + 2ε. Consider now the control ψ 2 = Φ 2 (τ 1 ), composed of jump instants t 1, t 2...t K in the interval [t, T, with jumps ξ 1, ξ 2...ξ K, and let y 1 ( ) and y 2 ( ) be the trajectories generated by (Φ 2, τ 1 ), or equivalently (ψ 2, τ 1 ), from y i (t) = x i, i = 1, 2. By Gronwall s lemma, we can say for all s [t, t 1, y 1 (s) y 2 (s) exp(c f (s t)) x 1 x 2. Looking more carefully at the first jump and using the Lipschitz continuity of g, we have y 1 (t + 1 ) y 2(t + 1 ) = (y 1 (t 1 )+g(t 1, y 1 (t 1 ), ξ 1)) (y 2 (t 1 )+g(t 1, y 2 (t 1 ), ξ 1)) (1 + c g ) y 1 (t 1 ) y 2(t 1 ) exp(c f (t 1 t))(1 + c g ) x 1 x 2. Repeating inductively the same argument, we have, for an impulse time t k, k {1, 2...K}, Finally y 1 (t + k ) y 2(t + k ) exp(c f (t k t))(1 + c g ) k x 1 x 2. for all s [t, T, y 1 (s) y 2 (s)) exp(c f (s t))(1 + c g ) K x 1 x 2, where we recall that K is the number of impulses in ψ 2. Then, J(t, x 1, Φ 2, τ 1 ) J(t, x 2, Φ 2, τ 1 ) T c L y 1 (s) y 2 (s) ds + c G y 1 (T ) y 2 (T ) t [ T (1 + c g ) K x 1 x 2 c L exp(c f (s t)) ds + c G exp(c f (T t)) t [ 1 (1 + c g ) K x 1 x 2 (exp(c f (T t 0 )) 1) + c G exp(c f (T t 0 )), c f 9
10 and we can say now W (t, x 1 ) W (t, x 2 ) C 1 x 1 x 2 + 2ε and, since ε was arbitrary, W (t, x 1 ) W (t, x 2 ) C 1 x 1 x 2. where the constant C 1 = (1 + c g ) K [ 1 (exp(c f (T t 0 )) 1) + c G exp(c f (T t 0 )), c f with K the maximum number of impulses according to the lemma. The roles of x 1, x 2 being symmetrical, we finally have W (t, x 1 ) W (t, x 2 ) C 1 x 1 x Lipschitz continuity in t Let us examine now the difference W (s 1, x) W (s 2, x). Assume first that s 2 > s 1. By the dynamic programming principle, W (s 1, x) [ W (s 2, x) = s2 inf sup L(s, y(s), τ(s)) ds + C(ξ k ) + W (s 2, y(s 2 )) Φ τ s 1 t k <s 2 W (s 2, y(s 1 )) We choose the strategy Φ where there is no impulse time between s 1 and s 2, then W (s 1, x) W (s 2, x) [ s2 sup L(s, y(s), τ(s)) ds + W (s 2, y(s 2 )) W (s 2, y(s 1 )) τ s 1 L s 2 s 1 + C 1 f s 2 s 1 c 2 s 2 s 1 with c 2 = L + C 1 f Assume now that s 1 > s 2. We denote by y 2 ( ) a trajectory generated by the controls considered from y(s 2 ) = x. By the dynamic programming principle, we have W (s 1, x) W (s 2, x) = s1 W (s 1, x) inf sup L(s, y 2 (s), τ(s)) ds + C(ξ k ) + W (s 1, y 2 (s 1 )). Φ τ s 2 k t k <s 1 10
11 Choose an arbirary ε > 0, and a strategy Φ ε, such that the inf Φ above is reached within ε. We have W (s 1, x) W (s 2, x) s1 W (s 1, x) sup L(s, y 2 (s), τ(s)) ds + C(ξ k ) + W (s 1, y 2 (s 1 )) + ε s 2 k t k <s 1 W (s 1, x) τ s1 s 2 L(s, y 2 (s), τ(s)) ds k t k <s 1 C(ξ k ) W (s 1, y 2 (s 1 )) + ε for any given τ( ). Let us, from now on fix a τ( ), and let Φ ε (τ( )) exhibit K jumps (ξ 1,..., ξ K ) in the interval [s 2, s 1 [. And let y 2 ( ) be the trajectory generated by these controls from y 2 (s 2 ) = x. We now bound W (s 1, x) using this sequence of jumps and Proposition 3.2. First, consider the effect of the multiple jump ξ = (ξ 1,..., ξ K ) at time s 1 on x. Let therefore z K = x + g(s 1, x, ξ), constructed as in the definition of multiple jumps, via z 0 = x and z k = z k 1 + g(s 1, z k 1, ξ k ). 1 Using Proposition 3.2, we have K W (s 1, x) W (s 1, z K ) + C(ξ k ) Therefore W (s 1, x) W (s 2, x) W (s 1, z K ) s1 s 2 k=1 L(s, y 2 (s), τ(s)) ds W (s 1, y 2 (s 1 )) + ε L s 1 s 2 + W (s 1, z K ) W (s 1, y 2 (s 1 )) + ε. We can write, for any τ( ), W (s 1, z K ) W (s 1, y 2 (s 1 )) = W (s 1, z K ) W (s 1, y 2 (t + K )) + W (s 1, y 2 (t + K )) W (s 1, y 2 (s 1 )) C 1 z K y 2 (t + K ) + C 1 f s 1 s 2. Now, for k {1,.., K}, we examine the difference k = y 2 (t + k ) z k. k+1 = y 2 (t k+1 ) + g(t k+1, y 2 (t k+1 ), ξ k+1) z k g(s 1, z k, ξ k+1 ) (1 + c g ) k + (1 + c g ) f t k+1 t k + c gg t k+1 s 1. 1 as opposed to the construction in subsection 2.2.1, here each ξ k may be a multiple jump itself. 11
12 We introduce δ k = (1 + c g ) f t k+1 t k + c gg t k+1 s 1, then which can be written as k k+1 (1 + c g ) k+1 k (1 + c g ) k + δ k (1 + c g ) k+1 (4) u k+1 u k + δ k (1 + c g ) k+1 with u k =. The strategy considered here is with K impulse times where (1 + c g ) k K is bounded. So, k=k 1 δ k u K u 1 + (1 + c g ) k+1 On the one hand, for any k {1,..K} k=1 δ k ((1 + c g ) f + c gg ) s 1 s 2 = δ = C 2 s 1 s 2. On the other hand, u 1 = c g, with Finally, and hence Therefore, where 1 = y 2 (t 1 ) + g(t 1, y 2 (t 1 ), ξ 1) y 2 (s 2 ) g(s 1, y 2 (s 2 ), ξ 1 ) δ We let ε 0 to obtain K 1 1 u K δ (1 + c g ) k+1 Kδ Kδ, k=0 K (1 + c g ) K Kδ (1 + c g ) K Kδ. W (s 1, x) W (s 2, x) C 3 s 1 s 2 + ε C 3 = L + C 1 [ f + C 2 (1 + c g ) K K. W (s 1, x) W (s 2, x) C 3 s 1 s 2. Since the inequality is symmetric in s 1 and s 2, it follows that W (s 1, x) W (s 2, x) C 3 s 1 s 2. We then proved that W is Lipschitz continuous with respect to time t, uniformly in x. 12
13 3.3 Terminal value Because of the possible jumps at the terminal time T, it is easy to see that, in general, W (t, x) does not tend to G(x) as t tends to T. Extend the set of multiple jumps to include jumps of zero multiplicity, meaning no jump. Call this extended set Ξ 0, extend trivially the operator M to a function independant from t, and let G(x) = inf ξ Ξ 0 [G(x + g(t, x, ξ)) + C(ξ) = min{g(x), MG(T, x)}. (5) We know from lemma 2.1 that G(x) is Lipschitz continuous. We claim Lemma 3.6 W (t, x) G(x) as t T. Proof Fix (t, x) and a strategy Φ. As in the previous proof, for each τ( ), gather all jumps of ψ = Φ(τ), if any, in a multiple jump ξ at the time T, and let z = x + g(t, x, ξ). The same argument as previously shows that there exists a constant C T such that, for all Φ and τ or J(t, x, Φ, τ) [G(z) + C( ξ) C T (T t). J(t, x, Φ, τ) = G(z) + C( ξ) + O(T t). The right hand side above only depends on ξ, not on τ( ) itself. It follows that inf sup J(t, x, Φ, τ) = inf (G(z) + C(ξ)) + O(T t) = G(x) + O(T t). φ ξ Ξ 0 τ The result follows letting t T. Remark 3.1 Because of the uniform convergence of W (t, x) to G(x) as t T, using the lemma 2.1 we have MW (t, x) M G(T, x) as t T, uniformly in R n, and using proposition 3.2, G(x) M G(T, x). 4 Isaacs quasi-variational inequality In this section we proove that the value function W is a viscosity solution of the Hamilton-Jacobi-Isaacs quasi-variational inequality, that we replace by an equivalent QVI easier to investigate. 13
14 4.1 Natural Isaacs Quasi-Variational Inequality In the domain [t 0, T [ R n, we consider the QVI { [ max min W τ K t W f(t, x, τ) L(t, x, τ), x } W (t, x) MW (t, x) = 0. (6) with the terminal condition: W (T, x) = G(x) in R n, where G is given by (5). Notice that it follows from hypothesis 1 that the term in square brackets in equation (6) above is continuous with respect to τ so that the minimum in τ over the compact K exists. Theorem 4.1 The function: (t, x) W (t, x) is a viscosity solution of the quasivariational inequality (6). Proof The proof will be in two parts (i) Let φ C 1 ([t 0, T [ R n ) and let ( t, x) be a local maximum for W φ. We have to prove that ( t, x) satisfies { [ max min φ τ t φ x f( t, x, τ) L( t, x, τ), } W ( t, x) inf [W ( t, x + g( t, x, ξ)) + C(ξ) 0. ξ (ii) Let φ C 1 ([t 0, T [ R n ) and let (t, x) be a local minimum for W φ. We have to prove that (t, x) satisfies { [ max min φ τ t φ f(t, x, τ) L(t, x, τ), x } W (t, x) inf [W (t, x + g(t, x, ξ)) + C(ξ) 0. ξ Let us prove (i). By Proposition 3.2, we have W ( t, x) inf ξ [W ( t, x + g( t, x, ξ)) + C(ξ) 0. Next for t > t, t close to t, we choose a strategy Φ where we have no jumps between t and t. By Proposition 3.1, we have [ t W ( t, x) sup τ( ) t L(s, y(s), τ(s)) ds + W (t, y(t )) 14.
15 By hypothesis, there exists a neighborhood N of ( t, x) such that W φ is maximum at ( t, x) in that neighborhood. For small enough t t > 0, since f is bounded, so is ẏ and then we can conclude that (t, y(t )) N. Hence we get thus This yields If 0 inf τ( ) 0 sup τ( ) [ t W (t, y(t )) W ( t, x) φ(t, y(t )) φ( t, x), t [ t L(s, y(s), τ(s)) ds + φ(t, y(t )) φ( t, x). t ( φ s φ y ) f(s, y(s), τ(s)) L(s, y(s), τ(s)) ds. (7) [ min φ τ t ( t, x) φ y f( t, x, τ) L( t, x, τ) > 0, then, this inequality is also true in a neighborhood of ( t, x), for any choice of τ K. Then t t ( φ s φ y ) f(s, y(s), τ(s)) L(s, y(s), τ(s)) ds > 0, which is in contradiction with inequality (7). Then, we conclude [ min φ τ t ( t, x) φ y f( t, x, τ) L( t, x, τ) 0. Let us now prove (ii). Let (t, x) be a local minimum for W φ. We have, by Proposition 3.2, If W (t, x) inf ξ W (t, x) inf ξ then we are done. Otherwise W (t, x) inf ξ But we have, by Proposition 3.3, W (t, x) = inf Φ sup τ [ t1 t [W (t, x + g(t, x, ξ)) + C(ξ) 0. [W (t, x + g(t, x, ξ)) + C(ξ) = 0, [W (t, x + g(t, x, ξ)) + C(ξ) < 0. (8) L(s, y(s), τ(s)) ds + MW (t 1, y(t 1 )), 15
16 t 1 being the first impulse time after t. For any ε > 0, there exists a strategy Φ ε where t ε 1 is the first impulse time after t, such that [ t ε 1 W (t, x) sup L(s, y(s), τ(s)) ds + MW (t ε 1, y(t ε 1)) ε. τ t We prove in the sequel that there exists δ > 0 such that t ε 1 t+δ for all ε. Assume to the contrary that there exists a sequence t εn 1 that converges towards t when ε n converges towards 0. Let η > 0 be any fixed real number. Since L is bounded, for n large enough, it holds that, for any τ( ), t εn 1 L(s, y(s), τ(s)) ds L t εn 1 t η. t Hence W (t, x) MW (t εn 1, y(tεn 1 )) ε n η. Now we choose ξ n such that MW (t εn 1, y(tεn 1 )) W (tεn 1, y(tεn 1 ) + g(tn 1, y(t εn 1 ), ξ n)) + C(ξ n ) η. Using the continuity of W and of g, it follows that, again for large enough n, W (t εn 1, y(tεn 1 ) + g(tn 1, y(t εn 1 ), ξ n)) W (t, x + g(t, x, ξ n )) η. Finally, we also have W (t, x + g(t, x, ξ n )) + C(ξ n ) MW (t, x). Combining the last four inequalities, we get, in the limit in n W (t, x) MW (t, x) 3η. This being true for any positive η, we conclude W (t, x) MW (t, x). This inequality is in contradiction with inequality (8). This means that t ε 1 t + δ. Now we choose t t, t + δ[. By Proposition 3.1, we have [ t W (t, x) = inf L(s, y(s), τ(s)) ds + W (t, y(t )) Φ τ Ω t [ t = sup L(s, y(s), τ(s)) ds + W (t, y(t )) τ Ω t t t L(s, y(s), τ) ds + W (t, y(t )), 16
17 for any constant τ K in the last line above. Since (t, x) is a local minimum of W φ, then for t close enough to t, for the same reasons than in (i), (t, y(t )) is in the neighborhood of (t, x), then we have that Then 0 0 t t t t L(s, y(s), τ) ds + φ(t, y(t )) φ(t, y), ( L(s, y(s), τ) + φ s + φ y Dividing by t t and letting t tend to t, we obtain ) f(s, y(s), τ) ds. φ t φ f(t, x, τ) L(t, x, τ) 0. y and then [ min φ τ t φ y f(t, x, τ) L(t, x, τ) An Equivalent Quasi-Variational Inequality In this section, we consider the new function Γ given by the classical change of variable Γ(t, x) = exp(t)w (t, x), for any t [t 0, T and x R n. Of course, the function Γ is bounded and uniformly continuous with respect to its arguments. A second property is given by the Proposition 4.2 W is a viscosity solution of (6) if and only if Γ is a viscosity solution to the following quasi-variational inequality in [t 0, T [ R n, { [ 0 = max min Γ Γ + Γ(t, x) f(t, x, τ) exp(t)l(t, x, τ), τ t } x (9) Γ(t, x) N Γ(t, x) where N Γ(t, x) = inf(γ(t, x + g(t, x, ξ)) + exp(t)c(ξ)). The terminal condition ξ for Γ is: Γ(T, x) = exp(t ) G(x) in R n. Proof It suffices to prove that if W is a viscosity solution to (6), then Γ is a viscosity solution to (9). The relation between W and Γ is symmetric and the proof the other way around would be the same. The proof will be in two parts 17
18 (i) Let φ C 1 ([t 0, T [ R n ) and let ( t, x) be a local maximum for Γ φ. We have to prove that ( t, x) satisfies { [ max min φ τ t ( t, x) + Γ( t, x) φ x ( t, x)f( t, x, τ) e t L( t, x, τ), } Γ( t, x) N Γ( t, x) 0 (ii) Let φ C 1 ([t 0, T [ R n ) and let (t, x) be a local minimum for Γ φ. We have to prove that (t, x) satisfies { [ max min φ φ (t, x) + Γ(t, x) τ t x (t, x)f(t, x, τ) et L(t, x, τ), } Γ(t, x) N Γ(t, x) 0 Let us prove (i). For (t, x) in the neighborhood of ( t, x), we have then M 1 = (Γ φ)( t, x) (Γ φ)(t, x) = exp(t)w (t, x) φ(t, x) 0 (W Φ)(t, x) where Φ(t, x) = exp( t)(φ(t, x) + M 1 ). Remark that ( t, x) is also a local maximum to W Φ. But since W is a viscosity solution to (6), then { [ max min exp( t)(φ( t, x) + M 1 ) exp( t) φ τ t ( t, x) exp( t) φ x ( t, x)f( t, x, τ) L( t, x, τ), } exp( t)γ( t, x) inf (exp( t)γ(t, x + g(t, x, ξ)) + C(ξ)) 0 ξ and then max { [ min Γ( t, x) φ τ t ( t, x) φ x ( t, x)f( t, x, τ) exp( t)l( t, x, τ), } Γ( t, x) inf ξ Γ(t, x + g(t, x, ξ)) + exp( t)c(ξ)) This means that Γ is a viscosity subsolution to (9). The idea of the proof is the same for point (ii). We then proved that Γ is a viscosity solution to (9). 0 18
19 5 Uniqueness We are now going to use the equivalent QVI to prove uniqueness of the viscosity solution. To do so, we introduce a variational inequality which we first study. 5.1 General Obstacle Problem In this section, we are interested by the following variational inequality where the second term under the max in the quasi-variational inequality (9) (Γ N Γ)(t, x) is replaced by the more general term (Γ P )(t, x) where P is in BUC([t 0, T [ R n ): for all (t, x) [t 0, T [ R n, 0 = max { [ min Γ τ t with the terminal condition: Γ + Γ(t, x) } x (Γ P )(t, x), f(t, x, τ) exp(t)l(t, x, τ), (10) x R n, Γ(T, x) = exp(t ) G(x). (11) For compatibility reasons, in order to have a solution for such obstacle problems, one has to assume that the function P satisfies P (T, x) exp(t ) G(x); by Remark 3.1, this is indeed the case for P := N Γ(t, x). In all the obstacle problems we consider below, we assume implicitly that such condition is fulfilled. Let Γ 1 and Γ 2 be respectively viscosity solutions of (10) in BUC, where L is equal to L 1, P is equal to P 1 and G is equal to G 1, respectively L is equal to L 2, P is equal to P 2 and G is equal to G 2, that is, { [ max min Γ 1 τ t + Γ 1(t, x) Γ 1 x f(t, x, τ) exp(t)l 1(t, x, τ), } (Γ 1 P 1 )(t, x) = 0, with the terminal condition: for all x R n, Γ 1 (T, x) = exp(t )G 1 (x), and { [ max min Γ 2 τ t + Γ 2(t, x) Γ 2 x f(t, x, τ) exp(t)l 2(t, x, τ), } (Γ 2 P 2 )(t, x) = 0, with the terminal condition: for all x R n, Γ 2 (T, x) = exp(t )G 2 (x). 19
20 Theorem 5.1 Assume that both Γ 1 and Γ 2 are bounded and uniformly continuous with respect to their arguments. If f, L 2, and G satisfy the assumptions 1, 3, and 5 then, (Γ 1 Γ 2 ) + } max {e T (L 1 L 2 ) +, (P 1 P 2 ) +, e T (G 1 G 2 ) +. (12) The end of this section is devoted to the proof of that theorem. We build the following test-function: let t, s [t 0, T and x, z R n, φ α,β,γ (t, x, s, z) = Γ 1 (t, x) Γ 2 (s, z) α( x 2 + z 2 ) x z 2 t s 2 β 2 γ 2. Let ( t, x, s, z) be a maximum point of φ α,β,γ which exists, since this is a continuous function going to infinity when x or z does, and s and t range over a compact set. We need the following lemma for the proof of Theorem (5.1). Lemma 5.2 Assume that M = sup [t0,t R n(γ 1 Γ 2 )(t, x) > 0, let ( t, x, s, z) be a maximal point of φ α,β,γ, and M α,β,γ = φ α,β,γ ( t, x, s, z). Then, for any ε > 0, there exists α 0, β 0 and γ 0 such that, for any α α 0, β β 0, γ γ 0, α( x 2 + z 2 ) + x z 2 β 2 + t s 2 γ 2 ε. (13) Proof By definition, ( t, x, s, z) satisfies, for all t, s [t 0, T, x, z R n Γ 1 ( t, x) Γ 2 ( s, z) α( x 2 + z 2 x z 2 ) β 2 t s 2 γ 2 Γ 1 (t, x) Γ 2 (s, z) α( x 2 + z 2 x z 2 t s 2 ) β 2 γ 2. Let us take t = s and x = z, then M α,β,γ Γ 1 (t, x) Γ 2 (t, x) 2α x 2. Let (t, x ) be a point where the sup(γ 1 Γ 2 ) is reached within δ, then Γ 1 (t, x ) Γ 2 (t, x ) M δ, where δ > 0 can be chosen arbitrarily small. We choose it and α such that M δ 2α x 2 > 0, which is always possible because x depends only on δ. We have in particular that M α,β,γ Γ 1 (t, x ) Γ 2 (t, x ) 2α x 2 M δ 2α x 2 > 0. (14) 20
21 Let r 2 = Γ 1 + Γ 2, then 0 < M δ 2α x 2 M α,β,γ r 2 α( x 2 + z 2 ) Thus, x z 2 β 2 t s 2 γ 2 x z rβ and t s rγ. (15) The function Γ 2 is assumed uniformly continuous in t and also in x. Let us therefore introduce the following functions: m(u) = n(v) = sup Γ 2 (t, x) Γ 2 (t, z), t, x z u Γ 2 (t, x) Γ 2 (s, x). sup t s v,x (16) Clearly, m and n decrease to 0 with their arguments. With these notations, and using (15), we have Γ 1 ( t, x) Γ 2 ( s, z) = Γ 1 ( t, x) Γ 2 ( t, x) + Γ 2 ( t, x) Γ 2 ( t, z) + Γ 2 ( t, z) Γ 2 ( s, z) M + m(rβ) + n(rγ). Place this in the definition of M α,β,γ and use (14): M δ 2α x 2 M α,β,γ or equivalently M + m(rβ) + n(rγ) α( x 2 + z 2 ) α( x 2 + z 2 ) + x z 2 β 2 t s 2 γ 2 x z 2 β 2 + t s 2 γ 2 δ + 2α x 2 + m(rβ) + n(rγ). Pick ε < 5M/3, take δ = ε/5, α 0 = ε/(5 x 2 ) if x 0, α 0 = 1 if x = 0, m(rβ 0 ) = n(rγ 0 ) = ε/5 to get (13). Let us now give the proof of Theorem (5.1). Let ( t, x, s, z) be a maximum point of φ α,β,γ. Case t and s different from T Assume first that both t and s are different from T. Then, for all x R n, t [t 0, T, Γ 1 ( t, x) Γ 2 ( s, z) α( x 2 + z 2 x z 2 ) β 2 t s 2 γ 2 Γ 1 (t, x) Γ 2 ( s, z) α( x 2 + z 2 x z 2 t s 2 ) β 2 γ 2. 21
22 Let φ 1 C 1 ([t 0, T [ R n ) be defined as φ 1 (t, x) = Γ 2 ( s, z) + α( x 2 + z 2 ) + x z 2 β 2 + t s 2 γ 2. This last inequality means that ( t, x) is a maximal point of Γ 1 (t, x) φ 1 (t, x). We also have, for all z R n, s [t 0, T Γ 1 ( t, x) Γ 2 ( s, z) α( x 2 + z 2 x z 2 ) β 2 t s 2 γ 2 Γ 1 ( t, x) Γ 2 (s, z) α( x 2 + z 2 x z 2 ) β 2 t s 2 γ 2. Let also φ 2 C 1 ([t 0, T [ R n ) be defined as φ 2 (s, z) = Γ 1 ( t, x) α( x 2 + z 2 ) x z 2 β 2 t s 2 γ 2. This inequality means that ( s, z) is a minimal point of Γ 2 (s, z) φ 2 (s, z). Then, { [ max min φ 1 τ t ( t, x) + Γ 1 ( t, x) φ 1 x ( t, x)f( t, x, τ) e t L 1 ( t, x, τ), } (17) Γ 1 ( t, x) P 1 ( t, x) 0, and { [ max min φ 2 τ s ( s, z) + Γ 2( s, z) φ 2 z ( s, z)f( s, z, τ) e s L 2 ( s, z, τ), } Γ 2 ( s, z) P 2 ( s, z) 0. (18) Inequality (17) can be decomposed into two inequalities (19), where we have set p β = 2( x z)/β 2, and (20): [ 2( t s) min τ γ 2 + Γ 1 ( t, x) p β + 2α x, f( t, x, τ) e t L 1 ( t, x, τ) 0, (19) Γ 1 ( t, x) P 1 ( t, x) 0, (20) Inequality (18) is equivalent to either inequality (21), or inequality (22): [ 2( t s) min τ γ 2 + Γ 2 ( s, z) p β 2α z, f( s, z, τ) e s L 2 ( s, z, τ) 0, (21) 22
23 Γ 2 ( s, z) P 2 ( s, z) 0. (22) In (21), min τ may be replaced by for all τ. Whenever it holds, [ Γ 1 ( t, x) Γ 2 ( s, z) sup p β, f( t, x, τ) f( s, z, τ) τ +2α x, f( t, x, τ) + 2α z, f( s, z, τ) +e t L 1 ( t, x, τ) e s L 2 ( s, z, τ) where sup[a + B + D + E + F + G τ A = p β, f( t, x, τ) f( t, z, τ), B = p β, f( t, z, τ) f( s, z, τ), D = 2α f ( x + z ), E = e t (L 1 ( t, x, τ) L 2 ( t, x, τ)), F = e t (L 2 ( t, x, τ) L 2 ( s, z, τ)), G = (e t e s )L 2 ( s, z, τ). It follows from the proof of Lemma 5.2 that, for α α 0, t s γ ε and x z β ε. L 2 is continuous in t, uniformly in (x, τ), and Lipschitz continuous in x with constant c L uniform in (t, τ). Choose γ 1 γ 0 such that, for t s γ 1 ε, e T L 2 (t, x, τ) L 2 (s, x, τ) ε for all α α 0, β β 0, and γ γ 1, and β 1 = min{β 0, e T ε/c L }. This insures that, for all α α 0, β β 1, γ γ 1, F 2ε. Finally, pick γ 2 = min{γ 1, e T ε/ L 2 }, so that for γ γ 2, we furthermore insure that G ε. We have A 2c f x z 2 /β 2 2c f ε for all α α 0, β β 0 and γ γ 0 and a fortiori if β β 1 and γ γ 2. Let us turn to B. Clearly, B p β f( t, z, τ) f( s, z, τ). Fix β = β 1. Then, for small α and γ, p β 2 ε/β 1. Since f is continuous in t, uniformly in (x, τ), using the bound t s γ ε, it is possible to find γ 3 γ 2 such that, for γ γ 3, f( t, z, τ) f( s, z, τ) εβ 1, hence B 2ε. Finally, it follows also from the lemma 5.2 that, for α, β, and γ as in the lemma, α x 2 ε, i.e. α x αε, and similarily for α z. Hence, keeping β = β 1, γ = γ 3, we may choose α 1 = min{α 0, ε f 2 } to insure that D 4ε. 23
24 Hence in the case under investigation, there exists a triple (α 1, β 2, γ 3 ) such that for this triple, Γ 1 ( t, x) Γ 2 ( s, z) e t sup[l 1 ( t, x, τ) L 2 ( t, x, τ) + aε τ e T (L 1 L 2 ) + + aε. (23) where a := 9 + 2c f does not depend on ε nor (α, β, γ). Let us assume now that (22) happened. Substracting (22) from (20), we obtain Γ 1 ( t, x) Γ 2 ( s, z) P 1 ( t, x) P 2 ( s, z). We use the bounds on t s and x z arising from the lemma and the uniform continuity of P 1 (or of P 2 ) to conclude that (α, β, γ) can be chosen small enough to insure Γ 1 ( t, x) Γ 2 ( s, z) (P 1 P 2 ) + + aε. In every cases, if t and s T, we have either (23) or the above inequality, thus, there always exist (α, β, γ) such that } Γ 1 ( t, x) Γ 2 ( s, z) max {e T (L 1 L 2 ) +, (P 1 P 2 ) + + aε. (24) Case t or s equal to T Now, we shall examine the case where t or s are equal to T. Then, (10) is not available. Assume that t = T. Provided that α α 0, β β 0, and γ γ 0, we may rewrite the definition (16) using the bounds (15) on t s and x z as Γ 2 (T, x) Γ 2 ( s, z) m(rβ) + n(rγ). From the boundary condition (11), we get Hence we obtain Γ 1 (T, x) = e T G 1 ( x) and Γ 2 (T, x) = e T G 2 ( x). Γ 1 (T, x) Γ 2 ( s, z) exp(t )(G 1 ( x) G 2 ( x)) + m(rβ) + n(rγ). The same reasoning holds for the case where s = T. Therefore, whenever t or s = T, we may again choose α α 0, β β 0, and γ γ 0 such that Γ 1 ( t, x) Γ 2 ( s, z) e T (G 1 G 2 ) + + aε. (25) And clearly, this also holds if both s = t = T, since then, using the continuity of G 2 and the bound (15), again if α α 0, β β 0, and γ γ 0, Γ 1 (T, x) Γ 2 (T, z) = e T (G 1 ( x) G 2 ( z)) e T (G 1 ( x) G 2 ( x)) + e T c G rβ, which is less than e T (G 1 G 2 ) + +aε provided that β min{β 0, aε/(e T c G r)}. 24
25 Synthesis Using (25) if either t = T or s = T (or both), and (24) otherwise, we may conclude that there always exist (α, β, γ) such that Γ 1 ( t, x) Γ 2 ( s, z) } max {e T (L 1 L 2 ) +, (P 1 P 2 ) +, e T (G 1 G 2 ) + + aε. (26) Finally, it follows from (14) that, for all (α, β, γ) used to get (26), M ε M δ 2α x 2 M α,β,γ Γ 1 ( t, x) Γ 2 ( s, z). Hence, for all (t, x) [t 0, T R n, using (26) we have Γ 1 (t, x) Γ 2 (t, x) M } max {e T (L 1 L 2 ) +, (P 1 P 2 ) +, e T (G 1 G 2 ) + + (a + 1)ε. And as ε was arbitrary, it follows (12), proving theorem Uniqueness of the Viscosity Solution of the New Quasi-variational Inequality In this section, we give the uniqueness result for the viscosity solution of the Hamilton-Jacobi-Isaacs quasi-variational inequality (9). Theorem 5.3 The quasi-variational inequality (9) has a unique bounded uniformly continuous viscosity solution. Proof Assume that (9) has two bounded uniformly continuous viscosity solutions Γ 1 and Γ 2. Let us first remark that for 0 < µ < 1, Θ = µγ 1 is a viscosity solution to the following quasi-variational inequality { [ 0 = max min Θ Θ + Θ(t, x) τ t x f(t, x, τ) et µl(t, x, τ), } (27) Θ(t, x) KΘ(t, x), where KΘ(t, x) = inf ξ (Θ(t, x + g(t, x, ξ)) + µe t C(ξ)), with the terminal condition: Θ(T, x) = µ exp(t ) G(x) in R n. We then apply the result (12) obtained in the previous section, with Θ = µγ 1 is the viscosity solution to the variational inequality (27) where the obstacle P 1 = KΘ and Γ 2 is a viscosity solution of (9) with the obstacle P 2 = N Γ 2. So, we have (Θ Γ 2 ) + max ( e T ((µ 1)L) +, KΘ N Γ 2 ) +, e T ((µ 1) G) + ) (28) 25
26 We have that, for any (t, x) [t 0, T R n (KΘ N Γ 2 )(t, x) sup(µγ 1 (t, x + g(t, x, ξ)) Γ 2 (t, x + g(t, x, ξ)) + ξ sup((µ 1) exp(t)c(ξ)) ξ We recall that C(ξ) inf ξ Hence C(ξ) = γ > 0: since µ < 1, this yields sup((µ 1) exp(t)c(ξ)) = (µ 1) exp(t)γ < γ < 0. ξ (KΘ N Γ 2 )(t, x) sup(µγ 1 (t, x + g(t, x, ξ)) Γ 2 (t, x + g(t, x, ξ))) + γ ξ and then (KΘ N Γ 2 )) + < (µγ 1 Γ 2 ) +. Equation (28) and the last inequality together imply (µγ 1 Γ 2 ) + max (e T ((µ 1)L) +, e T ((µ 1) G) ) +. Now, let µ 1, then since L and G are bounded, we finally obtain (Γ 1 Γ 2 ) + 0. This clearly implies that, for any (t, x) [t 0, T R n, Γ 1 (t, x) Γ 2 (t, x) 0. If Γ 1, Γ 2 are solutions of (9), we can exchange their role and obtain as well Γ 2 (t, x) Γ 1 (t, x), for all (t, x) [t 0, T R n. Finally Γ 1 (t, x) = Γ 2 (t, x) for any (t, x) [t 0, T R n. 6 Conclusion We can now give the result of this paper. Corollary 6.1 Under the assumptions of paragraph 2.1.2, the value function W is the unique bounded and uniformly continuous viscosity solution of the quasivariational inequality (6). 26
27 As an example of a use of this resut, one may consider the option pricing problem of references [4, 5. If the piecewise linear transaction costs are replaced by a more realistic piecewise affine cost, i.e. a fixed cost is charged for any transaction in addition to a variable part, then the problem at hand is exactly that considered here. This was actually the motivation for the present analysis. The problem with no fixed cost, investigated by other means in these references, leads to a more difficult problem in terms of uniqueness of the viscosity solution, since it corresponds to the case γ = 0 in this paper. As far as we know, the uniqueness of the bounded uniformly continuous viscosity solution in that case is still an open problem. References [1 Bardi M, Capuzzo-Dolcetta I (1997) Optimal and viscosity solutions of Hamilton-Jacobi-Bellman equations. Birkhaüser. Boston, Basel, Berlin [2 Barles G (1994) Solutions de viscosité des équations de Hamilton-Jacobi. Springer-Verlag, Mathématiques & Applications. Berlin, Heidelberg, New York [3 Barles G (1985) Deterministic impulse control problems. SIAM Journal on Control and Optimization 23: [4 Bernhard P (2005) A robust control approach to option pricing including transaction costs. Annals of the ISDG 7: Birkhaüser [5 Bernhard P, El Farouq N, Thiery S (2006) An Impulsive Differential Game Arising in Finance with Interesting Singularities. Annals of the ISDG 8: Birkhaüser [6 Crandall MG, Lions PL (1983) Viscosity solutions of Hamilton Jacobi equations. Transactions of the American Mathematical Society 177:1 42 [7 Dharmatti S, Shaiju AJ (2007) Infinite dimensional differential games with hybrid controls. Proceedings of Indian Academy of Sciences, Mathematics 117: [8 Dharmatti S, Ramaswamy M (2006) Zero-sum differential games involving hybrid controls. Journal of Optimization Theory and Applications 128: [9 Evans LC, Souganidis PE (1984) Differential games and representation formulas for the solution of Hamilton-Jacobi-Isaacs equations. Indiana University Journal of Mathematics 33:
28 [10 Fleming WH (1964) The convergence problem for differential games, 2. Annals of Mathematical Study 52: [11 Lions PL (1982) Generalized solutions of Hamilton-Jacobi equations. Pitman, Boston [12 Lions PL, Souganidis PE (1985) Differential games, optimal control and directional derivatives of viscosity solutions of Bellman s and Isaacs Equations. SIAM Journal on Control and Optimization 23: [13 Shaiju AJ and S. Dharmatti S (2005) Differential games with continuous, switching and impulse controls. Nonlinear Analysis 63:23 41 [14 Souganidis PE (1985) Max-min representations and product formulas for the viscosity solutions of Hamilton-Jacobi equations with applications to differential games. Nonlinear Analysis, Theory, Methods and Applications 9: [15 Yong JM (1994) Zero-sum differential games involving impulse controls. Applied Mathematics and Optimization 29:
Differential games withcontinuous, switching and impulse controls
Differential games withcontinuous, switching and impulse controls A.J. Shaiju a,1, Sheetal Dharmatti b,,2 a TIFR Centre, IISc Campus, Bangalore 5612, India b Department of Mathematics, Indian Institute
More informationProf. Erhan Bayraktar (University of Michigan)
September 17, 2012 KAP 414 2:15 PM- 3:15 PM Prof. (University of Michigan) Abstract: We consider a zero-sum stochastic differential controller-and-stopper game in which the state process is a controlled
More informationDifferential Games II. Marc Quincampoix Université de Bretagne Occidentale ( Brest-France) SADCO, London, September 2011
Differential Games II Marc Quincampoix Université de Bretagne Occidentale ( Brest-France) SADCO, London, September 2011 Contents 1. I Introduction: A Pursuit Game and Isaacs Theory 2. II Strategies 3.
More informationChain differentials with an application to the mathematical fear operator
Chain differentials with an application to the mathematical fear operator Pierre Bernhard I3S, University of Nice Sophia Antipolis and CNRS, ESSI, B.P. 145, 06903 Sophia Antipolis cedex, France January
More informationOn the Bellman equation for control problems with exit times and unbounded cost functionals 1
On the Bellman equation for control problems with exit times and unbounded cost functionals 1 Michael Malisoff Department of Mathematics, Hill Center-Busch Campus Rutgers University, 11 Frelinghuysen Road
More informationNonlinear Control Systems
Nonlinear Control Systems António Pedro Aguiar pedro@isr.ist.utl.pt 3. Fundamental properties IST-DEEC PhD Course http://users.isr.ist.utl.pt/%7epedro/ncs2012/ 2012 1 Example Consider the system ẋ = f
More informationOn the Multi-Dimensional Controller and Stopper Games
On the Multi-Dimensional Controller and Stopper Games Joint work with Yu-Jui Huang University of Michigan, Ann Arbor June 7, 2012 Outline Introduction 1 Introduction 2 3 4 5 Consider a zero-sum controller-and-stopper
More informationAn introduction to Mathematical Theory of Control
An introduction to Mathematical Theory of Control Vasile Staicu University of Aveiro UNICA, May 2018 Vasile Staicu (University of Aveiro) An introduction to Mathematical Theory of Control UNICA, May 2018
More informationA deterministic approach to the Skorokhod problem
Control and Cybernetics vol. 35 (26) No. 4 A deterministic approach to the Skorokhod problem by Piernicola Bettiol SISSA/ISAS via Beirut, 2-4 - 3413 Trieste, Italy e-mail: bettiol@sissa.it Abstract: We
More informationMin-Max Certainty Equivalence Principle and Differential Games
Min-Max Certainty Equivalence Principle and Differential Games Pierre Bernhard and Alain Rapaport INRIA Sophia-Antipolis August 1994 Abstract This paper presents a version of the Certainty Equivalence
More informationSébastien Chaumont a a Institut Élie Cartan, Université Henri Poincaré Nancy I, B. P. 239, Vandoeuvre-lès-Nancy Cedex, France. 1.
A strong comparison result for viscosity solutions to Hamilton-Jacobi-Bellman equations with Dirichlet condition on a non-smooth boundary and application to parabolic problems Sébastien Chaumont a a Institut
More informationHausdorff Continuous Viscosity Solutions of Hamilton-Jacobi Equations
Hausdorff Continuous Viscosity Solutions of Hamilton-Jacobi Equations R Anguelov 1,2, S Markov 2,, F Minani 3 1 Department of Mathematics and Applied Mathematics, University of Pretoria 2 Institute of
More informationNumerical Methods for Optimal Control Problems. Part I: Hamilton-Jacobi-Bellman Equations and Pontryagin Minimum Principle
Numerical Methods for Optimal Control Problems. Part I: Hamilton-Jacobi-Bellman Equations and Pontryagin Minimum Principle Ph.D. course in OPTIMAL CONTROL Emiliano Cristiani (IAC CNR) e.cristiani@iac.cnr.it
More informationMulti-dimensional Stochastic Singular Control Via Dynkin Game and Dirichlet Form
Multi-dimensional Stochastic Singular Control Via Dynkin Game and Dirichlet Form Yipeng Yang * Under the supervision of Dr. Michael Taksar Department of Mathematics University of Missouri-Columbia Oct
More informationUNCERTAINTY FUNCTIONAL DIFFERENTIAL EQUATIONS FOR FINANCE
Surveys in Mathematics and its Applications ISSN 1842-6298 (electronic), 1843-7265 (print) Volume 5 (2010), 275 284 UNCERTAINTY FUNCTIONAL DIFFERENTIAL EQUATIONS FOR FINANCE Iuliana Carmen Bărbăcioru Abstract.
More informationHOMEWORK ASSIGNMENT 6
HOMEWORK ASSIGNMENT 6 DUE 15 MARCH, 2016 1) Suppose f, g : A R are uniformly continuous on A. Show that f + g is uniformly continuous on A. Solution First we note: In order to show that f + g is uniformly
More informationOptimal Control and Viscosity Solutions of Hamilton-Jacobi-Bellman Equations
Martino Bardi Italo Capuzzo-Dolcetta Optimal Control and Viscosity Solutions of Hamilton-Jacobi-Bellman Equations Birkhauser Boston Basel Berlin Contents Preface Basic notations xi xv Chapter I. Outline
More informationThe Skorokhod problem in a time-dependent interval
The Skorokhod problem in a time-dependent interval Krzysztof Burdzy, Weining Kang and Kavita Ramanan University of Washington and Carnegie Mellon University Abstract: We consider the Skorokhod problem
More informationINFINITE TIME HORIZON OPTIMAL CONTROL OF THE SEMILINEAR HEAT EQUATION
Nonlinear Funct. Anal. & Appl., Vol. 7, No. (22), pp. 69 83 INFINITE TIME HORIZON OPTIMAL CONTROL OF THE SEMILINEAR HEAT EQUATION Mihai Sîrbu Abstract. We consider here the infinite horizon control problem
More informationExample 1. Hamilton-Jacobi equation. In particular, the eikonal equation. for some n( x) > 0 in Ω. Here 1 / 2
Oct. 1 0 Viscosity S olutions In this lecture we take a glimpse of the viscosity solution theory for linear and nonlinear PDEs. From our experience we know that even for linear equations, the existence
More informationOn differential games with long-time-average cost
On differential games with long-time-average cost Martino Bardi Dipartimento di Matematica Pura ed Applicata Università di Padova via Belzoni 7, 35131 Padova, Italy bardi@math.unipd.it Abstract The paper
More informationAn introduction to Birkhoff normal form
An introduction to Birkhoff normal form Dario Bambusi Dipartimento di Matematica, Universitá di Milano via Saldini 50, 0133 Milano (Italy) 19.11.14 1 Introduction The aim of this note is to present an
More informationUniqueness of unbounded viscosity solutions for impulse control problem
J. Math. Anal. Appl. 315 26 686 71 www.elsevier.com/locate/jmaa Uniqueness of unbounded viscosity solutions for impulse control problem Mythily Ramaswamy a,, Sheetal Dharmatti b a IISc-TIFR Mathematics
More informationAN OVERVIEW OF STATIC HAMILTON-JACOBI EQUATIONS. 1. Introduction
AN OVERVIEW OF STATIC HAMILTON-JACOBI EQUATIONS JAMES C HATELEY Abstract. There is a voluminous amount of literature on Hamilton-Jacobi equations. This paper reviews some of the existence and uniqueness
More informationViscosity Solutions of the Bellman Equation for Perturbed Optimal Control Problems with Exit Times 0
Viscosity Solutions of the Bellman Equation for Perturbed Optimal Control Problems with Exit Times Michael Malisoff Department of Mathematics Louisiana State University Baton Rouge, LA 783-4918 USA malisoff@mathlsuedu
More informationEXISTENCE THEOREMS FOR FUNCTIONAL DIFFERENTIAL EQUATIONS IN BANACH SPACES. 1. Introduction
Acta Math. Univ. Comenianae Vol. LXXVIII, 2(29), pp. 287 32 287 EXISTENCE THEOREMS FOR FUNCTIONAL DIFFERENTIAL EQUATIONS IN BANACH SPACES A. SGHIR Abstract. This paper concernes with the study of existence
More informationLimit value of dynamic zero-sum games with vanishing stage duration
Limit value of dynamic zero-sum games with vanishing stage duration Sylvain Sorin IMJ-PRG Université P. et M. Curie - Paris 6 sylvain.sorin@imj-prg.fr Stochastic Methods in Game Theory National University
More informationPassivity-based Stabilization of Non-Compact Sets
Passivity-based Stabilization of Non-Compact Sets Mohamed I. El-Hawwary and Manfredi Maggiore Abstract We investigate the stabilization of closed sets for passive nonlinear systems which are contained
More informationCLASSIFICATIONS OF THE FLOWS OF LINEAR ODE
CLASSIFICATIONS OF THE FLOWS OF LINEAR ODE PETER ROBICHEAUX Abstract. The goal of this paper is to examine characterizations of linear differential equations. We define the flow of an equation and examine
More informationA Monotone Scheme for Hamilton-Jacobi Equations via the Nonstandard Finite Difference Method
A Monotone Scheme for Hamilton-Jacobi Equations via the Nonstandard Finite Difference Method Roumen Anguelov, Jean M-S Lubuma and Froduald Minani Department of Mathematics and Applied Mathematics University
More information2 Statement of the problem and assumptions
Mathematical Notes, 25, vol. 78, no. 4, pp. 466 48. Existence Theorem for Optimal Control Problems on an Infinite Time Interval A.V. Dmitruk and N.V. Kuz kina We consider an optimal control problem on
More informationarxiv: v3 [math.ds] 22 Feb 2012
Stability of interconnected impulsive systems with and without time-delays using Lyapunov methods arxiv:1011.2865v3 [math.ds] 22 Feb 2012 Sergey Dashkovskiy a, Michael Kosmykov b, Andrii Mironchenko b,
More informationHJ equations. Reachability analysis. Optimal control problems
HJ equations. Reachability analysis. Optimal control problems Hasnaa Zidani 1 1 ENSTA Paris-Tech & INRIA-Saclay Graz, 8-11 September 2014 H. Zidani (ENSTA & Inria) HJ equations. Reachability analysis -
More informationRobustness for a Liouville type theorem in exterior domains
Robustness for a Liouville type theorem in exterior domains Juliette Bouhours 1 arxiv:1207.0329v3 [math.ap] 24 Oct 2014 1 UPMC Univ Paris 06, UMR 7598, Laboratoire Jacques-Louis Lions, F-75005, Paris,
More informationComputation of an Over-Approximation of the Backward Reachable Set using Subsystem Level Set Functions. Stanford University, Stanford, CA 94305
To appear in Dynamics of Continuous, Discrete and Impulsive Systems http:monotone.uwaterloo.ca/ journal Computation of an Over-Approximation of the Backward Reachable Set using Subsystem Level Set Functions
More informationA NOTE ON ALMOST PERIODIC VARIATIONAL EQUATIONS
A NOTE ON ALMOST PERIODIC VARIATIONAL EQUATIONS PETER GIESL AND MARTIN RASMUSSEN Abstract. The variational equation of a nonautonomous differential equation ẋ = F t, x) along a solution µ is given by ẋ
More informationInstitut für Mathematik
U n i v e r s i t ä t A u g s b u r g Institut für Mathematik Martin Rasmussen, Peter Giesl A Note on Almost Periodic Variational Equations Preprint Nr. 13/28 14. März 28 Institut für Mathematik, Universitätsstraße,
More informationAsymptotic Perron Method for Stochastic Games and Control
Asymptotic Perron Method for Stochastic Games and Control Mihai Sîrbu, The University of Texas at Austin Methods of Mathematical Finance a conference in honor of Steve Shreve s 65th birthday Carnegie Mellon
More informationThe Arzelà-Ascoli Theorem
John Nachbar Washington University March 27, 2016 The Arzelà-Ascoli Theorem The Arzelà-Ascoli Theorem gives sufficient conditions for compactness in certain function spaces. Among other things, it helps
More informationLinear Quadratic Zero-Sum Two-Person Differential Games Pierre Bernhard June 15, 2013
Linear Quadratic Zero-Sum Two-Person Differential Games Pierre Bernhard June 15, 2013 Abstract As in optimal control theory, linear quadratic (LQ) differential games (DG) can be solved, even in high dimension,
More informationMINIMAL TIME PROBLEM WITH IMPULSIVE CONTROLS
MINIMAL TIME PROBLEM WITH IMPULSIVE CONTROLS KARL KUNISCH AND ZHIPING RAO Abstract. Time optimal control problems for systems with impulsive controls are investigated. Sufficient conditions for the existence
More informationThe main motivation of this paper comes from the following, rather surprising, result of Ecker and Huisken [13]: for any initial data u 0 W 1,
Quasilinear parabolic equations, unbounded solutions and geometrical equations II. Uniqueness without growth conditions and applications to the mean curvature flow in IR 2 Guy Barles, Samuel Biton and
More informationNoncooperative continuous-time Markov games
Morfismos, Vol. 9, No. 1, 2005, pp. 39 54 Noncooperative continuous-time Markov games Héctor Jasso-Fuentes Abstract This work concerns noncooperative continuous-time Markov games with Polish state and
More informationSPACES ENDOWED WITH A GRAPH AND APPLICATIONS. Mina Dinarvand. 1. Introduction
MATEMATIČKI VESNIK MATEMATIQKI VESNIK 69, 1 (2017), 23 38 March 2017 research paper originalni nauqni rad FIXED POINT RESULTS FOR (ϕ, ψ)-contractions IN METRIC SPACES ENDOWED WITH A GRAPH AND APPLICATIONS
More informationThreshold behavior and non-quasiconvergent solutions with localized initial data for bistable reaction-diffusion equations
Threshold behavior and non-quasiconvergent solutions with localized initial data for bistable reaction-diffusion equations P. Poláčik School of Mathematics, University of Minnesota Minneapolis, MN 55455
More informationExam February h
Master 2 Mathématiques et Applications PUF Ho Chi Minh Ville 2009/10 Viscosity solutions, HJ Equations and Control O.Ley (INSA de Rennes) Exam February 2010 3h Written-by-hands documents are allowed. Printed
More informationClosed-Loop Impulse Control of Oscillating Systems
Closed-Loop Impulse Control of Oscillating Systems A. N. Daryin and A. B. Kurzhanski Moscow State (Lomonosov) University Faculty of Computational Mathematics and Cybernetics Periodic Control Systems, 2007
More informationResearch Article Almost Periodic Viscosity Solutions of Nonlinear Parabolic Equations
Hindawi Publishing Corporation Boundary Value Problems Volume 29, Article ID 873526, 15 pages doi:1.1155/29/873526 Research Article Almost Periodic Viscosity Solutions of Nonlinear Parabolic Equations
More informationSolution of Stochastic Optimal Control Problems and Financial Applications
Journal of Mathematical Extension Vol. 11, No. 4, (2017), 27-44 ISSN: 1735-8299 URL: http://www.ijmex.com Solution of Stochastic Optimal Control Problems and Financial Applications 2 Mat B. Kafash 1 Faculty
More informationON WEAKLY NONLINEAR BACKWARD PARABOLIC PROBLEM
ON WEAKLY NONLINEAR BACKWARD PARABOLIC PROBLEM OLEG ZUBELEVICH DEPARTMENT OF MATHEMATICS THE BUDGET AND TREASURY ACADEMY OF THE MINISTRY OF FINANCE OF THE RUSSIAN FEDERATION 7, ZLATOUSTINSKY MALIY PER.,
More informationViscosity Solutions of Fully Nonlinear Second Order Parabolic Equations with L 1 Dependence in Time and Neumann Boundary Conditions
Viscosity Solutions of Fully Nonlinear Second Order Parabolic Equations with L 1 Dependence in Time and Neumann Boundary Conditions Mariane BOURGOING Laboratoire de Mathématiques et Physique Théorique
More informationStability of Feedback Solutions for Infinite Horizon Noncooperative Differential Games
Stability of Feedback Solutions for Infinite Horizon Noncooperative Differential Games Alberto Bressan ) and Khai T. Nguyen ) *) Department of Mathematics, Penn State University **) Department of Mathematics,
More informationObserver design for a general class of triangular systems
1st International Symposium on Mathematical Theory of Networks and Systems July 7-11, 014. Observer design for a general class of triangular systems Dimitris Boskos 1 John Tsinias Abstract The paper deals
More informationConvergence of a first order scheme for a non local eikonal equation
Convergence of a first order scheme for a non local eikonal equation O. Alvarez, E. Carlini, R. Monneau, E. Rouy March 22, 2005 Abstract We prove the convergence of a first order finite difference scheme
More informationFunctional Analysis HW #3
Functional Analysis HW #3 Sangchul Lee October 26, 2015 1 Solutions Exercise 2.1. Let D = { f C([0, 1]) : f C([0, 1])} and define f d = f + f. Show that D is a Banach algebra and that the Gelfand transform
More informationLocal semiconvexity of Kantorovich potentials on non-compact manifolds
Local semiconvexity of Kantorovich potentials on non-compact manifolds Alessio Figalli, Nicola Gigli Abstract We prove that any Kantorovich potential for the cost function c = d / on a Riemannian manifold
More informationIdentification of Parameters in Neutral Functional Differential Equations with State-Dependent Delays
To appear in the proceedings of 44th IEEE Conference on Decision and Control and European Control Conference ECC 5, Seville, Spain. -5 December 5. Identification of Parameters in Neutral Functional Differential
More information1 Lyapunov theory of stability
M.Kawski, APM 581 Diff Equns Intro to Lyapunov theory. November 15, 29 1 1 Lyapunov theory of stability Introduction. Lyapunov s second (or direct) method provides tools for studying (asymptotic) stability
More informationNon-Zero-Sum Stochastic Differential Games of Controls and St
Non-Zero-Sum Stochastic Differential Games of Controls and Stoppings October 1, 2009 Based on two preprints: to a Non-Zero-Sum Stochastic Differential Game of Controls and Stoppings I. Karatzas, Q. Li,
More informationOptimal stopping for non-linear expectations Part I
Stochastic Processes and their Applications 121 (2011) 185 211 www.elsevier.com/locate/spa Optimal stopping for non-linear expectations Part I Erhan Bayraktar, Song Yao Department of Mathematics, University
More informationFurther Results on the Bellman Equation for Optimal Control Problems with Exit Times and Nonnegative Instantaneous Costs
Further Results on the Bellman Equation for Optimal Control Problems with Exit Times and Nonnegative Instantaneous Costs Michael Malisoff Systems Science and Mathematics Dept. Washington University in
More informationThe Minimum Speed for a Blocking Problem on the Half Plane
The Minimum Speed for a Blocking Problem on the Half Plane Alberto Bressan and Tao Wang Department of Mathematics, Penn State University University Park, Pa 16802, USA e-mails: bressan@mathpsuedu, wang
More informationu( x) = g( y) ds y ( 1 ) U solves u = 0 in U; u = 0 on U. ( 3)
M ath 5 2 7 Fall 2 0 0 9 L ecture 4 ( S ep. 6, 2 0 0 9 ) Properties and Estimates of Laplace s and Poisson s Equations In our last lecture we derived the formulas for the solutions of Poisson s equation
More informationRisk-Sensitive Control with HARA Utility
IEEE TRANSACTIONS ON AUTOMATIC CONTROL, VOL. 46, NO. 4, APRIL 2001 563 Risk-Sensitive Control with HARA Utility Andrew E. B. Lim Xun Yu Zhou, Senior Member, IEEE Abstract In this paper, a control methodology
More informationSTOCHASTIC PERRON S METHOD AND VERIFICATION WITHOUT SMOOTHNESS USING VISCOSITY COMPARISON: OBSTACLE PROBLEMS AND DYNKIN GAMES
STOCHASTIC PERRON S METHOD AND VERIFICATION WITHOUT SMOOTHNESS USING VISCOSITY COMPARISON: OBSTACLE PROBLEMS AND DYNKIN GAMES ERHAN BAYRAKTAR AND MIHAI SÎRBU Abstract. We adapt the Stochastic Perron s
More informationApproximation of Minimal Functions by Extreme Functions
Approximation of Minimal Functions by Extreme Functions Teresa M. Lebair and Amitabh Basu August 14, 2017 Abstract In a recent paper, Basu, Hildebrand, and Molinaro established that the set of continuous
More informationNotes on uniform convergence
Notes on uniform convergence Erik Wahlén erik.wahlen@math.lu.se January 17, 2012 1 Numerical sequences We begin by recalling some properties of numerical sequences. By a numerical sequence we simply mean
More informationLinear Quadratic Zero-Sum Two-Person Differential Games
Linear Quadratic Zero-Sum Two-Person Differential Games Pierre Bernhard To cite this version: Pierre Bernhard. Linear Quadratic Zero-Sum Two-Person Differential Games. Encyclopaedia of Systems and Control,
More informationThe principle of least action and two-point boundary value problems in orbital mechanics
The principle of least action and two-point boundary value problems in orbital mechanics Seung Hak Han and William M McEneaney Abstract We consider a two-point boundary value problem (TPBVP) in orbital
More informationHAMILTON-JACOBI EQUATIONS : APPROXIMATIONS, NUMERICAL ANALYSIS AND APPLICATIONS. CIME Courses-Cetraro August 29-September COURSES
HAMILTON-JACOBI EQUATIONS : APPROXIMATIONS, NUMERICAL ANALYSIS AND APPLICATIONS CIME Courses-Cetraro August 29-September 3 2011 COURSES (1) Models of mean field, Hamilton-Jacobi-Bellman Equations and numerical
More informationAC&ST AUTOMATIC CONTROL AND SYSTEM THEORY SYSTEMS AND MODELS. Claudio Melchiorri
C. Melchiorri (DEI) Automatic Control & System Theory 1 AUTOMATIC CONTROL AND SYSTEM THEORY SYSTEMS AND MODELS Claudio Melchiorri Dipartimento di Ingegneria dell Energia Elettrica e dell Informazione (DEI)
More informationPropagating terraces and the dynamics of front-like solutions of reaction-diffusion equations on R
Propagating terraces and the dynamics of front-like solutions of reaction-diffusion equations on R P. Poláčik School of Mathematics, University of Minnesota Minneapolis, MN 55455 Abstract We consider semilinear
More informationTHE SKOROKHOD OBLIQUE REFLECTION PROBLEM IN A CONVEX POLYHEDRON
GEORGIAN MATHEMATICAL JOURNAL: Vol. 3, No. 2, 1996, 153-176 THE SKOROKHOD OBLIQUE REFLECTION PROBLEM IN A CONVEX POLYHEDRON M. SHASHIASHVILI Abstract. The Skorokhod oblique reflection problem is studied
More informationDeterministic Dynamic Programming
Deterministic Dynamic Programming 1 Value Function Consider the following optimal control problem in Mayer s form: V (t 0, x 0 ) = inf u U J(t 1, x(t 1 )) (1) subject to ẋ(t) = f(t, x(t), u(t)), x(t 0
More informationProof. We indicate by α, β (finite or not) the end-points of I and call
C.6 Continuous functions Pag. 111 Proof of Corollary 4.25 Corollary 4.25 Let f be continuous on the interval I and suppose it admits non-zero its (finite or infinite) that are different in sign for x tending
More informationOn the Well-Posedness of the Cauchy Problem for a Neutral Differential Equation with Distributed Prehistory
Bulletin of TICMI Vol. 21, No. 1, 2017, 3 8 On the Well-Posedness of the Cauchy Problem for a Neutral Differential Equation with Distributed Prehistory Tamaz Tadumadze I. Javakhishvili Tbilisi State University
More informationPiecewise Smooth Solutions to the Burgers-Hilbert Equation
Piecewise Smooth Solutions to the Burgers-Hilbert Equation Alberto Bressan and Tianyou Zhang Department of Mathematics, Penn State University, University Park, Pa 68, USA e-mails: bressan@mathpsuedu, zhang
More informationNew Discretizations of Turbulent Flow Problems
New Discretizations of Turbulent Flow Problems Carolina Cardoso Manica and Songul Kaya Merdan Abstract A suitable discretization for the Zeroth Order Model in Large Eddy Simulation of turbulent flows is
More informationGLOBAL EXISTENCE RESULTS AND UNIQUENESS FOR DISLOCATION EQUATIONS
GLOBAL EXISTENCE RESULTS AND UNIQUENESS FOR DISLOCATION EQUATIONS GUY BARLES, PIERRE CARDALIAGUET, OLIVIER LEY & RÉGIS MONNEAU Abstract. We are interested in nonlocal Eikonal Equations arising in the study
More informationGENERAL EXISTENCE OF SOLUTIONS TO DYNAMIC PROGRAMMING PRINCIPLE. 1. Introduction
GENERAL EXISTENCE OF SOLUTIONS TO DYNAMIC PROGRAMMING PRINCIPLE QING LIU AND ARMIN SCHIKORRA Abstract. We provide an alternative approach to the existence of solutions to dynamic programming equations
More informationConverse Lyapunov theorem and Input-to-State Stability
Converse Lyapunov theorem and Input-to-State Stability April 6, 2014 1 Converse Lyapunov theorem In the previous lecture, we have discussed few examples of nonlinear control systems and stability concepts
More informationL2 gains and system approximation quality 1
Massachusetts Institute of Technology Department of Electrical Engineering and Computer Science 6.242, Fall 24: MODEL REDUCTION L2 gains and system approximation quality 1 This lecture discusses the utility
More informationLinear-Quadratic Stochastic Differential Games with General Noise Processes
Linear-Quadratic Stochastic Differential Games with General Noise Processes Tyrone E. Duncan Abstract In this paper a noncooperative, two person, zero sum, stochastic differential game is formulated and
More informationLecture notes on Ordinary Differential Equations. S. Sivaji Ganesh Department of Mathematics Indian Institute of Technology Bombay
Lecture notes on Ordinary Differential Equations S. Ganesh Department of Mathematics Indian Institute of Technology Bombay May 20, 2016 ii IIT Bombay Contents I Ordinary Differential Equations 1 1 Initial
More informationA maximum principle for optimal control system with endpoint constraints
Wang and Liu Journal of Inequalities and Applications 212, 212: http://www.journalofinequalitiesandapplications.com/content/212/231/ R E S E A R C H Open Access A maimum principle for optimal control system
More informationON THE PATHWISE UNIQUENESS OF SOLUTIONS OF STOCHASTIC DIFFERENTIAL EQUATIONS
PORTUGALIAE MATHEMATICA Vol. 55 Fasc. 4 1998 ON THE PATHWISE UNIQUENESS OF SOLUTIONS OF STOCHASTIC DIFFERENTIAL EQUATIONS C. Sonoc Abstract: A sufficient condition for uniqueness of solutions of ordinary
More informationParameter Dependent Quasi-Linear Parabolic Equations
CADERNOS DE MATEMÁTICA 4, 39 33 October (23) ARTIGO NÚMERO SMA#79 Parameter Dependent Quasi-Linear Parabolic Equations Cláudia Buttarello Gentile Departamento de Matemática, Universidade Federal de São
More informationViscosity Solutions for Dummies (including Economists)
Viscosity Solutions for Dummies (including Economists) Online Appendix to Income and Wealth Distribution in Macroeconomics: A Continuous-Time Approach written by Benjamin Moll August 13, 2017 1 Viscosity
More informationOn Stopping Times and Impulse Control with Constraint
On Stopping Times and Impulse Control with Constraint Jose Luis Menaldi Based on joint papers with M. Robin (216, 217) Department of Mathematics Wayne State University Detroit, Michigan 4822, USA (e-mail:
More informationLimit value of dynamic zero-sum games with vanishing stage duration
Limit value of dynamic zero-sum games with vanishing stage duration Sylvain Sorin IMJ-PRG Université P. et M. Curie - Paris 6 sylvain.sorin@imj-prg.fr Workshop on Analysis and Applications of Stochastic
More informationHomogenization of first order equations with (u/ε)-periodic Hamiltonians. Part I: local equations
Homogenization of first order equations with u/-periodic Hamiltonians. Part I: local equations Cyril Imbert, Régis Monneau February 23, 2006 Abstract. In this paper, we present a result of homogenization
More informationTHE CANTOR GAME: WINNING STRATEGIES AND DETERMINACY. by arxiv: v1 [math.ca] 29 Jan 2017 MAGNUS D. LADUE
THE CANTOR GAME: WINNING STRATEGIES AND DETERMINACY by arxiv:170109087v1 [mathca] 9 Jan 017 MAGNUS D LADUE 0 Abstract In [1] Grossman Turett define the Cantor game In [] Matt Baker proves several results
More informationWeak solutions of mean-field stochastic differential equations
Weak solutions of mean-field stochastic differential equations Juan Li School of Mathematics and Statistics, Shandong University (Weihai), Weihai 26429, China. Email: juanli@sdu.edu.cn Based on joint works
More informationOptimal stopping time formulation of adaptive image filtering
Optimal stopping time formulation of adaptive image filtering I. Capuzzo Dolcetta, R. Ferretti 19.04.2000 Abstract This paper presents an approach to image filtering based on an optimal stopping time problem
More informationMaster Thesis. Nguyen Tien Thinh. Homogenization and Viscosity solution
Master Thesis Nguyen Tien Thinh Homogenization and Viscosity solution Advisor: Guy Barles Defense: Friday June 21 th, 2013 ii Preface Firstly, I am grateful to Prof. Guy Barles for helping me studying
More informationSOME PROPERTIES ON THE CLOSED SUBSETS IN BANACH SPACES
ARCHIVUM MATHEMATICUM (BRNO) Tomus 42 (2006), 167 174 SOME PROPERTIES ON THE CLOSED SUBSETS IN BANACH SPACES ABDELHAKIM MAADEN AND ABDELKADER STOUTI Abstract. It is shown that under natural assumptions,
More informationA MAXIMUM PRINCIPLE FOR SEMICONTINUOUS FUNCTIONS APPLICABLE TO INTEGRO-PARTIAL DIFFERENTIAL EQUATIONS
Dept. of Math. University of Oslo Pure Mathematics ISBN 82 553 1382 6 No. 18 ISSN 0806 2439 May 2003 A MAXIMUM PRINCIPLE FOR SEMICONTINUOUS FUNCTIONS APPLICABLE TO INTEGRO-PARTIAL DIFFERENTIAL EQUATIONS
More informationNonlinear L 2 -gain analysis via a cascade
9th IEEE Conference on Decision and Control December -7, Hilton Atlanta Hotel, Atlanta, GA, USA Nonlinear L -gain analysis via a cascade Peter M Dower, Huan Zhang and Christopher M Kellett Abstract A nonlinear
More informationA TWO PARAMETERS AMBROSETTI PRODI PROBLEM*
PORTUGALIAE MATHEMATICA Vol. 53 Fasc. 3 1996 A TWO PARAMETERS AMBROSETTI PRODI PROBLEM* C. De Coster** and P. Habets 1 Introduction The study of the Ambrosetti Prodi problem has started with the paper
More informationEXISTENCE OF SOLUTIONS TO ASYMPTOTICALLY PERIODIC SCHRÖDINGER EQUATIONS
Electronic Journal of Differential Equations, Vol. 017 (017), No. 15, pp. 1 7. ISSN: 107-6691. URL: http://ejde.math.txstate.edu or http://ejde.math.unt.edu EXISTENCE OF SOLUTIONS TO ASYMPTOTICALLY PERIODIC
More information