ADJOINT METHODS FOR OBSTACLE PROBLEMS AND WEAKLY COUPLED SYSTEMS OF PDE. 1. Introduction

Similar documents
New Identities for Weak KAM Theory

Dynamical properties of Hamilton Jacobi equations via the nonlinear adjoint method: Large time behavior and Discounted approximation

VISCOSITY SOLUTIONS OF HAMILTON-JACOBI EQUATIONS, AND ASYMPTOTICS FOR HAMILTONIAN SYSTEMS

THREE SINGULAR VARIATIONAL PROBLEMS. Lawrence C. Evans Department of Mathematics University of California Berkeley, CA 94720

Theory of PDE Homework 2

Laplace s Equation. Chapter Mean Value Formulas

REGULARITY THEORY FOR HAMILTON-JACOBI EQUATIONS

Régularité des équations de Hamilton-Jacobi du premier ordre et applications aux jeux à champ moyen

Mean field games and related models

MINIMAL GRAPHS PART I: EXISTENCE OF LIPSCHITZ WEAK SOLUTIONS TO THE DIRICHLET PROBLEM WITH C 2 BOUNDARY DATA

Some notes on viscosity solutions

u xx + u yy = 0. (5.1)

UNIQUENESS OF POSITIVE SOLUTION TO SOME COUPLED COOPERATIVE VARIATIONAL ELLIPTIC SYSTEMS

Week 6 Notes, Math 865, Tanveer

Everywhere differentiability of infinity harmonic functions

Continuous dependence estimates for the ergodic problem with an application to homogenization

arxiv: v2 [math.ap] 28 Nov 2016

Stationary mean-field games Diogo A. Gomes

Some lecture notes for Math 6050E: PDEs, Fall 2016

AN OVERVIEW OF STATIC HAMILTON-JACOBI EQUATIONS. 1. Introduction

Landesman-Lazer type results for second order Hamilton-Jacobi-Bellman equations

Controlled Diffusions and Hamilton-Jacobi Bellman Equations

SOLUTION OF POISSON S EQUATION. Contents

Master Thesis. Nguyen Tien Thinh. Homogenization and Viscosity solution

Gradient Estimate of Mean Curvature Equations and Hessian Equations with Neumann Boundary Condition

Thuong Nguyen. SADCO Internal Review Metting

Lipschitz continuity for solutions of Hamilton-Jacobi equation with Ornstein-Uhlenbeck operator

ASYMPTOTIC BEHAVIOUR OF NONLINEAR ELLIPTIC SYSTEMS ON VARYING DOMAINS

P(E t, Ω)dt, (2) 4t has an advantage with respect. to the compactly supported mollifiers, i.e., the function W (t)f satisfies a semigroup law:

GLOBAL LIPSCHITZ CONTINUITY FOR MINIMA OF DEGENERATE PROBLEMS

Constrained Optimal Stopping Problems

ON NEUMANN PROBLEMS FOR NONLOCAL HAMILTON-JACOBI EQUATIONS WITH DOMINATING GRADIENT TERMS

REGULARITY FOR INFINITY HARMONIC FUNCTIONS IN TWO DIMENSIONS

REGULARITY RESULTS FOR THE EQUATION u 11 u 22 = Introduction

EXISTENCE OF SOLUTIONS TO ASYMPTOTICALLY PERIODIC SCHRÖDINGER EQUATIONS

1 Lyapunov theory of stability

Flux-limited solutions for quasi-convex Hamilton-Jacobi equations on networks

ON THE FOLIATION OF SPACE-TIME BY CONSTANT MEAN CURVATURE HYPERSURFACES

1. Introduction Boundary estimates for the second derivatives of the solution to the Dirichlet problem for the Monge-Ampere equation

The continuity method

Example 1. Hamilton-Jacobi equation. In particular, the eikonal equation. for some n( x) > 0 in Ω. Here 1 / 2

Green s Functions and Distributions

Integro-differential equations: Regularity theory and Pohozaev identities

C m Extension by Linear Operators

VISCOSITY SOLUTIONS OF ELLIPTIC EQUATIONS

Mean-Field Games with non-convex Hamiltonian

STOCHASTIC PERRON S METHOD AND VERIFICATION WITHOUT SMOOTHNESS USING VISCOSITY COMPARISON: OBSTACLE PROBLEMS AND DYNKIN GAMES

A LOCALIZATION PROPERTY AT THE BOUNDARY FOR MONGE-AMPERE EQUATION

THE NEUMANN PROBLEM FOR THE -LAPLACIAN AND THE MONGE-KANTOROVICH MASS TRANSFER PROBLEM

Minimal Surface equations non-solvability strongly convex functional further regularity Consider minimal surface equation.

We denote the space of distributions on Ω by D ( Ω) 2.

Homogenization of Neuman boundary data with fully nonlinear operator

PDE Methods for Mean Field Games with Non-Separable Hamiltonian: Data in Sobolev Spaces (Continued) David Ambrose June 29, 2018

On some weighted fractional porous media equations

Regularity for Poisson Equation

Mean Field Games on networks

Stability of Feedback Solutions for Infinite Horizon Noncooperative Differential Games

at time t, in dimension d. The index i varies in a countable set I. We call configuration the family, denoted generically by Φ: U (x i (t) x j (t))

A Perron-type theorem on the principal eigenvalue of nonsymmetric elliptic operators

Harmonic Functions and Brownian motion

for all subintervals I J. If the same is true for the dyadic subintervals I D J only, we will write ϕ BMO d (J). In fact, the following is true

Homogenization and error estimates of free boundary velocities in periodic media

The Dirichlet s P rinciple. In this lecture we discuss an alternative formulation of the Dirichlet problem for the Laplace equation:

u( x) = g( y) ds y ( 1 ) U solves u = 0 in U; u = 0 on U. ( 3)

SYMMETRY OF POSITIVE SOLUTIONS OF SOME NONLINEAR EQUATIONS. M. Grossi S. Kesavan F. Pacella M. Ramaswamy. 1. Introduction

On the Ladyzhenskaya Smagorinsky turbulence model of the Navier Stokes equations in smooth domains. The regularity problem

Optimal Transportation. Nonlinear Partial Differential Equations

POINTWISE BOUNDS ON QUASIMODES OF SEMICLASSICAL SCHRÖDINGER OPERATORS IN DIMENSION TWO

Analysis in weighted spaces : preliminary version

S chauder Theory. x 2. = log( x 1 + x 2 ) + 1 ( x 1 + x 2 ) 2. ( 5) x 1 + x 2 x 1 + x 2. 2 = 2 x 1. x 1 x 2. 1 x 1.

On a Fractional Monge Ampère Operator

Hamburger Beiträge zur Angewandten Mathematik

Stochastic Homogenization for Reaction-Diffusion Equations

VISCOSITY SOLUTIONS. We follow Han and Lin, Elliptic Partial Differential Equations, 5.

The heat equation in time dependent domains with Neumann boundary conditions

2 A Model, Harmonic Map, Problem

On John type ellipsoids

A Concise Course on Stochastic Partial Differential Equations

A CONVEX-CONCAVE ELLIPTIC PROBLEM WITH A PARAMETER ON THE BOUNDARY CONDITION

Regularity estimates for fully non linear elliptic equations which are asymptotically convex

Ahmed Mohammed. Harnack Inequality for Non-divergence Structure Semi-linear Elliptic Equations

Exam February h

ON WEAKLY NONLINEAR BACKWARD PARABOLIC PROBLEM

Level-set convex Hamilton-Jacobi equations on networks

OPTIMAL POTENTIALS FOR SCHRÖDINGER OPERATORS. 1. Introduction In this paper we consider optimization problems of the form. min F (V ) : V V, (1.

HOMOGENIZATION OF HAMILTON-JACOBI EQUATIONS NOTES AND REMARKS SON NGUYEN THAI TU. Department of Mathematics, University of Wisconsin at Madison

Homogenization of a Hele-Shaw-type problem in periodic time-dependent med

ξ,i = x nx i x 3 + δ ni + x n x = 0. x Dξ = x i ξ,i = x nx i x i x 3 Du = λ x λ 2 xh + x λ h Dξ,

The L p -dissipativity of first order partial differential operators

Sébastien Chaumont a a Institut Élie Cartan, Université Henri Poincaré Nancy I, B. P. 239, Vandoeuvre-lès-Nancy Cedex, France. 1.

EXPOSITORY NOTES ON DISTRIBUTION THEORY, FALL 2018

TRANSPORT IN POROUS MEDIA

A MAXIMUM PRINCIPLE FOR SEMICONTINUOUS FUNCTIONS APPLICABLE TO INTEGRO-PARTIAL DIFFERENTIAL EQUATIONS

Partial Differential Equations

AN ASYMPTOTIC MEAN VALUE CHARACTERIZATION FOR p-harmonic FUNCTIONS. To the memory of our friend and colleague Fuensanta Andreu

WEAK KAM THEORETIC ASPECTS FOR NONREGULAR COMMUTING HAMILTONIANS

Numerical Solutions of Geometric Partial Differential Equations. Adam Oberman McGill University

Topology of the set of singularities of a solution of the Hamilton-Jacobi Equation

The infinity-laplacian and its properties

A PROPERTY OF SOBOLEV SPACES ON COMPLETE RIEMANNIAN MANIFOLDS

Transcription:

ADJOINT METHODS FOR OBSTACLE PROBLEMS AND WEAKLY COPLED SYSTEMS OF PDE F. CAGNETTI, D. GOMES, AND H.V. TRAN Abstract. The adjoint method, recently introduced by Evans, is used to study obstacle problems, weakly coupled systems, cell problems for weakly coupled systems of Hamilton Jacobi equations, and weakly coupled systems of obstacle type. In particular, new results about the speed of convergence of common approximation procedures are derived.. Introduction In this paper we study the speed of convergence of certain approximations for obstacle problems and weakly coupled systems of Hamilton Jacobi equations, using the Adjoint Method. This technique, recently introduced by Evans (see [Eva0], and also [Tra] and [CGT]), is a very successful tool to understand several types of degenerate PDEs. It can be applied, for instance, to Hamilton Jacobi equations with non convex Hamiltonians, e.g. time dependent (see [Eva0]) and time independent (see [Tra]) Hamilton Jacobi equations, to weak KAM theory (see [Fat97a, Fat97b, Fat98a, Fat98b], [EG0], [EG02]), and to the infinity Laplacian equation (see [ES]). We address here several applications and propose some new open questions. Further results, which will not be discussed here, can be found in [Eva0] and [CGT]... Overview of the Adjoint Method. To apply the Adjoint Method to a (non linear) PDE, one has to consider the adjoint equation associated to the linearization of the original problem. In this way it is possible to prove new estimates, which can then be used to obtain additional information on the solution of the initial PDE. In order to give an idea of the technique, we show below how this was used in the context of Aubry-Mather theory, in the periodic setting (see [CGT]). To start with, we quote a fundamental result (see [LPV88]), stating existence and uniqueness of the effective Hamiltonian. Here with T n we denote the n-dimensional unit torus in R n, n N. Theorem. (Lions, Papanicolaou and Varadhan). Let H : T n R n R be smooth and coercive, i.e. lim H(x, p) = +. p +

2 F. CAGNETTI, D. GOMES, AND H.V. TRAN Then, for every P R n there exists a unique H(P ) R such that the equation H(x, P + D x u(x, P )) = H(P ) (.) admits a Z n -periodic viscosity solution u(, P ) : T n R. In the spirit of Theorem., we prove in [CGT] the analogue result for an elliptic regularization of equation (.). See also [Gom02] for similar results. Theorem.2 (See [CGT]). Let H : T n R n R be smooth and assume that ( ) lim p + 2 H(x, H(x, p) p) 2 + D x H(x, p) p = lim = +. p + p Then, for every η > 0 and every P R n, there exists a unique number H η (P ) R such that the equation η2 2 uη (x) + H(x, P + Du η (x)) = H η (P ) (.2) admits a unique (up to constants) Z n -periodic viscosity solution. Moreover, for every P R n lim H η (P ) = H(P ), η 0 + (up to subsequences) where H(P ) is given by Theorem.. In addition, we have u η u uniformly, (up to subsequences) where u is a Z n -periodic viscosity solution of (.). For every η > 0 and P R n, the formal linearized operator L η,p : C 2 (T n ) C(T n ) associated to equation (.2) is defined as L η,p v(x) := η2 2 v(x) + D ph(x, P + Du η (x)) Dv(x), v C 2 (T n ). As already mentioned, the main idea of the method consists in the introduction of the adjoint equation associated to L η,p : η2 2 ση (x) div(d p H(x, P + Du η (x))σ η (x)) = 0, in T n, σ η is non-negative, T n -periodic and σ η dx =. T n (.3) Then, exploiting the properties of the solution σ η of (.3), we can retrieve additional information about u η.

ADJOINT METHODS FOR OBSTACLE PROBLEMS AND WEAKLY COPLED SYSTEMS OF PDE 3 First of all, one can show new estimates that do not seem to be easily obtained in a classical way. As an example, define the function w η := Duη 2. Then, w η satisfies 2 D x H Du η + D p H Dw η = η2 2 wη η2 2 D2 u η 2. (.4) Multiplying the above relation by σ η and integrating by parts, we eventually get for some C > 0 independent of η. η 2 T n D 2 u η 2 σ η dx C, (.5) Relation (.5) gives information about the behavior of all the Hessian D 2 u η of u η in the support of σ η. We observe that, without passing to the adjoint equation, one can only conclude that η 2 u η C, thus obtaining a relation which involves just the Laplacian u η of u η. More generally, by considering functions of the form w η (x) = φ(x, P + Du η (x)) and studying the analogous of equation (.4), one can obtain further properties, using compensated compactness based estimates (see [Eva0], [CGT]). In addition, the Hamiltonian H is not required to be convex. When we have that H is uniformly convex in p (i.e. D 2 pph α for some α > 0), (.5) can be significantly improved (see [CGT]). Indeed, differentiating (.2) twice along a generic vector ξ R n with ξ = we have (here and always in the sequel, we use Einstein s convention for repeated indices in a sum) H ξξ + 2H ξpi u η ξx i + H pip j u η ξx i u η ξx j + D p H Du η ξξ = η2 2 uη ξξ. Then, multiplying by σ η and integrating by parts we get or more generally T n Du η ξ 2 σ η dx C, T n D 2 u η 2 σ η dx C, which is clearly stronger than (.5). The differences between convex and nonconvex setting can be also observed by investigating the existence of invariant Mather measures (see [CGT]), and by studying the nature of the shocks in Hamilton Jacobi equations (see [Eva0]). Finally, the treatment allows to analyze the speed of convergence of H η (P ) to H(P ). Indeed, classical arguments in elliptic regularization imply that u η and H η are smooth in η away from η = 0. Then, differentiating (.2) w.r.t. η D p H Du η η = H η η(p ) + η2 2 uη η + η u η,

4 F. CAGNETTI, D. GOMES, AND H.V. TRAN where we denoted the differentiation w.r.t. η with a subscript. Again, multiplying by σ η and integrating by parts we infer that ( ) /2 ( ) /2 H η η(p ) η u η σ η dx C η D 2 u η 2 σ η dx σ η dx C, T n T n T n where the latter inequality follows by using (.5). Thus, we conclude that H η (P ) H(P ) Cη, which shows that the speed of convergence is O(η). In the uniformly convex case, this result can be improved to O(η 2 ). Let us observe that, as pointed out in [Eva0] and by Fraydoun Rezakhanlou to us, these estimates on the speed of convergence can also be obtained via Maximum Principle..2. Outline of the paper. This paper contains four further sections concerning obstacle problems, weakly coupled systems, effective Hamiltonian for weakly coupled systems of Hamilton Jacobi equations, and weakly coupled systems of obstacle type, respectively. We use a common strategy to study all these problems. For this reason, we describe in more detail our approach just in the case of the obstacle problem. In all the paper, is an open bounded domain in R n with smooth boundary, n 2. Moreover, we will denote with ν the outer unit normal to. In Section 2 we consider an obstacle problem of the form max{u ψ, u + H(x, Du)} = 0 in, u = 0 on, (.6) where ψ : R and H : R n R are smooth, with ψ 0 on. This equation arises naturally in Optimal Control theory, in the study of optimal stopping (see [Lio82]). In [IY90], Ishii and Yamada consider similar problems in the special case in which H is uniformly convex in the second variable. Classically, in order to study (.6) one first modifies the equation, by adding a perturbation term that penalizes the region where u > ψ. Then, a solution is obtained as a limit of the solutions of the penalized problems. More precisely, let γ : R [0, + ) be smooth, such that and define γ : R [0, + ) as γ(s) = 0 for s 0, γ(s) > 0 for s > 0, 0 < γ (s) for s > 0, and lim γ(s) = +, s + ( s γ (s) := γ, for all s R, for all > 0. (.7) )

ADJOINT METHODS FOR OBSTACLE PROBLEMS AND WEAKLY COPLED SYSTEMS OF PDE 5 In some of the problems we discuss we also require γ to be convex in order to obtain improved results, but that will be pointed out where necessary. For every > 0, one can introduce the penalized PDE u + H(x, Du ) + γ (u ψ) = u in, (.8) u = 0 on. To avoid confusion, we stress the fact that here γ (u ψ) stands for the composition of the function γ with u ψ. nless otherwise stated, we will often simply write γ and (γ ) to denote γ (u ψ) and (γ ) (u ψ), respectively. Also, notice that in (.8) the parameter corresponds to η 2 in (.2). We made this choice in order to compare our results with existing estimates for the speed of convergence in literature. Thanks to [Lio82], for every > 0 there exists a smooth solution u to (.8). It is also well known that, up to subsequences, u converges uniformly to a viscosity solution u of (.6) (see also Section 2 for further details). In [IY90] Ishii and Yamada considered related problems when H(x, ) is uniformly convex, and studied the speed of convergence of the functions u to u. However, both the original problem, the regularized PDE, and their methods are different from ours. To the best of our knowledge, no results are available in literature concerning non convex Hamiltonians. We face here the problem requiring a coercivity assumption on H and a compatibility condition for equation (.6) (see hypotheses (H2.) and (H2.2), respectively), showing that the speed of convergence in the general case is O( /2 ). Theorem.3. Suppose conditions (H2.) and (H2.2) in Section 2 hold. Then, there exists a positive constant C, independent of, such that u u L C /2. At the end of the section, we give a dynamic and a stochastic interpretation of the problem (see Subsection 2.3 and Subsection 2.4, respectively). The proof of Theorem.3 consists of three steps. Step I: Preliminary Estimates. We first show that max x u (x) ψ(x) C, (.9) for some constant C > 0 independent of (see Lemma 2.2). This allows us to prove that see Proposition 2.. u L, Du L C,

6 F. CAGNETTI, D. GOMES, AND H.V. TRAN Step II: Adjoint Method. We consider the formal linearization of (.8), and then introduce the correspondent adjoint equation (see equation (2.6)). The study of this last equation for different values of the data allows us to obtain several useful estimates (see Lemma 2.3 and Lemma 2.4). Step III: Conclusion. We conclude the proof of Theorem.3 by showing that max u (x) C x, /2 u (x) := u (x), (.0) for some constant C > 0 independent of (see Lemma 2.5). The most delicate part of the proof of (.0) consists in controlling the term (see relation (2.)) γ(s) = s ( s ) 2 γ, s R, γ (s) := γ (s). (.) We underline that getting a bound for (.) can be extremely hard in general. In this context, this is achieved by differentiating equation (.8) w.r.t. (see equation (2.0)), and then by using relation (.9), Lemma 2.3 and Lemma 2.4. This means that we overcome the problem by essentially using the Maximum Principle and the monotonicity of γ (see estimates (2.2) and (2.3)). We were not able to obtain such a bound when dealing with homogenization or singular perturbation, where also similar terms appear. We believe it would be very interesting to find the correct way to apply the Adjoint Method in these situations. In Section 3 we study the weakly coupled system of Hamilton Jacobi equations c u + c 2 u 2 + H (x, Du ) = 0, in. (.2) c 2 u + c 22 u 2 + H 2 (x, Du 2 ) = 0, nder some coupling assumptions on the constants (see conditions (H3.2) and (H3.3)), Engler and Lenhart [EL9], Ishii and Koike [IK9] prove existence, uniqueness and stability for the viscosity solutions (u, u 2 ) of (.2), but they do not consider any approximation of the system. As before, we introduce perturbed problems (see (3.2)) and show that, under the same assumptions of [EL9], the speed of convergence of the corresponding solutions (u, u 2) to (u, u 2 ) is O( /2 ) (see Theorem 3.6). We observe that the coupling assumptions here play a crucial role and cannot be replaced. For the sake of simplicity, we just focus on a system of two equations, but the general case can be treated in a similar way. Section 4 is devoted to an analog of the effective Hamiltonian problem (.) introduced by Lions, Papanicolaou and Varadhan [LPV88], which is the following weakly coupled system of

ADJOINT METHODS FOR OBSTACLE PROBLEMS AND WEAKLY COPLED SYSTEMS OF PDE 7 Hamilton Jacobi equations: c u c u 2 + H (x, Du ) = H in T n. (.3) c 2 u + c 2 u 2 + H 2 (x, Du ) = H 2 Here c and c 2 are positive constants and H, H 2 : T n R n R are smooth, while u, u 2 : T n R and H, H 2 R are unknowns. Systems of this type have been studied by Camilli, Loreti and Yamada in [CL08] and [CLY09], for uniformly convex Hamiltonians in a bounded domain. They arise naturally in optimal control and in large deviation theory for random evolution processes. As in [CL08], we assume the system to be quasi-monotone and not necessarily monotone (see condition (H4.2)). In this context, by monotonicity we mean exactly the coupling assumptions of Engler and Lenhart in [EL9]. Moreover, we also require H, H 2 to be coercive (see condition (H4.)). As it happens for the cell problem in the framework of weak KAM theory, there is no hope here of a general uniqueness result for (u, u 2 ), even modulo the addition of constants. In the spirit of Theorem., studying a perturbation of (.3) (see (4.2)) we prove that there exist H, H 2 R such that the system above admits viscosity solutions u, u 2. Notice that H, H 2 are not unique as well, but c 2 H + c H 2 is. More precisely, we show that there exists a unique µ R such that c 2 H + c H 2 = µ, for every pair (H, H 2 ) R 2 for which viscosity solutions u, u 2 exist (see Theorem 4.2). This result is the analog of Theorem. by Lions, Papanicolaou and Varadhan in the case of systems, Finally, we prove that the speed of convergence of (u, u 2) to (H, H 2 ) is O() (see Theorem 4.4), where u and u 2 are the solutions of the approximating problem (4.2). In Section 5, we conclude the paper with the study of weakly coupled systems of obstacle type, namely max{u u 2 ψ, u + H (x, Du )} = 0 in, max{u 2 u ψ 2, u 2 + H 2 (x, Du 2 )} = 0 in. (.4) Problems of this type appeared in [CDE84] and [CLY09]. Here H, H 2 : R n R and ψ, ψ 2 : R are smooth, with ψ, ψ 2 α > 0. In this case, although the two equations in (.4) are coupled just through the difference u u 2 (weakly coupled system), the problem turns out to be considerably more difficult than the corresponding scalar equation (.6). Indeed, we cannot show now the analogous of estimate (.9) as in Section 2. For this reason, the hypotheses we require are stronger than in the scalar case. Together with the usual hypotheses of coercivity and compatibility (see conditions (H5.2) and (H5.4)), we have

8 F. CAGNETTI, D. GOMES, AND H.V. TRAN to assume that H (x, ) and H 2 (x, ) are convex (see (H5.)), and we also ask that D x H and D x H 2 are bounded (see (H5.3)). We were not able to relax these conditions. We believe it would be also very interesting to apply the Adjoint Method in this particular system with the above conditions relaxed. 2. Obstacle problem In this section, we study the following obstacle problem max{u ψ, u + H(x, Du)} = 0 u = 0 in on, (2.) where ψ : R and H : R n R are smooth, with ψ 0 on. We also assume that ( ) (H2.) lim p + 2 H(x, H(x, p) p) 2 + D x H(x, p) p = lim = + uniformly in x ; p + p (H2.2) there exists a function Φ C 2 () C () such that Φ ψ on, Φ = 0 on and Φ + H(x, DΦ) < 0 in. We observe that in the classical case H(x, p) = H(p) + V (x) with H(p) lim = +, p + p or when H is superlinear in p and D x H(x, p) C( + p ), then we immediately have (H2.). Assumption (H2.2) (stating, in particular, that Φ is a sub-solution of (2.)), will be used to derive the existence of solutions of (2.), and to give a uniform bound for the gradient of solutions of the penalized equation below. 2.. The classical approach. For every > 0, the penalized PDE is the equation given by u + H(x, Du ) + γ (u ψ) = u in, (2.2) u = 0 on, where γ is defined by (.7). From [Lio82] it follows that under conditions (H2.) and (H2.2), for every > 0 there exists a smooth solution u to (2.2). The first result we establish is a uniform bound for the C -norm of the sequence {u }. Proposition 2.. There exists a positive constant C, independent of, such that u L, Du L C. In order to prove Proposition 2., we need the following fundamental lemma:

ADJOINT METHODS FOR OBSTACLE PROBLEMS AND WEAKLY COPLED SYSTEMS OF PDE 9 Lemma 2.2. There exists a constant C > 0, independent of, such that max γ (u ψ) C, x max x u ψ C. Proof. We only need to show that max x γ (u ψ) C, since then the second estimate follows directly by the definition of γ. Since u ψ 0 on, we have max x γ (u ψ) = 0. Now, if max x γ (u ψ) = 0, then we are done. Thus, let us assume that there exists x such that max x γ (u ψ) = γ (u ψ)(x ) > 0. Since γ is increasing, we also have max x (u ψ) = u (x ) ψ(x ). Thus, using (2.2), by the Maximum principle (u (x ) ψ(x )) + γ (u (x ) ψ(x )) = u (x ) H(x, Du (x )) ψ(x ) ψ(x ) H(x, Dψ(x )) ψ(x ). Since u (x ) ψ(x ) > 0, γ (u (x ) ψ(x )) max( ψ + H(x, Dψ) + ψ(x) ) C, x for any <, and this concludes the proof. Proof of Proposition 2.. Suppose there exists x 0 such that u (x 0 ) = max x u (x). Then, since u (x 0 ) 0 and using the fact that γ 0 u (x 0 ) = u (x 0 ) H(x 0, 0) γ (u (x 0 ) ψ(x 0 )) H(x 0, 0) max ( H(x, 0)) C. x Let now x be such that u (x ) = min x u (x ). Then, using Lemma 2.2, u (x ) = u (x ) H(x, 0) γ (u (x ) ψ(x )) H(x, 0) γ (u (x ) ψ(x )) min ( H(x, 0) γ (u (x) ψ(x))) C. x This shows that u L is bounded. To prove that Du L is bounded independently of, we first need to prove that Du L ( ) is bounded by constructing appropriate barriers. Let Φ be as in (H2.2). For small enough, we have that Φ + H(x, DΦ) + γ (Φ ψ) < Φ, and Φ = 0 on. Therefore, Φ is a sub-solution of (2.2). By the comparison principle, u Φ in.

0 F. CAGNETTI, D. GOMES, AND H.V. TRAN Let now d(x) = dist(x, ). It is well-known that for some δ > 0 d C 2 ( δ ) and Dd = in δ, where δ := {x : d(x) < δ}. For µ > 0 large enough, the uniform bound on u L yields v := µd u on δ. Assumption (H2.) then implies v + H(x, Dv) + γ (v ψ) v H(x, µdd) Cµ 0, for µ is sufficiently large. So the comparison principle gives us that Φ u v in δ. Thus, since ν is the outer unit normal to, and Φ = u = v = 0 on, we have v u Φ (x) (x) ν ν ν (x), for x. Hence, we obtain Du L ( ) C. Next, let us set w = Du 2. By a direct computation one can see that 2 2( + (γ ) )w + D p H Dw + D x H Du (γ ) Du Dψ = w D 2 u 2. (2.3) If Du L max( Dψ L, Du L ( )) then we are done. Otherwise, max( Dψ L, Du L ( )) < Du L. We can choose x 2 such that w (x 2 ) = max x w (x). Then, using (2.3) D 2 u 2 (x 2 ) = w (x 2 ) 2w (x 2 ) D x H(x 2, Du (x 2 )) Du (x 2 ) + (γ ) ( Du (x 2 ) Dψ(x 2 ) Du 2 (x 2 ) ) (2.4) Moreover, for sufficiently small we have D x H(x 2, Du (x 2 )) Du (x 2 ). D 2 u 2 (x 2 ) 2 u (x 2 ) 2 = [u (x 2 ) + H(x 2, Du (x 2 )) + γ (u (x 2 ) ψ(x 2 ))] 2 2 H(x 2, Du (x 2 )) 2 C, (2.5) where we use Lemma 2.2 for the last inequality. Collecting (2.4) and (2.5) 2 H(x 2, Du (x 2 )) 2 + D x H(x 2, Du (x 2 )) Du (x 2 ) C. Recalling hypothesis (H2.), we must have Du L = Du (x 2 ) C. Thanks to Proposition 2. one can show that, up to subsequences, u converges uniformly to a viscosity solution u of the obstacle problem (2.).

ADJOINT METHODS FOR OBSTACLE PROBLEMS AND WEAKLY COPLED SYSTEMS OF PDE 2.2. Proof of Theorem.3. We now study the speed of convergence. To prove our theorem we need several steps. Adjoint method: The formal linearized operator L : C 2 () C() corresponding to (2.2) is given by L z := ( + (γ ) )z + D p H Dz z. We will now introduce the adjoint PDE corresponding to L. Let x 0 be fixed. We denote by σ the solution of: ( + (γ ) )σ div(d p Hσ ) = σ + δ x0, in, σ = 0, on, (2.6) where δ x0 stands for the Dirac measure concentrated in x 0. In order to show existence and uniqueness of σ, we have to pass to a further adjoint equation. Let f C() be fixed. Then, we denote by v the solution to ( + (γ ) )v + D p H Dv = v + f, in, v = 0, on. (2.7) When f 0, by using the Maximum Principle one can show that v 0 is the unique solution to (2.7). Thus, by the Fredholm Alternative we infer that (2.6) admits a unique solution σ. Moreover, one can also prove that σ C ( \ {x 0 }). Some additional properties of σ are given by the following lemma. Lemma 2.3 (Properties of σ ). Let ν denote the outer unit normal to. Then, (i) σ 0 on. In particular, σ 0 on. ν (ii) The following equality holds: ( + (γ ) ) σ σ dx = + ν ds. In particular, σ ν ds. Proof. First of all, consider equation (2.7) and observe that f 0 = v 0. (2.8) Indeed, assume f 0 and let x be such that v(x) = min v(x). x

2 F. CAGNETTI, D. GOMES, AND H.V. TRAN We can assume that x, since otherwise clearly v 0. Then, for every x (( + (γ ) )v(x) = v(x) + f(x) 0, and (2.8) follows, since + (γ ) > 0. Now, multiply equation (2.6) by v and integrate by parts, obtaining fσ dx = v(x 0 ). Taking into account (2.8), from last relation we infer that and this implies σ 0. To prove (ii), we integrate (2.6) over, to get ( + (γ ) )σ dx = = fσ dx 0 for every f 0, div(d p Hσ ) dx + (D p H ν)σ ds + σ dx + σ ν ds + = σ ds +, ν where we used the fact that σ = 0 on. sing the adjoint equation, we have the following new estimate. Lemma 2.4. There exists C > 0, independent of > 0, such that ( + (γ ) ) Du 2 σ dx + D 2 u 2 σ dx C. (2.9) 2 Proof. Multiplying (2.3) by σ and integrating by parts, using equation (2.6) we get ( + (γ ) ) Du 2 σ dx + D 2 u 2 σ dx 2 = w (x 0 ) [D x H Du (γ ) Dψ Du ] σ dx w σ ν ds. Thanks to Lemma 2.3 and Proposition 2. the conclusion follows. Relation (2.9) shows that we have a good control of the Hessian D 2 u in the support of σ. We finally have the following result, which immediately implies Theorem.3. Lemma 2.5. There exists C > 0, independent of, such that max u (x) C x. /2

ADJOINT METHODS FOR OBSTACLE PROBLEMS AND WEAKLY COPLED SYSTEMS OF PDE 3 Proof. Differentiating (2.2) w.r.t. we get ( + (γ ) )u + D p H Du + γ = u + u, in. (2.0) In addition, we have u (x) = 0 for all x, since u (x) = 0 on for every. So, we may assume that there exists x 2 such that u (x 2 ) = max x u (x). Consider the adjoint equation (2.6), and choose x 0 = x 2. Multiplying by σ both sides of (2.0) and integrating by parts, u (x 2 ) = γσ dx + u σ dx. Hence, u (x 2 ) γ σ dx + u σ dx. (2.) By Lemma 2.2, ( ) γ = ψ u u 2 γ ψ = u ψ (γ ) (u ψ) C(γ ). (2.2) Hence, thanks to Lemma 2.3 γ σ dx C (γ ) σ dx C, (2.3) while using (2.9) ( u σ dx Thus, by (2.), (2.3) and (2.4) ( C ) /2 ( ) /2 u 2 σ dx σ dx ) /2 ( D 2 u 2 σ dx ) /2 σ dx C. /2 (2.4) u (x 2 ) C, for <. (2.5) /2 2.3. Dynamic interpretation. We give now a dynamic interpretation of the measure σ. Thanks to the properties given by Lemma 2.3, and arguing as in [Eva0] and [CGT], we have the following theorem. Theorem 2.6. There exist (i) a measure µ on R n, such that φ(x, Du ) σ dx φ(x, p) dµ, φ C( R n ); R n

4 F. CAGNETTI, D. GOMES, AND H.V. TRAN (ii) a measure γ on R n, and a measure γ 2 on R n, such that ( + (γ ) ) φ(x, Du ) σ dx φ(x, p) dγ, φ C( R n ), R n and σ ν φ(x, Du ) ds φ(x, p) dγ 2, φ C( R n ), R n with the property that γ ( R n )+γ 2 ( R n ) = ; (iii) a nonnegative definite matrix of measures (m jk ), that we call matrix of dissipation measures, such that φ(x, Du )u x ix j u x ix k σ dx φ(x, p) dm jk, R n φ C( R n ), j, k =,..., n, where we used Einstein summation convention; (iv) a compact set K R n such that supp µ, supp γ, supp m jk K, supp γ 2 K. Next theorem gives a relation involving the measures µ, γ, γ 2, m kj. Theorem 2.7. For any φ C ( R n ) with φ(x, ) C 2 (R n ) and for any x, we have lim φ(x 0, Du ( (x 0 )) = Dp φ (Dψ p) + φ ) dγ + φ dγ 2 0 + R n R n ( + {H, φ} Dp φ Dψ ) dµ φ pjp k dm jk, R n R n where the symbol {, } stands for the Poisson bracket, that is {F, G} := D p F D x G D x F D p G, F, G C ( R n ). In particular, if φ(x, p) = φ(x) then φ dγ + φ dγ 2 + {H, φ} dµ = φ(x 0 ). R n R n R n Proof. Let us set ϕ (x) := φ(x, Du (x)) for every x. Then ϕ x i = φ xi + φ pj u x ix j, i =,..., n, and ϕ = x φ + 2φ xip j u x ix j + D p φ D( u ) + φ pjp k u x ix j u x ix k, where we used the notation x φ = i φ x ix i. Differentiating (2.2) w.r.t. x and computing the scalar product by D p φ, we get ( + (γ ) )D p φ Du (γ ) D p φ Dψ + D x H D p φ + H pj φ pi u x ix j = D p φ D( u ).

ADJOINT METHODS FOR OBSTACLE PROBLEMS AND WEAKLY COPLED SYSTEMS OF PDE 5 Thanks to the above calculation on ϕ x i and ϕ, ( + (γ ) )D p φ Du (γ ) D p φ Dψ + D x H D p φ + H pj (ϕ x j φ xj ) = ( ϕ x φ 2φ xip j u x ix j φ pjp k u x ix j u x ix k ). Hence, adding and subtracting the term D p φ Dψ ( + (γ ) )D p φ (Du Dψ)+D p φ Dψ {H, φ}+d p H Dϕ = ϕ ( x φ + 2φ xip j u x ix j ) φ pjp k u x ix j u x ix k. (2.6) Multiplying (2.6) by σ and integrating by parts over, (D p φ (Du Dψ) ϕ ) ( + (γ ) )σ dx+ (D p φ Dψ {H, φ}) σ dx = ϕ σ ν ds ϕ div(d p Hσ ) dx ϕ σ dx + ( x φ + 2φ xip j u x ix j ) σ dx Recalling equation (2.6) and the definition of ϕ, (D p φ (Du Dψ) ϕ ) ( + (γ ) )σ dx 0 + + = ϕ ( + (γ ) )σ dx (D p φ Dψ {H, φ}) σ dx + φ(x 0, Du (x 0 )) + φ pjp k u x ix j u x ix k σ dx. ( x φ + 2φ xip j u x ix j ) σ dx φ pjp k u x ix j u x ix k σ dx. Thanks to Hölder inequality and using (2.9), lim ( x φ + 2φ xip j u x ix j ) σ dx lim 0 + ϕ σ ν ds (2.7) C( + D 2 u ) σ dx lim 0 + C = 0. (2.8) Letting 0 +, using Theorem 2.6 and relation (2.8) we finally get lim φ(x 0, Du ( (x 0 )) = Dp φ (Dψ p) + φ ) dγ + φ dγ 2 0 + R n R n ( + {H, φ} Dp φ Dψ ) dµ φ pjp k dm jk. R n R n 2.4. Stochastic process interpretation. In this section we show that the problem could have been approached also by using stochastic processes. Let (Ω, F, P ) be a probability space, and let w t be a n-dimensional Brownian motion on Ω. Let > 0, and let u be a solution of (2.2). Let T > 0 and consider the solution x : [0, T ] R n of the SDE dx = D p H(x, Du (x )) dt + 2 dw t, x (0) = x, (2.9)

6 F. CAGNETTI, D. GOMES, AND H.V. TRAN with x arbitrary. Accordingly, the momentum variable is defined as p (t) = Du (x (t)). We then define the exit time as T := inf{t (0, T ] : x (t) }. Let us now recall some basic facts about stochastic calculus. Suppose z : [0, T ] R n is a solution to the SDE: dz i = a i dt + b ij w j t, i =,..., n, with a i and b ij bounded and progressively measurable processes. Let ϕ : R n R R be a smooth function. Then, ϕ(z, t) satisfies the Itô formula: dϕ = ϕ zi dz i + (ϕ t + 2 ) b ijb jk ϕ zizk dt. (2.20) An integrated version of the Itô formula is the Dynkin s formula: E [φ(z(t )) φ(z(0))] = E In the present situation, we have [ T 0 ( a i D zi φ(z(t)) + ) ] 2 b ijb jk Dz 2 iz k φ(z(t)) dt. a i = D pi H(x, D x u ), b ij = 2δ ij. Hence, recalling (2.9) and (2.20) dp = D 2 u dx + (Du )dt ( ) = D 2 u D p H + (Du ) dt + 2D 2 u dw t ( ) = D x H + ( + (γ ) )p (γ ) Dψ dt + 2D 2 u dw t, where in the last equality we used the identity obtained by differentiating (2.2) with respect to x. Thus, (x, p ) satisfies the following stochastic version of the Hamiltonian dynamics: dx = D p H(x, p ) dt + 2 dw t, ( ) dp = D x H(x, p ) + ( + (γ ) )p (γ ) Dψ dt + 2D 2 u dw t. (2.2) We are now going to study the behavior of the solutions u of equation (2.2) along the trajectory (x ( ), p ( )). Let φ C ( R n ). Thanks to the Itô formula and (2.2), the differential of the

ADJOINT METHODS FOR OBSTACLE PROBLEMS AND WEAKLY COPLED SYSTEMS OF PDE 7 function φ(x ( ), p ( )) is given by: ) dφ = D x φ dx + D p φ dp + ( x φ + 2φ xip j u xixj + φ pjp k u xixj u xixk dt ( ) = {φ, H} + ( + (γ ) )D p φ p (γ ) D p φ Dψ dt ) + ( x φ + 2φ xip j u xixj + φ pjp k u xixj u xixk dt + 2 ( D x φ + D 2 u D p φ ) dw t ( ) = {φ, H} + ( + (γ ) )D p φ (p Dψ) + D p φ Dψ dt ) + ( x φ + 2φ xip j u xixj + φ pjp k u xixj u xixk dt + 2 ( D x φ + D 2 u D p φ ) dw t. Thus, by Dynkin s formula we have the following equality E [φ(x (T ), p (T )) φ(x, Du (x))] [ ] T ) = E ({φ, H} + D p φ Dψ + x φ + 2φ xip j u xixj + φ pjp k u xixj u xixk dt 0 [ ] T + E ( + (γ ) )D p φ (p Dψ) dt, 0 which we may also write as [ ] T E (D p φ (p Dψ) φ) ( + (γ ) ) dt 0 + E + E = E [ T 0 [ T 0 [ T 0 ] ( ) D p φ Dψ {H, φ} dt + φ(x 0, Du (x 0 )) φ ( + (γ ) ) dt ] E [φ(x (T ), p (T ))] ] ) ( x φ + 2φ xip j u xixj + φ pjp k u xixj u xixk dt. (2.22) Relation (2.22) is the analogous of (2.7), and can be as well used (together with suitable estimates) to prove Theorem.3. 3. Weakly coupled systems of Hamilton Jacobi equations We study now the model of monotone weakly coupled systems of Hamilton-Jacobi equations considered by Engler and Lenhart [EL9], and by Ishii and Koike [IK9]. For the sake of simplicity, we will just focus on the following system of two equations: c u + c 2 u 2 + H (x, Du ) = 0 c 2 u + c 22 u 2 + H 2 (x, Du 2 ) = 0 in, (3.) with boundary conditions u = u 2 = 0 on. The general case of more equations or arbitrary boundary data can be treated in a similar way. We assume that the Hamiltonians H, H 2 : R n R are smooth satisfying

8 F. CAGNETTI, D. GOMES, AND H.V. TRAN (H3.) ( ) lim p + 2 H j(x, p) 2 H j (x, p) + D x H j (x, p) p = lim = + uniformly in x, p + p for every j =, 2. Following [EL9] and [IK9], we suppose further that (H3.2) c 2, c 2 0; (H3.3) there exists α > 0 such that c + c 2, c 2 + c 22 α > 0. We observe that, as a consequence, we also have c, c 22 > 0. Finally, we require that (H3.4) There exist Φ, Φ 2 C 2 () C () with Φ j = 0 on (j =, 2), and such that c Φ + c 2 Φ 2 + H (x, DΦ ) < 0 in, c 22 Φ 2 + c 2 Φ + H 2 (x, DΦ 2 ) < 0 in. Thanks to these conditions, the Maximum Principle can be applied and existence, comparison and uniqueness results hold true, as stated in [EL9]. We consider now the following regularized system (here > 0): c u + c 2 u 2 + H (x, Du ) = u in, (3.2) c 2 u + c 22 u 2 + H 2 (x, Du 2) = u 2 with boundary conditions u = u 2 = 0 on. Conditions (H3.), (H3.2), and (H3.3) yield the existence and uniqueness of the pair of solutions (u, u 2) in (3.2). Lemma 3.. There exists a positive constant C, independent of, such that u L, u 2 L C. Proof. First of all observe that u = u 2 = 0 on for every. Thus, it will be sufficient to show that u and u 2 are bounded in the interior of. Let us assume that there exists x such that We have max u j(x) = u (x). j=,2 x αu (x) c u (x) + c 2 u 2(x) H (x, 0) max ( H (x, 0)), x where we used (H3.3). Analogously, if x is such that then min u j(x) = u ( x), j=,2 x u ( x) c c + c 2 u ( x) + c 2 u c + c 2( x) H ( x, 0) 2 c + c 2 min c + c 2 x ( H (x, 0)).

ADJOINT METHODS FOR OBSTACLE PROBLEMS AND WEAKLY COPLED SYSTEMS OF PDE 9 Lemma 3.2. There exists a positive constant C, independent of, such that Du L, Du 2 L C. Proof. We will argue as in the proof of Proposition 2.. Step I: Bound on. We shall first show that max Du j(x) C, j=,2 x for some constant C independent of. As it was done in Section 2, we are going to construct appropriate barriers. For small enough, assumption (H3.4) implies that c Φ + c 2 Φ 2 + H (x, DΦ ) < Φ in, c 22 Φ 2 + c 2 Φ + H 2 (x, DΦ 2 ) < Φ 2 in, and Φ = Φ 2 = 0 on. Therefore, (Φ, Φ 2 ) is a sub-solution of (3.2). By the comparison principle, u j Φ j in, j =, 2. Let d(x), δ, and δ be as in the proof of Proposition 2.. For µ > 0 large enough, the uniform bounds on u L and u 2 L yield v := µd u j on δ, j =, 2, so that (c + c 2 )v + H (x, Dv) v H (x, µdd) µc in, (c 2 + c 22 )v + H 2 (x, Dv) v H 2 (x, µdd) µc in. Now, we have Φ j = u j = v = 0 on. Also, thanks to assumption (H3.), for µ > 0 large enough (c + c 2 )v + H (x, Dv) v 0 in, (c 2 + c 22 )v + H 2 (x, Dv) v 0 in, that is, the pair (v, v) is a super-solution for the system (3.2). Thus, the comparison principle gives us that Φ j u j v in δ, j =, 2. Then, from the fact that Φ j = u j = v = 0 on we get v ν (x) u j ν (x) Φ j ν Hence, we obtain Du j L ( ) C, j =, 2. (x), for x, j =, 2. Step II: Bound on. Setting wj = Du j 2, j =, 2, by a direct computation we have that 2 2c w + D p H Dw + c 2 Du Du 2 + D x H Du = w D 2 u 2, 2c 22 w2 + D p H 2 Dw2 + c 2 Du Du 2 + D x H 2 Du 2 = w2 D 2 u 2 2. (3.3)

20 F. CAGNETTI, D. GOMES, AND H.V. TRAN Assume now that there exists x such that Then, we have max wj(x) = w( x). j=,2 x D 2 u 2 ( x) = w ( x) 2c w ( x) c 2 Du ( x) Du 2( x) D x H Du ( x) Now, for sufficiently small 2(c + c 2 )w ( x) D x H Du ( x) D x H Du ( x). D 2 u ( x) 2 2 u ( x) 2 = [c u ( x) + c 2 u 2( x) + H ( x, Du ( x))] 2 2 H ( x, Du ( x)) 2 C. Collecting the last two relations we have 2 H ( x, Du ( x)) 2 + D x H ( x, Du ( x)) Du ( x) C. Recalling condition (H3.) the conclusion follows. Adjoint method. At this point, we introduce the adjoint of the linearization of system (3.2). The linearized operator corresponding to (3.2) is D p H (x, Du ) Dz + c z + c 2 z 2 z, L (z, z 2 ) := D p H 2 (x, Du 2) Dz 2 + c 22 z 2 + c 2 z z 2. Let us now identify the adjoint operator (L ). For every ν, ν 2 C c () we have (L ) (ν, ν 2 ), (z, z 2 ) := (ν, ν 2 ), L (z, z 2 ) = ν, [L (z, z 2 )] + ν 2, [L (z, z 2 )] 2 = [D p H (x, Du ) Dz + c z + c 2 z 2 z ] ν dx + = + [D p H 2 (x, Du 2) Dz 2 + c 22 z 2 + c 2 z z 2 ] ν 2 dx [ div(dp H ν ) + c ν + c 2 ν 2 ν ] z dx [ div(dp H 2 ν 2 ) + c 22 ν 2 + c 2 ν ν 2] z 2 dx. Then, the adjoint equations are: div(d p H σ, ) + c σ, + c 2 σ 2, = σ, + (2 i)δ x0 in, div(d p H 2 σ 2, ) + c 22 σ 2, + c 2 σ, = σ 2, + (i )δ x0 in, (3.4)

ADJOINT METHODS FOR OBSTACLE PROBLEMS AND WEAKLY COPLED SYSTEMS OF PDE 2 with boundary conditions σ, = 0 on, σ 2, = 0 on, where i {, 2} and x 0 will be chosen later. Existence and uniqueness of σ, and σ 2, follow by Fredholm alternative, by arguing as in Section 2, and we have σ,, σ 2, C ( \ {x 0 }). We study now further properties of σ, and σ 2,. Lemma 3.3 (Properties of σ,, σ 2, ). Let ν be the outer unit normal to. Then (i) σ j, 0 on. In particular, σj, 0 on (j =, 2). ν (ii) The following equality holds: 2 ( (c j + c j2 )σ j, σ j, ) dx ν ds =. In particular, j= 2 j= (c j + c j2 )σ j, dx. Proof. First of all, we consider the adjoint of equation (3.4): D p H (x, Du ) Dz + c z + c 2 z 2 z = f, D p H 2 (x, Du 2) Dz 2 + c 22 z 2 + c 2 z z 2 = f 2, (3.5) where f, f 2 C(), with boundary conditions z = z 2 = 0 on. Note that f, f 2 0 = min z j (x) 0. (3.6) Indeed, if the minimum is achieved for some x, then clearly z, z 2 0. Otherwise, assume for some x. sing condition (H3.2) Thanks to (H3.3), (3.6) follows. j=,2 x min z j (x) = z (x), j=,2 x (c + c 2 )z (x) c z (x) + c 2 z 2 (x) = z (x) + f (x) 0. Let us now multiply (3.4) and (3.4) 2 by the solutions z and z 2 of (3.5). relations obtained we have f σ, dx + f 2 σ 2, dx = (2 i)z (x 0 ) + ( i)z 2 (x 0 ). Thanks to (3.6), from last relation we conclude that f σ, dx + f 2 σ 2, dx 0, for every f, f 2 0, Adding up the

22 F. CAGNETTI, D. GOMES, AND H.V. TRAN and this implies that σ,, σ 2, 0. To prove (ii), it is sufficient to integrate equations (3.4) and (3.4) 2 over, and to add up the two relations obtained. Lemma 3.4. There exists a constant C > 0, independent of, such that D 2 u 2 σ, dx + D 2 u 2 2 σ 2, dx C. Proof. To show the lemma, one has first to multiply equations (3.3) and (3.3) 2 by σ, and σ 2, respectively. Then, adding up the relations obtained and using (3.4), thanks to Lemma 3. and Lemma 3.2 the conclusion follows. We now give the last lemma needed to estimate the speed of convergence. Here we use the notation u j, (x) := u j (x)/, j =, 2. Lemma 3.5. There exists a constant C > 0, independent of, such that max j=,2 x u j,(x) C /2. Proof. Differentiating (3.2) w.r.t we obtain the system c u, + c 2 u 2, + D p H Du, = u, + u, c 2 u, + c 22 u 2, + D p H 2 Du 2, = u 2, + u 2. (3.7) Since u, = u 2, = 0 on, we have Assume now that there exists x such that max x u,(x) = max x u 2,(x) = 0. max u j,(x) = u,( x), j=,2 x and let σ,, σ 2, be the solutions of system (3.4) with i = and x 0 = x. Multiplying equations (3.7) and (3.7) 2 by σ, and σ 2, respectively and adding up, thanks to (3.4) we obtain u,( x) = u σ, dx + u 2 σ 2, dx. Thanks to Lemma 3.4, and repeating the chain of inequalities in (2.4) one can show that u j σ j, dx C, j =, 2, /2 and from this the conclusion follows. We can now prove the following result on the speed of convergence.

ADJOINT METHODS FOR OBSTACLE PROBLEMS AND WEAKLY COPLED SYSTEMS OF PDE 23 Theorem 3.6. There exists C > 0, independent of, such that u u L, u 2 u 2 L C /2. Proof. The theorem is a direct consequence of Lemma 3.5. 4. Cell problem for Weakly coupled system of Hamilton Jacobi equations In this section, we study the following weakly coupled systems of Hamilton Jacobi equations: c u c u 2 + H (x, Du ) = H in T n, H, H 2 R, (4.) c 2 u + c 2 u 2 + H 2 (x, Du 2 ) = H 2 which is the analog of the cell problem for single equation introduced by Lions, Papanicolaou, and Varadhan [LPV88]. We will assume that H, H 2 C (T n R n ), and: ( ) (H4.) lim p + 2 H j(x, p) 2 + D x H j (x, p) p 8nc 2 j p 2 = + uniformly in x T n, j =, 2; (H4.2) c, c 2 > 0. It is easy to see that the coefficients of u, u 2 in this system do not satisfy the coupling assumptions of the previous section. Indeed, as it happens for the cell problem in the context of weak KAM theory, there is no hope of a uniqueness result for (4.). To find the effective Hamiltonians H, H 2 we use the same arguments as in [Tra]. First, for every > 0, let us consider the following regularized system: (c + )u c u 2 + H (x, Du ) = 2 u (c 2 + )u 2 c 2 u + H 2 (x, Du 2) = 2 u 2 in T n. (4.2) For every > 0 fixed, the coefficients of this new system satisfy the coupling assumptions (H3.2) and (H3.3) of the previous section. Thus, (4.2) admits a unique pair of smooth solutions u, u 2. In particular, this implies that u and u 2 are T n -periodic. The following result gives some a priori estimates. Theorem 4.. There exists C > 0, independent of, such that u L, u 2 L, Du L, Du 2 L C. Proof. Our proof is based on the Maximum Principle. Without loss of generality, we may assume that { max u j (x) } = u (x 0). j=,2 x T n for some x 0 T n. Applying the Maximum Principle to the first equation of (4.2), u (x 0) (c + )u (x 0) c u 2(x 0) H (x 0, 0) C, (4.3)

24 F. CAGNETTI, D. GOMES, AND H.V. TRAN and this shows the existence of a bound from above for u and u 2. sing a similar argument one can show that there is also a bound from below, so that u L, u 2 L C. (4.4) We observe that the previous inequality doesn t provide any bound for the difference u (x) u 2(x) in a generic point x T n. Nevertheless, thanks to (4.4) we have u (x 0) u 2(x 0) c H (x 0, 0) c u (x 0) C. (4.5) Then, (4.5) and (4.4) imply that c u (x 0) u 2(x 0) C. (4.6) In order to find a bound for the gradients, let us set wj = Du j 2, j =, 2. Then, by a direct 2 computation one can see that 2(c + )w + D p H Dw c Du Du 2 + D x H Du = 2 w 2 D 2 u 2 in T n. 2(c 2 + )w2 + D p H 2 Dw2 c 2 Du Du 2 + D x H 2 Du 2 = 2 w2 2 D 2 u 2 2 Without loss of generality, we may assume that there exists x T n such that Then, by the Maximum Principle { max w j (x) } = w(x ). j=,2 x T n 2 D 2 u (x ) 2 2(c + )w (x ) + c Du (x ) Du 2(x ) D x H Du (x ) D x H Du (x ). (4.7) Moreover, for sufficiently small 2 D 2 u (x ) 2 4 ( u (x )) 2 = [H (x, Du (x )) + (c + )u (x ) c u 2(x )] 2. (4.8) Also, thanks to (4.4) and (4.6) (c + )u (x ) c u 2(x ) u (x ) + c u (x ) u (x 0) + c u 2(x ) u 2(x 0) + c u (x 0) u 2(x 0) C + c u (x ) u (x 0) + c u 2(x ) u 2(x 0) C + 2c Du (x ) x x 0 C + 2c n Du (x ),

ADJOINT METHODS FOR OBSTACLE PROBLEMS AND WEAKLY COPLED SYSTEMS OF PDE 25 where we used the fact that the diameter of T n is n. Last relation, together with (4.8), gives that 2 D 2 u (x ) 2 [H (x, Du (x )) + (c + )u (x ) c u 2(x )] 2 2 H (x, Du (x )) 2 C 8nc 2 Du (x ) 2. sing last inequality and (4.7) we have 2 H (x, Du (x )) 2 + D x H Du (x ) 8nc 2 Du (x ) 2 C. This, thanks to condition (H4.), gives the conclusion. Thanks to Theorem 4., up to subsequences, u H, u 2 H 2, uniformly in T n, for some constants H, H 2 R. Furthermore, still up to subsequences, u min T n u u, u 2 min T n u 2 u 2, uniformly in T n, where u and u 2 are viscosity solutions of (4.). In general, H and H 2 are not unique. Indeed, let u, u 2 be viscosity solutions of (4.). Then, for every pair of constants C, C 2, the functions ũ := u + C and ũ 2 := u 2 + C 2 are still viscosity solutions of (4.), with new effective Hamiltonians H = H + c (C C 2 ), H2 = H 2 + c 2 (C 2 C ). Anyway, we have c 2 H + c H 2 = c 2 H + c H2. This suggests that, although H and H 2 may vary, the expression c 2 H + c H 2 is unique. Next theorem shows that this is the case. Theorem 4.2. There exists a constant µ R such that c 2 H + c H 2 = µ, for every pair (H, H 2 ) R 2 such that the system (4.) admits viscosity solutions u, u 2. Proof. Without loss of generality, we may assume c = c 2 =. Suppose, by contradiction, that there exist two pairs (λ, λ 2 ) R 2 and (µ, µ 2 ) R 2, and four functions u, u 2, ũ, ũ 2 C(T n ) such that λ + λ 2 < µ + µ 2 and u u 2 + H (x, Du ) = λ in T n, u + u 2 + H 2 (x, Du 2 ) = λ 2

26 F. CAGNETTI, D. GOMES, AND H.V. TRAN and ũ ũ 2 + H (x, Dũ ) = µ in T n. ũ + ũ 2 + H 2 (x, Dũ 2 ) = µ 2 By possibly substituting u and u 2 with functions û := u + C and û 2 := u 2 + C 2, for suitable constants C and C 2, we may always assume that λ < µ, λ 2 < µ 2. In the same way, by a further substitution u := u + C 3, u 2 := u 2 + C 3, with C 3 > 0 large enough, we may assume that u > ũ, u 2 > ũ 2. Then, there exists > 0 small enough such that ( + )u u 2 + H (x, Du ) < ( + )ũ ũ 2 + H (x, Dũ ) in T n. ( + )u 2 u + H 2 (x, Du 2 ) < ( + )ũ 2 ũ + H 2 (x, Dũ 2 ) Observe that the coefficients of the last system satisfy the coupling assumptions (H3.2) and (H3.3). Hence, applying the comparison theorem in [EL9] and [IK9], we conclude that u < ũ and u 2 < ũ 2, which gives a contradiction. In the sequel, all the functions will be regarded as functions defined in the whole R n Z n -periodic. Next lemma provides some a priori bounds on u and u 2. and Lemma 4.3. There exists a constant C > 0, independent of, such that u (x) u (y), u 2(x) u 2(y), u (x) u 2(y) C, x, y R n. Proof. The first two inequalities follow from the periodicity of u and u 2, and from the fact that Du and Du 2 are bounded. Let us now show the last inequality. Without loss of generality, we may assume that there exists x 0 T n such that { max u j (x) } = u (x 0). j=,2 x T n Combining the second inequality of the lemma with (4.5), u (x) u 2(y) u (x 0) u 2(x 0) + u 2(x 0) u 2(y) C, x, y R n. The proof can be concluded by repeating the same argument for min j=,2 x T n { u j (x) }. The following is the main theorem of the section. See also [Tra] for similar results. Theorem 4.4. There exists a constant C > 0, independent of, such that u H L, u 2 H 2 L C.

ADJOINT METHODS FOR OBSTACLE PROBLEMS AND WEAKLY COPLED SYSTEMS OF PDE 27 Adjoint method: Also in this case, we introduce the adjoint equations associated to the linearization of the original problem. We look for σ,, σ 2, which are T n -periodic and such that div(d p H σ, ) + (c + )σ, c 2 σ 2, = 2 σ, + (2 i)δ x0 in T n, (4.9) div(d p H 2 σ 2, ) + (c 2 + )σ 2, c σ, = 2 σ 2, + (i )δ x0 in T n, where i {, 2} and x 0 T n will be chosen later. The argument used in Section 2 gives also in this case existence and uniqueness for σ, and σ 2,. As before, we also have σ,, σ 2, C (T n \{x 0 }). The next two lemmas can be proven by using the same ways as in Lemma 3.3 and Lemma 3.4. Lemma 4.5 (Properties of σ,, σ 2, ). The functions σ,, σ 2, satisfy the following: (i) σ j, 0 on T n (j =, 2); (ii) Moreover, the following equality holds: 2 j= T n σ j, dx =. Lemma 4.6. There exists a constant C > 0, independent of, such that R 2 D 2 u 2 σ, dx C, n 2 R n D 2 u 2 2 σ 2, dx C. Finally, next lemma allows us to prove Theorem 4.4. Lemma 4.7. There exists a constant C > 0, independent of, such that max (u ), max 2) C. T n T n Proof. Differentiating (4.2) w.r.t., D p H Du, + (c + )u, + u c u 2, = 2 u, + 2 u, D p H 2 Du 2, + (c 2 + )u 2, + u 2 c 2 u, = 2 u 2, + 2 u 2, where we set u j, := u j /, j =, 2. Without loss of generality, we may assume that there exists x 2 T n such that (u { j(x)) = max u j, (x) + u j(x) } = u,(x 2 ) + u (x 2 ). j=,2 x T n x T n max j=,2 Choosing x 0 = x 2 in the adjoint equation (4.9), and repeating the same steps as in Theorem 2.5, we get u,(x 2 ) + u σ, dx + u 2σ 2, dx T n T n 2 u σ, dx + 2 T n T n u 2 σ 2, dx C, (4.0)

28 F. CAGNETTI, D. GOMES, AND H.V. TRAN where the latter inequality follows by repeating the chain of inequalities in (2.4) and thanks to Lemma 4.6. sing Lemma 4.3 and property (ii) of Lemma 4.5 we have u (x)σ, dx + u 2(x)σ 2, dx u (x 2 ) T n T n = (u (x) u (x 2 ))σ, dx + (u 2(x) u (x 2 ))σ 2, dx C. T n T n In view of the previous inequality, (4.0) becomes u,(x 2 ) + u (x 2 ) C, thus giving the bound from above. The same argument, applied to min j=,2 (u j (x)), allows to x T n prove the bound from below. Proof of Theorem 4.4. The theorem immediately follows by using Lemma 4.7. 5. weakly coupled systems of obstacle type In this last section we apply the Adjoint Method to weakly coupled systems of obstacle type. Let H, H 2 : R n R be smooth Hamiltonians, and let ψ, ψ 2 : R be smooth functions describing the obstacles. We assume that there exists α > 0 such that ψ, ψ 2 α in, (5.) and consider the system max{u u 2 ψ, u + H (x, Du )} = 0 in, max{u 2 u ψ 2, u 2 + H 2 (x, Du 2 )} = 0 in, (5.2) with boundary conditions u = u 2 = 0. We observe that (5.) guarantees the compatibility of the boundary conditions, since ψ, ψ 2 > 0 on. Although the two equations in (5.3) are coupled just through the difference u u 2, this problem turns out to be more difficult that the correspondent scalar equation (2.) studied in Section 2. For this reason, the hypotheses we require now are stronger. We assume that (H5.) H j (x, ) is convex for every x, j =, 2. (H5.2) Superlinearity in p: H j (x, p) lim = + uniformly in x, j =, 2. p p (H5.3) D x H j (x, p) C for each (x, p) R n, j =, 2.

ADJOINT METHODS FOR OBSTACLE PROBLEMS AND WEAKLY COPLED SYSTEMS OF PDE 29 (H5.4) There exist Φ, Φ 2 C 2 () C () with Φ j = 0 on (j =, 2), ψ 2 Φ Φ 2 ψ, and such that Φ j + H j (x, DΦ j ) < 0 in (j =, 2). Let > 0 and let γ : R [0, + ) be the function defined by (.7). We make in this section the additional assumption that γ is convex. We approximate (5.3) by the following system u + H (x, Du ) + γ (u u 2 ψ ) = u in, u 2 + H 2 (x, Du 2) + γ (u 2 u ψ 2 ) = u 2 in. (5.3) We are now ready to state the main result of the section. Theorem 5.. There exists a positive constant C, independent of, such that max{ u u L, u 2 u 2 L } C /2. In order to prove the theorem we need several lemmas. In the sequel, we shall use the notation θ := u u 2 ψ, θ2 := u 2 u ψ 2. The linearized operator corresponding to (5.3) is z + D p H (x, Du ) Dz + (γ ) θ L (z z 2 ) z, (z, z 2 ) := z 2 + D p H 2 (x, Du 2) Dz 2 + (γ ) θ 2 (z 2 z ) z 2. Then, the adjoint equations are: ( + (γ ) θ )σ, div(d p H σ, ) (γ ) θ 2 σ 2, = σ, + (2 i)δ x0 in, ( + (γ ) θ 2 )σ 2, div(d p H 2 σ 2, ) (γ ) θ σ, = σ 2, + (i )δ x0 in, (5.4) with boundary conditions σ, = 0 on, σ 2, = 0 on, where i {, 2} and x 0 will be chosen later. By repeating what was done in Section 2, we get the existence and uniqueness of σ, and σ 2, by Fredholm alternative. Furthermore, σ, and σ 2, are well defined and σ,, σ 2, C ( \ {x 0 }). In order to derive further properties of σ, and σ 2,, we need the following useful formulas. Lemma 5.2. For every ϕ, ϕ 2 C 2 () we have σ, (2 i)ϕ (x 0 ) = ν ϕ ds + (γ ) θ 2 ϕ σ 2, dx [ ( + (γ ) θ ) ϕ + D p H Dϕ ϕ ] σ, dx, (5.5)

30 F. CAGNETTI, D. GOMES, AND H.V. TRAN and (i )ϕ 2 (x 0 ) = + where ν is the outer unit normal to. σ 2, ν ϕ 2 ds (γ ) θ ϕ 2 σ, dx [ ( + (γ ) θ 2 ) ϕ 2 + D p H 2 Dϕ 2 ϕ 2 ] σ 2, dx, (5.6) Proof. The conclusion follows by simply multiplying by ϕ j (j =, 2) the two equations in (5.4) and integrating by parts. We can now prove the analogous of Lemma 2.3. Lemma 5.3 (Properties of σ,, σ 2, ). Let ν be the outer unit normal to. Then (i) σ j, 0 on. In particular, σj, 0 on (j =, 2). ν (ii) The following equality holds: 2 ( σ j, σ j, ) dx ν ds =. In particular, j= 2 j= σ j, dx. Proof. First of all, we consider the adjoint of equation (5.4): z + D p H (x, Du ) Dz + (γ ) θ (z z 2 ) z = f, z 2 + D p H 2 (x, Du 2) Dz 2 + (γ ) θ 2 (z 2 z ) z 2 = f 2, (5.7) where f, f 2 C(), with boundary conditions z = z 2 = 0 on. Note that Indeed, assume for some x. Then, f, f 2 0 = min z j (x) 0. (5.8) j=,2 x min z j (x) = z (x), j=,2 x z (x) z (x) + (γ ) θ (z (x) z 2 (x)) = z (x) + f (x) 0. Adding up relations (5.5) and (5.6) with ϕ = z and ϕ 2 = z 2 we get f σ, dx + f 2 σ 2, dx = (2 i)z (x 0 ) + ( i)z 2 (x 0 ). Thanks to (5.8), from last relation we conclude that f, f 2 0 = f σ, dx + f 2 σ 2, dx 0,

ADJOINT METHODS FOR OBSTACLE PROBLEMS AND WEAKLY COPLED SYSTEMS OF PDE 3 and this implies that σ,, σ 2, 0. The proof of property (ii) follows by choosing ϕ = ϕ 2 in (5.5) and (5.6), and summing up the relations obtained. We are now able to prove a uniform bound on u and u 2. Lemma 5.4. There exists a positive constant C, independent of, such that u L, u 2 L C. Proof. If the maximum max j=,2 {u j (x)} is attained on the boundary, then x max{u j(x)} = max{u j(x)} = 0. j=,2 x j=,2 x Otherwise, assume that there exists x be such that Then, max{u j(x)} = u (x). j=,2 x u (x) = u (x) H (x, 0) γ (u (x) u 2(x) ψ (x)) H (x, 0) sup ( H (x, 0)) C. x Analogously, to prove the lower bound suppose that min {u j(x)} = u ( x), j=,2 x for some x. Then, u ( x) u 2( x) ψ ( x) < 0 and so u ( x) = u ( x) H ( x, 0) γ (u ( x) u 2( x) ψ ( x)) = u ( x) H ( x, 0) H ( x, 0) inf x ( H (x, 0)) C. Next lemma will be used to give a uniform bound for Du and Du 2. Lemma 5.5. We have (γ ) θ σ, dx + (γ ) θ 2 σ 2, dx C, where C is a positive constant independent of.

32 F. CAGNETTI, D. GOMES, AND H.V. TRAN Proof. First of all, observe that condition (H5.) implies that H j (x, p) D p H j (x, p) p H j (x, 0), for every (x, p) R n, j =, 2. (5.9) In the same way, the convexity of γ implies Equation (5.3) gives ( s [ ( s )] s γ (s) [(γ ) (s)] s = γ (γ ) ) γ(0) = 0. (5.0) 0 = u + H (x, Du ) + γ θ u = u + D p H (x, Du ) Du u + H (x, Du ) D p H (x, Du ) Du + γ θ (γ ) θ θ + (γ ) θ (u u 2) (γ ) θ ψ. Multiplying last relation by σ,, integrating and using (5.9) and (5.0) (γ ) θ ψ σ, dx = [H (x, Du ) D p H (x, Du ) Du ] σ, dx + + + [ γ θ (γ ) θ θ ] σ, dx [( + (γ ) θ ) u + D p H (x, Du ) Du u (γ ) θ u 2] σ, dx H (x, 0) σ, dx [( + (γ ) θ ) u + D p H (x, Du ) Du u (γ ) θ u 2] σ, dx. Analogously, (γ ) θ 2 ψ 2 σ 2, dx + H 2 (x, 0) σ 2, dx [( + (γ ) θ 2 ) u 2 + D p H 2 (x, Du 2) Du 2 u 2 (γ ) θ 2 u 2] σ 2, dx. Summing up the last two relations and using (5.5) and (5.6) (γ ) θ ψ σ, dx + (γ ) θ 2 ψ 2 σ 2, dx (2 i)u (x 0 ) + (i )u 2(x 0 ) Thus, + H (, 0) L (γ ) θ σ, dx + + H (, 0) L α σ, dx + H 2 (, 0) L σ 2, dx. (γ ) θ 2 σ 2, dx 2 i α u (x 0 ) + i α u 2(x 0 ) σ, dx + H 2(, 0) L α where we used (5.), Lemma 5.3 and Lemma 5.4. σ 2, dx C, We can finally show the existence of a uniform bound for the gradients of u and u 2.