Partial Differential Equations 2 Variational Methods

Similar documents
LECTURE 1: SOURCES OF ERRORS MATHEMATICAL TOOLS A PRIORI ERROR ESTIMATES. Sergey Korotov,

Lectures on. Sobolev Spaces. S. Kesavan The Institute of Mathematical Sciences, Chennai.

Traces, extensions and co-normal derivatives for elliptic systems on Lipschitz domains

Partial Differential Equations

LECTURE # 0 BASIC NOTATIONS AND CONCEPTS IN THE THEORY OF PARTIAL DIFFERENTIAL EQUATIONS (PDES)

Second Order Elliptic PDE

Sobolev spaces. May 18

Functional Analysis. Franck Sueur Metric spaces Definitions Completeness Compactness Separability...

2. Function spaces and approximation

Sobolev Spaces. Chapter Hölder spaces

Sobolev Spaces. Chapter 10

We denote the space of distributions on Ω by D ( Ω) 2.

Numerical Solutions to Partial Differential Equations

4 Sobolev spaces, trace theorem and normal derivative

CONVERGENCE THEORY. G. ALLAIRE CMAP, Ecole Polytechnique. 1. Maximum principle. 2. Oscillating test function. 3. Two-scale convergence

Unbounded operators on Hilbert spaces

Laplace s Equation. Chapter Mean Value Formulas

Contents: 1. Minimization. 2. The theorem of Lions-Stampacchia for variational inequalities. 3. Γ -Convergence. 4. Duality mapping.

i=1 α i. Given an m-times continuously

Fourier Transform & Sobolev Spaces

t y n (s) ds. t y(s) ds, x(t) = x(0) +

Green s Functions and Distributions

Fact Sheet Functional Analysis

Notes on Sobolev Spaces A. Visintin a.a

Overview of normed linear spaces

Friedrich symmetric systems

Applied Analysis (APPM 5440): Final exam 1:30pm 4:00pm, Dec. 14, Closed books.

Chapter 1 Foundations of Elliptic Boundary Value Problems 1.1 Euler equations of variational problems

A metric space X is a non-empty set endowed with a metric ρ (x, y):

Lecture 17. Higher boundary regularity. April 15 th, We extend our results to include the boundary. Let u C 2 (Ω) C 0 ( Ω) be a solution of

THEOREMS, ETC., FOR MATH 515

A Concise Course on Stochastic Partial Differential Equations

Partial Differential Equations and Sobolev Spaces MAT-INF4300 autumn Snorre H. Christiansen November 10, 2016

The Dirichlet s P rinciple. In this lecture we discuss an alternative formulation of the Dirichlet problem for the Laplace equation:

Variational Formulations

An introduction to some aspects of functional analysis

Finite-dimensional spaces. C n is the space of n-tuples x = (x 1,..., x n ) of complex numbers. It is a Hilbert space with the inner product

Finite Elements. Colin Cotter. February 22, Colin Cotter FEM

JUHA KINNUNEN. Harmonic Analysis

CHAPTER VIII HILBERT SPACES

1 Continuity Classes C m (Ω)

Math The Laplacian. 1 Green s Identities, Fundamental Solution

SOBOLEV SPACES AND ELLIPTIC EQUATIONS

Your first day at work MATH 806 (Fall 2015)

Partial Differential Equations, 2nd Edition, L.C.Evans Chapter 5 Sobolev Spaces

Equations paraboliques: comportement qualitatif

Tools from Lebesgue integration

Math Tune-Up Louisiana State University August, Lectures on Partial Differential Equations and Hilbert Space

ON WEAKLY NONLINEAR BACKWARD PARABOLIC PROBLEM

Université de Metz. Master 2 Recherche de Mathématiques 2ème semestre. par Ralph Chill Laboratoire de Mathématiques et Applications de Metz

USING FUNCTIONAL ANALYSIS AND SOBOLEV SPACES TO SOLVE POISSON S EQUATION

Here we used the multiindex notation:

Functional Analysis. Martin Brokate. 1 Normed Spaces 2. 2 Hilbert Spaces The Principle of Uniform Boundedness 32

1. Introduction Boundary estimates for the second derivatives of the solution to the Dirichlet problem for the Monge-Ampere equation

Scientific Computing WS 2018/2019. Lecture 15. Jürgen Fuhrmann Lecture 15 Slide 1

Your first day at work MATH 806 (Fall 2015)

Heat Kernel and Analysis on Manifolds Excerpt with Exercises. Alexander Grigor yan

Simple Examples on Rectangular Domains

Appendix A Functional Analysis

PERTURBATION THEORY FOR NONLINEAR DIRICHLET PROBLEMS

Lecture Notes on PDEs

THE L 2 -HODGE THEORY AND REPRESENTATION ON R n

************************************* Applied Analysis I - (Advanced PDE I) (Math 940, Fall 2014) Baisheng Yan

Weak Convergence Methods for Energy Minimization

SYMMETRY RESULTS FOR PERTURBED PROBLEMS AND RELATED QUESTIONS. Massimo Grosi Filomena Pacella S. L. Yadava. 1. Introduction

Introduction to Real Analysis Alternative Chapter 1

l(y j ) = 0 for all y j (1)

2 A Model, Harmonic Map, Problem

Chapter 1. Introduction

Existence of minimizers for the pure displacement problem in nonlinear elasticity

Some lecture notes for Math 6050E: PDEs, Fall 2016

Functional Analysis I

Integral Jensen inequality

Lecture Notes Math 632, PDE Spring Semester Sigmund Selberg Visiting Assistant Professor Johns Hopkins University

SHARP BOUNDARY TRACE INEQUALITIES. 1. Introduction

EXPOSITORY NOTES ON DISTRIBUTION THEORY, FALL 2018

2.3 Variational form of boundary value problems

Notions such as convergent sequence and Cauchy sequence make sense for any metric space. Convergent Sequences are Cauchy

SPECTRAL PROPERTIES OF THE LAPLACIAN ON BOUNDED DOMAINS

Metric Spaces and Topology

Math 209B Homework 2

1.5 Approximate Identities

************************************* Partial Differential Equations II (Math 849, Spring 2019) Baisheng Yan

Locally convex spaces, the hyperplane separation theorem, and the Krein-Milman theorem

Exercises - Chapter 1 - Chapter 2 (Correction)

Some SDEs with distributional drift Part I : General calculus. Flandoli, Franco; Russo, Francesco; Wolf, Jochen

Numerical Analysis and Methods for PDE I

Sobolev Spaces 27 PART II. Review of Sobolev Spaces

L p Spaces and Convexity

Moscow, Russia; National Research University Higher School of Economics, Moscow, Russia

BOUNDARY VALUE PROBLEMS ON A HALF SIERPINSKI GASKET

Eigenvalues and Eigenfunctions of the Laplacian

Notes on Distributions

4 Hilbert spaces. The proof of the Hilbert basis theorem is not mathematics, it is theology. Camille Jordan

A variational weak weighted derivative: Sobolev spaces and degenerate elliptic equations

On some nonlinear parabolic equation involving variable exponents

Real Analysis Notes. Thomas Goller

CAPACITIES ON METRIC SPACES

Conservation law equations : problem set

ON THE REGULARITY OF SAMPLE PATHS OF SUB-ELLIPTIC DIFFUSIONS ON MANIFOLDS

Transcription:

Partial Differential Equations 2 Variational Methods Martin Brokate Contents 1 Variational Methods: Some Basics 1 2 Sobolev Spaces: Definition 9 3 Elliptic Boundary Value Problems 2 4 Boundary Conditions, Traces 29 5 Homogenization: Introduction 43 6 Averages and weak convergence 47 7 Periodic boundary conditions 53 8 Homogenization: The multidimensional case 59 9 Monotone Problems 69 1 The Bochner Integral 79 11 Linear parabolic equations 85 12 PDEs in Continuum Mechanics 1 Lecture Notes, Summer Term 216 Technical University of Munich, Department of Mathematics

1 Variational Methods: Some Basics Equations and minimization. The solution of equations is related to minimization. Suppose we want to find u R n with F (u) =, F : R n R n. If we can find a function J : R n R such that F = J, and if we can prove that J has a minimizer u, J(u) = min J(v), v R n, then we know that F (u) = J(u) =. Conversely, suppose we want to find a minimizer u of J. If we can find u R n with J(u) =, then u is a candidate for a minimizer of J. If in addition J is convex, all such candidates are minimizers; otherwise, second order conditions may be used. Dirichlet s principle for the Laplace operator. We consider the Dirichlet problem for the Poisson equation in an open bounded set R n, u = f u = g in auf. (1.1) We also consider the functional ( denotes the Euclidean norm in R n ) 1 J(v) = 2 v(x) 2 f(x)v(x) dx. (1.2) The functional J and the boundary value problem (1.1) are related as follows. Assume that f : R and g : R are continuous, let K = {v : v C 2 (), v(x) = g(x) for all x }. (1.3) Let u be a minimizer of J on K, that is, u K, J(u) = min J(v). (1.4) v K Let ϕ C () be given. We have u + λϕ K for all λ R. We consider J : R R, J(λ) = J(u + λϕ). (1.5) We have 1 J(λ) = 2 (u + λϕ)(x) 2 f(x)(u + λϕ)(x) dx = J(u) + λ u(x), ϕ(x) f(x)ϕ(x) dx + λ 2 1 2 ϕ(x) 2 dx. Since is a minimizer of the differentiable function J, we get = J () = u(x), ϕ(x) f(x)ϕ(x) dx = ( u(x) f(x))ϕ(x) dx. 1

(Here we have used partial integration and the fact that ϕ = on.) Since ϕ C () can be chosen arbitrarily, it follows that (we will prove this later) u(x) = f(x), for all x. (1.6) Thus, every minimizer of J on K solves (1.1). Conversely, let u C 2 () be a solution of (1.1), so in particular u K. Let v K be arbitrary. Then ( u(x) f(x))(u(x) v(x)) dx =. Since u v = on, partial integration gives u(x), u(x) v(x) dx f(x))(u(x) v(x)) dx =. It follows that (we drop the argument x of the functions) u 2 1 fu dx = u, v fv dx 2 u 2 + 1 2 v 2 fv dx, (1.7) since for arbitrary vectors y, z R n we have y, z y z 1 2 y 2 + 1 2 z 2. Subtracting (1/2) u 2 dx on both sides of (1.7), we arrive at so u is a minimizer of J on K. The equivalence J(u) J(v), u solves (1.1) u is a minimizer of J on K (1.8) is called Dirichlet s principle. In this context, the equation u = f is called the Euler equation or Euler-Lagrange equation corresponding to the functional J in (1.2). More general, this approach relates functionals of the form J(v) = L(x, v(x), v(x)) dx (1.9) to differential equations. These are ordinary differential equations in the case n = 1, R, and partial differential equations in the case n > 1. Minimizing functionals of the form (1.9) constitutes the basic problem of the calculus of variations. The Dirichlet principle yields equivalence, but not existence. 2

The existence problem is hard in general, because the domain of the functional J is a subset of a function space, which has infinite dimension. Therefore, functional analysis comes into play. Using its methods to prove existence of a minimizer is called the direct method of the calculus of variations. That such a minimizer solves the Euler equation is then proved as above by setting to zero the derivatives of J in all directions ϕ C (). Proving existence of a minimizer of J in (1.9) by proving first the existence of a solution to the Euler equation is called the indirect method. If J is not convex, however, solutions of the Euler equation for J in general need to have additional properties in order to be minimizers of J (as is the case in finite dimensions, where second derivatives of J play a role). Quadratic minimization: An abstract form of Dirichlet s principle. We consider the following situation. Problem 1.1 (Quadratic minimization problem) Let V be a vector space, K V convex, let a : V V R be a bilinear form and F : V R a linear form. We consider the problem min v K J(v), J(v) = 1 2 a(v, v) F (v). (1.1) Proposition 1.2 Consider the situation of Problem 1.1 and assume in addition that a is symmetric and positive definite. Then J is strictly convex, and for u K we have J(u) = min J(v) a(u, v u) F (v u) v K. (1.11) v K There exists at most one u K with this property. Proof: Let u, h V be arbitrary with h. We define We have J u,h (λ) = 1 2 J u,h : R R, J u,h (λ) = J(u + λh). λ2 a(u + λh, u + λh) F (u + λh) = J(u) + λ(a(u, h) F (h)) + a(h, h). (1.12) 2 Since a(h, h) >, the quadratic function J u,h is strictly convex, and so is J. Therefore J has at most one minimizer. : Let v K. Setting h = v u, from (1.12) and (1.11) it follows that J(v) J(u) = J(u + h) J(u) = (a(u, h) F (h)) + 1 a(h, h). 2 : Let v K. Again setting h = v u, we have u + λh K for λ 1 (since K is convex). Thus J u,h (λ) J u,h () = λ(a(u, h) F (h)) + λ2 a(h, h). (1.13) 2 3

Dividing by λ and passing to the limit λ yields a(u, h) F (h). The inequality system (one inequality for each v K) u K, a(u, v u) F (v u) v K, (1.14) is called a variational inequality. As (1.13) shows, the variational inequality says that the directional derivative of J in a minimizer is nonnegative for those directions h = v u which point into the admissible set K. Corollary 1.3 In the situation of Proposition 1.2 assume that K is an affine subspace of V, that is, K = v + U with v V, U subspace of V. Then we have J(u) = min J(v) a(u, w) = F (w) w U. (1.15) v K Proof: If u K we have U = {v u : v K} and therefore a(u, v u) F (v u) v K a(u, w) F (w) w U a(u, w) = F (w) w U, the latter equivalence holds because w U implies w U. The system of equations is called a variational equation. u v + U, a(u, v) = F (v) v U, (1.16) The situation (1.1) (1.4), which we considered at the beginning, becomes a special case of (1.16) if we set a(u, v) = u(x), v(x) dx, F (v) = f(x)v(x) dx. (1.17) The variational equation then reads u(x), v(x) dx = (So far we did not specify the space V and the subspace U.) If one starts from the Poisson equation f(x)v(x) dx, for all v in U. (1.18) u = f, in, (1.19) one may arrive at the variational formulation (1.18) immediately by multiplying (1.19) on both sides with functions v which are zero on, and performing partial integration. In this manner one bypasses the functional J. The variational equation. In order to obtain existence, the algebraic vector space structure by itself does not suffice. One uses the setting of normed spaces. Thus, topological aspects enter the picture. 4

Definition 1.4 (Bilinear form, continuity and ellipticity) Let V be a normed space. A bilinear form a : V V R is called continuous if there exists a C a > such that a(u, v) C a u v, for all u, v V, (1.2) it is called V -elliptic if there exists a c a > such that a(v, v) c a v 2, for all v V. (1.21) As an immediate consequence of this definition we see that a V -elliptic bilinear form is positive definite. In order to obtain existence, completeness of the normed space is required. Proposition 1.5 (Variational equation, solvability) Let (V, ) be a Banach space, let a : V V R be a symmetric, continuous and V -elliptic bilinear form, let F : V R be linear and continuous. Then there exists a unique solution u V of the variational equation Proof: For all v V we have Since c a >, a(u, v) = F (v) v V. (1.22) c a v 2 a(v, v) C a v 2. u, v a = a(u, v), v a = a(v, v), defines a scalar product whose associated norm a is equivalent to the original norm. Therefore, (V,, a ) is complete and thus a Hilbert space. Moreover, F : (V, a ) R is continuous. The representation theorem of Riesz (from functional analysis) states that there is a unique u V such that u, v a = F (v). Corollary 1.6 The solution u of the variational equation (1.22) in Proposition 1.5 satisfies u 1 c a F. (1.23) Proof: Inserting u for v in (1.22) yields c a u 2 a(u, u) = F (u) F u. 5

Proposition 1.5 and Corollary 1.6 state that, under the given assumptions, the solution of the variational equation is a well-posed problem in the sense that its solution exists, is unique and depends continuously upon the data (the right hand side specified by F ). The latter property has to interpreted as follows: the solution operator S : V V which maps F V (the dual space of V ) to the solution u of (1.22), is continuous. Indeed, it is linear (this follows immediately from (1.22)), and according to (1.23) we have S(F ) F c a, therefore S 1 c a. The variational inequality. In Proposition 1.5 we have assumed that the bilinear form is symmetric. It turns out that this assumption is superfluous. We present this extension in an even more general context, namely, for the variational inequality. Proposition 1.7 (Variational inequality, solvability) Let V be a Banach space, let K V be closed, convex and nonempty. Let a : V V R be a continuous V -elliptic bilinear form with a(v, v) c a v 2 v V, (1.24) let F : V R be linear and continuous. Then the variational inequality a(u, v u) F (v u) v K, (1.25) has a unique solution u K. Moreover, if ũ K is the solution of (1.25) corresponding to a linear and continuous F : V R in place of F, we have u ũ 1 c a F F. (1.26) Proof: We first prove (1.26). This also implies the uniqueness. Assume that u, ũ are solutions corresponding to F, F. Then Adding these inequalities yields so a(u, ũ u) F (ũ u), a(ũ, u ũ) F (u ũ). a(u ũ, ũ u) (F F )(ũ u), (1.27) c a u ũ 2 a(u ũ, u ũ) F F u ũ. (1.28) This shows that (1.26) holds. To prove existence we first consider the special case where a is symmetric. By Proposition 1.2 it suffices to show that the associated quadratic functional J(v) = 1 a(v, v) F (v) 2 has a minimizer on K. In order to prove this, we see that J(v) 1 2 c a v 2 F v = ( ca 2 v 6 ) 2 1 F 1 F 2 (1.29) 2c a 2c a

holds for all v V, therefore J is bounded from below. Define and let (u n ) n N be a sequence in K such that The sequence (u n ) n N is a Cauchy sequence, because d = inf J(v), (1.3) v K d J(u n ) d + 1 n. (1.31) c a u n u m 2 a(u n u m, u n u m ) ( 1 = 2a(u n, u n ) + 2a(u m, u m ) 4a 2 (u n + u m ), 1 ) 2 (u n + u m ) ( ) 1 = 4J(u n ) + 4J(u m ) 8J 2 (u n + u m ) ( 4 d + 1 ) ( + 4 d + 1 ) 8d n m ( 1 4 n + 1 ). m (1.32) As V is a Banach space, there exists a u V such that u n u. Since K is closed, we have u K. Since J is continuous, we have J(u) = d. This concludes the proof in the case where a is symmetric. Now let a be arbitrary. We utilize a continuation argument. We consider the family of bilinear forms a t (u, v) = a (u, v) + tb(u, v), t [, 1], (1.33) where a (u, v) = 1 (a(u, v) + a(v, u)), (1.34) 2 b(u, v) = 1 (a(u, v) a(v, u)). (1.35) 2 Then a is symmetric and a 1 = a. The bilinear forms a t are continuous and V -elliptic, and (1.24) holds for a t with the same constant c a as for a, because a t (u, u) = a(u, u) holds for all u V. As a is symmetric, it now suffices to prove the following claim: If the variational inequality a τ (u, v u) G(v u), v K, (1.36) has a unique solution u K for every G V, then the same is true if we replace τ by t with τ t τ + c a 2C a. (1.37) We prove this claim. Let t be given, satisfying (1.37). We only have to prove the existence of a solution. Let G : V R be linear and continuous. We look for a solution u t K satisfying a t (u t, v u t ) G(v u t ), v K. (1.38) 7

For this purpose, we consider for arbitrary given w V the variational inequality a τ (u, v u) F w (v u), v K, (1.39) where F w (v) = G(v) (t τ)b(w, v). (1.4) Let T : V K be the mapping which associates to each w V the solution u K of (1.39). T is a contraction, since T w 1 T w 2 1 F w1 F w2 1 t τ sup b(w 1 w 2, v) c a c a v =1 1 c a t τ C a w 1 w 2 1 2 w 1 w 2. (1.41) By Banach s fixed point theorem, T has a unique fixed point which we denote by u t. Then u t K, and for all v K we have a t (u t, v u t ) = a τ (u t, v u t ) + (t τ)b(u t, v u t ) F ut (v u t ) + (t τ)b(u t, v u t ) = G(v u t ). This finishes the proof. Corollary 1.8 (Lax-Milgram theorem) Let V be a Banach space, let a : V V R be a continuous V -elliptic bilinear form, let F : V R be linear and continuous. Then the variational equation has a unique solution u V. a(u, v) = F (v) v V, (1.42) Proof: We apply Proposition 1.7 with K = V. As in Corollary 1.3, the assertion follows from the equivalences a(u, v u) F (v u) v V a(u, v) F (v) v V a(u, v) = F (v) v V. (1.43) Assume that one wants to solve a given linear partial differential equation. This usually determines the form of the associated bilinear form a. If one wants to apply the Lax- Milgram theorem to prove existence and uniqueness, one has to choose the Banach space V such that a is continuous and V -elliptic. The spaces V for which this works are the Sobolev spaces. 8

2 Sobolev Spaces: Definition In this section always denotes an open subset of R n. Weak derivatives. Let u C 1 () and ϕ C (). Partial integration yields i u(x)ϕ(x) dx = u(x) i ϕ(x) dx, 1 i n. (2.1) The right side of this equation is well-defined for functions u whose restriction to the support of ϕ (a compact subset of ) is integrable. We thus consider the vector space L 1 loc() = {v : v K L 1 (K) for all compact subsets K }. (2.2) The functions in L 1 loc () are called locally integrable. Continuous functions u : R are locally integrable since they are bounded on every compact subset of ; they may or may not be integrable on. For example, u(x) = 1, u : (, 1) R, x belongs to L 1 loc (, 1), but not to L1 (, 1). We remark that L p () L 1 loc () for all p [1, ], no matter whether is bounded or not. For locally integrable functions one can use (2.1) in order to define a generalization of the classical derivative. Definition 2.1 (Weak partial derivative) Let u L 1 loc (). A function w L1 loc () which satisfies w(x)ϕ(x) dx = u(x) i ϕ(x) dx, for all ϕ C (), (2.3) is called an i-th weak partial derivative of u. We also denote it by i u. The notation i u for weak derivatives makes sense only if they are uniquely determined. This we will see later in Corollary 2.8. For u C 1 (), the classical partial derivatives i u of u are also weak derivatives of u. Consider the example For ϕ C (R) we have x ϕ (x) dx = xϕ (x) dx R = xϕ(x) x= x= = (sign x)ϕ(x) dx, R u : R R, u(x) = x. (2.4) xϕ (x) dx ϕ(x) dx xϕ(x) x= x= + ϕ(x) dx 9

therefore the weak derivative of u exists, and u = sign. (2.5) This equality has to be understood as an equality in L 1 loc (R), that is, u (x) = sign (x), a.e. in R. (2.6) For this example, the standard pointwise derivative of u exists in all points x, and is a.e. equal to the weak derivative. As a second example, consider the Heaviside funktion H : R R, { 1, x >, H(x) =, x. For ϕ C (R) we have R H(x)ϕ (x) dx = (2.7) ϕ (x) dx = ϕ(). (2.8) If w L 1 loc (R) were a weak derivative of H, we would have w(x)ϕ(x) dx = ϕ(), for all ϕ C (R). (2.9) R But this is not possible (see an exercise), thus the Heaviside function does not have a weak derivative in the sense of Definition 2.1. One can of course consider the pointwise derivative of H, which is zero except at x =. However, this function is a.e. equal to zero, so (when it appears under an integral sign) it does not distinguish between H and a constant function. But one can use (2.8) to generalize the concept of a weak derivative even further and consider the linear functional T : C (R) R defined by T (ϕ) = ϕ() (2.1) as a generalized derivative of H. This is the starting point of distribution theory; the corresponding derivative is also called distributional derivative. The theory of distributions was initiated in the 4s of the previous century, the main proponent being Laurent Schwartz (1915 22). Approximation by smooth functions. Given a function u : R, there are various ways to define smooth approximations u ε of u with u ε u as ε. Here we mainly use approximation by convolution. Let us recall the definition of the convolution of two functions u and v, (u v)(x) = u(x y)v(y) dy. (2.11) R n It is defined for u, v L 1 (R n ) and yields a function u v L 1 (R n ). It has the properties u v = v u, u v 1 u 1 v 1. (2.12) 1

We want to approximate a given function u by functions η ε u; they will turn out to be smooth if η ε is smooth. In order to achieve this, we define { exp ( ) 1 t, t >, ψ : R R, ψ(t) = (2.13), t, where α > is chosen such that ψ : R R, ψ(r) = ψ(1 r 2 ), (2.14) η 1 : R n R, η 1 (x) = α ψ( x ), (2.15) R n η 1 (x) dx = 1. (2.16) For given ε > we define the functions η ε : R n R, η ε (x) = 1 ( x ) ε η n 1. (2.17) ε These functions are radially symmetric (that is, they only depend upon x ), and we have η ε C (R n ), supp (η ε ) = K(; ε), η ε, η ε (x) dx = 1. (2.18) R n The function η 1, or the family (η ε ) ε>, is called a mollifier. Given u : R, by ũ we denote its extension to R n, { u(x), x, ũ(x) =, x /. Given u L 1 loc (), we define functions uε : R n R by We then have u ε (x) = η ε (x y)ũ(y) dy = R n We define the ε-neighbourhood of a subset U of R n by (2.19) u ε = η ε ũ. (2.2) η ε (x y)u(y) dy, for all x R n. (2.21) U ε = {x : x R n, dist (x, U) < ε}. (2.22) Since η ε (x y) = if x y ε, we see from (2.21) that supp (u ε ) ε, if u L 1 loc(). (2.23) To denote higher order partial derivatives we use multi-indices. Let α = (α 1,..., α n ) N n be a multi-index. We set α = α 1 1 α 2 2 αn n, α = n α j. (2.24) j=1 11

Proposition 2.2 Let R n be open, u L p (), 1 p, ε >. Then we have u ε C (R n ), supp (u ε ) ε, (2.25) α u ε (x) = α η ε (x y)u(y) dy, for all x R n. (2.26) u ε L p (R n ) u L p (). (2.27) Proof: Let α be a multi-index. Fix x R n. For x, y R n with x x 1 we have we have { u(y) α α η ε η ε (x y) u(y), y x 1 + ε,, otherwise, since x y > ε if y x > 1 + ε. The function on the right side is integrable and does not depend on x. Therefore, we can interchange α with the convolution integral in the neighbourhood of every point x R n. This proves (2.25) and (2.26). Let us now prove (2.27). In the case p =, (2.27) follows from the estimate u ε (x) η ε (x y) u(y) dy u η ε (x y) dy u, x R n. In the case 1 < p <, 1 p + 1 q = 1, we obtain for all x Rn, using Hölder s inequality, u ε (x) = η ε (x y)u(y) dy ( u(y) p η ε (x y) dx u(y) (η ε (x y)) 1 p (ηε (x y)) 1 q dx ) 1 p ( ) 1 q η ε (x y) dx } {{ } 1. (2.28) From this we obtain, now for 1 p < u ε (x) p dx u(y) p η ε (x y) dy dx = R n R n u(y) p dx. u(y) p R n η ε (x y) dx dy Definition 2.3 Let R n be open. We say that U R n is compactly embedded in, if U is compact and U. We write U. (2.29) Let us denote by C () = {v : v C(), supp (v) } (2.3) the set of all continuous functions on whose support is a compact subset of. We have C () L p () for all p [1, ]. 12

Proposition 2.4 The set C (R n ) is a dense subspace of L p (R n ) for all p [1, ), that is, for every u L p (R n ) and every δ > there exists a v C (R n ) such that u v p δ. Proof: One possibility is to consider first the special case u = 1 B, where B is an arbitrary measurable subset of, using the regularity of the Lebesgue measure. Another possibility is to use Lusin s theorem (which says that measurable functions can be approximated uniformly by continuous functions on subsets of almost full measure). This might be discussed in the exercises. Proposition 2.5 Let u C (R n ). Then u ε u uniformly in R n. Proof: We have for all x R n u ε (x) u(x) = η ε (x y)(u(y) u(x)) dy η ε (x y) u(y) u(x) dy R n R n η ε (x y) dy max u(z) u(y) = max u(z) u(y) R n z y ε z y ε as ε, since u is uniformly continuous on the compact set supp (u). Proposition 2.6 Let u L p (), 1 p <. Then we have u ε u in L p () for ε. Proof: Let δ >. We extend u to ũ L p (R n ), setting ũ to outside. According to Proposition 2.4 we can find v C (R n ) with ũ v L p (R n ) δ, thus u v L p () δ. Then u u ε L p () u v L p () + v vε L p () + vε u ε L p () (2.31) 2δ + v v ε L p (), (2.32) because and, by Proposition 2.2 v ε u ε = η ε v η ε u = (v u) ε (v u) ε L p () v ũ L p (R n ) δ. Since v C (R n ), we have v ε v uniformly by Proposition 2.5, so v ε v in L p () and therefore lim sup u u ε L p () 2δ. ε As δ > was arbitrary, the assertion follows. The next result (or a variant of it) is also called the fundamental lemma of the calculus variations. Proposition 2.7 Let R n be open, let u L 1 loc (), assume that u(x)ϕ(x) dx =, for all ϕ C (). (2.33) Then u = a.e. in. 13

Proof: Let x be arbitrary. If ε is small enough, the support of the function ϕ(y) = η ε (x y) is compactly embedded in, so ϕ C (). From (2.33) it follows that u ε (x) = η ε (x y)u(y) dy = u(y)ϕ(y) dy =. R n Consequently, lim u ε (x) =, for all x. ε On the other hand, for every ball B we have u L 1 (B), so u ε u in L 1 (B) by Proposition 2.6. Thus u = a.e. in B and therefore u = a.e. in. Corollary 2.8 Weak partial derivatives of a function u L 1 loc () are uniquely determined. Proof: If w, w L 1 loc () satisfy w(x)ϕ(x) dx = u(x) i ϕ(x) dx = w(x)ϕ(x) dx for all ϕ C (), Proposition 2.7 implies that w w = a.e. in. Sobolev spaces. We have already defined weak derivatives of first order via the partial integration rule. The same can be done for higher derivatives. Definition 2.9 Let u L 1 loc (), α a multi-index. A function w L1 loc () which satisfies w(x)ϕ(x) dx = ( 1) α u(x) α ϕ(x) dx, for all ϕ C (), (2.34) is called an α-th weak partial derivative of u and denoted by α u. Again, it follows from the fundamental lemma of the calculus of variations (Proposition 2.7) that there can be at most one α-th weak partial derivative of u. Definition 2.1 (Sobolev space) Let R n be open, k N, 1 p. We define In particular, we have W k,p () = {v : v L p (), α v L p () for all α k}. (2.35) W,p () = L p (). (2.36) The spaces W k,p () are vector spaces. Indeed, if u, v L 1 loc () have weak derivatives α u and α v, we have (β α u(x) + γ α v(x))ϕ(x) dx = ( 1) α (βu(x) + γv(x)) α ϕ(x) dx for all β, γ R and all ϕ C (). Therefore, α (βu + γv) exists as a weak partial derivative, and α (βu + γv) = β α u + γ α v. (2.37) 14

Proposition 2.11 Let R n be open, k N, 1 p. The space W k,p () is a Banach space when equipped with the norm 1 1 v W k,p () = p α v p p = p α v(x) p dx, 1 p <, (2.38) α k α k v W k, () = α v, p =. (2.39) α k Proof: Consider first p <. The triangle inequality holds because 1 u + v W k,p () = p α u + α v p p ( α u p + α v p ) p α k α u p p α k 1 p α k + α v p p α k 1 p 1 p = u W k,p () + v W k,p (). All other properties of the norm follow immediately from the definitions. It remains to prove the completeness. Let (u n ) n N be a Cauchy sequence in W k,p (). Due to α u n α u m p u n u m W k,p () the sequences ( α u n ) n N are Cauchy sequences in L p () for all multi-indices α k. Therefore there exist u L p (), u α L p () satisfying u n u, α u n u α (2.4) in L p (). Let now ϕ C (). We have u(x) α ϕ(x) dx = lim u n (x) n α ϕ(x) dx = lim ( 1) α n α u n (x)ϕ(x) dx = ( 1) α u α (x)ϕ(x) dx, therefore α u = u α for all α k and thus u W k,p (). For the case p = the proofs are analogous. We have seen that for u L p (), the functions u ε defined by u ε (x) = η ε (x y)u(y) dy converge to u in L p (). This does not carry over to W k,p (); near the boundary of, the situation is more complicated. This is related to the fact that for u W k,p (), the extension ũ = u on, ũ = outside of, in general does not belong to W k,p () (compare the example of the Heaviside function). However, if we stay away from the boundary, this problem does not arise. Consider open sets, U R n with U. Then dist (U, ) := inf x y >, (2.41) x U,y since U is compact. (In the case = R n we set dist (U, ) =.) 15

Lemma 2.12 Let, U R n be open, U. let u W k,p (), 1 p <. Then u ε C (U) W k,p (U) holds for all ε satisfying < ε < dist (U, ). Moreover, u ε u in W k,p (U) for ε. Proof: Let ε < dist (U, ). By Proposition 2.2 we have u ε C (U). Moreover, for all x U α u ε (x) = α η ε (x y)u(y) dy = ( 1) α y α η ε (x y)u(y) dy = η ε (x y) α u(y) dy = (η ε α u)(x). Since α u L p (), by Proposition 2.2 we get α u ε L p (U). From Proposition 2.6 we get that α u ε α u in L p (U). While in general it does not hold that u ε u in W k,p (), we nevertheless can construct smooth approximations in W k,p () from the interior. Proposition 2.13 Let R n be open, let v W k,p (), 1 p <. Then there exists a sequence (v n ) n N in C () W k,p () with v n v in W k,p (). Proof: We first consider the case R n. We define U j = {x : x, dist (x, ) > 1 j and x < j}, V j = U j+3 \ U j+1, j 1, (2.42) and V = U 3. We have = Let (β j ) j be a partition of unity for with V j. j= β j 1, β j C (V j ), β j = 1. (2.43) j= Since v W k,p (), one also has β j v W k,p () (see the exercises), and supp (β j v) V j. Let now δ > be arbitrary. We choose ε j > sufficiently small so that for w j = η εj (β j v), W j = U j+4 \ U j, j 1, W = U 4, we have from Lemma 2.12, applied to β j v and W j in place of u and U, supp (w j ) W j, (2.44) w j β j v W k,p () = w j β j v W k,p (W j ) 2 (j+1) δ. (2.45) We set w = w j. j= 16

By construction, on each W j only finitely many summands are nonzero. Therefore we have w C (), as w j C () for each j according to Lemma 2.12. It now follows that w v W k,p () = w j β j v w j β j v W k,p () δ 2 (j+1) = δ. j= j= W k,p () As δ > was arbitrary, the assertion follows. In the case = R n the proof proceeds analogously, with the choice U j = B(, j), the open ball around with radius j. Density results of this kind can be used to extend to Sobolev spaces properties known to hold for smooth functions. For example, we know that j= i (uv) = ( i u)v + u i v, u, v C 1 (). (2.46) Consider now u W 1,p (), v C 1 (). According to Proposition 2.13, choose u n W 1,p () C () such that u n u in W 1,p (). Let ϕ C () be arbitrary. We have u n v i ϕ dx = i (u n v)ϕ dx = (( i u n )v + u n i v)ϕ dx. Since u n u and i u n i u in L p () as n, we obtain uv i ϕ dx = (( i u)v + u i v)ϕ dx. This shows that i (uv) exists as a weak derivative and that (2.46) holds. If moreover v and its derivatives are bounded, then uv W 1,p (). Definition 2.14 Let R n be open, let 1 p <, k N. We define W k,p () W k,p () by W k,p () = C (), (2.47) the closure being taken with respect to the norm in W k,p (). W k,p () is a closed subspace of W k,p () and therefore itself a Banach space, if equipped with the norm of W k,p (). W k,p () is a function space whose elements are zero on in a certain weak sense; it is not clear at this point whether the statement v = on makes sense for a general v W k,p (), because equality in W k,p () means equality almost everywhere, and the statement v = a.e. on does not give any information when has measure in R n (which usually is the case). Definition 2.15 Let R n be open, let k N. We define H k () = W k,2 (), H k () = W k,2 (). (2.48) j= 17

Proposition 2.16 Let R n be open, k N. The spaces H k () and H k () are Hilbert spaces when equipped with the scalar product u, v H k () = α u, α v L 2 () = α u(x) α v(x) dx. (2.49) α k α k We have v W k,2 () = v, v H k (), v Hk (). (2.5) Proof: The properties of the scalar product (2.49) immediately follow from the corresponding properties of the scalar product in L 2 (). Obviously we have (2.5), and by Proposition 2.11 the space H k () is complete. Setting v H k () = one obtains a seminorm on H k () with α =k α v(x) 2 dx 1 2, (2.51) v H k () v H k (). (2.52) For k >, this is not a norm since v H k () = whenever v is a polynomial of degree less than k. The following inequality (2.53) is known as the Poincaré-Friedrichs inequality. Proposition 2.17 Let [ R, R] n be open. Then for every v H() 1 we have v(x) 2 dx 4R 2 j v(x) 2 dx, 1 j n, (2.53) and therefore with C = 4R 2 v(x) 2 dx C v(x) 2 dx. (2.54) Proof: Consider first v C (). Extending v by zero on all of R n we have v C (R n ). Let x = (x 1,..., x n ). Then we have x1 v(x) = v( R, x 2,..., x n ) + }{{} 1 v(t, x 2,..., x n ) dt, R = and therefore, by the Cauchy-Schwarz inequality, v(x) 2 = ( x1 2R R R R 1 1 v(t, x 2,..., x n ) dt) 2 ( 1 v(t, x 2,..., x n )) 2 dt. 18 x1 R 1 2 dt x1 R ( 1 v(t, x 2,..., x n )) 2 dt

Since the right hand side does not depend on x 1, it follows that R R R v(x) 2 dx 1 4R 2 ( 1 v(t, x 2,..., x n )) 2 dt. (2.55) Integrating with respect to the other coordinates x 2,..., x n yields v(x) 2 dx 4R 2 ( 1 v(x)) 2 dx, R Since we may choose any other j in places of 1, (2.53) is proved for v C (). For arbitrary v H(), 1 we choose a sequence (v k ) k N in C () with v k v in H(). 1 Then (2.53) holds for v k and, passing to the limit k, for v also. Proposition 2.18 Let R n be open and bounded. Then there exists a C > such that v H k () C v H k (), for all v H k (). (2.56) Proof: Let v H k (), let α be a multi-index with α = j 1, 1 j k. Then we have α v H 1 (), and it follows from Proposition 2.17 that α v 2 2 4R2 1 α v 2 2. Consequently, so and finally α v 2 2 4R2 1 α v 2 2 4R2 α v 2 2, α =j 1 α =j 1 α =j v 2 H j 1 () 4R2 v 2 H j (), k k v 2 H k () = v 2 H j () (4R 2 ) k j v 2 H k (). j= j= Corollary 2.19 The space H k () becomes a Hilbert space when equipped with the scalar product corresponding to H (), and k H k () defines a norm on H k () which is equivalent to H (). k With an analogous proof, one can also obtain the Poincaré-Friedrichs inequality for an exponent p (1, ) instead of 2, v L p () C v 1,p L p (), for all v W (). (2.57) This yields an equivalent norm on W 1,p () in the same manner as in Corollary 2.19. The optimal (that is, smallest possible) C in (2.57) is called Poincaré constant, it depends on and p. It is not easy to determine it. 19

3 Elliptic Boundary Value Problems We want to solve the boundary value problem Lu = f, in, (3.1) u =, on. (3.2) Here, R n is open, f : R, and L is an elliptic differential operator of the form n n Lu = j (a ij (x) i u) + b i (x) i u + c(x)u (3.3) i,j=1 i=1 = div (A(x) T u) + b(x), u + c(x)u. One says that L is written in divergence form. The variational formulation of (3.1), (3.2) is obtained multiplying (3.1) by an arbitary test function ϕ C () and integrating over. After partial integration, the left side of (3.1) then becomes n n Lu(x)ϕ(x) dx = a ij (x) i u(x) j ϕ(x) dx + b i (x) i u(x)ϕ(x) dx i,j=1 + c(x)u(x)ϕ(x) dx. i=1 Thus, the corresponding bilinear form is given by n n a(u, v) = a ij (x) i u(x) j v(x) dx + b i (x) i u(x)v(x) dx i,j=1 i=1 + c(x)u(x)v(x) dx (3.4) = u(x), A(x) v(x) dx + b(x), u(x) v(x) dx + c(x)u(x)v(x) dx. It is symmetric if A is symmetric and b =. Assumption 3.1 We assume that there exists a > such that n ξ T A(x)ξ = ξ i a ij (x)ξ j a ξ 2, for all ξ R n, x. (3.5) i,j=1 (In this case, the differential operator L is called uniformly elliptic.) We moreover assume that a ij, b i, c L () for all i, j. Definition 3.2 (Weak solution) Assume that 3.1 holds and that f L 2 (). A function u H() 1 is called a weak solution of the boundary value problem (3.1), (3.2) if u solves the variational equation where a is defined by (3.4) and a(u, v) = F (v), for all v H 1 (), (3.6) F (v) = f(x)v(x) dx. (3.7) 2

Lemma 3.3 Let R n be open, assume that 3.1 holds. Then (3.4) defines a continuous bilinear form on H 1 (). Moreover, (3.7) defines a continuous linear form on H 1 () if f L 2 (). Proof: Setting a = n ess sup x 1 i,j n we get for all u, v H 1 () a(u, v) u(x), A(x) v(x) dx + a ij (x), b = n ess sup b(x) x b(x), u(x) v(x) dx + c(x)u(x)v(x) dx a u 2 v 2 + b u 2 v 2 + c u 2 v 2 (a + b + c ) u H 1 () v H 1 (). Let us now consider the special case b = c =, that is, the boundary value problem becomes i,j j (a ij (x) i u) = f, in, (3.8) u =, on. (3.9) Proposition 3.4 Let R n be open and bounded, let f L 2 (), assume 3.1. Then (3.8), (3.9) has a unique weak solution u H 1 (), and u H 1 () C F. (3.1) Here, the norm F of the functional (3.7) is taken in the dual space of H 1 (), and C does not depend on F. Proof: In order to apply Lax-Milgram, Proposition 1.8, in view of Lemma 3.3 it remains to show that a is H()-elliptic. 1 According to our assumptions we have a(v, v) = v(x), A(x) v(x) dx a v(x) 2 dx = a v 2 H 1 () a C v 2 H 1 (), the last inequality is a consequence of the Poincaré-Friedrichs inequality, see Proposition 2.18. Since we have F (v) Therefore, from (3.1) we obtain the estimate f(x)v(x) dx f 2 v 2 f 2 v H 1 (), (3.11) F f 2. (3.12) u H 1 () C f 2. (3.13) We now consider the general form of the operator L from (3.3), that is, the first-order and zero-order terms are also included. 21

Lemma 3.5 Let R n be open, assume 3.1. Then there exists c such that the bilinear form (3.4) satisfies for all v H 1 (). a(v, v) a 2 v 2 2 c v 2 2, (3.14) This inequality (or a variant of it) is called Gårding s inequality. Proof: We have, with b as in the proof of Lemma 3.3, a v 2 2 v(x), A(x) v(x) dx = a(v, v) b(x), v(x) v(x) + c(x)v(x) 2 dx a(v, v) + b v(x) v(x) dx + c v 2 2. We have v(x) v(x) dx = ( ) ( ) 1 ε v(x) ε v(x) dx ε 2 v 2 2 + 1 2ε v 2 2. Setting ε = a /b, we get a 2 v 2 2 a(v, v) + ( (b ) 2 2a ) + c v 2 2. From this, the assertion follows, setting c = (b ) 2 /2a + c. The bilinear form a is in general not H 1 ()-elliptic; it is if (3.5) holds for c =. However, the bilinear form a µ (u, v) = a(u, v) + µ u, v L 2 () (3.15) is even H 1 ()-elliptic for µ > µ := c since we then have by virtue of Lemma 3.5 a µ (v, v) a 2 v 2 2 + (µ c ) v 2 2. The bilinear form a µ is associated to the boundary value problem Lu + µu = f, in, (3.16) u =, on, (3.17) the operator L being taken from (3.3). The equation Lu + µu = f is also called the Helmholtz equation. The foregoing considerations together with the Lax-Milgram theorem 1.8 yield the following result. Proposition 3.6 Let R n be open, assume 3.1, let f L 2 (). Then there exists a µ R such that, for every µ > µ, the boundary value problem (3.16), (3.17) has a unique weak solution u H(). 1 22

Note that µ may be negative if Gårding s inequality holds for some c <. The derivation of the variational formulation (3.1) (3.4) shows that every classical solution of the boundary value problem is a weak solution (we refrain from presenting an exact formulation of this issue). On the other hand, the existence and uniqueness results of the variational theory only yield weak solutions. One then asks whether these weak solutions possess additional smoothness properties (thus, possibly, one might arrive at classical solutions). Such results are called regularity results. Let us come back to the boundary value problem u = f, in, (3.18) u =, auf. (3.19) Let us assume for the moment that u is a solution such that u L 2 () and the following partial integration is valid: n n u(x) 2 dx = i 2 u(x) j 2 u(x) dx = j i u(x) i j u(x) dx This would imply that = i,j=1 n ( i j u(x)) 2 dx. i,j=1 i,j=1 u 2 H 2 () = u 2 2 = f 2 2. (3.2) One therefore may hope that a solution u of (3.18), (3.19) has to be an element of H 2 () if the right side f belongs to L 2 (). An estimate like u H 2 () C f 2 obtained as above from manipulation of the problem equations is called an a priori estimate. This is an important heuristic technique; in a second step (usually the difficult part) one has to prove that the solution actually has the asserted regularity. For our elliptic boundary value problem, one way of proving H 2 regularity is to replace u by a difference quotient approximation for u and to obtain the existence of weak second derivatives from a compactness argument. Our function spaces have infinite dimension; it turns out that the appropriate notion is weak compactness. We are interested in the relation between weak derivatives and difference quotients. Let v : R and h be given. We consider the difference quotient (Dj h v)(x) = v(x + he j) v(x) h Here, e j denotes the j-th unit vector. We set, 1 j n. (3.21) D h v = (D h 1v,..., D h nv). (3.22) Lemma 3.7 Let, U R n be open, let U, p [1, ). Then D h j v L p (U) jv L p (), 1 j n, (3.23) for all v W 1,p () and all h with h < dist (U, ). 23

Proof: First, assume that v C () W 1,p (). For all x U and h < dist (U, ) we get 1 1 v(x + he j ) v(x) = j v(x + the j )h dt h j v(x + the j ) dt, thus U Dj h v(x) p dx = U 1 ( 1 p j v(x + the j ) dt) dx U j v(x + the j ) p dx dt 1 U 1 j v(x + the j ) p dt dx j v p L p () dt = jv p L p (). Let now v W 1,p () be arbitrary. According to Proposition 2.13 we choose a sequence (v k ) k N in C () W 1,p () satisfying v k v in W 1,p (). For fixed h, (3.23) holds for all v k. Since Dj h v k Dj h v in W 1,p (U) as well as j v k j v in L p (), passing to the limit k we obtain (3.23). Conversely, we want to derive the existence of a weak derivative j v from the uniform boundedness of the difference quotients Dj h v. We need a result of functional analysis: if (f k ) k N is a bounded sequence in L p (), 1 < p <, then there exists a subsequence (f km ) m N which converges weakly to some f L p (). This means that lim m f km (x)ϕ(x) dx = f(x)ϕ(x) dx, for all ϕ L q (), (3.24) where q is the exponent dual to p, 1 p + 1 q = 1. Moreover, the weak limit has the property f L p () lim inf m f k m L p (). (3.25) We write for weak convergence. f km f Proposition 3.8 Let, U R n be open, U, 1 < p <, v L p (), j {1,..., n}. Assume that there exists C > and h dist (U, ) such that Dj h v C (3.26) L p (U) for all h, h < h. Then we have j v L p (U), and j v L p (U) C. (3.27) Proof: Let ϕ C (U) be arbitrary. For h < h we have v(x)dj h ϕ(x) dx = v(x) ϕ(x + he j) ϕ(x) dx = h = (D h j v)(x)ϕ(x) dx. v(x) v(x he j ) ϕ(x) dx h 24

We choose a sequence (h m ) m N such that h m, h m, and a w L p (U) satisfying D hm j v w. Since D hm j ϕ j ϕ uniformly in, for sufficiently small h it follows that U so w = j v. v(x) j ϕ(x) dx = v(x) j ϕ(x) dx = lim m [ ] = lim (D hm j v)(x)ϕ(x) dx = m v(x)(d hm j ϕ)(x) dx w(x)ϕ(x) dx = U w(x)ϕ(x) dx, We come back to the question of the regularity of solutions of the elliptic equation Lu = f, in, (3.28) where L is the operator given in divergence form as in (3.3). We present a result on interior regularity, that is, on the regularity of a weak solution in parts of away from the boundary. Such a weak solution satisfies a(u, v) = f(x)v(x) dx, for all v H(), 1 (3.29) where a is the bilinear form given by (3.4). Proposition 3.9 Let, U be open, U, let f L 2 (). In addition to 3.1, assume that a ij C 1 () for 1 i, j n. Let u H 1 () be a weak solution of Lu = f in the sense of (3.29). Then we have u H 2 (U), and u H 2 (U) C( f L 2 () + u L 2 () ). (3.3) Here, C depends only on, U and on the coefficients of L. Proof: The idea of the proof is to estimate Dk h u L 2 () and then apply Proposition 3.8 to conclude that the weak second derivatives of u exist and that (3.3) holds. The variational equation (3.29) uses variations v defined on. In order to obtain results on U, one uses the technique of localization by cut-off functions. For this purpose, we choose open sets W, Y R n such that Let ζ C () such that U W Y. (3.31) ζ 1, ζ U = 1, ζ ( \ W ) =. (3.32) Such a ζ is called a cut-off function. For every v H() 1 we have u(x), A(x) v(x) dx = [f(x) b(x), u(x) c(x)u(x)]v(x) dx. (3.33) 25

Let h, h sufficiently small, let k {1,..., n}. As a test function in (3.33) we choose v = D h k (ζ2 D h ku). (3.34) This function replaces u as a test function. The difference quotient is defined as before, (D h ku)(x) = 1 h (u(x + he k) u(x)) (3.35) First, we estimate the left side of (3.33) from below. We have n [ u(x), A(x) v(x) dx = a ij (x) i u(x) j D h k (ζ2 Dku)(x) ] h dx = i,j=1 n i,j=1 = A 1 + A 2, ( D h k (a ij i u) ) (x) ( j (ζ 2 D h ku))(x) dx (3.36) where we set n A 1 = a ij (x + he k )(Dk( h i u))(x) ζ 2 (x)(dk( h j u))(x) dx, (3.37) i,j=1 and n A 2 = a ij (x + he k )(Dk( h i u))(x) (Dku)(x) h 2ζ(x) j ζ(x) dx i,j=1 + n i,j=1 i u D h ka ij [ζ 2 D h k( j u) + 2ζ( j ζ)(d h ku)] dx. (3.38) Since L is uniformly elliptic, we get (see 3.1) A 1 = (ζ(x)d h k u)(x), A(x + he k )ζ(x)(dk u)(x) h dx a (D h k u)(x) 2 ζ 2 (x) dx. (3.39) Next, we want to estimate A 2 from above. As a ij C 1 (), we have a ij C 1 (U) and D h ka ij (x) k a ij. (3.4) In the following computations will appear constants C i ; they depend on U, W, Y,, a ij, b i, c, ζ, but not on f, u and h. For A 2 we get, because ζ = outside of W, A 2 C 1 ζ(x) [ Dk u(x) h Dku(x) h + Dk u(x) h u(x) + Dku(x) h u(x) ] dx. W Therefore, for every ε > we get A 2 εc 1 ζ 2 (x) Dk u(x) h 2 dx + C 1 ε + C 1 Dku(x) h 2 + u(x) 2 dx. W 26 W D h ku(x) 2 + u(x) 2 dx (3.41)

Setting ε = a /(2C 1 ), we obtain A 2 a ζ 2 (x) D h 2 k u(x) 2 dx + C 2 By Lemma 3.7 we have W Dku(x) h 2 dx W Y D h ku(x) 2 + u(x) 2 dx. (3.42) u(x) 2 dx. (3.43) Altogether, the left side of (3.33) can be estimated as u(x), A(x) v(x) dx a ζ 2 (x) D h 2 k u(x) 2 dx C 3 Y u(x) 2 dx. (3.44) Secondly, we consider the right side of (3.33) and denote it by B. Because v = outside of Y, we have B C 4 ( f + u + u ) v dx. (3.45) Using once more Lemma 3.7, since ζ = outside of W, it follows that v 2 dx = D h k (ζ2 Dku) h 2 dx (ζ 2 Dku) h 2 dx Y W = 2ζ 4 Dk u h 2 + 8ζ 2 ζ 2 Dku h 2 dx W 2ζ 2 Dk u h 2 dx + C 5 u 2 dx. It follows that B ε W 2ζ 2 D h k u 2 dx + C 6 ε Y Y f 2 + u 2 + u 2 dx + C 5 Y Y u 2 dx. (3.46) Setting ε = a /8, we get B a 4 ζ 2 D h k u 2 dx + C 7 Y f 2 + u 2 + u 2 dx. (3.47) All in all, we now have obtained from (3.33) the estimate a D h 4 k u 2 dx a ζ 2 D h 4 k u 2 dx C 8 f 2 + u 2 + u 2 dx. (3.48) U It now follows from Proposition 3.8, since C 8 does not depend on h, that k u L 2 (U), k u 2 L 2 (U) C 9 f 2 + u 2 + u 2 dx, (3.49) and therefore u H 2 (U), as k was arbitrary. Finally, we want to estimate u 2 dx Y 27 Y Y

from above. Choose ζ C () such that ζ Y = 1. We test the variational equation with v = ζ 2 u. This yields as above (but simpler) a ζ 2 u 2 dx u, Aζ 2 u dx = u, A v dx u, 2Auζ ζ dx C 1 ( f + u + u ) u ζ dx a ζ 2 u 2 dx + C 11 f 2 + u 2 dx. 2 Consequently, Combining (3.49) and (3.5) we obtain Y u 2 dx C 12 u 2 H 2 (U) C 13( f 2 L 2 () + u 2 L 2 () ) f 2 + u 2 dx. (3.5) and then, again because of (3.5), the assertion. Note that the boundary behaviour of the solution u does not matter for this result. Proposition 3.9 shows that, in the given situation, the solution u gains two orders of differentiability with respect to the right side f. (This is optimal, since one certainly cannot expect more.) This also applies when f H k (). Proposition 3.1 In the situation of Proposition 3.9, assume in addition that for some k N we have a ij, b i, c C k+1 () and f H k (). Then u H k+2 (U) and u H k+2 (U) C( f H k () + u L 2 () ), (3.51) where C depends only on k,, U and on the coefficients of L. Proof: One uses induction with respect to k. Proposition 3.9 both yields the induction basis k = and the essential tool for the induction step. It is carried out in detail in the book L. Evans (Partial Differential Equations), section 6.3. 28

4 Boundary Conditions, Traces Inhomogeneous Dirichlet boundary conditions. The only boundary condition we have discussed so far is u = on, the so-called homogeneous Dirichlet boundary condition. Let us now consider the problem with an inhomogeneous Dirichlet boundary condition, u = f, in, u = g, on. (4.1) In the homogeneous case we did not treat u = on as a constraining equation, but instead incorporated it into the underlying function space by requiring a solution u to belong to H 1 (). This can be extended to (4.1) as follows. Assume that g H 1 () is given. Instead of u = g on we impose the condition u g H 1 (). (4.2) As H() 1 = C () by definition, this means that there exists a sequence ϕ n C () with ϕ n u g in H(). 1 As before, the weak formulation of the differential equation reads fϕ dx = ( u)ϕ dx = u, ϕ dx, for all ϕ C (), or, with a and F as before, This is equivalent to a(u, ϕ) = F (ϕ), for all ϕ C (). (4.3) a(u, v) = F (v), for all v H 1 (), (4.4) because for every v H 1 () we can find ϕ n C () with ϕ n v in H 1 (), and we may pass to the limit in (4.3). The variational formulation (4.2), (4.4) can be rewritten in terms of w = u g. We then have the problem to find w, w H 1 (), a(w, v) = F (v) a(g, v), for all v H 1 (). (4.5) By the results of the previous chapter, this problem has a unique solution w H 1 (). These considerations apply to the more general problem Lu = f, in, u = g, on. (4.6) Proposition 4.1 Let R n be open and bounded, let f L 2 (), g H 1 (), assume 3.1. Then (4.6) has a unique weak solution u H 1 (), and for some constant C independent from f and g. u H 1 () C( f L 2 () + g H 1 () ) (4.7) 29

Proof: From Proposition 3.4, applied to (4.5), we get that w H 1 () C 1(C 2 f L 2 () + C 3 g H 1 () ) Since u H 1 () w H 1 () + g H 1 (), the assertion follows. The procedure above gives rise to the question: How are functions g : R related to functions g H 1 ()? If v : R is continuous, then its restrictions to and are continuous. One would like to define an operator γ which maps H 1 () to a suitable class of functions defined on, such that for continuous functions it coincides with their restrictions. Traces on hyperplanes. One first considers a function v C (R n ) and its restriction to the hyperplane R n 1 {}. This defines a function and an operator γv : R n 1 R, (γv)(x ) = v(x, ), x = (x 1,..., x n 1 ) R n 1, (4.8) γ : C (R n ) C (R n 1 ). (4.9) The function γv is called the trace of v on the hyperplane R n 1 {}, and the operator γ is called a trace operator. Let X be a Banach space of functions over R n such that C () is dense in X. One wants to find a Banach space Y of functions over R n 1 such that γ can be extended to a linear and continuous operator from X to Y. It turns out that for X = H 1 (R n ) one can choose Y = L 2 (R n 1 ). However, γ(h 1 (R n )) turns out to be a proper subset of L 2 (R n 1 ); in order to describe the class of functions which are traces of functions in H 1, one has to introduce Sobolev spaces of fractional order. One way to define those spaces makes use of the Fourier transform. For v C () the Fourier transform is defined by (Fv)(ξ) = ˆv(ξ) = (2π) n 2 v(x)e i x,ξ dx, ξ R n. (4.1) R n It has the properties j v(ξ) = iξ jˆv(ξ), x j v(ξ) = i jˆv(ξ). (4.11) Since u(x)v(x) dx = R n û(ξ)ˆv(ξ) dξ, R n u, v C (), (4.12) it can be extended to an isometry F : L 2 (R n ) L 2 (R n ). Lemma 4.2 For every k N there exists C k > such that 1 (1 + ξ 2 ) k C k α k ξ 2α C k (1 + ξ 2 ) k, for all ξ R n. (4.13) Proof: Exercise. 3

Proposition 4.3 Let k N. Then u, v k,f = (1 + ξ 2 ) k û(ξ)ˆv(ξ) dξ (4.14) R n defines a scalar product on C (R n ). The associated norm v 2 k,f = (1 + ξ 2 ) k ˆv(ξ) 2 dξ (4.15) R n on C (R n ) is equivalent to the Sobolev norm H k (R ), that is, there exists C n k > such that 1 (1 + ξ 2 ) k ˆv(ξ) 2 dξ v 2 H C k (R n ) C k (1 + ξ 2 ) k ˆv(ξ) 2 dξ (4.16) k R n R n for all v C (). Proof: Let v C (R n ). Since v(x) 2 dx = R n ˆv(ξ) 2 dξ, R n (4.17) we get v 2 H k (R n ) = α v(x) 2 dx = α v(ξ) 2 dξ α k R n α k R n = ξ αˆv(ξ) 2 dξ = ˆv(ξ) 2 ξ 2α dξ. R n R n α k α k From Lemma 4.2 we now obtain (4.16). Proposition 4.4 (Sobolev space with fractional norm) Let s. Then u, v s,f = (1 + ξ 2 ) s û(ξ)ˆv(ξ) dξ, (4.18) R n v s,f = v, v s,f, (4.19) defines a scalar product on the space and H s (R n ) is a Hilbert space. H s (R n ) = {v : v L 2 (R n ), v s,f < }, (4.2) The spaces H s (R n ) are called Bessel potential spaces. Proof: We consider X = L 2 (R n ; µ) with the measure µ = fλ, f(ξ) = (1 + ξ 2 ) s, 31

that is, w dµ = (1 + ξ 2 ) s w(ξ) dξ. R n R n X is a Hilbert space, and H s (R n ) = {v : v L 2 (R n ), ˆv L 2 (R n ; µ)}. Since the Fourier transform is linear, H s (R n ) is a vector space, and u, v s,f = û, ˆv X, v s,f = ˆv X, defines a scalar product and its associated norm on H s (R n ). For v H s (R n ) we have v L 2 = ˆv L 2 ˆv X = v s,f. Let {v m } be a Cauchy sequence in H s (R n ). Then {ˆv m } is a Cauchy sequence in X, and {v m } is a Cauchy sequence in L 2 (R n ). Therefore, there exist w X and v L 2 (R n ) such that ˆv m w in X and v m v in L 2 (R n ), thus moreover ˆv m w in L 2 (R n ) and ˆv m ˆv in L 2 (R n ). It follows that w = ˆv and therefore v H s (R n ). Lemma 4.5 Let k N. Then C (R n ) is dense in H k (R n ) with respect to H k (R n ), that is, we have H k (R n ) = H k (R n ). Proof: Since C (R n ) H k (R n ) is dense in H k (R n ) by Proposition 2.13, it suffices to find for every v C (R n ) H k (R n ) a sequence (v m ) m N in C (R n ) such that v m v in H k (R n ). To this purpose, choose ψ C (R n ) such that ψ(x) = 1 for x 1 and ψ(x) = for x 2. We set ( x v m (x) = ψ v(x). (4.21) m) Then for all multi-indices α with α k we have { ( α (v v m )(x) = α (1 ψ( x ) m ))v(x) C β α β v(x), x > m, =, x m, where C does not depend on m. It follows that R n α (v v m )(x) 2 dx C β α x >m β v(x) 2 dx for m, since β v L 2 (R n ) for all β k. Lemma 4.6 For s the space C (R n ) is dense in H s (R n ) with respect to s,f. Proof: This is proved, for example, in the book of Wloka: Partial Differential Equations (in German and in English), on p.96 of the German edition. With the aid of Lemma 4.5 and Lemma 4.6 one can prove that, for integer k, the two definitions of H k (R n ) (based on weak derivatives, or on the Fourier transform) yield the same space and that the corresponding norms are equivalent on H k (R n ) (and not only on the subspace C (R n ), as was obtained in Proposition 4.3). The proof is analogous to that of Proposition 4.4. 32

Lemma 4.7 For s t we have The embedding j : H t (R n ) H s (R n ) is continuous. H t (R n ) H s (R n ). (4.22) Proof: From (1 + ξ 2 ) s (1 + ξ 2 ) t we obtain v s,f v t,f. These results now enable us to define the trace operator γ on H s (R n ) such that γv : R n 1 R, (γv)(x ) = v(x, ), x = (x 1,..., x n 1 ) R n 1, (4.23) for smooth functions v. Proposition 4.8 (Trace theorem) Let s R, s > 1. There exists a unique linear and continuous mapping 2 such that (4.23) holds for all v C (R n ). γ : H s (R n ) H s 1 2 (R n 1 ) (4.24) Proof: It suffices to prove that there exists C > such that γϕ H s 1 2 (R n 1 ) C ϕ H s (R n ), for all ϕ C (R n ), (4.25) because then γ is linear and continuous on the dense subset C (R n ) of H s (R n ), and therefore can be extended uniquely to a linear continuous mapping on H s (R n ), by a theorem of functional analysis. Let ϕ C (R n ). We define We fix x R n 1 and define g(x ) = ϕ(x, ), x R n 1. (4.26) ψ(x n ) = ϕ(x, x n ), x n R. (4.27) Applying the Fourier transform in R, we get ˆψ(ξ n ) = 1 ϕ(x, x n )e ixnξn dx n, (4.28) 2π R thus g(x ) = ϕ(x, ) = ψ() = 1 2π = 1 2π R 1 2π R R ˆψ(ξ n )e iξn dξ n ϕ(x, x n )e ixnξn dx n dξ n. 33