The optimal partial transport problem

Similar documents
PARTIAL REGULARITY OF BRENIER SOLUTIONS OF THE MONGE-AMPÈRE EQUATION

Local semiconvexity of Kantorovich potentials on non-compact manifolds

C 1 regularity of solutions of the Monge-Ampère equation for optimal transport in dimension two

REGULARITY OF MONOTONE TRANSPORT MAPS BETWEEN UNBOUNDED DOMAINS

Optimal Transportation. Nonlinear Partial Differential Equations

FULL CHARACTERIZATION OF OPTIMAL TRANSPORT PLANS FOR CONCAVE COSTS

Optimal transportation on non-compact manifolds

Recall that if X is a compact metric space, C(X), the space of continuous (real-valued) functions on X, is a Banach space with the norm

Optimization and Optimal Control in Banach Spaces

Topological properties

Lecture Notes in Advanced Calculus 1 (80315) Raz Kupferman Institute of Mathematics The Hebrew University

Course 212: Academic Year Section 1: Metric Spaces

Centre for Mathematics and Its Applications The Australian National University Canberra, ACT 0200 Australia. 1. Introduction

Maths 212: Homework Solutions

MAT 570 REAL ANALYSIS LECTURE NOTES. Contents. 1. Sets Functions Countability Axiom of choice Equivalence relations 9

3 (Due ). Let A X consist of points (x, y) such that either x or y is a rational number. Is A measurable? What is its Lebesgue measure?

1 )(y 0) {1}. Thus, the total count of points in (F 1 (y)) is equal to deg y0

1 Directional Derivatives and Differentiability

g 2 (x) (1/3)M 1 = (1/3)(2/3)M.

Analysis Finite and Infinite Sets The Real Numbers The Cantor Set

+ 2x sin x. f(b i ) f(a i ) < ɛ. i=1. i=1

On the regularity of solutions of optimal transportation problems

A few words about the MTW tensor

INVERSE FUNCTION THEOREM and SURFACES IN R n

Chapter 1. Measure Spaces. 1.1 Algebras and σ algebras of sets Notation and preliminaries

LECTURE 15: COMPLETENESS AND CONVEXITY

The Monge problem on non-compact manifolds

Problem Set. Problem Set #1. Math 5322, Fall March 4, 2002 ANSWERS

Integration on Measure Spaces

A LOCALIZATION PROPERTY AT THE BOUNDARY FOR MONGE-AMPERE EQUATION

On a Class of Multidimensional Optimal Transportation Problems

Problem Set 2: Solutions Math 201A: Fall 2016

Lecture Notes Introduction to Ergodic Theory

CHAPTER 6. Differentiation

MATH 202B - Problem Set 5

P-adic Functions - Part 1

Real Analysis Math 131AH Rudin, Chapter #1. Dominique Abdi

Functional Analysis. Franck Sueur Metric spaces Definitions Completeness Compactness Separability...

Topological properties of Z p and Q p and Euclidean models

CHAPTER I THE RIESZ REPRESENTATION THEOREM

(U) =, if 0 U, 1 U, (U) = X, if 0 U, and 1 U. (U) = E, if 0 U, but 1 U. (U) = X \ E if 0 U, but 1 U. n=1 A n, then A M.

2 (Bonus). Let A X consist of points (x, y) such that either x or y is a rational number. Is A measurable? What is its Lebesgue measure?

Banach Spaces V: A Closer Look at the w- and the w -Topologies

3 Integration and Expectation

2. Function spaces and approximation

2 Lebesgue integration

A STRATEGY FOR NON-STRICTLY CONVEX TRANSPORT COSTS AND THE EXAMPLE OF x y p IN R 2

arxiv: v2 [math.ag] 24 Jun 2015

Existence and Uniqueness

6. Duals of L p spaces

Calculus of Variations

Introduction to Real Analysis Alternative Chapter 1

Contents Ordered Fields... 2 Ordered sets and fields... 2 Construction of the Reals 1: Dedekind Cuts... 2 Metric Spaces... 3

Thus, X is connected by Problem 4. Case 3: X = (a, b]. This case is analogous to Case 2. Case 4: X = (a, b). Choose ε < b a

Division of the Humanities and Social Sciences. Supergradients. KC Border Fall 2001 v ::15.45

Math 426 Homework 4 Due 3 November 2017

Convexity in R n. The following lemma will be needed in a while. Lemma 1 Let x E, u R n. If τ I(x, u), τ 0, define. f(x + τu) f(x). τ.

Chapter 2. Metric Spaces. 2.1 Metric Spaces

Chapter 6. Integration. 1. Integrals of Nonnegative Functions. a j µ(e j ) (ca j )µ(e j ) = c X. and ψ =

and BV loc R N ; R d)

Behaviour of Lipschitz functions on negligible sets. Non-differentiability in R. Outline

L p Spaces and Convexity

CHAPTER 9. Embedding theorems

1/12/05: sec 3.1 and my article: How good is the Lebesgue measure?, Math. Intelligencer 11(2) (1989),

On John type ellipsoids

Lebesgue Integration: A non-rigorous introduction. What is wrong with Riemann integration?

Topology of the set of singularities of a solution of the Hamilton-Jacobi Equation

Exercise 1. Let f be a nonnegative measurable function. Show that. where ϕ is taken over all simple functions with ϕ f. k 1.

Representation of the polar cone of convex functions and applications

Contents: 1. Minimization. 2. The theorem of Lions-Stampacchia for variational inequalities. 3. Γ -Convergence. 4. Duality mapping.

Moment Measures. Bo az Klartag. Tel Aviv University. Talk at the asymptotic geometric analysis seminar. Tel Aviv, May 2013

1. Introduction Boundary estimates for the second derivatives of the solution to the Dirichlet problem for the Monge-Ampere equation

Sobolev regularity for the Monge-Ampère equation, with application to the semigeostrophic equations

Measurable functions are approximately nice, even if look terrible.

Laplace s Equation. Chapter Mean Value Formulas

Remarks on the Gauss-Green Theorem. Michael Taylor

SMSTC (2017/18) Geometry and Topology 2.

MATH 426, TOPOLOGY. p 1.

Geometry and topology of continuous best and near best approximations

Weak convergence. Amsterdam, 13 November Leiden University. Limit theorems. Shota Gugushvili. Generalities. Criteria

fy (X(g)) Y (f)x(g) gy (X(f)) Y (g)x(f)) = fx(y (g)) + gx(y (f)) fy (X(g)) gy (X(f))

Vector Integration and Disintegration of Measures. Michael Taylor

ESTIMATES FOR THE MONGE-AMPERE EQUATION

Notes on Distributions

EXPOSITORY NOTES ON DISTRIBUTION THEORY, FALL 2018

3 Measurable Functions

1 Cheeger differentiation

Sobolev spaces. May 18

n [ F (b j ) F (a j ) ], n j=1(a j, b j ] E (4.1)

5 Measure theory II. (or. lim. Prove the proposition. 5. For fixed F A and φ M define the restriction of φ on F by writing.

1 The Local-to-Global Lemma

SEPARABILITY AND COMPLETENESS FOR THE WASSERSTEIN DISTANCE

Real Analysis Notes. Thomas Goller

MAT 257, Handout 13: December 5-7, 2011.

U e = E (U\E) e E e + U\E e. (1.6)

MATHS 730 FC Lecture Notes March 5, Introduction

Analysis Comprehensive Exam Questions Fall F(x) = 1 x. f(t)dt. t 1 2. tf 2 (t)dt. and g(t, x) = 2 t. 2 t

CHAPTER 7. Connectedness

Math 54: Topology. Syllabus, problems and solutions. Pierre Clare. Dartmouth College, Summer 2015

PHASE TRANSITIONS: REGULARITY OF FLAT LEVEL SETS

Transcription:

The optimal partial transport problem Alessio Figalli Abstract Given two densities f and g, we consider the problem of transporting a fraction m [0, min{ f L 1, g L 1}] of the mass of f onto g minimizing a transportation cost. If the cost per unit of mass is given by x y 2, we will see that uniqueness of solutions holds for m [ f g L 1, min{ f L 1, g L 1}]. This extends the result of Caffarelli and McCann in [8], where the authors consider two densities with disjoint supports. The free boundaries of the active regions are shown to be (n 1)-rectifiable (provided the supports of f and g have Lipschitz boundaries), and under some weak regularity assumptions on the geometry of the supports they are also locally semiconvex. Moreover, assuming f and g supported on two bounded strictly convex sets Ω, Λ R n, and bounded away from zero and infinity on their respective supports, C 0,α loc regularity of the optimal transport map and local C1 regularity of the free boundaries away from Ω Λ are shown. Finally, the optimal transport map extends to a global homeomorphism between the active regions. 1 Introduction In a recent paper [8], Caffarelli and McCann studied the following variant of the Monge- Kantorovich problem: let f, g L 1 (R n ) be two nonnegative functions, and denote by Γ (f, g) the set of nonnegative finite Borel measures on R n R n whose first and second marginals are dominated by f and g respectively, i.e. γ(a R n ) f(x) dx, γ(r n A) g(y) dy A for all A R n Borel. Denoting by M (γ) the mass of γ (i.e. M (γ) := R n R dγ), fix a n certain amount m [0, min{ f L 1, g L 1}] which represents the mass one wants to transport, and consider the following partial transport problem: minimize C(γ) := x y 2 dγ(x, y) R n R n among all γ Γ (f, g) with M (γ) = m. Using weak topologies, one can easily prove existence of minimizers for any fixed amount of mass m [0, min{ f L 1, g L 1}] (see Section 2). We denote by Γ o (m) the set of such minimizers. Centre de Mathématiques Laurent Schwartz, UMR 7640, Ecole Polytechnique - 91128 Palaiseau, France. e-mail: figalli@math.polytechnique.fr A 1

In general one cannot expect uniqueness of minimizers. Indeed, if m R f g (where (f n g)(x) := min{f(x), g(x)}), any γ supported on the diagonal {x = y} with marginals dominated by f g is a minimizer with zero cost. To ensure uniqueness, in [8] the authors assume f and g to have disjoint supports. Under this assumption they are able to prove (as in the classical Monge-Kantorovich problem) that there exists a (unique) convex function ψ such that the unique minimizer is concentrated on the graph of ψ (see [8, Section 2]). This ψ is also shown to solve in a weak sense a Monge-Ampère double obstacle problem (see [8, Section 4]). Moreover, strengthening the disjointness assumption into the hypothesis on the existence of a hyperplane separating the supports of the two measures, the authors prove a semiconvexity result on the free boundaries (see [8, Sections 5] 1 ). Furthermore, under some classical regularity assumptions on the measures and on their supports, local C 1,α regularity of ψ (which is equivalent to local C 0,α regularity of the transport map) and on the free boundaries of the active regions is shown (see [8, Sections 6-7]). The aim of this paper is to understand what happens if we remove the disjointness assumption. In Section 2 we will see that, although minimizers are non-unique for m < R f g (but in n this case the set of minimizers can be trivially described), uniqueness holds for any m R f g. n Moreover, exactly as in [8], the unique minimizer is concentrated on the graph of the gradient of a convex function. In Remark 2.11 we will also see that our argument for the uniqueness of minimizers extends to more general cost functions on R n, and also to the case where f and g are two densities on a Riemannian manifold with c(x, y) = d(x, y) 2, d(x, y) being the Riemannian distance. Then, in Section 3 we will prove that the marginals of the minimizers always dominate the common mass f g (that is all the common mass is both source and target). This property, which has an interest in its own, will also play a crucial role in the regularity results of Section 4. Indeed, thanks to this domination property, in Paragraph 4.1 we can prove a local semiconvexity result on the free boundaries, which reduces to the Caffarelli-McCann result in the disjoint case (see Propositions 4.4 and 4.5 for a precise statement). Paragraph 4.2 is devoted to the regularity properties of the transport map and the free boundary. First, as in [8], we will prove local C 0,α regularity of the transport map (see Theorem 4.8). On the other hand we will see that in our case something completely different happens: usually, assuming C regularity on the density of f and g (together with some convexity assumption on their supports), one can show that the transport map is C too. In our case we will show that C 0,α loc regularity is in some sense optimal: we can find two C densities on R, supported on two bounded intervals and bounded away from zero on their supports, such that the transport map is not C 1 (see Remark 4.9). Regarding the regularity of the free boundaries, we will prove a local C 1 regularity away from supp(f) supp(g) (see Theorem 4.11 for a precise statement). Furthermore, as in [8, Section 6], we will see that the transport map extends to a global homeomorphism between the active regions (see Theorem 4.10). Finally we will show how one can adapt the proofs in [8, Section 6] to deduce properties like 1 In [8] the authors speak about the semiconcavity of the free boundary, but this is equivalent to semiconvexity up to a change of coordinates. 2

the path-connectedness of the active regions, or the fact that free boundary never maps to free boundary. In Remark 4.15 we also discuss a possible improvement of the Cloc 1 regularity of the free boundaries away from supp(f) supp(g) into a C 1,α loc regularity. 1.1 Preliminaries on measure theory and convex functions We first recall some definitions which will play an important role in the paper: Definition 1.1 (Push-forward) Let X, Y be complete separable metric spaces, µ a finite Borel measure on X, and F : X Y a Borel map. The push-forward F # µ is the measure on Y defined by F # µ(b) = µ(f 1 (B)) for any Borel set B Y. Definition 1.2 (Marginals) Let X, Y be complete separable metric spaces, and let γ be a finite Borel measure on X Y. We say that µ and ν are, respectively, the first and the second marginals of γ if h 1 (x) dγ(x, y) = h 1 (x) dµ(x), h 2 (y) dγ(x, y) = h 2 (y) dν(y), X Y X X Y for all bounded continuous functions h 1 : X R, h 2 : Y R. Definition 1.3 (Minumum and maximum of measures) Let X be a complete separable metric spaces, µ, ν be two finite Borel measure on X. We define µ ν and µ ν by µ ν(b) := inf { µ(b 1 ) + ν(b 2 ) : B 1 B 2 =, B 1 B 2 = B, B 1, B 2 Borel } B X Borel, µ ν(b) := sup { µ(b 1 ) + ν(b 2 ) : B 1 B 2 =, B 1 B 2 = B, B 1, B 2 Borel } B X Borel. Moreover, we say that µ ν if µ(b) ν(b) for all B X Borel. It is not difficult to check that µ ν and µ ν are still finite Borel measures. Moreover the equality µ ν + µ ν = µ + ν holds. Indeed, given a Borel set B, assume for simplicity that we have µ ν(b) = µ(b 1 ) + ν(b 2 ) for some partition B 1, B 2 of B. Then, given any other partition B 1, B 2 of B, we have µ(b 2 )+ν(b 1 ) = µ(b) µ(b 1 ) + ν(b) ν(b 2 ) µ(b) µ(b 2) + ν(b) ν(b 1) = µ(b 1) + ν(b 2), that is µ ν(b) = µ(b 2 ) + ν(b 1 ), which gives µ ν(b) + µ ν(b) = µ(b) + ν(b) as wanted (in the general case when the infimum is not attained, it suffices to consider B 1 and B 2 such that µ ν(b) µ(b 1 ) + ν(b 2 ) ε for some ε > 0 arbitrarily small). Let us also recall the so-called Disintegration Theorem (see for instance [4, Theorem 5.3.1], [10, III-70]). We state it in a particular case, which is however sufficient for our purpose: Theorem 1.4 (Disintegration of measures) Let X, Y be complete separable metric spaces, and let γ be a finite Borel measure on X Y. Denote by µ and ν the marginals of γ on the first Y 3

and second factor respectively. Then there exists two measurable families of probability measures (γ x ) x X and (γ y ) y Y such that γ(dx, dy) = γ x (dy) dµ(x) = γ y (dx) dν(y), i.e. X Y ϕ(x, y) dγ(x, y) = X ( Y ) ( ) ϕ(x, y) dγ x (y) dµ(x) = ϕ(x, y) dγ y (x) dν(y) Y X for all bounded continuous functions ϕ : X Y R. We now recall some classical definitions on convex functions: Definition 1.5 (Subdifferential and Legendre transform) Let φ : R n R {+ } be a convex function. If φ(x) is finite, the subdifferential of φ at x is defined as φ(x) := {p R n : φ(z) φ(x) p, z x z R n }. Moreover, for A R n, we define φ(a) := x A φ(x). The Legendre transform φ : R n R {+ } of φ is the convex function defined as φ (y) := sup x R n y, x φ(x). It is well-known that a convex function and its Legendre transform are related by the following properties: y φ(x) x φ (y), φ(x) + φ (y) x, y x, y R n with equality if and only if y φ(x). Since convex functions are locally Lipschitz in the interior of their domain, by Rademacher s Theorem they are differentiable a.e. Indeed, a stronger result holds (see for instance [2] or [17, Chapter 14, First Appendix]): Theorem 1.6 (Alexandrov) Let A R n be an open set, φ : A R be a convex function. Then φ is twice differentiable (Lebesgue) a.e., that is for almost every x A there exists a linear map 2 φ(x) : R n R n such that φ(x + v) = φ(x) + φ(x), v + 1 2 2 φ(x) v, v + o( v 2 ) v R n. Moreover φ(x) is differentiable a.e. in A, and its differential coincides with 2 φ(x). 4

1.2 Notation In this paper, we will repeatedly adopt the following notation and conventions: - If µ is a (Borel) measure on R n with density h with respect to the Lebesgue measure L n, we often use h in place of µ = hl n. For example, we write F # h in place of F # (hl n ). - If γ Γ (f, g), we denote by f γ and g γ the densities of the first and the second marginal of γ respectively (observe that the constraint γ Γ (f, g) implies that both marginals are absolutely continuous with respect to the Lebesgue measure). - If h 1, h 2 : R n R are Borel functions, we will often consider the Borel set {h 1 > h 2 }. We remark that if h 1, h 2 are defined up to a set of (Lebesgue) measure zero, then {h 1 > h 2 } is well-defined up to a set of measure zero. - If B 1 and B 2 are two Borel set (possibly defined up to a set of measure zero), we say that a.e. B 1 B 2 if the inclusion B 1 B 2 holds up to a set of measure zero. 2 Properties of minimizers As we already said in the introduction, we consider f, g L 1 (R n ) two nonnegative Borel function, and we denote by Γ (f, g) the set of nonnegative finite Borel measures on R n R n whose first and second marginals are dominated by f and g respectively. We will always assume for simplicity that both f and g are compactly supported, although many results holds under more general assumptions (for example if f and g have finite second moments). Let m max := min{ f L 1, g L 1}. We fix m [0, m max ], and we consider the minimization problem C(m) := min C(γ), (2.1) γ Γ (f,g), M (γ)=m where M (γ) = R n R n dγ is the mass of γ, and C(γ) := R n R n x y 2 dγ is the cost of γ. Since the set Γ(m) := {γ Γ (f, g), M (γ) = m} is non-empty and weakly compact, it is simple to prove by the direct method of the calculus of variations existence of minimizers (see [17, Chapter 4] or [8, Lemma 2.2]). Let Γ o (m) Γ(m) denotes the set of minimizers. We want to understand their structure. Let us define m min := R n f g. We observe that for m m min, given any density 0 h f g with mass m (i.e. R n h = m), the plan γ := (Id Id) # h is optimal, since its cost is zero. Moreover, since all minimizers have zero cost, they are clearly of this form. Thus the set of minimizers is not a singleton except for m = m min, in which case the unique minimizer is given by (Id Id) # (f g). We now want to study the case m > m min. The main difference between our strategy and the one developed in [8] is the following: in [8, Section 2] the authors introduce a Lagrange multiplier for the mass constraint, add a point at infinity which acts as a tariff-free reservoir, and study the relations given by classical duality 5

theorems. In this way they are able to deduce existence and uniqueness of minimizers when the supports of f and g are disjoint. Our strategy is instead to attack directly the minimization problem by studying the convexity properties of the function m C(m), and then looking at the consequences that follow from them. In this way we will prove that there exists a unique minimizer which is concentrated on the graph of the gradient of a convex function. In particular, if f and g have disjoint support, we recover the uniqueness result in [8]. The proof of our result is divided in two steps: 1. The function m C(m) is convex on [0, m max ]. Moreover, if m 0 is a point of strict convexity for C(m), then Γ o (m 0 ) is a singleton. 2. The function m C(m) is strictly convex on (m min, m max ]. Combining the above steps, we immediately deduce the uniqueness of minimizers for m > m min. We remark that the proof of the second step will require an analysis of the structure of minimizers, that will be obtained applying in a careful way Brenier s Theorem (see Theorem 2.3). 2.1 Step 1: properties of C(m) Here we show some elementary properties of the function m C(m). Lemma 2.1 (Convexity of C(m)) The function m C(m) is identically zero on [0, m min ], and it is convex on [0, m max ]. Proof. The fact that C(m) = 0 for m [0, m min ] follows from the observation that any γ supported on the diagonal with marginals dominated by f g is a minimizer with zero cost. The convexity of C(m) is a simple consequence of the linearity of the functional and the convexity of the constraints: if γ 1 Γ o (m 1 ) and γ 2 Γ o (m 2 ), for any λ [0, 1] we have λγ 1 + (1 λ)γ 2 Γ(λm 1 + (1 λ)m 2 ). Therefore we obtain C ( ) ( ) λm 1 + (1 λ)m 2 C λγ1 + (1 λ)γ 2 = λc(γ 1 ) + (1 λ)c(γ 2 ) = λc(m 1 ) + (1 λ)c(m 2 ) λ [0, 1]. Proposition 2.2 (Strict convexity implies uniqueness) Let γ 1, γ 2 Γ o (m 0 ) (possibly γ 1 = γ 2 ), and define γ := γ 1 γ 2, γ + := γ 1 γ 2. If we denote m = M (γ 1 γ 2 ) and m + = M (γ 1 γ 2 ), then C(m) is affine on [m, m + ]. In particular, if C(m) its strictly convex at m 0, then m = m + and Γ o (m 0 ) is a singleton. Proof. As γ + γ + = γ 1 + γ 2 (recall the discussion after Definition 1.3), and γ, γ + Γ (f, g) thanks to Theorem 2.6 below, we have C(m ) + C(m + ) C(γ ) + C(γ + ) = C(γ 1 ) + C(γ 2 ) = 2C(m 0 ), 6

that is C(m + ) + C(m ) 2 C(m 0 ). Since m ++m 2 = m 0 and C(m) is convex, we deduce that C(m) is affine on [m, m + ]. Therefore, if C(m) its strictly convex at m 0, then m = m +, which implies γ = γ +. Thus γ 1 = γ 2, and by the arbitrariness of γ 1, γ 2 Γ o (m 0 ) we deduce that Γ o (m 0 ) is a singleton. 2.2 Graph property of minimizers We will need the following result for the classical Monge-Kantorovich problem (see [5, 6, 14, 15]): Theorem 2.3 Let f, g L 1 (R n ) be nonnegative compactly supported functions such that R f = n R g, and consider the Monge-Kantorovich problem: n minimize x y 2 dγ(x, y) R n R n among all γ which have f and g as first and second marginals, respectively. Then there exists a unique optimal γ o. Moreover there exists a globally Lipschitz convex function ψ : R n R such that ψ(x) supp(g ) for a.e. x R n, and γ o = (Id ψ) # f = ( ψ Id) # g, where ψ denotes the Legendre transform of ψ. This implies in particular ψ ( ψ(x)) = x f -a.e., ψ( ψ (y)) = y g -a.e. (2.2) Finally ψ solves the Monge-Ampère equation det( 2 ψ)(x) = f (x) g ( ψ(x)) f -a.e. (2.3) Conversely, if ϕ : R n R is a convex function such that ϕ # f = g, then (Id ϕ) # f solves the Monge-Kantorovich problem. Let us first prove the following key result, which shows that solutions of the partial optimal transport problem are solutions of an optimal transport problem: Proposition 2.4 (Minimizers solve an optimal transport problem) Let m [0, m max ], γ Γ o (m), and consider the Monge-Kantorovich problem: minimize C(γ) = x y 2 dγ(x, y) R n R n among all γ which have f + (g g γ ) and g + (f f γ ) as first and second marginals, respectively. Then the unique minimizer is given by γ + (Id Id) # ( (f f γ ) + (g g γ ) ). (2.4) 7

Remark 2.5 Although using Theorem 2.3 the above proposition could be proved in a simpler way, we prefer to give a proof independent of it to show that the minimizing property of the plan defined by (2.4) holds true whenever c(x, y) is a nonnegative cost function. Proof. Let γ have marginals f = f + (g g γ ) and ḡ = g + (f f γ ). The idea is to prove that, since R f = m + n R (f n f γ ), γ has to send at least an amount m of the mass of f onto g. In particular there exists γ γ such that M (γ ) m and γ Γ (f, g). From this fact the result will follow easily. Let us therefore prove the existence of γ. We consider the disintegration of γ with respect to its first and second marginals respectively, that is γ(dx, dy) = γ x (dy) f(x) dx = γ y (dx) ḡ(y) dy (see Theorem 1.4). Then we define γ (dx, dy) := γ x (dy) f(x) dx, and we denote by f and g its marginals. It is clear that f = f, g g + (f f γ ). Since M (γ ) = R n f = m + R n (f f γ ), it is not difficult to see that R n g g m; indeed g g g ( g + (f f γ ) ) g (f f γ ) g (f f γ ) = m. R n R n R n R n R n Thus we immediately get that γ (dx, dy) := γ y (dx) (g g)(y) dy is the desired subplan. Now, since γ Γ (f, g) and M (γ ) m, we get C(γ ) C(m). As C(γ) C(γ ) and γ was arbitrary, we have proved that the infimum in the Monge-Kantorovich problem is greater or equal than C(m). On the other hand, if we consider the plan γ defined by Equation (2.4), then C( γ) = C( γ) = C(m). Thus γ is a minimizer. Thanks to the above proposition, we can prove the following: Theorem 2.6 (Graph structure of minimizers) Let m [0, m max ]. There exists a globally Lipschitz convex function ψ such that ψ(x) {g > 0} for a.e. x R n, and all γ Γ o (m) are concentrated on the graph of ψ. Moreover ψ is injective f-a.e., and for any γ Γ o (m) ψ(x) = x for a.e. x {f γ < f} {g γ < g}, (2.5) and f γ (x) = g γ (x) for a.e. x {f γ < f} {g γ < g}. (2.6) 8

Remark 2.7 The key point of the above statement is that the function ψ is the same for all minimizers. Since we will prove later that for m [m min, m max ] there exists a unique minimizer, while for m [0, m min ] the minimizers are all concentrated on the graph of the identity map, the fact that ψ is independent of the minimizer is not interesting in itself. However we preferred to state the theorem in this form, because we believe that the strategy of the proof is interesting and could be used in other situations in which uniqueness of minimizers fails. Proof. It is simple to see that Γ o (m) is compact with respect to the weak topology of measures. In particular Γ o (m) is separable, and we can find a dense countable subset (γ n ) n=1 Γo (m). Denote by γ := n=1 1 2 γ n n. Since the minimization problem (2.1) is linear and the constraints are convex, γ Γ o (m). The idea now is that, if we prove that γ is concentrated on a graph, then all γ Γ o (m) have to be concentrated on such a graph. To prove the graph property of γ we apply Proposition 2.4: we know that the plan γ := γ + (Id Id) # ( (f f γ ) + (g g γ ) ), (2.7) solves the Monge-Kantorovich problem: minimize C(γ) = x y 2 dγ(x, y) R n R n among all γ which have f + (g g γ ) and g + (f f γ ) as first and second marginals, respectively. By Theorem 2.3 we deduce that γ is concentrated on the graph of the gradient of a convex function ψ, that is γ = (Id ψ) # ( f + (g g γ ) ), (2.8) and by (2.2) we deduce that ψ is injective f-a.e. Combining (2.8) with (2.7) we deduce that γ is both concentrated on the graph of Id and on the graph of ψ for a.e. x {f γ < f} {g γ < g}, which implies ψ(x) = x for a.e. x {f γ < f} {g γ < g}. (2.9) Fix now any γ Γ o (m), and let us prove that ψ(x) = x for a.e. x {f γ < f} (the case x {g γ < g} being analogous). By (2.9) we know that ψ(x) = x for a.e. x n( {fγn < f} {g γn < g} ). Thus it suffices to prove that, for any k N, { f f γ > 1 } a.e. k { f f γn > 1 }. 2k n=1 First of all we observe that, thanks to the density of (γ n ) n=1 in Γo (m), the set (f γn ) n=1 is dense in {f γ : γ Γ o (m)} with respect to the the weak topology of measures. On the other hand, since f γ f for all γ Γ o (m), the density of (f γn ) n=1 holds also with respect to the the weak topology of L 1. Therefore, if by contradiction there exists a Borel set A, with A > 0, such that A {f f γ > 1 k } and A {f f γ n > 1 2k } = for all n 1, we would obtain f f γ 1 k A, f f γn 1 A n 1, 2k A A 9

which contradicts the density of (f γn ) n=1 in {f γ : γ Γ o (m)} in the weak topology of L 1. This proves (2.5). Observing that ψ # f γ = g γ, applying (2.3) with f = f γ and g = g γ, we get det( 2 ψ)(x) = f γ(x) g γ ( ψ(x)) f γ -a.e. It is a standard measure theory result that, if x is a density point for the set { ψ(x) = x}, and ψ(x) is differentiable at x, then 2 ψ( x) = I n, where I n denotes the identity matrix on R n (see [3, Proposition 3.73(c)]). By this fact and (2.5), (2.6) follows easily. We now use the above theorem to show a domination property of minimizers which has an interest in itself, and which will play a crucial role in the regularity of the free boundary of the active regions. Proposition 2.8 (Common mass is both source and target) Let m 0 [m min, m max ], γ Γ o (m). Then f γ f g, g γ f g (that is, all the common mass is both source and target for every minimizer). Proof. For m 0 = m min the result is clear since, as we already said, the unique minimizer is given by (Id Id) # (f g). So we can assume m 0 > m min. Applying Theorem 2.6 we know that we can write γ = (Id ψ) # f γ, with ψ invertible f γ -a.e. We observe that, since {f γ < f g} {f γ < f} and {g γ < f g} {g γ < g}, by (2.5) we get ψ(x) = x for a.e. x {f γ < f g} {g γ < f g}. Moreover, by (2.6), f γ (x) = g γ (x) for a.e. x {f γ < f g} {g γ < f g}. Thus either both f γ and g γ are greater or equal of f g, or {f γ < f g} = {g γ < f g} = and f γ = g γ on that set. Suppose by contradiction that h := [f g f γ ] + = [f g g γ ] + is not identically zero. Let m h > 0 denote the mass of h, and consider the plan γ h = (Id Id) # h. Since f γ+γh = f γ + f γh = f γ + [f g f γ ] + f and g γ+γh = g γ + g γh = g γ + [f g g γ ] + g, we have γ + γ h Γ (f, g). Observing that C(γ h ) = 0, we get C(m 0 + m h ) C(γ + γ h ) = C(γ) = C(m 0 ). As C(m) is convex and increasing, it has to be constant on the interval [0, m 0 + m h ], and since C(0) = 0, C(m) 0 on [0, m 0 + m h ]. This is impossible if m 0 m min, since it would imply that a mass m 0 + m h > m min should stay at rest. This contradiction gives the desired result. 2.3 Step 2: strict convexity of C(m) In order to prove the strict convexity property, we first need to show that a linear part in the graph of C(m) would imply a monotonicity result on minimizers. Lemma 2.9 Let m min m 1 < m 2 m max, and assume that C(m) is linear on [m 1, m 2 ]. Fix γ 1 Γ o (m 1 ). Then there exists γ 2 Γ o (m 2 ) such that γ 1 γ 2. 10

Proof. Let γ 2 Γ o (m 2 ), and assume γ 1 γ 2. We will modify γ 2 into γ 2 so that γ 2 Γ o (m 2 ) and γ 1 γ 2. Let us consider γ 1+γ 2 2. Since C(m) is linear on [m 1, m 2 ], we immediately get γ 1+γ 2 2 Γ o ( m 1+m 2 2 ). In particular we can apply Theorem 2.6 to deduce the existemce of a convex function ψ such that both γ 1 and γ 2 are concentrated on the graph of ψ. We now define γ = γ 1 γ 2 = (Id ψ) # (f γ1 f γ2 ), and we write γ 1 = γ + γ 1, γ 2 = γ + γ 2, with γ 1 = (Id ψ) # ( [fγ1 f γ2 ] + ), γ2 = (Id ψ) # ( [fγ2 f γ1 ] + ). Let λ (0, 1) be such that M (λ γ 2 ) = M ( γ 1 ). Since the function ψ is injective f-a.e., it is simple to see that γ + γ 1 + γ 2 Γ (f, g) (indeed, its marginals are dominated by f γ1 f γ2 and g γ1 g γ2 respectively). Thanks to the optimality of γ 1 and γ 2 we have C(γ 1 ) = C(γ + γ 1 ) C(γ + λ γ 2 ), C(γ 2 ) = C(γ + γ 2 ) C(γ + γ 1 + (1 λ) γ 2 ), which implies C( γ 1 ) = C(λ γ 2 ), and therefore C(γ + γ 2 ) = C(γ + γ 1 + (1 λ) γ 2 ). Since γ 1 + (1 λ) γ 2 = γ + γ 1 + (1 λ) γ 2 γ + γ 1 + γ 2 Γ (f, g), we see that γ 2 := γ 1 + (1 λ) γ 2 Γ o (m 2 ) is the desired minimizer. Theorem 2.10 (Strict convexity of C(m)) The function m C(m) is strictly convex on [m min, m max ]. Proof. Assume by contradiction that there exist m min m 1 < m 2 m max such that that C(m) is linear on [m 1, m 2 ]. Thanks to Lemma 2.9 we can find γ 1 Γ o (m 1 ) and γ 2 Γ o (m 2 ) such that γ 1 γ 2. Let us define f := f γ2 f γ1, g = g γ2 g γ1. We are now interested in the minimization problem C(m) := min γ Γ ( f, g), M (γ)=m C(γ). (2.10) Let m := m 2 m 1 = f R = n R g, and define γ := γ n 2 γ 1. Since γ 1 + λ γ Γ o (m 1 + λ m) for λ [0, 1], this easily implies that C(λ γ) = C(λ m), i.e. λ γ is optimal in the minimization problem (2.10) for all λ [0, 1]. We now want to prove that this is impossible. Fix any λ (0, 1), and as in the proof of Theorem 2.6 consider the Monge-Kantorovich problem: minimize C(γ) = x y 2 dγ(x, y) R n R n 11

among all γ which have λ f +(1 λ) g and λ g +(1 λ) f as first and second marginals respectively (observe that (1 λ) g = g g λ γ, (1 λ) f = f f λ γ ). By Proposition 2.4 γ λ := λ γ + (Id Id) # ( (1 λ)[ f + g] ), solves the Monge-Kantorovich problem. Thus, applying Theorem 2.3, we deduce that γ λ is concentrated on the graph of the gradient of a globally Lipschitz convex function ψ. In particular, since λ (0, 1), γ is concentrated on the graph of ψ and ψ(x) = x for a.e. x { f + g > 0}. This clearly implies f = g and γ = (Id Id) # f. Therefore 0 = C( γ) = C(m2 ) C(m 1 ), and so C(m) is constant on [m 1, m 2 ]. Since m 1 m min, as in the proof of Proposition 2.8 this is impossible. Remark 2.11 (Extension to more general cost functions) We observe that all the above arguments do not really use that the cost is quadratic: all we need is (1) c(x, y) 0 and c(x, y) = 0 only for x = y; (2) whenever both the source and the target measure are compactly supported and absolutely continuous with respect to the Lebesgue measure, the Monge-Kantorovich problem with cost c(x, y) has a unique solution which is concentrated on the graph of a function T ; (2 ) T is injective a.e. on the support of the source measure; (2 ) T is differentiable a.e. on the support of the source measure. Let us remark that condition (2 ) was used in the proof of Lemma 2.9, while (2 ) is needed for deducing (2.6) from (2.5). Some simple assumptions on the cost function which ensure the validity of (2)-(2 )-(2 ) are: (a) c C 2 (R n R n ); (b) the map y x c(x, y) is injective for all x R n ; (b ) the map x y c(x, y) is injective for all y R n ; (c) det( x,y c) 0 for all x, y R n. To understand why (a)-(b)-(c) imply (2)-(2 ), we recall that (a)-(b) ensure that the existence of an optimal transport map T. Moreover this map is implicitly defined by the identity x c(x, T (x)) = φ(x), where φ(x) is given by φ(x) = inf c(x, y) C y, y supp(ν) 12

with y C y locally bounded (see [17, Chapter 10]). By (a) we know that φ is locally semiconcave, which implies that its gradient (which exists a.e.) is differentiable a.e. Since by (a)-(b)-(c) the map (x, p) (x, [ x c(x, )] 1 p) is a smooth diffeomorphism, we obtain that also T is differentiable a.e. Finally, (a)-(b ) give the existence of an optimal transport map also for the Monge-Kantorovich problem with cost ĉ(x, y) := c(y, x). From this fact (2 ) follows easily. If c(x, y) = d(x, y) 2 with d(x, y) a Riemannian distance on a manifold, although c(x, y) does not satisfies the above assumptions, existence and uniqueness of the optimal transport map is still true [16, 12], and the Jacobian identity det( T ) = holds almost everywhere [9, 13] (although in the non-compact case, one has to define the gradient of T in an appropriate weak sense). So our existence and uniqueness result for the partial transport problem applies also to this case. 3 Properties of the active regions Let us denote by Ω and Λ the Borel sets {f > 0} and {g > 0} respectively. We have {f g > 0} = Ω Λ. For m [m min, m max ] we denote by γ m = (Id ψ m ) # f m = ( ψ m Id) # g m the unique minimizer of the minimization problem (2.1), where f m and g m denote the two marginals of γ m. We define the active source and the active target as F m := set of density points of {f m > 0}, G m := set of density points of {g m > 0}. We want to study the regularity properties of these sets. First of all, by Proposition 2.8, we have a.e. a.e. F m Ω Λ, G m Ω Λ, which will be a key fact to prove the regularity of the boundary of the free regions. We now prove an interior ball condition, which is the analogous in our formalism of [8, Corollary 2.4]. Proposition 3.1 (Structure of the active regions) There exists a set Γ m R n R n on which γ m is concentrated such that a.e. F m = (Ω Λ) { x Ω : x ȳ 2 < x ȳ 2}, ( x,ȳ) Γ m G m a.e. = (Ω Λ) ( x,ȳ) Γ m f g T { y Λ : x y 2 < x ȳ 2}. Proof. Let us recall that γ m = (Id ψ m ) # f m = ( ψm Id) # g m. We denote by D ψm and D ψ m the sets where ψ m and ψm are respectively differentiable. We recall that the gradient of a convex function is continuous on its domain of definition. We define Γ m := (Id ψ m )(F m D ψm ) ( ψ m Id)(G m D ψ m ) = { (x, y) : y = ψ m (x) with x F m D ψm and x = ψ m(y) with y G m D ψ m }. (3.1) 13

Since D ψm and D ψ m have full measure for f m and g m respectively, it is clear that γ m is concentrated on Γ m. We will prove the thesis only for F m (the case of G m being analogous). Since D ψm has full measure for f m, the result will follow from the inclusions F m D ψm a.e. (Ω Λ) ( x,ȳ) Γ m { x Ω : x ȳ 2 < x ȳ 2} a.e F m. (3.2) Let us first prove the left inclusion. Let x F m D ψm \ Ω Λ. Then obviously ψ m (x) x. Define v x := ψ m (x) x 0. Since x is a density point for the set {f m > 0}, there exists a sequence of points (x k ) such that (x k, ψ m (x k )) Γ m, x k x and x x k, v x 1 2 x x k v x (the idea is that we want x x k almost parallel to v x ). Thanks to the choice of x k it is simple to check that, since ψ m (x k ) ψ m (x), we have x ψ m (x k ) 2 < x k ψ m (x k ) 2 for k large enough. a.e. Since F m Ω, this implies the desired inclusion. We now have to prove the right inclusion in (3.2). The heuristic idea is simple: if a point x Ω is such that x ȳ 2 < x ȳ 2 but x F m, then we can replace ( x, ȳ) with (x, ȳ) to obtain a measure on R n R n with the same mass than γ m but which pays less, and this would contradict the optimality of γ m. Here is the rigorous argument: let us consider the set E := { x Ω : x ȳ 2 < x ȳ 2} \ Ω Λ. ( x,ȳ) Γ m First of all, since the set Γ m is separable, we can find a dense countable subset ( (x k, y k ) ) k N Γ m, so that E = k N { x Ω : x yk 2 < x k y k 2} \ Ω Λ. Assume by contradiction that E a.e. F m. This implies that there exist k N and a Borel set A with A > 0 such that Fix ε > 0. First of all, since A { x Ω : x y k 2 < x k y k 2} \ Ω Λ, A F m =. A ε := A { x Ω : x y k 2 < x k y k 2 ε } A as ε 0 (i.e. χ Aε χ A in L 1 as ε 0), by monotone convergence there exists ε > 0 small such that A ε > 0. Fix η > 0 small (the smallness will depend on ε, as we will see below). Since 14

y k = ψ m (x k ), ψ m is continuous at x k, and x k is a density point for F m, we can find a small ball B δ (x k ) such that x ψ m (x) 2 x k y k 2 ε 8, ψ m(x) ψ m (x k ) 2 η x B δ (x k ) F m D ψm. (3.3) Therefore, for η = η(ε) small enough, ( ) 2 x y 2 xk y k 2 ε + η xk y k 2 ε 4 Up to choosing δ smaller, we can assume A ε f B δ (x k ) f m > 0. density g δ := ψ m# (f m χ Bδ (x k )), we can replace the plan γ δ := (Id ψ m ) # (f m χ Bδ (x k )) x A ε, y ψ m (B δ (x k ) F m D ψm ). (3.4) Thus, if we consider the with any plan γ d such that f γd fχ Aε and g γd = g δ. In this way, since M ( γ δ ) = M (γ δ ), thanks to (3.3) and (3.4) we get ( C( γ δ ) x k y k 2 ε ) ( M ( γ δ ) < x k y k 2 ε ) M (γ δ ) C(γ δ ) 4 8 Since γ m γ δ + γ δ Γ(m), the above inequality would contradict the optimality of γ m. Thus E F m and also the second inclusion is proved. Remark 3.2 (The case of general cost) The above result can be generalized to the cost functions considered in Remark 2.11, obtaining that a.e. F m = (Ω Λ) {x Ω : c(x, ȳ) c( x, ȳ)}, ( x,ȳ) Γ m G m a.e. = (Ω Λ) ( x,ȳ) Γ m {y Λ : c(x, ȳ) c( x, ȳ)}, with Γ m defined as (Id T )(F m D T ) (T 1 Id)(G m D T 1), where T is the optimal transport map, and D T and D T 1 denote the set of continuity points for T and T 1 respectively. Remark 3.3 (Equality everywhere) Let us define U m := (Ω Λ) B x ȳ (ȳ), ( x,ȳ) Γ m V m := (Ω Λ) ( x,ȳ) Γ m B x ȳ ( x). By the proof of above proposition it is not difficult to deduce that, if Ω and Λ are open sets, then F m D ψm U m Ω F m 15

(and analogously for G m ). Since F m D ψm is of full Lebesgue measure inside F m, and U m is open, recalling the definition of F m one easily obtains F m = U m Ω, that is the equality is true everywhere and not only up to set of measure zero (and analogously G m = V m Λ). In particular, the inclusions F m Ω Λ and G m Ω Λ hold. Remark 3.4 (Monotone expansion of the active regions) Thanks to the uniqueness of minimizers for m [m min, m max ], we can apply [8, Theorem 3.4] to show the monotone expansion of the active regions. More precisely one obtains so that in particular f m1 f m2 and g m1 g m2 for m min m 1 m 2 m max, F m1 F m2 and G m1 G m2 for m min m 1 m 2 m max. 4 Regularity results By the duality theory developed in [8], as for m m min there exists a unique optimal transport map ψ m, one could prove that ψ m is the unique Brenier solution to a Monge-Ampère obstacle problem 2 (see [8, Section 4]). However we prefer here to not make the link with the Monge- Ampère obstacle problem (which would require to use the duality theory in [8, Section 2] in order to construct the obstacles), and instead we concentrate directly on the regularity of the maps ψ m and of the free boundaries of the active regions F m and G m. Since for m m min the map ψ m is trivial (just take ψ m (x) = x 2 /2 so that ψ m (x) = x everywhere), from now on we will consider only the case m > m min. 4.1 Partial semiconvexity of the free boundary We will assume Ω = {f > 0} and Λ = {g > 0} to be open and bounded sets. We define the free boundaries of F m and G m as F m Ω and G m Λ respectively. Thanks to Remark 3.3, they respectively coincide with U m Ω and V m Λ. Therefore, to prove regularity results on the free boundaries, we need to study the regularity properties of the sets U m and V m inside Ω and Λ, respectively. Definition 4.1 Let E and F be open sets. We say that E F is locally semiconvex if, for each x E F, there exists a ball B r (x) F such that E B r (x) can be written in some system of coordinates as the graph of a semiconvex function, and E B r (x) is contained in the epigraph of such a function. In [8, Section 5] the authors prove the semiconvexity of the free boundaries of the active regions assuming the existence of a hyperplane which separates the supports of f and g. Their 2 In [8] the authors adopt the terminology of weak solution, instead of Brenier solution. 16

proof is based on an analogue of Proposition 3.1. Indeed the existence of a separating hyperplane, together with the fact that in their case x ψ(x) δ > 0, allows to write U m Ω as the epigraph of a function u : R n 1 R such that, for each point x = (x, u(x )) U m Ω, there exists a ball of radius δ which touches the graph of u at x from below. Thanks to this property, they can show the semiconvexity of the free boundary. In particular they deduce that the Lebesgue measure of U m and U m is zero, and so f m and g m give no mass to U m and V m respectively (this property is crucial to apply Caffarelli s regularity theory in this context, see [8, Theorem 6.3] and the proof of Theorem 4.8 in the next paragraph). Remark 4.2 The assumption of the existence of a separating hyperplane plays an important role in the semiconvexity property. Indeed assume that g is supported on the ball B 1 (0) = { x 1}, and f has a connected support which contains the points x + := (2, 0,..., 0) and x := ( 2, 0,..., 0). It could be possible that the maps ψ send the points x + and x into two (distinct) points y + and y such that x + y + = x y = 2. In this case the active target region V m contains A := B 1 (0) ( B 2 (x + ) B 2 (x ) ), and so V m could be not a graph near the origin (and so it would not be locally semiconvex). However it is still possible to prove that V m is (n 1)-rectifiable, so that in particular it has zero Lebesgue measure (see Proposition 4.4 below). In our case everything becomes a priori much difficult for two reasons: first of all, since we do not want to assume the supports of f and g to be disjoint, of course we cannot have a separating hyperplane. Furthermore, as we do not have any lower bound on the quantity x ψ(x), the condition of having a ball touching from inside at each point of the boundary becomes a priori useless. We will solve these problems in two ways, depending on which of the two following results we want to prove: (a) The free boundaries have zero Lebesgue measure (resp. are (n 1)-rectifiable, i.e. countable union of Lipschitz hypersurfaces). (b) Each free boundary can be written as the union of a locally semiconvex hypersurface together with a closed subset of a Lipschitz hypersurface. The idea behind both results is the following: by what we already proved, the active regions contain Ω Λ (Proposition 3.1). Therefore, since x ψ(x) can be 0 only for x Ω Λ, at the points where radius of the ball touching from inside goes to 0 we will use the information that the set Ω Λ is touching U m from inside. This observation is more or less sufficient to prove (a). On the other hand, by Remark 4.2, it is clear that to prove (b) we need some geometric assumption on the supports f and g. To deal with this problem, we will assume that there exists an open convex set C such that Λ C and Ω \ Λ R n \ C. This hypothesis, which generalizes the assumption of a separating hyperplane, is satisfied for instance if Λ is convex, since it suffices to take C = Λ. We will need the following: 17

Lemma 4.3 Let A be a bounded Borel subset of R n, and define E := x A B r(x) (x), where x r(x) (0, + ) is a Borel function. Assume that there exist δ, R > 0 such that δ r(x) R for all x A. Then E is a (n 1)-rectifiable set, and in particular has zero Lebesgue measure. Proof. First of all, let Γ R n R denotes the graph of x r(x). Then E = B r (x). (x,r) Γ Denoting by Γ the closure of Γ, it is simple to check that E = B r (x). (4.1) (x,r) Γ Indeed, if y (x,r) Γ B r(x), then there exist (x, r) Γ such that y B r (x). Since the ball B r (x) is open, it is clear that if (x k, r(x k )) is a sequence of points in Γ that converges to (x, r), then y B r(xk )(x k ) for k large enough. This proves (4.1). Let now y E. This implies that for all ε > 0 the following holds: y x r ε (x, r) Γ and there exists (x ε, r ε ) Γ such that y x ε r ε + ε. Since Γ is compact (recall that A is bounded and r [δ, R] for all (x, r) Γ), it is simple to deduce that there exists (x 0, r 0 ) Γ such that y x 0 = r 0, that is y B r0 (x 0 ). As r 0 δ, we have proved that at each point y E we can find a ball of radius δ touching E at y from the interior. This condition, called Interior Ball Property, implies that E is (n 1)-rectifiable 3. On the other hand we remark that, to prove just that E has zero Lebesgue measure, it suffices to show that E has no density points. Indeed we recall that, given a Borel set A, it is a well-known fact that a.e. point x A is a density point (see for instance [11, Paragraph 1.7.1, Corollary 3]). To see that E has no density points it suffices to observe that if y E then y B y x (x) for some ball B y x (x) E, and so L n( E B r (y) ) lim r 0 L n( B r (y) ) lim L n( B r (y) \ B y x (x) ) r 0 L n( B r (y) ) = 1 2. 3 Using [1, Lemma 2.1] one can prove for instance a stronger result: let {v 1,..., v kn } be a subset of the unit sphere of R n such that, for each vector p R n with unit norm, there exists i {1,..., k n} with v i, p 1/2. Then, if to any x E we associate a unit vector p x R n such that B δ (x δp x) E, we can decompose E as the union of A i, with A i := {x E : v i, p x 1/2}, and thanks to [1, Lemma 2.1] one can show that each A i is locally semiconvex. 18

We can now prove (a). Proposition 4.4 (No mass on the free boundary) Assume Ω Λ open, with L n( (Ω Λ) ) = 0. Let (Ω Λ) ε := {x R n : dist(x, Ω Λ) < ε}, and assume that for any ε > 0 dist ( Ω \ (Ω Λ) ε, Λ \ Ω ) =: δ(ε) > 0. (4.2) Then L n( U m ) = 0, and in particular both fm and g m give no mass to U m. Moreover, if (Ω Λ) is Lipschitz, then U m is (n 1)-rectifiable. Proof. where Decompose the boundary of U m as U e m := U m \ (Ω Λ), U m = U e m U b m, U b m := U m (Ω Λ). Obviously U b m has measure zero. Since by definition U m Ω Λ, and Ω Λ is open, we easily get U m (Ω Λ) =, that is U m cannot enter inside Ω Λ. Therefore we can write U e m as the increasing union of the sets B e ε := U e m \ (Ω Λ) ε. Consider the open set E ε := ( x,ȳ) Γ m, x (Ω Λ) ε B x ȳ (ȳ), where Γ m is the set defined in (3.1). Thanks to (4.2), we can apply Lemma 4.3 to deduce that L n( E ε ) = 0. Observing that U e m \ (Ω Λ) ε E ε, we obtain that L n( Bε) e = 0 for all ε > 0, and so L n ( Um) e = 0. Suppose now that (Ω Λ) is Lipschitz. Then Um b is clearly (n 1)-rectifiable. Moreover, by Lemma 4.3, also E ε is (n 1)-rectifiable for all ε > 0. From this we conclude easily. Let us now prove (b). Proposition 4.5 (Local semiconvexity away from the intersection of the supports) Assume Ω open, and suppose there exists an open convex set C such that Λ C and Ω \ Λ R n \ C. We decompose the boundary of U m inside Ω as U m Ω = ( U e m Ω ) ( U b m Ω ), where U e m := U m \ C, Then U e m is locally semiconvex inside Ω. U b m := U m C. 19

Proof. By the definition of C, one can easily check that Ω Λ = Ω C. In particular Ω Λ is open. Therefore, since by Remark 3.3 we have U m Ω = F m Ω Λ = Ω C, we easily obtain ( U m Ω) C =, that is the free boundary cannot enter inside C. Thus we can write Um e as the increasing union of the sets B e l := U e m {x R n : dist(x, C) 1/l}. Let us write C as the intersection of a countable set of halfspaces: C = k H k, H k = {x R n : x, v k < 0}. We see that each set B e l can be written as B e l = k B e l,k, Be l,k := U e m {x R n : dist(x, H k ) 1/l}. We now remark that, since for any k N the set Ω {dist(x, H k ) 1/l} is separated from Λ by a hyperplane, by the same proof in [8, Section 5] we deduce that Bl,k e Ω is contained in a semiconvex graph. This easily implies that the set Um e Ω is locally semiconvex. Remark 4.6 If we exchange the role of Ω and Λ in the above propositions, the above regularity properties hold for V m in place of U m. 4.2 Regularity of the transport map and of the free boundary By the results of Section 2, we know that the optimal plan γ m is induced by a convex function ψ m via γ m = (Id ψ m ) # f m. Since ψ m# f m = g m, we will say that ψ m is a Brenier solution to the Monge-Ampère equation det( 2 ψ m )(x) = f m (x) g m ( ψ m (x)) on F m, ψ m (F m ) G m. (4.3) We recall that the function ψ m was constructed applying Theorem 2.3 to the Monge-Kantorovich problem: minimize C(γ) = x y 2 dγ(x, y) R n R n among all γ which have f + (g g m ) and g + (f f m ) as first and second marginals, respectively (see Proposition 2.4). In particular ψ m is unique f-a.e. Therefore, if f is strictly positive on an open connected set Ω R n, then ψ m is unique on Ω up to additive constants. We now want to deduce regularity properties on ψ m and F m. Exchanging the role of f and g, all results on ψ m and F m will be true also for ψ m and G m. We recall that, thanks to Remark 3.3, F m = U m Ω, G m = V m Λ. Assumption 1: we assume that f and g are supported on two bounded open sets Ω and Λ respectively. Moreover we assume that f and g are bounded away from zero and infinity on Ω and Λ respectively. 20

In order to prove regularity results on transport maps arising from the Monge-Kantorovich problem it is well known that one needs to assume at least convexity of the target domain (see [7]), but even assuming Λ to be convex one cannot expect G m Λ to be convex. However we can adapt the strategy used in [8, Section 6] in order to prove local C 1,α regularity of ψ m. On the other hand in our case, even assuming f and g to be C, we cannot expect ψ m to be C in the interior of Ω, while this is the case if Ω and Λ are disjoint (see Remark 4.9 below). The idea to prove interior regularity for ψ m is to apply Caffarelli s regularity theory. To this aim we have to prove, as in [8, Section 6], that ψ m solves a Monge-Ampère equation with convex target domain. Let us recall the interior regularity result of Caffarelli [7]: Theorem 4.7 Let f and g be nonnegative densities supported on two bounded open sets Ω and Λ respectively. Assume Λ convex, and that f and g are bounded away from zero and infinity on Ω and Λ respectively. If ψ : R n (, + ] is a convex function such that ψ # f = g, then there exists α > 0 such that ψ C 1,α loc (Ω). Moreover ψ is strictly convex on Ω. Let us therefore assume Λ convex. Since ψ m solves the Monge-Kantorovich problem from f m +(f f m )+(g g m ) to g m +(f f m )+(g g m ), and ψ m (x) = x for x {f > f m } {g > g m }, we have ψ m# (f m + (g g m )) = g. (4.4) This implies (see Theorem 2.3) that ψ m solves a Monge-Kantorovich problem where the target measure g is bounded from away from zero and infinity on the bounded open convex set Λ. Since, under Assumption 1, f m + (g g m ) is bounded from above, in order to apply Caffarelli s interior regularity theory we need to prove the existence of an open bounded set on which this density is concentrated, and to show that it is bounded away from zero on this set. We observe that ψ m (x) = x and f m (x) = g m (x) = (f g)(x) for a.e. x {f f m > 0} {g g m > 0} (see Theorem 2.6). Thus This easily implies that f m = g m = f g if 0 < f m < f or 0 < g m < g. f m + (g g m ) = f or f m + (g g m ) = g if f m = f or f m = f g, f m + (g g m ) = g if f m = g m = 0. Therefore f m + (g g m ) = { f or g in Um Ω, g in Λ \ V m. (4.5) Thus f m + (g g m ) is bounded away from zero on the domain (U m Ω) (Λ \ V m ), and this domain has full mass since V m has zero Lebesgue measure (see Proposition 4.4 and Remark 4.6). Combining this with Theorem 4.7, we obtain: 21

Theorem 4.8 (Interior C 1,α regularity) Suppose that Assumption 1 holds, and that Λ is convex. If ψ m is a Brenier solution to (4.3), then ψ m C 1,α loc (U m Ω) and is strictly convex on U m Ω. Remark 4.9 (Smooth densities need not to have a smooth solution) In [8], the authors show that, if f and g are C with disjoint supports, and if the support of g is convex, then ψ m C. This property strongly relies on the fact that the supports are disjoint. Indeed, even if we assume f (resp. g) to be C we cannot expect f m (resp. g m ) to be continuous on its support. In particular we cannot expect for higher regularity results. Consider for instance the following 1-dimensional example: let h 0 be a even function on R of class C with support contained in [ 1, 1]. Define f(x) = χ [ 4,4] (x) + h(x + 2), g(x) = χ [ 4,4] (x) + h(x 2) (where χ A denotes the indicator function of a set A). If m = m min = 8, then all the common mass stay at rest. If now m = 8 + ε with 0 < ε < 1 1 h, then the following happens: let δ = δ(ε) (0, 2) be such that 1+δ 1 h = 1 1 δ h = ε, and let T δ denote the optimal transport map (for the classical Monge-Kantorovich problem) which sends fχ [ 1 δ,1+δ] into gχ [ 1 δ,1+δ]. Then it is not difficult to see that x if x [ 4, 1 δ], ψ m(x) = T δ (x) if x [ 1 δ, 1 + δ], x if x [1 + δ, 4], f m = χ [ 4,4] (x) + h(x + 2)χ [ 1 δ, 1] (x), g m = χ [ 4,4] (x) + h(x 2)χ [1,1+δ] (x). Therefore, although f and g are C on their supports (which are smooth and convex), f m and g m are not continuous. Since ψ m = f m g m ψ m on [ 4, 4], and ψ m is continuous on [ 4, 4] (so that ψ m( 1 δ) = 1 δ), we get that lim x ( 1 δ) ψ m(x) = + and so ψ m is not C 2. lim x ( 1 δ) + lim x ( 1 δ) ψ m(x) = 1, f m (x) f( 1 δ) g m (ψ m(x)) = = 1 + h(1 δ) > 1, g( 1 δ) We now want to prove global C 1 regularity of ψ m up to the free boundary. Moreover we will prove, following the lines of the proof of [8, Theorem 6.3], that the free boundary is locally a C 1 hypersurface away from Ω Λ, and the transport map displaces along it in the perpendicular direction. To this aim, we will need to assume strict convexity of the domains: 22

Assumption 2: Ω and Λ are strictly convex. We remark that Assumption 2 implies in particular that Ω is connected. Thus ψ m is uniquely defined on Ω up to additive constants, and it makes sense to speak about the Brenier solution to (4.3). Theorem 4.10 (Global C 1 regularity) Suppose that Assumptions 1 and 2 hold. Let ψ m be the Brenier solution to (4.3). Then there exists ψ m C 1 (R n ) C 1,α loc (U m Ω) such that ψ m = ψ m on U m Ω, ψ m (x) = x on Λ \ V m, and ψ m (R n ) = Λ. Moreover ψ m : U m Ω V m Λ is a homeomorphism. Proof. Since the assumptions on f and g are symmetric, by Proposition 2.4 we deduce that ψm is the unique optimal transport map for the Monge-Kantorovich problem from g m + (f f m ) to f. By Theorem 4.8 applied to ψm we know that ψm C 1,α loc (V m Λ) and is strictly convex on V m Λ. Since Ω and Λ are both convex, we can apply Proposition 4.5 (with C = Ω) to deduce that ( V m \ Ω) Λ is locally semiconvex. Moreover, as ψm(y) = y for a.e. y Λ \ V m {g > g m }, we obtain ψm(y) = y 2 2 + C on each connected component of the open set Λ \ V m. Thus it is not difficult to see that ψm is strictly convex on the full domain Λ, and thanks to the strict convexity of Λ ψm is strictly convex also on Λ. Let us consider the strictly convex function { ψ φ m = m on Λ, + on R n \ Λ, and define ψ m := φ m. Exactly as in [8, Theorem 6.3], the strict convexity of φ m implies that ψ m C 1 (R n ). Furthermore, since y ψ m (x) if and only if x φ m (y), we deduce that ψ m (R n ) Λ, which implies that ψ m is globally Lipschitz. Finally, as φ m ψm with equality on Λ, we deduce that ψ m (x) ψ m (x) x R n with equality if ψ m (x) Λ. Since by (4.4) and (4.5) ψ m (x) Λ for a.e. x (U m Ω) (Λ \ V m ), we deduce that ψ m gives the desired extension of ψ m. This implies in particular that ψ m : U m Ω Λ extends to a continuous map from U m Ω to V m Λ. Indeed the extension cannot takes values outside V m since f m does not vanish on U m Ω and ψ m# f m = g m is supported on V m Λ. By the symmetry of the assumptions on f and g, the above argument implies that also : V m Λ Ω extends to a continuous map from V m Λ to U m Ω. Since by (2.2) ψm( ψ m (x)) = x a.e. inside U m Ω, and ψ m ( ψm(y)) = y a.e. inside V m Λ, by continuity both equalities hold everywhere inside U m Ω and V m Λ respectively. Thus ψ m : U m Ω V m Λ is a homeomorphism with inverse given by ψm. Since both maps extend continuously up to the boundary, we deduce that ψ m : U m Ω V m Λ is a homeomorphism. We can now prove the C 1 regularity of the free boundary away from Ω Λ. Theorem 4.11 (Free boundary is C 1 away from Ω Λ) Suppose that Assumptions 1 and 2 hold. Let U m Ω = ( U e m Ω ) ( U b m Ω ) be the decomposition provided by Proposition 4.5 23