Stochastic Optimization Lecture notes for part I. J. Frédéric Bonnans 1

Size: px
Start display at page:

Download "Stochastic Optimization Lecture notes for part I. J. Frédéric Bonnans 1"

Transcription

1 Stochastic Optimization Lecture notes for part I Optimization Master, University Paris-Saclay Version of January 25, 2018 J. Frédéric Bonnans 1 1 Centre de Mathématiques Appliquées, Inria, Ecole Polytechnique, Palaiseau, France. Updated versions: bonnans/notes.html

2

3 These lecture notes originated in Master courses given for several years in the University of Paris VI and in the University of Paris-Saclay. They give an introduction to convex analysis, in a Banach space setting, and its application to stochastic programming problems, i.e., to optimization problem where the decision must be taken in the presence of uncertainties, with a focus on dynamical problems. This is an active subject of research that covers many applications, especially in economics. There are already valuable classical textbooks on the subject, as Birge and Louveaux [7], Kall and Wallace [28]. The book [64] by Wallace and Ziemba is dedicated to applications. Some more advanced material is presented in Ruszczynski and Shapiro [53], Shapiro, Dentcheva and Ruszczynski [58], Föllmer and A. Schied [22]. Let us mention the historical review paper by Wets [65]. The basic tool for studying such problems is the combination of convex analysis with measure theory. Classical sources in convex analysis are Rockafellar [45], Ekeland and Temam [19]. An introduction to integration and probability theory is given in Malliavin [35]. The notes have the following structure. Chapter 1 present the basic tools in convex analysis: Fenchel conjugates, subdifferential, Lagrangian and perturbation duality, without neglecting linear programming duality and Hoffman s bounds. Chapter 2 recalls integration theory for a general measured space, and applies it to the computation of the conjugate of an integral functional. It ends with the study of optimization problems with integral cost function and constraints. Chapter 3 presents the concept of convergence in law, and the associated asymptotics (delta theorems). It applies these results to the computation of the asymptotic law of a sampled stochastic problem (including the case of constraints in expectations). Chapter 4 presents the notions of conditional expectation and variance, and applies them to the analysis of the optimality conditions of a dynamic, stochastic optimization problem with information constraints. The last chapter discusse some algorithmic questions. 3

4

5 Contents Chapter 1. A convex optimization toolbox 9 1. Convex functions Optimization problems Separation of convex sets Weak duality and saddle points Linear programming and Hoffman bounds Conjugacy Duality theory Perturbation duality Subdifferential calculus Minimax theorems Calmness Specific structures, applications Maxima of bounded functions Linear conical optimization Polyhedra Infimal convolution Recession functions and the perspective function Duality for nonconvex problems Convex relaxation Applications of the Shapley-Folkman theorem First order optimality conditions Multiobjective optimization Notes 70 Chapter 2. An integration toolbox Measure theory Measurable space Measures Kolmogorov s extension of measures Limits of measurable functions Integration L p spaces Bochner integrals Integral functionals Minimization of Carathéodory integrals 93 5

6 6 CONTENTS 2.2. Measurable multimappings Convex integrands Conjugates of integral functionals Deterministic decision in R m Constrained random decision Linear programming with simple recourse Applications of the Shapley-Folkman theorem Integrals of multimappings Constraints on integral terms Examples and exercices Notes 111 Chapter 3. Sampling and optimizing Examples and motivation Maximum likelihood Convergence in law and related asymptotics Probabilities over metric spaces Convergence in law Central limit theorems Delta theorem First order delta theorem Second order delta theorem Solving equations Error estimates Empirical distribution Minimizing over a sample Uniform convergence of values Asymptotic law Expectation constraints Large deviations Principle of large deviations Error estimates in stochastic programming Notes 133 Chapter 4. Dynamic stochastic optimization Conditional expectation Functional dependency Construction of the conditional expectation Conditional expectation of non integrable functions Computation in some simple cases Convergence theorems Conditional variance Compatibility with a subspace Compatibility with measurability constraints No recourse 146

7 CONTENTS 7 2. Dynamic stochastic programming Dynamic uncertainty Abstract optimality conditions The growing information framework The standard full information framework Independent noises Elementary examples Application to the turbining problem Notes 153 Chapter 5. Algorithms Stochastic dual dynamic programming (SDDP) Static case: Kelley s algorithm [29] Deterministic dual dynamic programming Stochastic case Introduction to linear decision rules About the Frobenius norm Setting Linear programming reformulation Linear conic reformulation Dual bounds in a conic setting 164 Bibliography 167 Index 171

8

9 1 A convex optimization toolbox Contents 1. Convex functions Optimization problems Separation of convex sets Weak duality and saddle points Linear programming and Hoffman bounds Conjugacy Duality theory Perturbation duality Subdifferential calculus Minimax theorems Calmness Specific structures, applications Maxima of bounded functions Linear conical optimization Polyhedra Infimal convolution Recession functions and the perspective function Duality for nonconvex problems Convex relaxation Applications of the Shapley-Folkman theorem First order optimality conditions Multiobjective optimization Notes Convex functions 9

10 10 1. A CONVEX OPTIMIZATION TOOLBOX 1.1. Optimization problems The language of minimization problems. Denote the set of extended real numbers by R := R { } {+ }. A minimization problem is of the form (P f,k ) Min x f(x); x K, where K is a subset of some set X, and f : X R (we say that f is extended real valued). The domain of f is (1.1) dom(f) := {x X; f(x) < + }. We say that f is proper is its domain is not empty, and if f(x) >, for all x X. The feasible set and value of (P f,k ) are resp. (1.2) F (P f,k ) := dom(f) K; val(p f,k ) := inf{f(x); x F (P )}. Since the infimum over the empty set is +, we have that val(p f,k ) < + iff F (P f,k ). The solution set of (P f,k ) is defined as (1.3) S(P f,k ) := {x F (P f,k ); f(x) = val(p f,k )}. Note that S(P f,k ) = when F (P f,k ) =. A metric over X is a function say d : X X R + such that d(x, y) = 0 iff x = y, that is symmetric: d(x, y) = d(y, x) and that satisfies the triangle inequality (1.4) d(x, z) d(x, y) + d(y, z). We say that X is a metric space if it is endowed with a metric. In that case, we say that f is lower semi continuous, or l.s.c., if for all x X, f(x) lim inf k f(x k ) whenever x k x. A minimizing sequence for problem (P f,k ) is a sequence x k in F (P f,k ) such that f(x k ) val(p f,k ). Such a sequence exists iff F (P f,k ) is non empty. Any (infinite) subsequence of a minimizing sequence is itself a minimizing sequence. If X is a metric space, K is closed and f is l.s.c., then any limit point of a minimizing sequence is solution of (P f,k ). Example 1.1. Consider the problem of minimizing the exponential function over R. The value if finite, but the solution set is empty. Note that minimizing subsequences have no limit point in R Operations on extended real valued functions. In the context of minimization problems, that f(x) = + is just a way to express that x is not feasible. Therefore the following algebraic rules for extended real valued functions are to be used: if f and g are extended real valued functions over X, then h := f + g is the extended real valued function over X defined by { (1.5) h(x) = + if max(f(x), g(x)) = +, f(x) + g(x) otherwise. Note that there is no ambiguity in this definition (taking the usual addition rules in the presence of ± ). The domain of the sum is the intersection of the domains.

11 1. CONVEX FUNCTIONS 11 Example 1.2. With a subset K of X we associate the indicatrix function I K : X R defined by { 0 if x K, (1.6) I K (x) := + otherwise. Let f K (x) := f(x) + I K (x). Then (P f,k ) has the same feasible set, value and set of solutions as (P fk,x). If f is an extended real valued function and λ > 0, we may define λf by the natural rule (1.7) (λf)(x) = λf(x), for all x X. Observe that (P λf,k ) has the same feasible set and set of solutions as (P f,x ), and the values are related by (1.8) val(p λf,k ) = λ val(p f,k ). For λ = 0, we must think of 0f as the limit of λf when λ 0, and therefore set { f(x) if f(x) = ±, (1.9) (0f)(x) = 0 otherwise. Then (P 0f,K ) has the same feasible set as (P f,x ), and its set of solutions is F (P f,x ) if f is proper. Example 1.3. Consider the entropy function f(x) = x log x (with the convention that 0 log 0 = 0) if x 0, and + otherwise. Then 0f is the indicatrix of R +. More generally, if f is proper, then 0f is the indicatrix of its domain Maximization problems. For a maximization problem (D g,k ) Max g(x); x x K we have similar conventions, adapted to the maximization framework. In particular the domain of g is dom(g) = {x X; g(x) > }. The domain of the sum is still the intersection of the domains. Note that (D g,k ) is essentially the same problem as (D g,k ) Min x g(x); x K. Indeed these two problems have the same feasible set and set of solutions, and they have opposite values Convex sets and functions. Let X be a vector space. We say that K X is convex if (1.10) For any x and y in K, and α (0, 1), we have that αx + (1 α)y K. We say that f : X R is convex if { For any x and y in dom(f), and α (0, 1), we have that (1.11) f(αx + (1 α)y) αf(x) + (1 α)f(y). We see that a convex function has a convex domain.

12 12 1. A CONVEX OPTIMIZATION TOOLBOX The epigraph of f : X R is the set (1.12) epi(f) := {(x, α) X R; α f(x)}. Its projection over X is dom(f). One easily checks the following: Lemma 1.4. Let f : X R. Then (i) Its epigraph is convex iff f is convex, (ii) if X is a metric space, its epigraph is closed iff f is l.s.c. Example 1.5. The epigraph of the indicatrix of K X is K R Separation of convex sets. We recall without proof the Hahn-Banach theorem, valid in a vector space setting, and deduce from it some results of separation of convex sets in normed vector spaces The Hahn-Banach theorem. Let X be a vector space. We say that p : X R is positively homogeneous and subadditive if it satisfies { (i) p(αx) = αp(x), for all x X and α > 0, (1.13) (ii) p(x + y) p(x) + p(y), for all x and y in X. Remark 1.6. (a) Taking x = 0 in (1.13)(i), we obtain that p(0) = 0, and so we could as well take α = 0 in (1.13)(i). (b) If β (0, 1), combining the above relations, we obtain that (1.14) p(βx + (1 β)y) βp(x) + (1 β)p(y), i.e., p is convex. Conversely, it is easily checked that a positively homogeneous (finite-valued) convex function is subadditive. The analytical form of the Hahn-Banach theorem, a nontrivial consequence of Zorn s lemma, is as follows (see [10] for a proof): Theorem 1.7. Let p satisfy (1.13), X 1 be a vector subspace of X, and λ be a linear form defined on X 1, that is dominated by p in the sense that (1.15) λ(x) p(x), for all x X 1. Then there exists a linear form µ on X, dominated by p, whose restriction on X 1 coincides with λ. We say that a real vector space X is a normed space when endowed with a mapping X R, x x satisfying the three axioms x 0, with equality iff x = 0, (1.16) αx = α x, for all α R, x X, x + x x + x, (triangle inequality). Then (x, y) x y is a metric over X. We denote the norm of Euclidean spaces, i.e., finite dimensional spaces endowed with the norm ( i x2 i )1/2, by x. A sequence x k in a normed vector space X is said to be a Cauchy sequence if x p x q 0 when p, q. We say that X is a Banach space if every Cauchy sequence has a (necessarily unique) limit.

13 1. CONVEX FUNCTIONS 13 The topological dual X of the normed vector space X is the set of continuous linear forms (applications X R) on X. In the sequel, by dual space we will mean the topological dual. We denote the duality product between x X and x X by x, x X or simply x, x. Note that a linear form say l over X is continuous iff it is continuous at 0, which holds iff sup{l(x); x 1} <. So we may endow X with the norm (1.17) x := sup{ x, x ; x 1}. It is easily checked that X is a Banach space. The dual of R n (space of vertical vectors) is denoted by R n (space of horizontal vectors). In the sequel we may denote the dual norm by x. If X and Y are Banach spaces, we denote by L(X, Y ) the Banach space of linear continuous mappings X Y, endowed with the norm A := sup{ Ax ; x 1}. We denote by B X (resp. B X ) the open (resp. closed) unit ball of X. If x 1 is a continuous linear form on a linear subspace X 1 of X, is norm is defined accordingly: (1.18) x 1 1, = sup{ x, x ; x X 1, x 1}. Here are some other corollaries of the Hahn-Banach theorem. Corollary 1.8. Let x 1 be a continuous linear form on a linear subspace X 1 of the normed space X. Then there exists x X, whose restriction on X 1 coincides with x 1, and such that (1.19) x = x 1 1,. Proof. Apply theorem 1.7 with p(x) := x 1 1, x. Since x, ±x p(x), we have that x 1 implies x, ±x x 1 1,. The result follows. Corollary 1.9. Let x 0 belong to the normed vector space X. Then there exists x X such that x = 1 and x, x 0 = x 0. Proof. Apply corollary 1.8 with X 1 = Rx 0 and x 1 (tx 0) = t x 0, for t R. The orthogonal of E X is the closed subspace of X defined by (1.20) E := {x X ; x, x = 0, for all x E}. Lemma Let E be a subspace of X. Then E = {0} iff E is dense. Proof. a) If E is dense, given x X, there exists a sequence x k in E, x k x and hence, for all x E, x, x = lim k x, x k = 0, proving that x = 0. b) If E is not dense, let x 0 Ē (closure of E). We may assume that x 0 = 1 and that B(x 0, ε) E = for some ε > 0. Let E 0 := E (Rx 0 ) denote the space spanned by E 0 and x 0. Consider the linear form λ on E 0 defined by (1.21) λ(e + αx 0 ) = α, for all e E and α R. Since any x E 0 has a unique decomposition as x = e + αx 0 with e E and α R, the linear form is well-defined. Let such an x satisfy α 0. Since e := e/α does not belong to B(x 0, ε), we have that x = α x 0 e ε α, and hence, λ(x) = α x /ε. If α = 0 we still have λ(x) x /ε. By corollary 1.8, λ has an extension to a continuous linear form on X, which is a nonzero element of E.

14 14 1. A CONVEX OPTIMIZATION TOOLBOX Bidual space, reflexivity. Given x X, the mapping l x : X R, x x, x is by (1.17) linear continuous. Since x, x x x, its norm l x (in the bidual space X ) is not greater than x, and as a consequence of corollary 1.9, is equal to x : the mapping x l x is isometric. This allows to identify X with a closed subspace of X. We say that X is reflexive if X = X. The Hilbert spaces are reflexive, see [10] Separation theorems. We assume here that X is a normed vector space. A (topological) hyperplane of X is a set of the form: (1.22) H x,α := {x X; x, x = α, for some (x, α) X R, x 0}. We call (closed) half-space of X a set of the form (1.23) {x X; x, x α}, where x 0. Definition Let A and B be two subsets of X. We say that the hyperplane H x,α separates A and B if (1.24) x, a α x, b, for all (a, b) A B. We speak of a strict separation if (1.25) x, a < x, b, for all (a, b) A B, and of a strong separation if, for some ε > 0, (1.26) x, a + ε α x, b ε, for all (a, b) A B. We say that x X (nonzero) separates A and B if (1.24) holds for some α, strictly separates A and B if (1.25) holds, and strongly separates A and B if (1.26) holds for some ε > 0 and α. If A is the singleton {a}, then we say that x separates a and B, etc. Given two subsets A and B of a vector space X, we define their Minkowski sum and difference as { A + B = {a + b; a A, b B}, (1.27) A B = {a b; a A, b B}. The first geometric form of the Hahn-Banach theorem is as follows: Theorem Let A and B be two nonempty subsets of the normed vector space X, with empty intersection. If A B is convex and has a nonempty interior, then there exists an hyperplane H x,α separating A and B, such that (1.28) x, a < x, b, whenever (a, b) A B and a b int(a B). Note that A B has a nonempty interior whenever either A or B has a nonempty interior. The proof needs the following concept. Definition Let C be a convex subset of X whose interior contains 0. The gauge function of C is (1.29) g C (x) := inf{β > 0; β 1 x C}. Example If C is the closed unit ball of X, then g C (x) = x for all x X.

15 1. CONVEX FUNCTIONS 15 A gauge function is obviously positively homogeneous and finite. If B(0, ε) C for some ε > 0, then: (1.30) g C (x) x /ε, for all x X, so it is bounded over bounded sets. In addition, for any β > g C (x) and γ > 0, since x βc and B(0, γε) γc, we get x+b(0, γε) (β+γ)c, so that g C (y) g C (x)+γ, for all y B(x, γε). We have proved that (1.31) If B(0, ε) C, then g C is Lipschitz with constant 1/ε. It easily follows that (1.32) {x X; g C (x) < 1} = int(c) C = {x X; g C (x) 1}, Lemma A gauge is subadditive and convex. Proof. Let x and y belong to X. For all β x > g C (x) and β y > g C (y), we have that (β x ) 1 x C and (β y ) 1 y C, so that (1.33) x + y β x + β y = β x β x + β y (β x ) 1 x + β y β x + β y (β y ) 1 y C. Therefore, g C (x+y) β x +β y. Since this holds for any β x > g C (x) and β y > g C (y), we obtain that g C is subadditive. Since g C is positively homogeneous, it easily follows that g c is convex. Proof of theorem Let x 0 int(b A); since A B =, x 0 0. Set (1.34) C := {a b + x 0, b B, a A}. We easily check that 0 int(c). Obviously, x separates A and B iff it separates C and {x 0 }. Let λ be the linear form defined on X 1 := Rx 0 by λ(tx 0 ) = t, for t R. Since A B =, x 0 C, and hence, g C (x 0 ) 1. It easily follows that λ is dominated by g C on X 1. By theorem 1.7, there exists a (possibly not continuous) linear form x on X, dominated by g C, whose restriction on X 1 coincides with λ. Since 0 int(c) we have that (1.30) holds for some ε > 0, and hence, being dominated by g C, x is continuous. It follows that x, x 1, for all x C, or equivalently (1.35) x, a x, b + x, x 0 1, for all (a, b) A B, whereas x, x 0 = 1. Therefore x separates A and B. In addition, if a b int(a B), say B(a b, ε) A B for some ε > 0, then x, a b+e 0 whenever e ε; maximizing over e B(0, ε) we obtain that x, a b ε x. Relation (1.28) follows. Corollary Let E be a closed convex subset of the normed space X. Then there exists an hyperplane that strongly separates any x 0 E and E. Proof. For ε > 0 small enough, the open convex set A := B(x 0, ε) has empty intersection with E. By theorem 1.12, there exists x 0 separating A and E, that is (1.36) x, x 0 + ε x = sup{ x, a ; a A} inf{ x, b ; b E}. The conclusion follows.

16 16 1. A CONVEX OPTIMIZATION TOOLBOX Remark Corollary 1.16 can be reformulated as follows: any closed convex subset of a normed space is the intersection of half spaces in which it is contained. The following example shows that, even in a Hilbert space, one cannot in general separate two convex sets with empty intersection. Example Let X = l 2 be the space of real sequences whose sum of squares of coefficients is summable. Let C be the subset of X of sequences with finitely many nonzero coefficients, the last one being positive. Then C is a convex cone that does not contain 0. Let x separate 0 and C. We can identify the Hilbert space X with its dual, and therefore x with an element of X. Since each element e i of the natural basis belongs to the cone C, we must have x i 0 for all i, and x j > 0 for some j. For any ε > 0 small enough, x := e j + εe j+1 belongs to C, but x, x = x j + εx j+1 < 0. This shows that one cannot separate the convex sets So, 0 and C cannot be separated Relative interior. Again let X be a normed vector space, and E be a convex subset of X. We denote by affhull(e) the intersection of affine spaces containing E, and by affhull(e) its closure; the latter is the smallest closed affine space containing E. The relative interior of E, denoted by rint(e), is the interior of E viewed as a subset of affhull(e). Proposition Let A and B be two nonempty subsets of X, with empty intersection. If A B is convex and has a nonempty relative interior, then there exists an hyperplane H x,α separating A and B, and such that (1.37) x, a < x, b, whenever (a, b) A B and a b rint(a B). Proof. Set E := B A and Y := affhull(e). By theorem 1.12, there exists y in Y separating 0 and E, with strict inequality for rint(e). By theorem 1.7, there exists x X whose restriction to Y is y, and the conclusion holds with x. Remark By the previous proposition when B = {b} is a singleton, noting that rint(a b) = rint(a) b, we obtain that when A is convex, if b rint(a) then there exists x X such that (1.38) x, a < x, b, whenever a rint(a). Since any convex subset of a finite dimensional subspace, not reduced to a singleton, has a nonempty relative interior 1, we deduce the following: Corollary Let A and B be two convex and nonempty subsets of an Euclidean space, with empty intersection. Then there exists an hyperplane H x,α separating A and B, such that (1.37) holds. 1 Except maybe when the set is a singleton and then the dimension is zero, where this is a matter of definition. However the case when A B reduces to a singleton means that both A and B are singletons and then it is easy to separate them.

17 1. CONVEX FUNCTIONS Weak duality and saddle points. Let X and Y be two sets and let L : X Y R. Then we have the weak duality inequality (1.39) sup y Y inf x X Indeed, let (x 0, y 0 ) X Y. Then L(x, y) inf sup x X y Y L(x, y). (1.40) inf L(x, y 0) L(x 0, y 0 ) sup L(x 0, y). x X Removing the middle term, maximizing the left side w.r.t. y 0 and minimizing the right side w.r.t. x 0, we obtain (1.39). We next define the primal and dual values, resp., for x X and y Y, by (1.41) p(x) := sup L(x, y); y Y and the primal and dual problem by (P ) Min x X p(x) (D) Max y Y d(y). y Y d(y) := inf L(x, y), x X The weak duality inequality says that val(d) val(p ). We say that ( x, ȳ) X Y is a saddle point of L over X Y if (1.42) L( x, y) L( x, ȳ) L(x, ȳ), for all (x, y) X Y. An equivalent relation is (1.43) sup L( x, y) = L( x, ȳ) = inf L(x, ȳ). y Y x X Minorizing the left hand term by changing x into the infimum w.r.t. x X and majorizing symetrically the right hand term, we obtain (1.44) inf sup x X y Y L(x, y) L( x, ȳ) sup y Y inf x X L(x, y) which, combined with the weak duality inequality, shows that x S(P ), ȳ S(D) and (1.45) val(d) = val(p ) = L( x, ȳ). But we have more, in fact if we denote by SP (L) the set of saddle points, then: Lemma The following holds: (1.46) {val(d) = val(p ) is finite} SP (L) = S(P ) S(D). Proof. Indeed, let x S(P ) and ȳ S(D). Then (1.47) val(d) = inf x X L(x, ȳ) L( x, ȳ) sup L( x, y) = val(p ). y Y If val(d) = val(p ), then these inequalities are equalities, so that (1.43) holds, and therefore, ( x, ȳ) is a saddle point. The converse implication was already obtained.

18 18 1. A CONVEX OPTIMIZATION TOOLBOX 1.4. Linear programming and Hoffman bounds Linear programming. We assume in this section that X is a vector space (with no associated topology; this abstract setting does not make the proofs more complicated). Consider the infinite-dimensional linear program (LP ) Min c, x ; a i, x b i, i = 1,..., p; x X where c and a i, i = 1,..., p, are linear forms over X, b R p, and, denotes the action of a linear form over X. The associated Lagrangian function L : X R p R is defined as p (1.48) L(x, λ) := c, x + λ i ( a i, x b i ), where the multiplier λ has to belong to R p +. The primal value satisfies { c, x if x F (LP ), (1.49) p(x) = sup L(x, λ) = + otherwise. λ R p + i=1 Therefore (LP ) and the primal problem (of minimizing p(x)) have the same value and set of solutions. Since L(x, λ) = c + p i=1 λ ia i, x λb, we have that { λb if c + p (1.50) d(λ) = inf L(x, λ) = i=1 λ ia i = 0, x otherwise. The dual problem has therefore the same value and set of solutions than the following problem, called dual to (LP ): p (LD) Max λb; c + λ i a i = 0. λ R p + For x F (LP ), we denote the associated set of active constraints by (1.51) I(x) := {1 i p; a i, x = b i }. Consider the optimality system { (i) c + p (1.52) i=1 λ ia i = 0, (ii) λ i 0, a i, x b i, λ i ( a i, x b i ) = 0, i = 1,..., p. Lemma The pair (x, λ) F (LP ) F (LD) is a saddle point of the Lagrangian iff (1.52) holds. Proof. Let x F (LP ) and λ F (LD). Then (1.52)(i) holds, implying that the difference of cost function is equal to p (1.53) c, x + λb = λ i (b i a i, x ). i=1 This sum of nonnegative terms is equal to zero iff the last relation of (1.52)(ii) holds, and then x S(LP ) and λ S(LD), proving that (x, λ) is a saddle point. The converse implication is easily obtained. i=1 We next deal with the existence of solutions.

19 1. CONVEX FUNCTIONS 19 Lemma If (LP ) (resp. (LD)) has a finite value, then its set of solutions is non empty. Proof. The proof of the two cases being similar, it suffices to prove the first statement. Since (LP ) has a finite value, there exists a minimizing sequence x k. Extracting if necessary a subsequence, we may assume that I(x k ) is constant say equal to J. Among such minimizing sequences we may assume that J is of maximal cardinal. If c, x k has, for large enough k, a constant value, then the corresponding x k are solution of (LP ). Otherwise, extracting if necessary a subsequence, we may assume that c, x k+1 < c, x k for all k. Set d k := x k+1 x k, and consider the set E k := {ρ 0; x k + ρd k F (LP )}. Since c, d k < 0 and val(lp ) >, this set is bounded. The maximal element is (1.54) ρ k := argmin{(b i a i, x k )/ a i, d k ; a i, d k > 0}. i We have that y k := x k + ρ k d k F (LP ), c, y k < c, x k and J I(y k ) strictly. Extracting from y k a minimizing sequence with constant set of active constraints, strictly containing J, we contradict the definition of J. Lemma Given y 1,..., y q in R n, the set E := { q i=1 λ iy i, λ 0} is closed. Proof. (i) Assume first that the y i are linearly independent, and denote by Y the generated vector space, of which they are a basis. The coefficients λ i represent the coordinate in this basis, and so, if e k e in E, its coordinates converge to those of e and the result follows. (ii) We next deal with the general case. Let e k e in E. Let us associate with each e k some λ k R q + such that ek = q i=1 λk i yi, and that λ k has minimal support (the support being the set of nonzero components). Taking if necessary a subsequence, we may assume that this support J is constant along the sequence. We claim that {y i, i J} is linearly independent, for if q i=1 µ iy i = 0 with µ 0, and µ i = 0 when λ k i = 0, we can find β k so that µ k := λ k + β k µ is nonnegative and has a support smaller than J, and e k = q i=1 µk i yi, in contradiction with the definition of J. Since {y i, i J} is linearly independent, we deduce as in step (i) that the λ k converge to some λ and that e = q λ i=1 i y i. The conclusion follows. We next prove a strong duality result. Lemma If val(lp ) is finite, then val(lp ) = val(ld) and both S(LP ) and S(LD) are non empty. Proof. By lemma 1.26, (LP ) has a solution x. Set J := I( x), j := J (the cardinal of J); we may assume that J = 1,..., j. Consider the set (1.55) C := {d X; a i, d 0, i J}. Obviously, if d C, then x + ρd F (LP ) for small enough ρ > 0, and consequently (1.56) c, d 0, for all d C.

20 20 1. A CONVEX OPTIMIZATION TOOLBOX Consider the mapping Ax := ( a 1, x,..., a j, x, c, x ) over X, with image E 1 R j+1. We claim that the point z := (0,..., 0, 1) does not belong to the set E 2 := E 1 + R j+1 +. Indeed, otherwise we would have z Ax, for some x X, and so a i, x 0, for all i J, whereas z j+1 = 1 contradicts (1.56). Let z 1,..., z q be a basis of the vector space E 1. Then E 2 is the set of nonnegative linear combinations of {±z 1,..., ±z q, e 1,..., e j+1 }, where by e i we denote the elements of the natural basis of R j+1. By lemma 1.25, E 2 is closed. Corollary 1.16, allows to strictly separate z and E 2. That is, (1.57) λ j+1 = λz < inf y E 2 λy, for some nonzero λ R j+1. Since E 2 is a cone, the above infimum is zero, whence λ j+1 > 0. Changing λ into λ/λ j+1 if necessary, we may assume that λ j+1 = 1. Since E 2 = E 1 + R j+1 + it follows that λ R j+1 +, and since E 1 E 2 we deduce that 0 i J λ ia i + c, d for all d X, meaning that i J λ ia i + c = 0. Let us now set λ i = 0 for i p, i J. Then (1.52) holds at the point x. By lemma 1.23, ( x, λ) is a saddle point and the conclusion follows. Remark It may happen, even in a finite dimensional setting, that both the primal and dual problem are unfeasible, so that they have value + and resp.; consider for instance the problem Min x R { x; 0 x = 1; x 0}, whose dual is Max λ R 2{ λ 1 ; 1 λ 2 = 0; λ 2 0} Hoffman bounds. As an application of linear programming duality we present Hoffman s lemma [27]. Lemma Given a Banach space X, a 1,..., a p in X, and b R p, set (1.58) C b := {x X; a i, x b i, i = 1,..., p}. Then there exists an Hoffman constant M > 0, not depending on b, such that, if C b, then p (1.59) dist(x, C b ) M ( a i, x b i ) +, for all x X. i=1 Proof. a) Define A L(X, R p ) by Ax = ( a 1, x,..., a p, x ). Let x 1,..., x q be elements of X such that (Ax 1,..., Ax q ) is a basis of Im(A). Then q p, and the family {x 1,..., x q } is linearly independent. Denote by H the vector space with basis x 1,..., x q. The Euclidean norm over H is equivalent to the one induced by X, since all norms are equivalent on finite dimensional spaces. b) Let x X. We may express the coordinates of Ax on the basis of the {Ax j } as functions of x, i.e., write Ax = q j=1 α j(x)ax j for some linear function α : X R q, that is continuous. Indeed, by the equivalence of norms in a finite dimensional space, for some positive c not depending on x: (1.60) α(x) c Ax c A x. Setting Cx = q j=1 α j(x)x j, and Bx := x Cx, we may write x = Bx + Cx, and

21 1. CONVEX FUNCTIONS 21 (1.61) ABx = Ax q α j (x)ax j = 0, j=1 i.e., AB = 0. So, C b (assumed to be non empty) is invariant by the addition of an element of the image of B. In particular it contains an element ˆx such that ˆx x H, that is, (1.62) ˆx = x + q β j x j. Specifically, let ˆx be such an element of C b for which β is minimum, that is, β is solution of the problem 1 (1.63) Min γ R q 2 γ 2 ; a i, x + j=1 q γ j x j b i, i = 1,..., p. j=1 The following optimality conditions hold: there exist λ R q + such that (1.64) γ j + as well a p λ i a i, x j = 0, j = 1,..., q. i=1 (1.65) λ i a i, x + q γ j x j b i = 0, i = 1,..., p. j=1 Using first (1.64) and then (1.65), we obtain (1.66) q q γ 2 = λ i a i, γ j x j = λ i ( a i, x b i ) λ j=1 γ 2 j = i,j i=1 q ( a i, x b i ) +. c) Among all possible multipliers λ we may take one with minimal support; From (1.64) we deduce the existence of M 1 > 0 not depending on x and b, such that (1.67) λ M 1 γ. Combining with (1.66), we deduce that q (1.68) γ M 1 ( a i, x b i ) +. i=1 The conclusion follows since, as noticed before, the Euclidean norm on H is equivalent to the one induced by the norm of X. i=1

22 22 1. A CONVEX OPTIMIZATION TOOLBOX Open mapping theorem. We will generalize the previous result, and for this need the following fundamental result in functional analysis. Theorem 1.29 (Open mapping theorem). Let X and Y be Banach spaces, and let A L(X, Y ) be surjective. Then αb Y AB X, for some α > 0. Proof. See e.g. [10]. Corollary Let A and α be as in theorem Then Im(A ) is closed, and (1.69) A λ α λ, for all λ Y. Proof. By the open mapping theorem, we have that (1.70) A λ = sup λ, Ax X α sup λ, y Y = α λ, x 1 y 1 proving (1.69). Let us now check that Im(A ) is closed. Let x k in Im(A ) converge to x. There exists a sequence λ k Y such that x k = A λ k. In view of (1.69), λ k is a Cauchy sequence and hence has a limit λ Y. Therefore x = A λ Y. The conclusion follows. Proposition Let X and Y be Banach spaces, and A L(X, Y ). Then Im(A ) (KerA), with equality if A has closed range. Proof. (a) Let x Im(A ), i.e., x = A y, for some y Y, and x Ker A. Then x, x X = y, Ax Y = 0. Therefore, Im(A ) (Ker A). (b) Assume now that A has closed range. Let x (Ker A). For y Y, set (1.71) v(y) := x, x, where x X satisfies Ax = y. Since x (Ker A), any x such that Ax = y gives the same value of x, x X, and therefore v(y) is well-defined. It is easily checked that it is a linear function. By the open mapping theorem, applied to the restriction of A from X to its image (the latter being a Banach space by hypothesis), there exists x α 1 y Y B X such that Ax = y, so that v(y) α 1 x y Y. So, v is a linear and continuous mapping, i.e., there exists y Y such that v(y) = y, y Y. For all x X, we have therefore x, x X = y, Ax Y = A y, x X, so that x = A y, as was to be proved. Remark See in example another proof, based on duality theory. Example Let X := L 2 (0, 1), Y := L 1 (0, 1), and A L(X, Y ) be the injection of X into Y. Then Ker A is reduced to 0, and therefore its orthogonal is X. On the other hand, we have that for y L (0, 1), A y is the operator in X defined by x 1 0 y (t)x(t)dt. So the image of A is a dense subspace of X, but A is not surjective, and therefore its image is not closed. We next give a useful generalization of Hoffman s lemma 1.28 in the homogeneous case.

23 1. CONVEX FUNCTIONS 23 Lemma Given Banach spaces X and Y, A L(X, Y ) surjective, and a 1,..., a p in X, set C := {x X; Ax = 0; a i, x 0, i = 1,..., p}. Then there exists an Hoffman constant M > 0 such that ( ) p (1.72) dist(x, C) M Ax + ( a i, x ) +, for all x X. i=1 Proof. By the open mapping theorem 1.29, there exists x Ker A such that x x α 1 Ax, where α is given by theorem Therefore for some M > 0 not depending on x: ( (1.73) ai, x ) + ( a i, x ) + + M Ax. Applying lemma 1.28 to x, with Ker A in place of X, we obtain the desired conclusion Conjugacy Basic properties. Let X be a Banach space and f : X R. Its (Legendre- Fenchel) conjugate is the function f : X R defined by (1.74) f (x ) := sup x, x f(x). x X This can be motivated as follows. Let us look for an affine minorant of f of the form x, x β. For given x, the best (i.e., minimal) value of β is precisely f (x ). Being a supremum of affine functions, f is l.s.c. convex. We obviously have the Fenchel-Young inequality (1.75) f (x ) x, x f(x), for all x X and x X. Remark We have that (1.76) f (x ) := for each x if dom(f) =. (1.77) f (x ) > for each x if dom(f). Since the supremum over an empty set is, we may always express f by maximizing over dom(f): (1.78) f (x ) := sup x, x f(x). x dom(f) If f(x) is finite we can write the Fenchel-Young inequality inequality in the more symmetric form (1.79) x, x f(x) + f (x ). Lemma Let f be proper. Then the symmetric form (1.79) of the Fenchel- Young inequality inequality is valid for any (x, x ) in X X. Proof. By (1.77), f (x ) >, and f(x) >, so that (1.79) makes sense and is equivalent to the Fenchel-Young inequality.

24 24 1. A CONVEX OPTIMIZATION TOOLBOX Example Let X be an Hilbert space, and set f : X R by f(x) := 1 2 x 2. We identify X with its dual. Then it happens that f (x) = f(x) for all x in X, leading to the well-known inequality (where the l.h.s. is the scalar product between x and y): (1.80) (x, y) 1 2 x y 2, for all x, y in X. Example Let p > 1. Define f : R R by f(x) := x p /p. For y R, the maximum in x xy f(x) is attained at 0 if y = 0, and otherwise for some x 0 of the same sign as y such that x p 1 = y. Introducing the conjugate exponent p such that 1/p + 1/p = 1, we get f (y) = y p /p, so that xy x p /p + y p /p, for all x, y in R. Similarly, for some p > 1, let f : R n R be defined by f(x) := x p p/p, where x p p = n i=1 x i p. We easily obtain that f (y) = y p p /p, leading to the Young inequality (1.81) n i=1 x i y i 1 p x p p + 1 p y p p, for all x, y in R. Exercice Let A be a symmetric, positive definite n n matrix. (i) Check that the conjugate of f(x) := 1 2 x Ax is f (y) := 1 2 y A 1 y. Taking x = y, deduce the Young inequality (1.82) x x Ax x A 1 x, for all x R n. Conclude that (1.83) A + A 1 2I is positive semidefinite, if A is symmetric and positive definite. Exercice Check that the conjugate of the indicatrix of the (open or closed) unit ball of X is the dual norm. Exercice Let f(x) := αg(x) with α > 0 and g : X R. Show that (1.84) f (x ) = αg (x /α). Exercice Show that the Fenchel conjugate of the exponentiel is the entropy function H with value H(x) = x(log x 1) if x > 0, H(0) = 0, and H(x) = if x < 0. Deduce the inequality xy e x + y(log y 1), for all x R and y > 0. The biconjugate of f is the function f : X R defined by (1.85) f (x) := sup x X x, x f (x ). Proposition The biconjugate f is the supremum of affine minorants of f. Proof. Let x X. If f (x ) =, then f has no affine minorant with slope x. Otherwise, as we already observed, x, x f (x ) is an affine minorant of f with the best possible constant term. The conclusion follows.

25 1. CONVEX FUNCTIONS 25 An hyperplane ( x, β) X R separating (x 0, α 0 ) from epi(f) is such that (1.86) x, x 0 + βα 0 x, x + βα, for all (x, α) epi(f). If dom(f), then for some x dom(f), we can take α + and it follows that β 0. If β = 0, we say that the hyperplane is vertical and the above relation reduces to (1.87) x, x 0 x, x for all x dom(f), i.e., x separates x 0 from dom(f). Otherwise, we say that the separating hyperplane is oblique. We may then assume that β = 1 and we obtain that (1.88) x, x 0 + α 0 x, x + f(x), for all x dom(f). This is equivalent to say that x x, x x 0 + α 0 is an affine minorant of f. Theorem Let f : X R be proper, l.s.c. convex. Then f = f. Proof. a) It suffices to prove that any (x 0, α 0 ) epi(f) can be strictly separated from epi(f) by an oblique hyperplane. Indeed, the corresponding affine minorant guarantees then that (x 0, α 0 ) epi(f ). Since f f, it follows that epi(f) = epi(f ) as is to be proved. b) Since epi(f) is a closed convex subset of X R, by corollary 1.16, it can be strictly separated from (x 0, α 0 ). Note that not all separating hyperplanes are vertical, since otherwise f would have value over its (nonempty) domain. It follows that f has an affine minorant, say x x, x + γ. If the hyperplane strictly separating (x 0, α 0 ) and epi(f) is oblique, it provides an affine minorant of f with value greater than α 0 at the point x 0, as required. If, on the contrary, the hyperplane strictly separating (x 0, α 0 ) and epi(f) is vertical, say y, x x 0 + ε 0 for all x dom(f), with y 0 and ε > 0, then we have that for any β > 0 and all x dom(f): (1.89) f(x) β( y, x x 0 + ε) + x, x + γ, meaning that the above r.h.s. is an affine minorant of f. At the same time, its value at x 0 is βε + x, x 0 + γ, which for β > 0 large enough is larger than α 0. So this r.h.s. is an oblique hyperplane separating (x 0, α 0 ) from epi(f). The conclusion follows. Definition (i) Let E X. The convex hull conv(e) is the smallest convex set containing E, i.e., the set of finite convex combinations (linear combinations with nonnegative weights whose sum is 1) of elements of E. The convex closure of E denoted by conv(e) is the smallest closed convex set containing E (i.e., the intersection of closed convex set containing E). (ii) Let f : X R. The convex closure of f is the function conv(f) : X R whose epigraph is conv(epi(f)) (note that conv(f) is the supremum of l.s.c. convex minorants of f). We have obviously that f = conv(f) iff f is convex and l.s.c.

26 26 1. A CONVEX OPTIMIZATION TOOLBOX Theorem 1.46 (Fenchel-Moreau-Rockafellar). Let f : X R. We have the following alternative: either (i) f = identically, conv(f) has no finite value, and has value at some point, or (ii) f = conv(f) and conv(f)(x) >, for all x X. Proof. If f is identically equal to +, the conclusion is obvious. So we may assume that dom(f). Since f is a l.s.c. convex minorant of f, we have that f conv(f). So, if conv(f)(x 1 ) = for some x 1 X, then f has no affine minorant and f =. In addition, since conv(f) is l.s.c. convex, for any x dom(conv(f)), setting x θ := θx + (1 θ)x 1, we have that (1.90) conv(f)(x) lim θ 1 conv(f)(x θ ) θ conv(f)(x) + (1 θ) conv(f)(x 1 ) =, so that (i) holds. On the contrary, if (i) does not hold, then f has a continuous affine minorant, so that then conv(f)(x) >, for all x X. Being proper, l.s.c. and convex, conv(f) is by theorem 1.44 the supremum of its affine minorants, which coincide with the affine minorants of f. The conclusion follows. Corollary Let f be convex X R. Then (i) conv(f)(x) = lim inf x x f(x ), for all x X, (ii) if f is finite valued and l.s.c. at some x 0 X, then f(x 0 ) = f (x 0 ). Proof. (i) Set g(x) := lim inf x x f(x ). It is easily checked that g is a l.s.c. convex minorant of f, and therefore g conv(f). On the other hand, since conv(f) is a l.s.c. minorant of f, we have that conv(f)(x) lim inf x x f(x ) = g(x), proving (i). (ii) By point (i), since f is finite valued and l.s.c. at x 0, we have that f(x 0 ) = conv(f)(x 0 ) >, and we conclude with theorem Example Let K be a nonempty closed convex subset of X, and set f(x) = if x K, and f(x) = + otherwise. Then f is l.s.c. convex, and f has value everywhere, so that f f Conjugacy in dual spaces. Let X be a Banach space, and g : X R. Its (Legendre-Fenchel) conjugate (in the dual sense) is the function g : X R defined by (1.91) g (x) := sup x X x, x g(x ). So we have the dual Fenchel-Young inequality (1.92) g (x) x, x g(x ), for all x X and x X. Being a supremum of affine functions, g is l.s.c. convex. Its biconjugate g is the Legendre-Fenchel conjugate (in the sense of section 1.5) of g, i.e. (1.93) g (x ) := sup x, x g (x). x X Let us call affine function a function X R of the form x x, x + α, with (x, α) X R; note that this excludes the affine functions of the form x

27 1. CONVEX FUNCTIONS 27 x, x + α, where x X \ X. We call affine minorants of g the affine functions that minorize g. By the same arguments as in section 1.5, we obtain that (1.94) g is the supremum of affine minorants of g, and so we get the following result: Lemma Let g : X R. We have that g = g iff g is a supremum of affine functions. Remark For any f : X R, the Fenchel conjugate f is a supremum of affine functions. It follows that (1.95) f = f. Example Remember the definition (1.6) of the indicatrix function. The support function σ K : X R (sometimes also denoted by σ(, K)) is defined by (1.96) σ K (x ) := sup{ x, x ; x K}. Clearly σ K is the conjugate of I K. If K is closed and convex, then I K is proper, l.s.c. and convex, and hence, is equal to its biconjugate, so that the conjugate of σ K is I K. Otherwise, let K be the smallest closed convex set containing K. It is easily checked that I K = σ K. Since I K is l.s.c. convex and proper it is equal to its biconjugate. We proved that (1.97) I K and σ K are conjugate one to each other Continuity and subdifferentiability. Let f : X R have a finite value at some x X. We define the subdifferential of f at x as the set (1.98) f(x) := {x X ; f(x ) f(x) + x, x x, for all x X}. Equivalently, f(x) is the set of slopes of affine minorants of f that are exact (i.e., equal to f) at the point x. The inequality in (1.98) may be written as (1.99) x, x f(x) x, x f(x ), for all x X. Therefore x f(x) iff x attains the maximum in the definition of f (x ), i.e., we have that (1.100) {x f(x)} {f (x ) + f(x) = x, x }. In other words, x f(x) iff it gives an equality in the Fenchel-Young inequality (1.79). By (1.85), f is the supremum of its affine minorants which are of the form x, x β, β f (x ), for x dom(f ). It follows that the affine minorants that are exact at x for f are those which attain the supremum in (1.85), i.e. (1.101) {x f (x)} {f (x ) + f (x) = x, x }. Also, if f(x), then the corresponding affine minorants, exact at x, are also minorants of f exact at x, and therefore (1.102) { f(x) } {f (x) = f(x)} { f (x) = f(x)}.

28 28 1. A CONVEX OPTIMIZATION TOOLBOX We also may define the subdifferential of a function g : X R with finite value at x as (1.103) g(x ) := {x X; g(y ) g(x ) + y x, x, for all y X }. Similarly to what was done before we can express the above inequality as (1.104) x, x g(x ) y, x g(y ), for all y X. This means that x g(x ) iff x attains the maximum in the definition of g (x), i.e., we have that (1.105) {x g(x )} {g(x ) + g (x) = x, x }. With similar arguments we obtain that (1.106) {x g (x )} {g (x ) + g (x) = x, x }. When g is itself a conjugate function we deduce the following. Lemma Let f : X R have a finite value at some x X. That the equality holds in the Fenchel-Young inequality (1.79) implies that x f (x ); the converse holds if f is proper, l.s.c. convex. Proof. If (1.79) holds with equality, we know that f(x) = f (x) and so, by (1.105) applied to g = f, x f (x ) holds. Conversely, if x f (x ), then by (1.105) applied to g = f, we have that f (x ) + f (x) = x, x. When f is proper, l.s.c. convex, f (x) = f(x), so that the equality holds in the Fenchel-Young inequality, as was to be proved. Remark So, if f is proper, l.s.c. convex, we have that (1.107) x f(x) iff x f (x ). In this sense the Fenchel Legendre transform is an extension of the property of the classical Legendre transform which, under certain conditions, associates with a smooth function f over R n another smooth function ˆf over R n such that y = f (x) iff x = ˆf (y). In the analysis of stochastic problems we will need the following sensitivity analysis results for linear programs. Example Given d R m, b R p and matrices A and M of size p m and p n resp., let f : R n R be defined by (1.108) f(x) := inf {dy; Ay = b + Mx}. y R m + This is the value of a linear program whose dual is (D x ) Max λ(b + Mx); d + λa 0. λ Rp The next lemma gives an expression of f(x). Lemma Let f have a finite value at x X. Then f( x) is non empty and satisfies (1.109) f( x) = { λm; λ S(D x )}.

29 Proof. The conjugate of f is 1. CONVEX FUNCTIONS 29 (1.110) f (x ) = sup x;y 0 {x x dy; Ay = b + Mx} = inf x;y 0 {dy x x; Ay = b + Mx}. Since f( x) is finite, the linear program involved in the above r.h.s. is feasible. By linear programmming duality (lemma 1.26) it has the same value as its dual, and hence, (1.111) f (x ) = sup λ R p { λb; x = λm; d + λa 0}. The Fenchel Young inequality implies (1.112) 0 f( x) + f (x ) x x = f( x) x x + inf λ R p {λb; x = λm; d + λa 0}. Since f( x) is the finite value of a feasible linear program, it is equal to val(d x ). So, let λ S(D x ). The Fenchel Young inequality (1.112) is equivalent to (1.113) λ(b + M x) x x + inf λ R p {λb; x = λm; d + λa 0}. When equality holds, the linear program on the r.h.s. has a solution say λ, and x x = λm x, so that the equality holds iff (1.114) λ(b + M x) = min λ R p {λ(b + M x); x = λm; d + λa 0}. Remember that this is the case of equality in the Fenchel Young inequality, and therefore it holds iff x f( x). Since the cost function and last constraint correspond to those of (D x ), it follows that any solution ˆλ of the linear program on the r.h.s. belongs to S(D x ). We have proved that, if x f(x), then x = λm for some λ S(D x ). The converse obviously holds in view of (1.114). Remark Consider the particular case when b = 0 and M is the opposite of the identity. Rewriting as b the variable x, we obtain that the function (1.115) f(b) := inf {dy; Ay + b = 0} y R m + has, over its domain, a subdifferential equal to the set of solutions of the dual problem (1.116) Max λb; d + λa 0. λ Rp We now show that, for convex functions, a local uniform upper bound implies a Lipschitz property as well as subdifferentiability. Lemma Let f : X R be convex, finitely valued at x 0, and uniformly upper bounded near x 0, i.e., for some a R and r > 0: (1.117) f(x) a whenever x x 0 r. Then f is Lipschitz of constant say L on B(x 0, 1 2 r).

30 30 1. A CONVEX OPTIMIZATION TOOLBOX Proof. Let ε ]0, 1[ and h X, h = εr. Since x 0 ± ε 1 h B(x 0, r), we have by convexity of f that It follows that f(x 0 + h) (1 ε)f(x 0 ) + εf(x 0 + ε 1 h) (1 ε)f(x 0 ) + εa, f(x 0 + h) (1 + ε)f(x 0 ) εf(x 0 ε 1 h) (1 + ε)f(x 0 ) εa. (1.118) f(x 0 + h) f(x 0 ) ε(a f(x 0 )) = r 1 (a f(x 0 )) h. Therefore, for all x B(x 0, r), we have that f(x) b, with b := f(x 0 ) (a f(x 0 )). Let x 1 B(x 0, r 1 ), where r 1 := 1 2 r. Then b f(x) a, for all x B(x 1, r 1 ). Applying (1.118) at the point x 1, with r = r 1, we get (1.119) f(x 1 + h) f(x 1 ) r 1 1 (a b) h, for all h < r 1. Therefore f is Lipschitz with constant r 1 1 (a b) over B(x 0, r 1 ), as was to be proved. Lemma Let f : X R be convex, and Lipschitz of constant L near x 0 X. Then f(x 0 ) is non empty and included in B(0, L). Proof. Let ˆx B(x 0, 1 2r). Set E = {(x, γ) X R; γ > f(x)}. Since f is continuous at x 0, for ε > 0 small enough, (1.120) B(x 0, ε) [f(x 0 ) + 1, ) E, so that E has a non empty interior. By theorem 1.12, there exists (λ, α) X R separating (ˆx, f(ˆx)) and E, i.e., such that (1.121) λ, ˆx + αf(ˆx) λ, x + αγ, for all x dom(f), γ > f(x). Taking x = x 0 and γ > f(x), γ +, we see that α 0. The separating hyperplane cannot be vertical, since ˆx int(dom(f)), so that we may take α = 1. Minimizing w.r.t. γ we obtain that f(x) f(ˆx) λ, x ˆx, proving that λ f(ˆx). We now check that f(x) B(0, L). Assume that x f(x), with x > L. Then there exists d X with d = 1 and x, d > L. Therefore by the definition of a subdifferential, f(x + σd) f(x) (1.122) lim x, d > L, σ 0 σ in contradiction with the fact that L is a local Lipschitz constant. Example Consider the entropy function f(x) = x log x if x 0 (with value 0 at zero), and f(x) = + otherwise. Then f is l.s.c. convex, and the subdifferential is empty at x = 0. So, in general, even in an Euclidean space, a l.s.c. convex function may have an empty subdifferential at some points of its domain. We next introduce a concept that in some sense is an algebraic variant of the interior. Definition Let S X, where X is a vector space. Then we say that x S belongs to the core of S and write x core(s) if, for each h X, there exists ε > 0 such that [x εh, x + εh] S.

Parcours OJD, Ecole Polytechnique et Université Pierre et Marie Curie 05 Mai 2015

Parcours OJD, Ecole Polytechnique et Université Pierre et Marie Curie 05 Mai 2015 Examen du cours Optimisation Stochastique Version 06/05/2014 Mastère de Mathématiques de la Modélisation F. Bonnans Parcours OJD, Ecole Polytechnique et Université Pierre et Marie Curie 05 Mai 2015 Authorized

More information

Optimization and Optimal Control in Banach Spaces

Optimization and Optimal Control in Banach Spaces Optimization and Optimal Control in Banach Spaces Bernhard Schmitzer October 19, 2017 1 Convex non-smooth optimization with proximal operators Remark 1.1 (Motivation). Convex optimization: easier to solve,

More information

Chapter 1. Preliminaries. The purpose of this chapter is to provide some basic background information. Linear Space. Hilbert Space.

Chapter 1. Preliminaries. The purpose of this chapter is to provide some basic background information. Linear Space. Hilbert Space. Chapter 1 Preliminaries The purpose of this chapter is to provide some basic background information. Linear Space Hilbert Space Basic Principles 1 2 Preliminaries Linear Space The notion of linear space

More information

Math 273a: Optimization Subgradients of convex functions

Math 273a: Optimization Subgradients of convex functions Math 273a: Optimization Subgradients of convex functions Made by: Damek Davis Edited by Wotao Yin Department of Mathematics, UCLA Fall 2015 online discussions on piazza.com 1 / 42 Subgradients Assumptions

More information

On duality theory of conic linear problems

On duality theory of conic linear problems On duality theory of conic linear problems Alexander Shapiro School of Industrial and Systems Engineering, Georgia Institute of Technology, Atlanta, Georgia 3332-25, USA e-mail: ashapiro@isye.gatech.edu

More information

Continuity of convex functions in normed spaces

Continuity of convex functions in normed spaces Continuity of convex functions in normed spaces In this chapter, we consider continuity properties of real-valued convex functions defined on open convex sets in normed spaces. Recall that every infinitedimensional

More information

GEOMETRIC APPROACH TO CONVEX SUBDIFFERENTIAL CALCULUS October 10, Dedicated to Franco Giannessi and Diethard Pallaschke with great respect

GEOMETRIC APPROACH TO CONVEX SUBDIFFERENTIAL CALCULUS October 10, Dedicated to Franco Giannessi and Diethard Pallaschke with great respect GEOMETRIC APPROACH TO CONVEX SUBDIFFERENTIAL CALCULUS October 10, 2018 BORIS S. MORDUKHOVICH 1 and NGUYEN MAU NAM 2 Dedicated to Franco Giannessi and Diethard Pallaschke with great respect Abstract. In

More information

2. Dual space is essential for the concept of gradient which, in turn, leads to the variational analysis of Lagrange multipliers.

2. Dual space is essential for the concept of gradient which, in turn, leads to the variational analysis of Lagrange multipliers. Chapter 3 Duality in Banach Space Modern optimization theory largely centers around the interplay of a normed vector space and its corresponding dual. The notion of duality is important for the following

More information

Dedicated to Michel Théra in honor of his 70th birthday

Dedicated to Michel Théra in honor of his 70th birthday VARIATIONAL GEOMETRIC APPROACH TO GENERALIZED DIFFERENTIAL AND CONJUGATE CALCULI IN CONVEX ANALYSIS B. S. MORDUKHOVICH 1, N. M. NAM 2, R. B. RECTOR 3 and T. TRAN 4. Dedicated to Michel Théra in honor of

More information

Chapter 2: Preliminaries and elements of convex analysis

Chapter 2: Preliminaries and elements of convex analysis Chapter 2: Preliminaries and elements of convex analysis Edoardo Amaldi DEIB Politecnico di Milano edoardo.amaldi@polimi.it Website: http://home.deib.polimi.it/amaldi/opt-14-15.shtml Academic year 2014-15

More information

8. Conjugate functions

8. Conjugate functions L. Vandenberghe EE236C (Spring 2013-14) 8. Conjugate functions closed functions conjugate function 8-1 Closed set a set C is closed if it contains its boundary: x k C, x k x = x C operations that preserve

More information

Chapter 2 Convex Analysis

Chapter 2 Convex Analysis Chapter 2 Convex Analysis The theory of nonsmooth analysis is based on convex analysis. Thus, we start this chapter by giving basic concepts and results of convexity (for further readings see also [202,

More information

THE UNIQUE MINIMAL DUAL REPRESENTATION OF A CONVEX FUNCTION

THE UNIQUE MINIMAL DUAL REPRESENTATION OF A CONVEX FUNCTION THE UNIQUE MINIMAL DUAL REPRESENTATION OF A CONVEX FUNCTION HALUK ERGIN AND TODD SARVER Abstract. Suppose (i) X is a separable Banach space, (ii) C is a convex subset of X that is a Baire space (when endowed

More information

Convex Functions. Pontus Giselsson

Convex Functions. Pontus Giselsson Convex Functions Pontus Giselsson 1 Today s lecture lower semicontinuity, closure, convex hull convexity preserving operations precomposition with affine mapping infimal convolution image function supremum

More information

ZERO DUALITY GAP FOR CONVEX PROGRAMS: A GENERAL RESULT

ZERO DUALITY GAP FOR CONVEX PROGRAMS: A GENERAL RESULT ZERO DUALITY GAP FOR CONVEX PROGRAMS: A GENERAL RESULT EMIL ERNST AND MICHEL VOLLE Abstract. This article addresses a general criterion providing a zero duality gap for convex programs in the setting of

More information

Helly's Theorem and its Equivalences via Convex Analysis

Helly's Theorem and its Equivalences via Convex Analysis Portland State University PDXScholar University Honors Theses University Honors College 2014 Helly's Theorem and its Equivalences via Convex Analysis Adam Robinson Portland State University Let us know

More information

Locally convex spaces, the hyperplane separation theorem, and the Krein-Milman theorem

Locally convex spaces, the hyperplane separation theorem, and the Krein-Milman theorem 56 Chapter 7 Locally convex spaces, the hyperplane separation theorem, and the Krein-Milman theorem Recall that C(X) is not a normed linear space when X is not compact. On the other hand we could use semi

More information

The proximal mapping

The proximal mapping The proximal mapping http://bicmr.pku.edu.cn/~wenzw/opt-2016-fall.html Acknowledgement: this slides is based on Prof. Lieven Vandenberghes lecture notes Outline 2/37 1 closed function 2 Conjugate function

More information

Chapter 3. Characterization of best approximations. 3.1 Characterization of best approximations in Hilbert spaces

Chapter 3. Characterization of best approximations. 3.1 Characterization of best approximations in Hilbert spaces Chapter 3 Characterization of best approximations In this chapter we study properties which characterite solutions of the approximation problem. There is a big difference in the treatment of this question

More information

LECTURE 25: REVIEW/EPILOGUE LECTURE OUTLINE

LECTURE 25: REVIEW/EPILOGUE LECTURE OUTLINE LECTURE 25: REVIEW/EPILOGUE LECTURE OUTLINE CONVEX ANALYSIS AND DUALITY Basic concepts of convex analysis Basic concepts of convex optimization Geometric duality framework - MC/MC Constrained optimization

More information

BASICS OF CONVEX ANALYSIS

BASICS OF CONVEX ANALYSIS BASICS OF CONVEX ANALYSIS MARKUS GRASMAIR 1. Main Definitions We start with providing the central definitions of convex functions and convex sets. Definition 1. A function f : R n R + } is called convex,

More information

Convex Analysis Background

Convex Analysis Background Convex Analysis Background John C. Duchi Stanford University Park City Mathematics Institute 206 Abstract In this set of notes, we will outline several standard facts from convex analysis, the study of

More information

Convex Optimization Notes

Convex Optimization Notes Convex Optimization Notes Jonathan Siegel January 2017 1 Convex Analysis This section is devoted to the study of convex functions f : B R {+ } and convex sets U B, for B a Banach space. The case of B =

More information

Math 341: Convex Geometry. Xi Chen

Math 341: Convex Geometry. Xi Chen Math 341: Convex Geometry Xi Chen 479 Central Academic Building, University of Alberta, Edmonton, Alberta T6G 2G1, CANADA E-mail address: xichen@math.ualberta.ca CHAPTER 1 Basics 1. Euclidean Geometry

More information

MAT 570 REAL ANALYSIS LECTURE NOTES. Contents. 1. Sets Functions Countability Axiom of choice Equivalence relations 9

MAT 570 REAL ANALYSIS LECTURE NOTES. Contents. 1. Sets Functions Countability Axiom of choice Equivalence relations 9 MAT 570 REAL ANALYSIS LECTURE NOTES PROFESSOR: JOHN QUIGG SEMESTER: FALL 204 Contents. Sets 2 2. Functions 5 3. Countability 7 4. Axiom of choice 8 5. Equivalence relations 9 6. Real numbers 9 7. Extended

More information

The Subdifferential of Convex Deviation Measures and Risk Functions

The Subdifferential of Convex Deviation Measures and Risk Functions The Subdifferential of Convex Deviation Measures and Risk Functions Nicole Lorenz Gert Wanka In this paper we give subdifferential formulas of some convex deviation measures using their conjugate functions

More information

Functional Analysis. Franck Sueur Metric spaces Definitions Completeness Compactness Separability...

Functional Analysis. Franck Sueur Metric spaces Definitions Completeness Compactness Separability... Functional Analysis Franck Sueur 2018-2019 Contents 1 Metric spaces 1 1.1 Definitions........................................ 1 1.2 Completeness...................................... 3 1.3 Compactness......................................

More information

Introduction to Functional Analysis

Introduction to Functional Analysis Introduction to Functional Analysis Carnegie Mellon University, 21-640, Spring 2014 Acknowledgements These notes are based on the lecture course given by Irene Fonseca but may differ from the exact lecture

More information

Some Properties of the Augmented Lagrangian in Cone Constrained Optimization

Some Properties of the Augmented Lagrangian in Cone Constrained Optimization MATHEMATICS OF OPERATIONS RESEARCH Vol. 29, No. 3, August 2004, pp. 479 491 issn 0364-765X eissn 1526-5471 04 2903 0479 informs doi 10.1287/moor.1040.0103 2004 INFORMS Some Properties of the Augmented

More information

Appendix PRELIMINARIES 1. THEOREMS OF ALTERNATIVES FOR SYSTEMS OF LINEAR CONSTRAINTS

Appendix PRELIMINARIES 1. THEOREMS OF ALTERNATIVES FOR SYSTEMS OF LINEAR CONSTRAINTS Appendix PRELIMINARIES 1. THEOREMS OF ALTERNATIVES FOR SYSTEMS OF LINEAR CONSTRAINTS Here we consider systems of linear constraints, consisting of equations or inequalities or both. A feasible solution

More information

Lecture Note 5: Semidefinite Programming for Stability Analysis

Lecture Note 5: Semidefinite Programming for Stability Analysis ECE7850: Hybrid Systems:Theory and Applications Lecture Note 5: Semidefinite Programming for Stability Analysis Wei Zhang Assistant Professor Department of Electrical and Computer Engineering Ohio State

More information

Semicontinuous functions and convexity

Semicontinuous functions and convexity Semicontinuous functions and convexity Jordan Bell jordan.bell@gmail.com Department of Mathematics, University of Toronto April 3, 2014 1 Lattices If (A, ) is a partially ordered set and S is a subset

More information

In Progress: Summary of Notation and Basic Results Convex Analysis C&O 663, Fall 2009

In Progress: Summary of Notation and Basic Results Convex Analysis C&O 663, Fall 2009 In Progress: Summary of Notation and Basic Results Convex Analysis C&O 663, Fall 2009 Intructor: Henry Wolkowicz November 26, 2009 University of Waterloo Department of Combinatorics & Optimization Abstract

More information

(convex combination!). Use convexity of f and multiply by the common denominator to get. Interchanging the role of x and y, we obtain that f is ( 2M ε

(convex combination!). Use convexity of f and multiply by the common denominator to get. Interchanging the role of x and y, we obtain that f is ( 2M ε 1. Continuity of convex functions in normed spaces In this chapter, we consider continuity properties of real-valued convex functions defined on open convex sets in normed spaces. Recall that every infinitedimensional

More information

Overview of normed linear spaces

Overview of normed linear spaces 20 Chapter 2 Overview of normed linear spaces Starting from this chapter, we begin examining linear spaces with at least one extra structure (topology or geometry). We assume linearity; this is a natural

More information

Division of the Humanities and Social Sciences. Supergradients. KC Border Fall 2001 v ::15.45

Division of the Humanities and Social Sciences. Supergradients. KC Border Fall 2001 v ::15.45 Division of the Humanities and Social Sciences Supergradients KC Border Fall 2001 1 The supergradient of a concave function There is a useful way to characterize the concavity of differentiable functions.

More information

Extreme points of compact convex sets

Extreme points of compact convex sets Extreme points of compact convex sets In this chapter, we are going to show that compact convex sets are determined by a proper subset, the set of its extreme points. Let us start with the main definition.

More information

A NICE PROOF OF FARKAS LEMMA

A NICE PROOF OF FARKAS LEMMA A NICE PROOF OF FARKAS LEMMA DANIEL VICTOR TAUSK Abstract. The goal of this short note is to present a nice proof of Farkas Lemma which states that if C is the convex cone spanned by a finite set and if

More information

Elements of Convex Optimization Theory

Elements of Convex Optimization Theory Elements of Convex Optimization Theory Costis Skiadas August 2015 This is a revised and extended version of Appendix A of Skiadas (2009), providing a self-contained overview of elements of convex optimization

More information

DEFINABLE VERSIONS OF THEOREMS BY KIRSZBRAUN AND HELLY

DEFINABLE VERSIONS OF THEOREMS BY KIRSZBRAUN AND HELLY DEFINABLE VERSIONS OF THEOREMS BY KIRSZBRAUN AND HELLY MATTHIAS ASCHENBRENNER AND ANDREAS FISCHER Abstract. Kirszbraun s Theorem states that every Lipschitz map S R n, where S R m, has an extension to

More information

Convex Analysis and Optimization Chapter 2 Solutions

Convex Analysis and Optimization Chapter 2 Solutions Convex Analysis and Optimization Chapter 2 Solutions Dimitri P. Bertsekas with Angelia Nedić and Asuman E. Ozdaglar Massachusetts Institute of Technology Athena Scientific, Belmont, Massachusetts http://www.athenasc.com

More information

CONTENTS. 4 Hausdorff Measure Introduction The Cantor Set Rectifiable Curves Cantor Set-Like Objects...

CONTENTS. 4 Hausdorff Measure Introduction The Cantor Set Rectifiable Curves Cantor Set-Like Objects... Contents 1 Functional Analysis 1 1.1 Hilbert Spaces................................... 1 1.1.1 Spectral Theorem............................. 4 1.2 Normed Vector Spaces.............................. 7 1.2.1

More information

Convex Optimization Theory. Chapter 5 Exercises and Solutions: Extended Version

Convex Optimization Theory. Chapter 5 Exercises and Solutions: Extended Version Convex Optimization Theory Chapter 5 Exercises and Solutions: Extended Version Dimitri P. Bertsekas Massachusetts Institute of Technology Athena Scientific, Belmont, Massachusetts http://www.athenasc.com

More information

Victoria Martín-Márquez

Victoria Martín-Márquez A NEW APPROACH FOR THE CONVEX FEASIBILITY PROBLEM VIA MONOTROPIC PROGRAMMING Victoria Martín-Márquez Dep. of Mathematical Analysis University of Seville Spain XIII Encuentro Red de Análisis Funcional y

More information

A Unified Analysis of Nonconvex Optimization Duality and Penalty Methods with General Augmenting Functions

A Unified Analysis of Nonconvex Optimization Duality and Penalty Methods with General Augmenting Functions A Unified Analysis of Nonconvex Optimization Duality and Penalty Methods with General Augmenting Functions Angelia Nedić and Asuman Ozdaglar April 16, 2006 Abstract In this paper, we study a unifying framework

More information

Lecture: Duality of LP, SOCP and SDP

Lecture: Duality of LP, SOCP and SDP 1/33 Lecture: Duality of LP, SOCP and SDP Zaiwen Wen Beijing International Center For Mathematical Research Peking University http://bicmr.pku.edu.cn/~wenzw/bigdata2017.html wenzw@pku.edu.cn Acknowledgement:

More information

An introduction to some aspects of functional analysis

An introduction to some aspects of functional analysis An introduction to some aspects of functional analysis Stephen Semmes Rice University Abstract These informal notes deal with some very basic objects in functional analysis, including norms and seminorms

More information

Math 273a: Optimization Subgradients of convex functions

Math 273a: Optimization Subgradients of convex functions Math 273a: Optimization Subgradients of convex functions Made by: Damek Davis Edited by Wotao Yin Department of Mathematics, UCLA Fall 2015 online discussions on piazza.com 1 / 20 Subgradients Assumptions

More information

2) Let X be a compact space. Prove that the space C(X) of continuous real-valued functions is a complete metric space.

2) Let X be a compact space. Prove that the space C(X) of continuous real-valued functions is a complete metric space. University of Bergen General Functional Analysis Problems with solutions 6 ) Prove that is unique in any normed space. Solution of ) Let us suppose that there are 2 zeros and 2. Then = + 2 = 2 + = 2. 2)

More information

Subdifferential representation of convex functions: refinements and applications

Subdifferential representation of convex functions: refinements and applications Subdifferential representation of convex functions: refinements and applications Joël Benoist & Aris Daniilidis Abstract Every lower semicontinuous convex function can be represented through its subdifferential

More information

AN INTRODUCTION TO CONVEXITY

AN INTRODUCTION TO CONVEXITY AN INTRODUCTION TO CONVEXITY GEIR DAHL NOVEMBER 2010 University of Oslo, Centre of Mathematics for Applications, P.O.Box 1053, Blindern, 0316 Oslo, Norway (geird@math.uio.no) Contents 1 The basic concepts

More information

Optimality Conditions for Nonsmooth Convex Optimization

Optimality Conditions for Nonsmooth Convex Optimization Optimality Conditions for Nonsmooth Convex Optimization Sangkyun Lee Oct 22, 2014 Let us consider a convex function f : R n R, where R is the extended real field, R := R {, + }, which is proper (f never

More information

Functional Analysis. Martin Brokate. 1 Normed Spaces 2. 2 Hilbert Spaces The Principle of Uniform Boundedness 32

Functional Analysis. Martin Brokate. 1 Normed Spaces 2. 2 Hilbert Spaces The Principle of Uniform Boundedness 32 Functional Analysis Martin Brokate Contents 1 Normed Spaces 2 2 Hilbert Spaces 2 3 The Principle of Uniform Boundedness 32 4 Extension, Reflexivity, Separation 37 5 Compact subsets of C and L p 46 6 Weak

More information

Semi-infinite programming, duality, discretization and optimality conditions

Semi-infinite programming, duality, discretization and optimality conditions Semi-infinite programming, duality, discretization and optimality conditions Alexander Shapiro School of Industrial and Systems Engineering, Georgia Institute of Technology, Atlanta, Georgia 30332-0205,

More information

Optimization Theory. A Concise Introduction. Jiongmin Yong

Optimization Theory. A Concise Introduction. Jiongmin Yong October 11, 017 16:5 ws-book9x6 Book Title Optimization Theory 017-08-Lecture Notes page 1 1 Optimization Theory A Concise Introduction Jiongmin Yong Optimization Theory 017-08-Lecture Notes page Optimization

More information

Local strong convexity and local Lipschitz continuity of the gradient of convex functions

Local strong convexity and local Lipschitz continuity of the gradient of convex functions Local strong convexity and local Lipschitz continuity of the gradient of convex functions R. Goebel and R.T. Rockafellar May 23, 2007 Abstract. Given a pair of convex conjugate functions f and f, we investigate

More information

Self-equilibrated Functions in Dual Vector Spaces: a Boundedness Criterion

Self-equilibrated Functions in Dual Vector Spaces: a Boundedness Criterion Self-equilibrated Functions in Dual Vector Spaces: a Boundedness Criterion Michel Théra LACO, UMR-CNRS 6090, Université de Limoges michel.thera@unilim.fr reporting joint work with E. Ernst and M. Volle

More information

B. Appendix B. Topological vector spaces

B. Appendix B. Topological vector spaces B.1 B. Appendix B. Topological vector spaces B.1. Fréchet spaces. In this appendix we go through the definition of Fréchet spaces and their inductive limits, such as they are used for definitions of function

More information

CHAPTER 5. Birkhoff Orthogonality and a Revisit to the Characterization of Best Approximations. 5.1 Introduction

CHAPTER 5. Birkhoff Orthogonality and a Revisit to the Characterization of Best Approximations. 5.1 Introduction CHAPTER 5 Birkhoff Orthogonality and a Revisit to the Characterization of Best Approximations 5.1 Introduction The notion of orthogonality in an arbitrary normed space, with the norm not necessarily coming

More information

Aliprantis, Border: Infinite-dimensional Analysis A Hitchhiker s Guide

Aliprantis, Border: Infinite-dimensional Analysis A Hitchhiker s Guide aliprantis.tex May 10, 2011 Aliprantis, Border: Infinite-dimensional Analysis A Hitchhiker s Guide Notes from [AB2]. 1 Odds and Ends 2 Topology 2.1 Topological spaces Example. (2.2) A semimetric = triangle

More information

Linear Programming. Larry Blume Cornell University, IHS Vienna and SFI. Summer 2016

Linear Programming. Larry Blume Cornell University, IHS Vienna and SFI. Summer 2016 Linear Programming Larry Blume Cornell University, IHS Vienna and SFI Summer 2016 These notes derive basic results in finite-dimensional linear programming using tools of convex analysis. Most sources

More information

Appendix B Convex analysis

Appendix B Convex analysis This version: 28/02/2014 Appendix B Convex analysis In this appendix we review a few basic notions of convexity and related notions that will be important for us at various times. B.1 The Hausdorff distance

More information

Lecture 3. Optimization Problems and Iterative Algorithms

Lecture 3. Optimization Problems and Iterative Algorithms Lecture 3 Optimization Problems and Iterative Algorithms January 13, 2016 This material was jointly developed with Angelia Nedić at UIUC for IE 598ns Outline Special Functions: Linear, Quadratic, Convex

More information

Strongly convex functions, Moreau envelopes and the generic nature of convex functions with strong minimizers

Strongly convex functions, Moreau envelopes and the generic nature of convex functions with strong minimizers University of Wollongong Research Online Faculty of Engineering and Information Sciences - Papers: Part B Faculty of Engineering and Information Sciences 206 Strongly convex functions, Moreau envelopes

More information

Convex Functions and Optimization

Convex Functions and Optimization Chapter 5 Convex Functions and Optimization 5.1 Convex Functions Our next topic is that of convex functions. Again, we will concentrate on the context of a map f : R n R although the situation can be generalized

More information

On duality gap in linear conic problems

On duality gap in linear conic problems On duality gap in linear conic problems C. Zălinescu Abstract In their paper Duality of linear conic problems A. Shapiro and A. Nemirovski considered two possible properties (A) and (B) for dual linear

More information

Iterative Convex Optimization Algorithms; Part One: Using the Baillon Haddad Theorem

Iterative Convex Optimization Algorithms; Part One: Using the Baillon Haddad Theorem Iterative Convex Optimization Algorithms; Part One: Using the Baillon Haddad Theorem Charles Byrne (Charles Byrne@uml.edu) http://faculty.uml.edu/cbyrne/cbyrne.html Department of Mathematical Sciences

More information

Introduction to Convex and Quasiconvex Analysis

Introduction to Convex and Quasiconvex Analysis Introduction to Convex and Quasiconvex Analysis J.B.G.Frenk Econometric Institute, Erasmus University, Rotterdam G.Kassay Faculty of Mathematics, Babes Bolyai University, Cluj August 27, 2001 Abstract

More information

Math 5052 Measure Theory and Functional Analysis II Homework Assignment 7

Math 5052 Measure Theory and Functional Analysis II Homework Assignment 7 Math 5052 Measure Theory and Functional Analysis II Homework Assignment 7 Prof. Wickerhauser Due Friday, February 5th, 2016 Please do Exercises 3, 6, 14, 16*, 17, 18, 21*, 23*, 24, 27*. Exercises marked

More information

Best approximations in normed vector spaces

Best approximations in normed vector spaces Best approximations in normed vector spaces Mike de Vries 5699703 a thesis submitted to the Department of Mathematics at Utrecht University in partial fulfillment of the requirements for the degree of

More information

Functional Analysis I

Functional Analysis I Functional Analysis I Course Notes by Stefan Richter Transcribed and Annotated by Gregory Zitelli Polar Decomposition Definition. An operator W B(H) is called a partial isometry if W x = X for all x (ker

More information

LECTURE SLIDES ON BASED ON CLASS LECTURES AT THE CAMBRIDGE, MASS FALL 2007 BY DIMITRI P. BERTSEKAS.

LECTURE SLIDES ON BASED ON CLASS LECTURES AT THE CAMBRIDGE, MASS FALL 2007 BY DIMITRI P. BERTSEKAS. LECTURE SLIDES ON CONVEX ANALYSIS AND OPTIMIZATION BASED ON 6.253 CLASS LECTURES AT THE MASSACHUSETTS INSTITUTE OF TECHNOLOGY CAMBRIDGE, MASS FALL 2007 BY DIMITRI P. BERTSEKAS http://web.mit.edu/dimitrib/www/home.html

More information

A Dual Condition for the Convex Subdifferential Sum Formula with Applications

A Dual Condition for the Convex Subdifferential Sum Formula with Applications Journal of Convex Analysis Volume 12 (2005), No. 2, 279 290 A Dual Condition for the Convex Subdifferential Sum Formula with Applications R. S. Burachik Engenharia de Sistemas e Computacao, COPPE-UFRJ

More information

Chapter 1. Introduction

Chapter 1. Introduction Chapter 1 Introduction Functional analysis can be seen as a natural extension of the real analysis to more general spaces. As an example we can think at the Heine - Borel theorem (closed and bounded is

More information

Inequality Constraints

Inequality Constraints Chapter 2 Inequality Constraints 2.1 Optimality Conditions Early in multivariate calculus we learn the significance of differentiability in finding minimizers. In this section we begin our study of the

More information

CHAPTER 2: CONVEX SETS AND CONCAVE FUNCTIONS. W. Erwin Diewert January 31, 2008.

CHAPTER 2: CONVEX SETS AND CONCAVE FUNCTIONS. W. Erwin Diewert January 31, 2008. 1 ECONOMICS 594: LECTURE NOTES CHAPTER 2: CONVEX SETS AND CONCAVE FUNCTIONS W. Erwin Diewert January 31, 2008. 1. Introduction Many economic problems have the following structure: (i) a linear function

More information

Chapter 1. Preliminaries

Chapter 1. Preliminaries Introduction This dissertation is a reading of chapter 4 in part I of the book : Integer and Combinatorial Optimization by George L. Nemhauser & Laurence A. Wolsey. The chapter elaborates links between

More information

LECTURE SLIDES ON CONVEX OPTIMIZATION AND DUALITY THEORY TATA INSTITUTE FOR FUNDAMENTAL RESEARCH MUMBAI, INDIA JANUARY 2009 PART I

LECTURE SLIDES ON CONVEX OPTIMIZATION AND DUALITY THEORY TATA INSTITUTE FOR FUNDAMENTAL RESEARCH MUMBAI, INDIA JANUARY 2009 PART I LECTURE SLIDES ON CONVEX OPTIMIZATION AND DUALITY THEORY TATA INSTITUTE FOR FUNDAMENTAL RESEARCH MUMBAI, INDIA JANUARY 2009 PART I BY DIMITRI P. BERTSEKAS M.I.T. http://web.mit.edu/dimitrib/www/home.html

More information

On John type ellipsoids

On John type ellipsoids On John type ellipsoids B. Klartag Tel Aviv University Abstract Given an arbitrary convex symmetric body K R n, we construct a natural and non-trivial continuous map u K which associates ellipsoids to

More information

Some Background Material

Some Background Material Chapter 1 Some Background Material In the first chapter, we present a quick review of elementary - but important - material as a way of dipping our toes in the water. This chapter also introduces important

More information

Shiqian Ma, MAT-258A: Numerical Optimization 1. Chapter 4. Subgradient

Shiqian Ma, MAT-258A: Numerical Optimization 1. Chapter 4. Subgradient Shiqian Ma, MAT-258A: Numerical Optimization 1 Chapter 4 Subgradient Shiqian Ma, MAT-258A: Numerical Optimization 2 4.1. Subgradients definition subgradient calculus duality and optimality conditions Shiqian

More information

Recall that if X is a compact metric space, C(X), the space of continuous (real-valued) functions on X, is a Banach space with the norm

Recall that if X is a compact metric space, C(X), the space of continuous (real-valued) functions on X, is a Banach space with the norm Chapter 13 Radon Measures Recall that if X is a compact metric space, C(X), the space of continuous (real-valued) functions on X, is a Banach space with the norm (13.1) f = sup x X f(x). We want to identify

More information

Extreme Abridgment of Boyd and Vandenberghe s Convex Optimization

Extreme Abridgment of Boyd and Vandenberghe s Convex Optimization Extreme Abridgment of Boyd and Vandenberghe s Convex Optimization Compiled by David Rosenberg Abstract Boyd and Vandenberghe s Convex Optimization book is very well-written and a pleasure to read. The

More information

I teach myself... Hilbert spaces

I teach myself... Hilbert spaces I teach myself... Hilbert spaces by F.J.Sayas, for MATH 806 November 4, 2015 This document will be growing with the semester. Every in red is for you to justify. Even if we start with the basic definition

More information

POLARS AND DUAL CONES

POLARS AND DUAL CONES POLARS AND DUAL CONES VERA ROSHCHINA Abstract. The goal of this note is to remind the basic definitions of convex sets and their polars. For more details see the classic references [1, 2] and [3] for polytopes.

More information

Lecture Notes in Advanced Calculus 1 (80315) Raz Kupferman Institute of Mathematics The Hebrew University

Lecture Notes in Advanced Calculus 1 (80315) Raz Kupferman Institute of Mathematics The Hebrew University Lecture Notes in Advanced Calculus 1 (80315) Raz Kupferman Institute of Mathematics The Hebrew University February 7, 2007 2 Contents 1 Metric Spaces 1 1.1 Basic definitions...........................

More information

Almost Convex Functions: Conjugacy and Duality

Almost Convex Functions: Conjugacy and Duality Almost Convex Functions: Conjugacy and Duality Radu Ioan Boţ 1, Sorin-Mihai Grad 2, and Gert Wanka 3 1 Faculty of Mathematics, Chemnitz University of Technology, D-09107 Chemnitz, Germany radu.bot@mathematik.tu-chemnitz.de

More information

Lecture 6 - Convex Sets

Lecture 6 - Convex Sets Lecture 6 - Convex Sets Definition A set C R n is called convex if for any x, y C and λ [0, 1], the point λx + (1 λ)y belongs to C. The above definition is equivalent to saying that for any x, y C, the

More information

arxiv:math/ v1 [math.fa] 4 Feb 1993

arxiv:math/ v1 [math.fa] 4 Feb 1993 Lectures on Maximal Monotone Operators R. R. Phelps Dept. Math. GN 50, Univ. of Wash., Seattle WA 98195; phelps@math.washington.edu (Lectures given at Prague/Paseky Summer School, Czech Republic, August

More information

ON GAP FUNCTIONS OF VARIATIONAL INEQUALITY IN A BANACH SPACE. Sangho Kum and Gue Myung Lee. 1. Introduction

ON GAP FUNCTIONS OF VARIATIONAL INEQUALITY IN A BANACH SPACE. Sangho Kum and Gue Myung Lee. 1. Introduction J. Korean Math. Soc. 38 (2001), No. 3, pp. 683 695 ON GAP FUNCTIONS OF VARIATIONAL INEQUALITY IN A BANACH SPACE Sangho Kum and Gue Myung Lee Abstract. In this paper we are concerned with theoretical properties

More information

Spectral Theory, with an Introduction to Operator Means. William L. Green

Spectral Theory, with an Introduction to Operator Means. William L. Green Spectral Theory, with an Introduction to Operator Means William L. Green January 30, 2008 Contents Introduction............................... 1 Hilbert Space.............................. 4 Linear Maps

More information

In English, this means that if we travel on a straight line between any two points in C, then we never leave C.

In English, this means that if we travel on a straight line between any two points in C, then we never leave C. Convex sets In this section, we will be introduced to some of the mathematical fundamentals of convex sets. In order to motivate some of the definitions, we will look at the closest point problem from

More information

Structural and Multidisciplinary Optimization. P. Duysinx and P. Tossings

Structural and Multidisciplinary Optimization. P. Duysinx and P. Tossings Structural and Multidisciplinary Optimization P. Duysinx and P. Tossings 2018-2019 CONTACTS Pierre Duysinx Institut de Mécanique et du Génie Civil (B52/3) Phone number: 04/366.91.94 Email: P.Duysinx@uliege.be

More information

Set, functions and Euclidean space. Seungjin Han

Set, functions and Euclidean space. Seungjin Han Set, functions and Euclidean space Seungjin Han September, 2018 1 Some Basics LOGIC A is necessary for B : If B holds, then A holds. B A A B is the contraposition of B A. A is sufficient for B: If A holds,

More information

Convex Analysis and Economic Theory AY Elementary properties of convex functions

Convex Analysis and Economic Theory AY Elementary properties of convex functions Division of the Humanities and Social Sciences Ec 181 KC Border Convex Analysis and Economic Theory AY 2018 2019 Topic 6: Convex functions I 6.1 Elementary properties of convex functions We may occasionally

More information

Convex Optimization Conjugate, Subdifferential, Proximation

Convex Optimization Conjugate, Subdifferential, Proximation 1 Lecture Notes, HCI, 3.11.211 Chapter 6 Convex Optimization Conjugate, Subdifferential, Proximation Bastian Goldlücke Computer Vision Group Technical University of Munich 2 Bastian Goldlücke Overview

More information

USING FUNCTIONAL ANALYSIS AND SOBOLEV SPACES TO SOLVE POISSON S EQUATION

USING FUNCTIONAL ANALYSIS AND SOBOLEV SPACES TO SOLVE POISSON S EQUATION USING FUNCTIONAL ANALYSIS AND SOBOLEV SPACES TO SOLVE POISSON S EQUATION YI WANG Abstract. We study Banach and Hilbert spaces with an eye towards defining weak solutions to elliptic PDE. Using Lax-Milgram

More information

Handout 2: Elements of Convex Analysis

Handout 2: Elements of Convex Analysis ENGG 5501: Foundations of Optimization 2018 19 First Term Handout 2: Elements of Convex Analysis Instructor: Anthony Man Cho So September 10, 2018 As briefly mentioned in Handout 1, the notion of convexity

More information

A SET OF LECTURE NOTES ON CONVEX OPTIMIZATION WITH SOME APPLICATIONS TO PROBABILITY THEORY INCOMPLETE DRAFT. MAY 06

A SET OF LECTURE NOTES ON CONVEX OPTIMIZATION WITH SOME APPLICATIONS TO PROBABILITY THEORY INCOMPLETE DRAFT. MAY 06 A SET OF LECTURE NOTES ON CONVEX OPTIMIZATION WITH SOME APPLICATIONS TO PROBABILITY THEORY INCOMPLETE DRAFT. MAY 06 CHRISTIAN LÉONARD Contents Preliminaries 1 1. Convexity without topology 1 2. Convexity

More information

A Geometric Framework for Nonconvex Optimization Duality using Augmented Lagrangian Functions

A Geometric Framework for Nonconvex Optimization Duality using Augmented Lagrangian Functions A Geometric Framework for Nonconvex Optimization Duality using Augmented Lagrangian Functions Angelia Nedić and Asuman Ozdaglar April 15, 2006 Abstract We provide a unifying geometric framework for the

More information