LECTURE 12 LECTURE OUTLINE. Subgradients Fenchel inequality Sensitivity in constrained optimization Subdifferential calculus Optimality conditions

Similar documents
LECTURE 25: REVIEW/EPILOGUE LECTURE OUTLINE

LECTURE 10 LECTURE OUTLINE

LECTURE 4 LECTURE OUTLINE

LECTURE 3 LECTURE OUTLINE

LECTURE 13 LECTURE OUTLINE

Lecture 1: Background on Convex Analysis

Convex Optimization Theory. Chapter 5 Exercises and Solutions: Extended Version

LECTURE 9 LECTURE OUTLINE. Min Common/Max Crossing for Min-Max

Convex Optimization Theory

LECTURE SLIDES ON BASED ON CLASS LECTURES AT THE CAMBRIDGE, MASS FALL 2007 BY DIMITRI P. BERTSEKAS.

Subgradients. subgradients and quasigradients. subgradient calculus. optimality conditions via subgradients. directional derivatives

LECTURE SLIDES ON CONVEX ANALYSIS AND OPTIMIZATION BASED ON LECTURES GIVEN AT THE MASSACHUSETTS INSTITUTE OF TECHNOLOGY CAMBRIDGE, MASS

LECTURE SLIDES ON CONVEX OPTIMIZATION AND DUALITY THEORY TATA INSTITUTE FOR FUNDAMENTAL RESEARCH MUMBAI, INDIA JANUARY 2009 PART I

Subgradients. subgradients. strong and weak subgradient calculus. optimality conditions via subgradients. directional derivatives

Optimality Conditions for Nonsmooth Convex Optimization

Convex Functions. Pontus Giselsson

Extended Monotropic Programming and Duality 1

Shiqian Ma, MAT-258A: Numerical Optimization 1. Chapter 4. Subgradient

Convex Analysis and Optimization Chapter 4 Solutions

Convex Optimization Theory. Chapter 3 Exercises and Solutions: Extended Version

Lecture 8. Strong Duality Results. September 22, 2008

Coordinate Update Algorithm Short Course Subgradients and Subgradient Methods

Optimization and Optimal Control in Banach Spaces

Chapter 2: Preliminaries and elements of convex analysis

Convex Optimization. (EE227A: UC Berkeley) Lecture 4. Suvrit Sra. (Conjugates, subdifferentials) 31 Jan, 2013

Solutions Chapter 5. The problem of finding the minimum distance from the origin to a line is written as. min 1 2 kxk2. subject to Ax = b.

Duality for almost convex optimization problems via the perturbation approach

In Progress: Summary of Notation and Basic Results Convex Analysis C&O 663, Fall 2009

8. Conjugate functions

EE 546, Univ of Washington, Spring Proximal mapping. introduction. review of conjugate functions. proximal mapping. Proximal mapping 6 1

Almost Convex Functions: Conjugacy and Duality

Convex Optimization Theory. Chapter 4 Exercises and Solutions: Extended Version

Convex Analysis and Optimization Chapter 2 Solutions

GEOMETRIC APPROACH TO CONVEX SUBDIFFERENTIAL CALCULUS October 10, Dedicated to Franco Giannessi and Diethard Pallaschke with great respect

Some Properties of the Augmented Lagrangian in Cone Constrained Optimization

Math 273a: Optimization Convex Conjugacy

BASICS OF CONVEX ANALYSIS

Characterizations of the solution set for non-essentially quasiconvex programming

Convex Optimization Theory. Athena Scientific, Supplementary Chapter 6 on Convex Optimization Algorithms

1 Convex Optimization Models: An Overview

Nonlinear Programming 3rd Edition. Theoretical Solutions Manual Chapter 6

Math 273a: Optimization Subgradients of convex functions

A New Fenchel Dual Problem in Vector Optimization

Local strong convexity and local Lipschitz continuity of the gradient of convex functions

Sequential Pareto Subdifferential Sum Rule And Sequential Effi ciency

Subgradient. Acknowledgement: this slides is based on Prof. Lieven Vandenberghes lecture notes. definition. subgradient calculus

Primal/Dual Decomposition Methods

18.657: Mathematics of Machine Learning

Dedicated to Michel Théra in honor of his 70th birthday

A Dual Condition for the Convex Subdifferential Sum Formula with Applications

On duality theory of conic linear problems

Sum of two maximal monotone operators in a general Banach space is maximal

On the use of semi-closed sets and functions in convex analysis

Lecture 3: Lagrangian duality and algorithms for the Lagrangian dual problem

A Geometric Framework for Nonconvex Optimization Duality using Augmented Lagrangian Functions

Math 273a: Optimization Subgradients of convex functions

The proximal mapping

Epiconvergence and ε-subgradients of Convex Functions

ON A CLASS OF NONSMOOTH COMPOSITE FUNCTIONS

Convex Optimization and Modeling

Convex Optimization Theory. Athena Scientific, Supplementary Chapter 6 on Convex Optimization Algorithms

The sum of two maximal monotone operator is of type FPV

Convex Analysis Background

Quasi-relative interior and optimization

Convex Analysis and Economic Theory AY Elementary properties of convex functions

Victoria Martín-Márquez

Asteroide Santana, Santanu S. Dey. December 4, School of Industrial and Systems Engineering, Georgia Institute of Technology

Convex Optimization M2

Convex Optimization & Lagrange Duality

Lecture: Duality of LP, SOCP and SDP

The local equicontinuity of a maximal monotone operator

Handout 2: Elements of Convex Analysis

On Gap Functions for Equilibrium Problems via Fenchel Duality

STABLE AND TOTAL FENCHEL DUALITY FOR CONVEX OPTIMIZATION PROBLEMS IN LOCALLY CONVEX SPACES

CHAPTER 2: CONVEX SETS AND CONCAVE FUNCTIONS. W. Erwin Diewert January 31, 2008.

Enhanced Fritz John Optimality Conditions and Sensitivity Analysis

SHORT COMMUNICATION. Communicated by Igor Konnov

Extreme Abridgment of Boyd and Vandenberghe s Convex Optimization

On the relations between different duals assigned to composed optimization problems

Continuity of convex functions in normed spaces

Convex analysis and profit/cost/support functions

ON THE RANGE OF THE SUM OF MONOTONE OPERATORS IN GENERAL BANACH SPACES

Lower semicontinuous and Convex Functions

ALGORITHMS FOR MINIMIZING DIFFERENCES OF CONVEX FUNCTIONS AND APPLICATIONS

Convex Optimization Conjugate, Subdifferential, Proximation

Analysis and Linear Algebra. Lectures 1-3 on the mathematical tools that will be used in C103

Optimality conditions and complementarity, Nash equilibria and games, engineering and economic application

Convexity, Duality, and Lagrange Multipliers

arxiv:math/ v1 [math.fa] 4 Feb 1993

On duality gap in linear conic problems

Convex Optimization Notes

Robust Duality in Parametric Convex Optimization

Convex Feasibility Problems

Subdifferential representation of convex functions: refinements and applications

Lecture 8: Basic convex analysis

Convex Optimization and an Introduction to Congestion Control. Lecture Notes. Fabian Wirth

1. Gradient method. gradient method, first-order methods. quadratic bounds on convex functions. analysis of gradient method

LP Duality: outline. Duality theory for Linear Programming. alternatives. optimization I Idea: polyhedra

Introduction to Convex and Quasiconvex Analysis

A set C R n is convex if and only if δ C is convex if and only if. f : R n R is strictly convex if and only if f is convex and the inequality (1.

Transcription:

LECTURE 12 LECTURE OUTLINE Subgradients Fenchel inequality Sensitivity in constrained optimization Subdifferential calculus Optimality conditions Reading: Section 5.4 All figures are courtesy of Athena Scientific, and are used with permission. 1

SUBGRADIENTS f(z) ( g, 1) ( x, f(x) z Let f : R n (, ] be a convex function. A vector g R n is a subgradient of f at a point x dom(f) if f(z) f(x)+(z x) g, apple z R n Support Hyperplane Interpretation: g is a subgradient if and only if f(z) z g f(x) x g, apple z R n so g is a subgradient at x if and only if the hyperplane in ( R n+1 that has normal ( g, 1) and passes through x, f(x) supports the epigraph of f. The set of all subgradients at x is the subdifferential of f at x, denoted f(x). By convention f(x) =Ø for x/dom(f). 2

EXAMPLES OF SUBDIFFERENTIALS Some examples: f(x) = x f(x) 1 x x f(x) =max {, (1/2)(x 2 1) -1 f(x) -1 1 x -1 1 x If f is differentiable, then f(x) ={ f(x)}. Proof: If g f(x), then f(x + z) f(x)+g z, apple z R n. ( Apply this with z = f(x) g, R, and use 1st order Taylor series expansion to obtain f(x) g 2 o( )/, apple < 3

EXISTENCE OF SUBGRADIENTS Let f : R n (, ] be proper convex. Consider MC/MC with M = epi(f x ), f x (z) =f(x + z) f(x) f(z) f x (z) Epigraph of f ( g, 1) Translated Epigraph of f ( g, 1) ( x, f(x) z z By 2nd MC/MC Duality Theorem, f(x) is nonempty and compact if and only if x is in the interior of dom(f). ( More generally: for every x ri dom(f)), where: f(x) =S + G, S is the subspace that is parallel to the a ne hull of dom(f) G is a nonempty and compact set. 4

EXAMPLE: SUBDIFFERENTIAL OF INDICATOR Let C be a convex set, and C be its indicator function. For x / C, C (x) =Ø (by convention). For x C, we have g C (x) iff C (z) C (x)+g (z x), apple z C, or equivalently g (z x) for all z C. Thus C (x) is the normal cone of C at x, denoted N C (x): N C (x) = g g (z x), apple z C. x C x C N C (x) N C (x) 5

EXAMPLE: POLYHEDRAL CASE C a 1 x N C (x) a 2 For the case of a polyhedral set we have C = {x a i x b i,i=1,...,m}, N C (x) = { {} if x int(c), cone ( {a i a i x = b i} if x / int(c). Proof: Given x, disregard inequalities with a i x<b i, and translate C to move x to, so it becomes a cone. The polar cone is N C (x). 6

FENCHEL INEQUALITY Let f : R n (, ] be proper convex and let f be its conjugate. Using the definition of conjugacy, we have Fenchel s inequality: x y f(x)+f (y), apple x R n,y R n. Conjugate Subgradient Theorem: The following two relations are equivalent for a pair of vectors (x, y): (i) x y = f(x)+f (y). (ii) y f(x). If f is closed, (i) and (ii) are equivalent to (iii) x f (y). f(x) f (y) Epigraph of f Epigraph of f ( y, 1) x ( x, 1) f (y) y 2 f(x) 7

MINIMA OF CONVEX FUNCTIONS Application: Let f be closed proper convex and let X be the set of minima of f over R n. Then: (a) X = f (). ( (b) X is nonempty if ri dom(f ). (c) X is nonempt ( y and compact int dom(f ). if and only if Proof: (a) We have x X iff f(x) f(x ) for all x R n. So x X iff f(x ) iff x f () where: 1st relation follows from the subgradient inequality 2nd relation follows from the conjugate subgradient theorem ( (b) f () is nonempty if ri dom(f ). (c) f () ( is nonempty and compact if int dom(f ). Q.E.D. if and only 8

SENSITIVITY INTERPRETATION Consider MC/MC for the case M = epi(p). Dual function is q(µ) = inf p(u)+µ u = p ( µ), u R m where p is the conjugate of p. Assume p is proper convex and strong duality holds, so p() = w = q = sup µ Rm p ( µ). Let Q be the set of dual optimal solutions, Q = µ p() + p ( µ )=. From Conjugate Subgradient Theorem, µ Q if and only if µ p(), i.e., Q = p(). If p is convex and differentiable at, p() is equal to the unique dual optimal solution µ. Constrained optimization example: p(u) = inf f(x), x X, g(x) u If p is convex and differentiable, p() µ j =, j =1,...,r. u j 9

EXAMPLE: SUBDIFF. OF SUPPORT FUNCTION Consider the support function X (y) of a set X. To calculate X (y) at some y, we introduce r(y) = X (y + y), y R n. We have X (y) = r() = arg min x We have r (x) = sup y R R n{y x r(y) }, or n r (x). r (x) = sup {y x X (y + y) } = (x) y x, y R n where is the indicator function of cl ( conv(x). Hence X (y) = arg min x n (x) y x, or R X (y) = arg ( max y x x cl conv(x) σ X (y 2 ) y 2 X σ X (y 1 ) y 1 1

EXAMPLE: SUBDIFF. OF POLYHEDRAL FN Let f(x) = max{ a 1 x + b 1,...,a rx + b r }. f(x) ( g, 1) r(x) ( g, 1) Epigraph of f x x x For a fixed x R n, consider A x = j a j x + b j = f(x) and the function r(x) = max a j x j A x. It can be seen that f(x) = r(). Since r is the support function of the finite set {a j j A x }, we see that ( f(x) = r() = conv {a j j A x } 11

CHAIN RULE Let f : R m (, ] be convex, and A be a matrix. Consider F (x) = f(ax) and assume that F is proper. If either f is polyhedral or else Range(A) ri(dom(f)) = Ø, then F (x) =A f(ax), apple x R n. Proof: Showing F (x) A f(ax) is simple and does not require the relative interior assumption. For the reverse inclusion, let d F (x) so F (z) F (x)+(z x) d or f(az) z d f(ax) x d for all z, so (Ax, x) solves minimize f(y) z d subject to y dom(f), Az = y. If R(A) ri(dom(f)) = Ø, by strong duality theorem, there is a dual optimal solution, such that (Ax, x) arg min f(y) z d + (Az y) y R m,z R n Since the min over z is unconstrained, we have d = A, so Ax arg min y R m f(y) y, or f(y) f(ax)+ (y Ax), apple y R m. Hence f(ax), so that d = A A f(ax). It follows that F (x) A f(ax). In the polyhedral case, dom(f) is polyhedral. Q.E.D. 12

SUM OF FUNCTIONS Let f i : R n (, ], i =1,...,m, be proper convex functions, and let F = f 1 + + f m. Assume that m 1=1 ri ( dom(f i) = Ø. Then F (x) = f 1 (x)+ + f m (x), apple x R n. Proof: We can write F in the form F (x) =f(ax), where A is the matrix defined by Ax =(x,...,x), and f : R mn (, ] is the function f(x 1,...,x m )=f 1 (x 1 )+ + f m (x m ). Use the proof of the chain rule. Extension: If for some k, the functions f i, i = 1,...,k, are polyhedral, it is su cient to assume ) k i=1 dom(f i) m i=k+1 ri( dom(f i ) ) = Ø. 13

CONSTRAINED OPTIMALITY CONDITION Let f : R n (, ] be proper convex, let X be a convex subset of R n, and assume that one of the following four conditions holds: (i) ri ( dom(f) ri(x) = Ø. (ii) f is polyhedral and dom( ) ri( ) = Ø ( f X. (iii) X is polyhedral and ri dom(f) X = Ø. (iv) f and X are polyhedral, and dom(f) X = Ø. Then, a vector x minimizes f over X iff there exists g f(x ) such that g belongs to the normal cone N X (x ), i.e., g (x x ), apple x X. Proof: x minimizes F (x) =f(x)+ X (x) if and only if F (x ). Use the formula for subdifferential of sum. Q.E.D. 14

ILLUSTRATION OF OPTIMALITY CONDITION Level Sets of f N C (x ) N C (x ) Level Sets of f C f(x ) x C f(x ) g x In the figure on the left, f is differentiable and the condition is that which is equivalent to f(x ) N C (x ), f(x ) (x x ), apple x X. In the figure on the right, f is nondifferentiable, and the condition is that g N C (x ) for some g f(x ). 15

MIT OpenCourseWare http://ocw.mit.edu 6.253 Convex Analysis and Optimization Spring 212 For information about citing these materials or our Terms of Use, visit: http://ocw.mit.edu/terms.