Lecture 3 Introduction to optimality conditions

Similar documents
Lecture 2 - Unconstrained Optimization Definition[Global Minimum and Maximum]Let f : S R be defined on a set S R n. Then

Second Order Optimality Conditions for Constrained Nonlinear Programming

TMA 4180 Optimeringsteori KARUSH-KUHN-TUCKER THEOREM

Zangwill s Global Convergence Theorem

MATH 5720: Unconstrained Optimization Hung Phan, UMass Lowell September 13, 2018

9 Sequences of Functions

Structural and Multidisciplinary Optimization. P. Duysinx and P. Tossings

Optimality conditions for unconstrained optimization. Outline

1 Convexity, concavity and quasi-concavity. (SB )

Unconstrained Optimization

Constrained Optimization and Lagrangian Duality

Chapter 2: Preliminaries and elements of convex analysis

ECE580 Exam 1 October 4, Please do not write on the back of the exam pages. Extra paper is available from the instructor.

DO NOT OPEN THIS QUESTION BOOKLET UNTIL YOU ARE TOLD TO DO SO

1 Convexity, Convex Relaxations, and Global Optimization

UNDERGROUND LECTURE NOTES 1: Optimality Conditions for Constrained Optimization Problems

Examination paper for TMA4180 Optimization I

1 Overview. 2 A Characterization of Convex Functions. 2.1 First-order Taylor approximation. AM 221: Advanced Optimization Spring 2016

SWFR ENG 4TE3 (6TE3) COMP SCI 4TE3 (6TE3) Continuous Optimization Algorithm. Convex Optimization. Computing and Software McMaster University

Lecture 7 Monotonicity. September 21, 2008

Lecture 5 : Projections

Lecture 12 Unconstrained Optimization (contd.) Constrained Optimization. October 15, 2008

MATHEMATICAL ECONOMICS: OPTIMIZATION. Contents

Mathematics 530. Practice Problems. n + 1 }

Optimization Theory. A Concise Introduction. Jiongmin Yong

Optimization Tutorial 1. Basic Gradient Descent

Optimality Conditions for Constrained Optimization

EC9A0: Pre-sessional Advanced Mathematics Course. Lecture Notes: Unconstrained Optimisation By Pablo F. Beker 1

Constrained Optimization Theory

Lecture 3. Optimization Problems and Iterative Algorithms

Assignment 1: From the Definition of Convexity to Helley Theorem

SECTION C: CONTINUOUS OPTIMISATION LECTURE 11: THE METHOD OF LAGRANGE MULTIPLIERS

Lec3p1, ORF363/COS323

ISM206 Lecture Optimization of Nonlinear Objective with Linear Constraints

CHAPTER 7. Connectedness

Existence of minimizers

Math (P)refresher Lecture 8: Unconstrained Optimization

Lecture 7: Semidefinite programming

Division of the Humanities and Social Sciences. Supergradients. KC Border Fall 2001 v ::15.45

ON GENERALIZED-CONVEX CONSTRAINED MULTI-OBJECTIVE OPTIMIZATION

2.098/6.255/ Optimization Methods Practice True/False Questions

Contents: 1. Minimization. 2. The theorem of Lions-Stampacchia for variational inequalities. 3. Γ -Convergence. 4. Duality mapping.

Subgradients. subgradients and quasigradients. subgradient calculus. optimality conditions via subgradients. directional derivatives

Optimization and Optimal Control in Banach Spaces

Econ Lecture 3. Outline. 1. Metric Spaces and Normed Spaces 2. Convergence of Sequences in Metric Spaces 3. Sequences in R and R n

c 1998 Society for Industrial and Applied Mathematics

The Dirichlet s P rinciple. In this lecture we discuss an alternative formulation of the Dirichlet problem for the Laplace equation:

Optimality Conditions for Nonsmooth Convex Optimization

CHAPTER 2: CONVEX SETS AND CONCAVE FUNCTIONS. W. Erwin Diewert January 31, 2008.

Chapter 2: Unconstrained Extrema

6.252 NONLINEAR PROGRAMMING LECTURE 10 ALTERNATIVES TO GRADIENT PROJECTION LECTURE OUTLINE. Three Alternatives/Remedies for Gradient Projection

Economics 204 Summer/Fall 2017 Lecture 7 Tuesday July 25, 2017

The general programming problem is the nonlinear programming problem where a given function is maximized subject to a set of inequality constraints.

Numerical Optimization

Lectures 9 and 10: Constrained optimization problems and their optimality conditions

Best approximations in normed vector spaces

Lecture 8 Plus properties, merit functions and gap functions. September 28, 2008

Lecture 1 Introduction to optimization

Approximate Farkas Lemmas in Convex Optimization

March 25, 2010 CHAPTER 2: LIMITS AND CONTINUITY OF FUNCTIONS IN EUCLIDEAN SPACE

Lecture 5: September 12

Introduction to Machine Learning Lecture 7. Mehryar Mohri Courant Institute and Google Research

3 Measurable Functions

Contents Real Vector Spaces Linear Equations and Linear Inequalities Polyhedra Linear Programs and the Simplex Method Lagrangian Duality

Algorithms for constrained local optimization

Inequality Constraints

LECTURE 25: REVIEW/EPILOGUE LECTURE OUTLINE

Subgradients. subgradients. strong and weak subgradient calculus. optimality conditions via subgradients. directional derivatives

OPTIMISATION 2007/8 EXAM PREPARATION GUIDELINES

Econ 508-A FINITE DIMENSIONAL OPTIMIZATION - NECESSARY CONDITIONS. Carmen Astorne-Figari Washington University in St. Louis.

Nonlinear Programming (NLP)

Some Properties of the Augmented Lagrangian in Cone Constrained Optimization

Introduction and Preliminaries

Convex Functions. Pontus Giselsson

Convex Optimization Lecture 6: KKT Conditions, and applications

Math 273a: Optimization Basic concepts

GENERALIZED CONVEXITY AND OPTIMALITY CONDITIONS IN SCALAR AND VECTOR OPTIMIZATION

Econ Lecture 7. Outline. 1. Connected Sets 2. Correspondences 3. Continuity for Correspondences

ON SPACE-FILLING CURVES AND THE HAHN-MAZURKIEWICZ THEOREM

Convex Optimization and Modeling

SECTION C: CONTINUOUS OPTIMISATION LECTURE 9: FIRST ORDER OPTIMALITY CONDITIONS FOR CONSTRAINED NONLINEAR PROGRAMMING

Nonlinear Programming and the Kuhn-Tucker Conditions

The Relation Between Pseudonormality and Quasiregularity in Constrained Optimization 1

Gradient Descent. Dr. Xiaowei Huang

Convex Analysis and Economic Theory AY Elementary properties of convex functions

CSCI5654 (Linear Programming, Fall 2013) Lectures Lectures 10,11 Slide# 1

IE 521 Convex Optimization

1 Strict local optimality in unconstrained optimization

Solutions to Tutorial 7 (Week 8)

1 Review of last lecture and introduction

Course Notes for EE227C (Spring 2018): Convex Optimization and Approximation

Constrained Optimization

Lecture 8. Strong Duality Results. September 22, 2008

ON LICQ AND THE UNIQUENESS OF LAGRANGE MULTIPLIERS

Lecture 2: Convex Sets and Functions

Preprint February 19, 2018 (1st version October 31, 2017) Pareto efficient solutions in multi-objective optimization involving forbidden regions 1

MATH 51H Section 4. October 16, Recall what it means for a function between metric spaces to be continuous:

x +3y 2t = 1 2x +y +z +t = 2 3x y +z t = 7 2x +6y +z +t = a

Maths 212: Homework Solutions

Local strong convexity and local Lipschitz continuity of the gradient of convex functions

Transcription:

TMA947 / MMG621 Nonlinear optimisation Lecture 3 Introduction to optimality conditions Emil Gustavsson October 31, 2013 Local and global optimality We consider an optimization problem which is that to minimize f(x), subject to x S, (1a) (1b) where S R n is a nonempty set, and f : R n R {+ } is a given function. When n = 1, we f(x) 1 2 3 4 5 6 7 S x know that an optimal solution will be found at boundary points of S, stationary points, that is, wheref (x) = 0, discontinuities of f or f. Definition (global minimum). x S is a global minimum of f overs if f(x ) f(x), x S The next important definition is of a local minimum. 1

Definition (local minimum). x S is a local minimum off overs if ε > 0 such thatf(x ) f(x), x S B ε (x ), whereb ε (x ) := {y R n y x < ε} is the Euclidean ball with radiusεcentered atx. x S is a strict local minimum of f over S if f(x ) < f(x) holds above for x x. One of the most important theorems in optimization is the following. Theorem (Fundamental Theorem of global optimality). Consider the problem (1), where S is a convex set andf is convex on S. Then, every local minimum of f overs is also a global minimum Proof. See Theorem 4.3 in the book. Existence of optimal solutions First some basic notations. We say that a set S R n is open if for every x S there exists an ε > 0 such that B ε (x) := {y R n y x < ε} S. We say that a set S R n is closed if R n \S is open. A limit point of a set S R n is a point x such that there exists a sequence {x k } k=1 S fulfilling x k x. We can then define a closed set as a set which contains all its limit points. Is a set is both closed and bounded, we call it compact. Two important definitions we need to formulate Weierstrass Theorem are the following. Definition (weakly coercive function). A function f is said to be weakly coercive with respect to the set S if eithers is bounded or lim f(x) = x x S Definition (lower semi-continuity). A function f is said to be lower semi-continuous at x if the value f(x) is less than or equal to every limit of f asx k x In other words,f is lower semi-continuous atx S if x k x = f(x) liminf k f(x k) Now we can formulate Weierstrass Theorem which guarantees the existence of optimal solutions to an optimization problem as long as a few assumptions are satisfied. 2

f x Figure 1: A lower semi-continuous function in one variable Theorem (Weierstrass Theorem)). Consider the problem (1), where S is a nonempty and closed set and f is lower semi-continuous on S. If f is weakly coercive with respect to S, then there exists a nonempty, closed and bounded (thus compact) set of optimal solutions to the problem (1). Proof. See Theorem 4.6 in the book. One way to remember what the assumptions in Weierstrass Theorem are is to imagine what can go wrong, i.e., when does a problem not have an optimal solution. One example of an optimization problem where the solution set is empty is whenf(x) = 1/x and S = R n. Optimality conditions whens = R n When S = R n, i.e., the problem is an unconstrained optimization problem, then the following theorem holds. Theorem (necessary condition for optimality, C 1 ). If f C 1 on R n, then x is a local minimum of f on R n = f(x ) = 0 Proof. See Theorem 4.13. ( ) Note that f(x) = f(x) T. x 1,..., f(x) x n The opposite of the theorem is, however, not true. Take f(x) = x 3 andx = 0. We can strengthen the theorem by assuming that f is also inc 2. Theorem (necessary condition for optimality, C 2 ). If f C 2 on R n, then x is a local minimum of f on R n = { f(x ) = 0 2 f(x) 0 Proof. See Theorem 4.16. Remember that for a matrix A R n n, the notion A 0 (A positive semidefinite) means that x T Ax 0, for all x R n. Now once again, the opposite direction in the theorem is not true. 3

However, by assuming positive definiteness of the Hessian off, we can obtain a sufficient condition. Theorem (sufficient condition for optimality, C 2 ). Iff C 2 on R n, then { f(x ) = 0 2 = x f(x) 0 is a strict local minimum of f on R n Proof. See Theorem 4.17. To get sufficient conditions for a point to be a local minimum, we need to assume convexity of the function f. Theorem (necessary and sufficient condition for optimality, C 1 ). If f C 1 is convex on R n, then x is a global minimum off on R n f(x ) = 0 Proof. See Theorem 4.18. Optimality conditions for S R n WhenS = R n, the directions we could move from a pointxand still stay feasible wherer n itself. When we consider cases wheres R n this might not hold. Definition (feasible direction). Let x S. A vectorp R n defines a feasible direction at x if δ > 0 : x+αp S, for all α [0,δ]. So the feasible directions at a point x S describes the directions in which we can move without becoming infeasible. Definition (descent direction). Let x R n. A vector p defines a descent direction with respect to f atxis δ > 0 : f(x+αp) < f(x), for all α (0,δ]. Suppose that f C 1 around a point x R n, and that p R n. If f(x) T p < 0 then the vector p defines a direction of descent with respect to f at x. We can now state necessary optimality conditions for cases when S R n. Theorem (necessary optimality conditions). Suppose thats R n and thatf C 1. a) If x S is a local minimum of f overs, then holds for all feasible directionspatx. f(x ) T p 0 b) Suppose thats is convex. If x is a local minimum of f overs, then f(x ) T (x x ) 0, x S (2) 4

Proof. See Proposition 4.22 in the book. We refer to (2) as a variational inequality and we can now extend the notion of stationary points by denoting them as points fulfilling (2). This our first out of four definition of a stationary point. Now the necessary and sufficient conditions for optimality can be stated in the following theorem. Theorem (necessary and sufficient optimality conditions). Suppose S R n is a convex set and that f C 1 is a convex function. Then x is a global minimum of f overs f(x ) T (x x ) 0, x S. Proof. See Theorem 4.23 in the book. Note that whens = R n, the expression to the right just becomes f(x ) = 0. Why? We will now present three additional definitions of a stationary point which are all equivalent to (2). The first one we get by taking the minimum of the left-hand-side of (2) and then realizing that the optimal value must be zero, i.e., min x S f(x ) T (x x ) = 0. (3) Convince yourself that (2) and (3) are equivalent! Now we claim that (2) and (3) are also equivalent with x = Proj S [x f(x )]. (4) The equation (4) states that if you stand in a stationary point and take a step in the direction of the negative gradient and then project back to the feasible set, you should end up in the same point. The details for showing this equivalence can be found in the book (pp. 100 101). See also Figure 3. For the last equivalent definition of a stationary point, we need to introduce the normal cone Definition (normal cone). Suppose the set S is closed and convex. Let x S. Then the normal cone to S atxis the set N S (x) := { p R n p T (y x) 0, y S }. Think of the normal cone at a point x as all direction pointing straight out from the set. See Figure 3. Now the fourth equivalent definition of a stationary point is that That (5) is equivalent to (2) is trivial. f(x ) N S (x ). (5) 5

N S (x ) x f(x ) x 01 01 01 y X Figure 2: Illustration of the normal cone and the projection operator. Summary of optimality condition for convex S R n Definition (stationary point). Suppose that S is convex and that f C 1. A point x S fulfilling the four equivalent statements a) b) are called a stationary point. a) b) c) d) f(x ) T (x x ) 0, x S, min x S f(x ) T (x x ) = 0, x = Proj S [x f(x )], f(x ) N S (x ). The two important theorems which will be utilized throughout the whole course are the following. Theorem (necessary optimality conditions). Suppose thats is convex and thatf C 1. Then x is a local minimum off overs = x is stationary Theorem (necessary and sufficient optimality conditions). Suppose thats is convex and thatf C 1 is convex. Then x is a global minimum of f overs x is stationary As we will see later in the course, the last definition (the inclusion (5)) is the only one that can be extended to the case of non-convex sets S. 6

The separation theorem Now we have the tools to prove the separation theorem. Theorem (the separation theorem). Suppose that S R n is closed and convex, and that the point y does not lie in S. Then there exists a vectorπ 0 and a scalarα R such thatπ T y > α, and π T x α for all x S. Proof. See Theorem 4.28. x 2 2 y = (1.5,1.5) T π T x = α x 1 +x 2 = 2 P 2 x 1 Figure 3: Illustration of the separation theorem. 7