Nonlinear Optimization

Similar documents
Chap 2. Optimality conditions

4TE3/6TE3. Algorithms for. Continuous Optimization

Nonlinear Programming

Finite Dimensional Optimization Part III: Convex Optimization 1

Numerical Optimization

Generalization to inequality constrained problem. Maximize

Constrained Optimization

The Karush-Kuhn-Tucker (KKT) conditions

Optimization Problems with Constraints - introduction to theory, numerical Methods and applications

More on Lagrange multipliers

Convex Optimization & Lagrange Duality

I.3. LMI DUALITY. Didier HENRION EECI Graduate School on Control Supélec - Spring 2010

Extreme Abridgment of Boyd and Vandenberghe s Convex Optimization

Introduction to Machine Learning Lecture 7. Mehryar Mohri Courant Institute and Google Research

UNDERGROUND LECTURE NOTES 1: Optimality Conditions for Constrained Optimization Problems

Introduction to Optimization Techniques. Nonlinear Optimization in Function Spaces

Lecture 18: Optimization Programming

Optimality, Duality, Complementarity for Constrained Optimization

ON LICQ AND THE UNIQUENESS OF LAGRANGE MULTIPLIERS

Lecture: Duality of LP, SOCP and SDP

Constrained Optimization Theory

Constrained Optimization and Lagrangian Duality

Convex Optimization Lecture 6: KKT Conditions, and applications

Introduction to Machine Learning Prof. Sudeshna Sarkar Department of Computer Science and Engineering Indian Institute of Technology, Kharagpur

Optimality Conditions for Constrained Optimization

5. Duality. Lagrangian

The Karush-Kuhn-Tucker conditions

2.098/6.255/ Optimization Methods Practice True/False Questions

CO 250 Final Exam Guide

Convex Optimization M2

LECTURE 10 LECTURE OUTLINE

Constrained optimization

Shiqian Ma, MAT-258A: Numerical Optimization 1. Chapter 4. Subgradient

Optimization. A first course on mathematics for economists

Lecture 6: Conic Optimization September 8

Convex Optimization and Modeling

Nonlinear Programming and the Kuhn-Tucker Conditions

CONSTRAINED NONLINEAR PROGRAMMING

Support Vector Machines: Maximum Margin Classifiers

Lectures 9 and 10: Constrained optimization problems and their optimality conditions

Lecture 2: Linear SVM in the Dual

Chapter 3: Constrained Extrema

Title Problems. Citation 経営と経済, 65(2-3), pp ; Issue Date Right

Topic one: Production line profit maximization subject to a production rate constraint. c 2010 Chuan Shi Topic one: Line optimization : 22/79

Lecture 3. Optimization Problems and Iterative Algorithms

Optimization. Yuh-Jye Lee. March 28, Data Science and Machine Intelligence Lab National Chiao Tung University 1 / 40

Constrained maxima and Lagrangean saddlepoints

Duality. Lagrange dual problem weak and strong duality optimality conditions perturbation and sensitivity analysis generalized inequalities

Convex Optimization Boyd & Vandenberghe. 5. Duality

Course Notes for EE227C (Spring 2018): Convex Optimization and Approximation

Seminars on Mathematics for Economics and Finance Topic 5: Optimization Kuhn-Tucker conditions for problems with inequality constraints 1

Additional Homework Problems

TMA 4180 Optimeringsteori KARUSH-KUHN-TUCKER THEOREM

Solving generalized semi-infinite programs by reduction to simpler problems.

Optimisation in Higher Dimensions

Duality Theory of Constrained Optimization

Karush-Kuhn-Tucker Conditions. Lecturer: Ryan Tibshirani Convex Optimization /36-725

Linear programming: Theory

In view of (31), the second of these is equal to the identity I on E m, while this, in view of (30), implies that the first can be written

Enhanced Fritz John Optimality Conditions and Sensitivity Analysis

Bindel, Spring 2017 Numerical Analysis (CS 4220) Notes for So far, we have considered unconstrained optimization problems.

Inequality Constraints

Math 5311 Constrained Optimization Notes

Convex Optimization Theory. Chapter 5 Exercises and Solutions: Extended Version

SECTION C: CONTINUOUS OPTIMISATION LECTURE 9: FIRST ORDER OPTIMALITY CONDITIONS FOR CONSTRAINED NONLINEAR PROGRAMMING

CONSTRAINT QUALIFICATIONS, LAGRANGIAN DUALITY & SADDLE POINT OPTIMALITY CONDITIONS

5 Handling Constraints

Mathematical Economics. Lecture Notes (in extracts)

Date: July 5, Contents

Lecture: Duality.

Lagrange Relaxation and Duality

IE 5531: Engineering Optimization I

Support vector machines

Motivation. Lecture 2 Topics from Optimization and Duality. network utility maximization (NUM) problem:

IE 521 Convex Optimization Homework #1 Solution

EE/AA 578, Univ of Washington, Fall Duality

Applied Lagrange Duality for Constrained Optimization

UNCLASSIFIED AD NUMBER LIMITATION CHANGES

A SHIFTED PRIMAL-DUAL PENALTY-BARRIER METHOD FOR NONLINEAR OPTIMIZATION

The general programming problem is the nonlinear programming problem where a given function is maximized subject to a set of inequality constraints.

A DUALITY THEOREM FOR NON-LINEAR PROGRAMMING* PHILIP WOLFE. The RAND Corporation

Quiz Discussion. IE417: Nonlinear Programming: Lecture 12. Motivation. Why do we care? Jeff Linderoth. 16th March 2006

Mathematical Economics: Lecture 16

Miscellaneous Nonlinear Programming Exercises

Assignment 1: From the Definition of Convexity to Helley Theorem

8 Barrier Methods for Constrained Optimization

CSCI : Optimization and Control of Networks. Review on Convex Optimization

A Brief Review on Convex Optimization

Lecture 3: Lagrangian duality and algorithms for the Lagrangian dual problem

Subgradient. Acknowledgement: this slides is based on Prof. Lieven Vandenberghes lecture notes. definition. subgradient calculus

2.3 Linear Programming

Outline. Roadmap for the NPP segment: 1 Preliminaries: role of convexity. 2 Existence of a solution

Optimization for Communications and Networks. Poompat Saengudomlert. Session 4 Duality and Lagrange Multipliers

Recita,on: Loss, Regulariza,on, and Dual*

AM 205: lecture 19. Last time: Conditions for optimality, Newton s method for optimization Today: survey of optimization methods

OPTIMALITY AND STABILITY OF SYMMETRIC EVOLUTIONARY GAMES WITH APPLICATIONS IN GENETIC SELECTION. (Communicated by Yang Kuang)

Lecture 7: Convex Optimizations

Differentiable exact penalty functions for nonlinear optimization with easy constraints. Takuma NISHIMURA

Convex Optimization Overview (cnt d)

Concave programming. Concave programming is another special case of the general constrained optimization. subject to g(x) 0

Transcription:

Nonlinear Optimization Etienne de Klerk (UvT)/Kees Roos e-mail: C.Roos@ewi.tudelft.nl URL: http://www.isa.ewi.tudelft.nl/ roos Course WI3031 (Week 4) February-March, A.D. 2005 Optimization Group 1

Outline for today NECESSARY optimality conditions for convex constrained optimization; The Farkas Lemma; The Lagrange function; Saddle points of the Lagrange function; The Karusch Kuhn Tucker (KKT) optimality conditions revisited; Examples. In the course notes: 2.2.3 till the end of Chapter 2. Optimization Group 2

The generic constrained convex problem (CO) (CO) min f(x) s.t. g j (x) 0, j = 1,, m x C, where C R n is a convex set; f, g 1,, g m are convex functions on C (or on an open set that contains the set C); The set of feasible solutions will be denoted by F, hence F = {x C g j (x) 0, j = 1,, m}. Optimization Group 3

Slater points Recall from last week: Definition: A vector (point) x 0 C 0 is called a Slater point of (CO) if g j (x 0 ) < 0, for all j where g j is nonlinear, g j (x 0 ) 0, for all j where g j is linear. If a Slater point exists we say that (CO) is Slater regular, or (CO) satisfies the Slater condition, or (CO) satisfies the Slater constraint qualification. Optimization Group 4

Slater points: singular constraints Some constraint functions g j (x) might take the value zero for all feasible points. Such constraints are called singular while the others are called regular. We define the index sets J r for the singular and J r for the singular sonstraints: J s = {j J g j (x) = 0 for all x F}, J r = J \ J s = {j J g j (x) < 0 for some x F}. Remark: Note, that if (CO) is Slater regular, then all singular constraints must be linear. Optimization Group 5

Ideal Slater points Definition: A point x C 0 is called an Ideal Slater point of (CO) if g j (x ) < 0 for all j J r, Lemma: If (CO) is Slater regular then there exists an ideal Slater point x F. NB: An ideal Slater point is in the relative interior of F. Optimization Group 6

Example Consider the optimization problem x 2 min f(x) s.t. x 2 1 + x2 2 4 x 1 x 2 2 x 2 1 C = R 2. F x 1 The point (1, 1) is a Slater point, but not an ideal Slater point. The point ( 3 2, 3 4 ) is an ideal Slater point. Optimization Group 7

Convex Farkas Lemma The number a is a lower bound for the optimal value of (CO) if and only if the inequality system has no solution. f(x) < a g j (x) 0, x C j = 1,, m (1) Lemma 2.22 (Farkas) If the inequality system (1) satisfies the Slater condition then it has no solution if and only if there exists a vector y = (y 1,, y m ) 0 such that f(x) + y j g j (x) a for all x C. (2) The systems (1) and (2) are called alternative systems, because exactly one of them has a solution. Optimization Group 8

Example: application of the Farkas lemma Let us consider the convex optimization problem (CO) min 1 + x s.t. x 2 1 0 x R. Then (CO) is Slater regular (why?) The optimal value is 0. Hence, the system 1 + x < 0 x 2 1 0 x R has no no solution. By the Farkas lemma there must exist a real number y 0 such that 1 + x + y ( x 2 1 ) 0, x R. Indeed, taking y = 1 2 we get g(x) = x + 1 + y(x 2 1) = 1 2 x2 + x + 1 2 = 1 2 (x + 1)2 0. Optimization Group 9

Exercise 2.9 Let A R m n and b R m. Exactly one of the following alternative systems (I) or (II) is solvable: or In this case, f(x) = b T x, (I) A T x 0, x 0, b T x < 0, (II) Ay b, y 0. g j (x) = (A T x) j = ( a j) T x, j = 1,, m where a j denotes column j of A, and C is the positive orthant (nonnegative vectors) of R m. Optimization Group 10

Langrangean The Lagrange function (or Langrangean) of (CO) is defined by: L(x, y) := f(x) + y j g j (x) where x C and y 0. Note that the Lagrangean is convex in x and linear in y. If x F then g j (x) 0 for each j. Since y 0 we then have m y j g j (x) 0. So L(x, y) f(x), x F, y 0. Equality holds if and only if m y j g j (x) = 0 which is equivalent to y j g j (x) = 0, j = 1,..., m. Another consequence is that for x F we have sup L(x, y) f(x). y 0 Optimization Group 11

Saddle points Definition A vector pair (x, y ), with x C and y 0 is called a saddle point of the Lagrange function L(x, y) if L(x, y) L(x, y ) L(x, y ), x C, y 0. Due to the definition of the Lagrangean this means f(x ) + y j g j (x ) f(x ) + y jg j (x ) f(x) + y jg j (x), x C, y 0, or, equivalently m y j g j (x ) m y j g j(x ), y 0 f(x ) + m y j g j(x ) f(x) + m y j g j(x), x C. N.B. Skip Lemma 2.26 in the book. It is false!! Optimization Group 12

Karush-Kuhn-Tucker Theorem m y jg j (x ) m y j g j(x ), y 0 f(x ) + m y j g j(x ) f(x) + m y j g j(x), x C. Theorem 2.27 The problem (CO) is given. Assume that the Slater regularity condition is satisfied. The vector x is an optimal solution of (CO) if and only if there is a vector y such that (x, y ) is a saddle point of the Lagrange function L. As we will see, the if part does not require regularity or convexity: a saddle point of the Lagrangian always corresponds to an optimal solution! The proof of the only if part will make clear that we need convexity and Slater regularity to ensure that the Lagrangian has a saddle point if (CO) has an optimal solution. A corollary of the proof of the if part is that for any saddle point (x, y ) we have the so called complementarity property: y j g j(x ) = 0, j = 1,..., m. Optimization Group 13

Karush-Kuhn-Tucker Theorem: if proof m y jg j (x ) m y j g j(x ), y 0 f(x ) + m y j g j(x ) f(x) + m y j g j(x), x C. Let (x, y ) be a saddle point. Then x C, y 0, and L(x, y) L(x, y ) L(x, y ) for all x C and all y 0. The first inequality gives y j g j (x ) y j g j(x ), y 0. The right hand side is fixed. If g j (x ) > 0 for some j, we can let the right hand side go to infinity, by taking y i = 0 for j i and y j. This contradiction makes clear that g j (x ) 0. Thus it follows that x F. Since y 0 and g j (x ) 0, we see that the rhs is 0. By taking y = 0 the lhs equals 0. Hence we must have y j g j(x ) = 0. The second inequality now gives that x is optimal since for x F C one has f(x ) f(x) + y j g j(x) f(x). Optimization Group 14

Karush-Kuhn-Tucker Theorem: only if proof Let us take an optimal solution x of (CO). Then the inequality system f(x) < f(x ), g j (x) 0 (j = 1,..., m), x C, is infeasible. By the Convex Farkas Lemma there must exist a y 0 such that f(x) + y j g j(x) f(x ), x C. Taking x = x we obtain m y j g j(x ) 0. The converse also holds, hence y j g j(x ) = 0. Consequently, f(x ) = L(x, y ). Now we have, for all x C and y 0, L(x, y) = f(x ) + y j g j (x ) f(x ) f(x) + proving that (x, y ) is a saddle point of L(x, y). y j g j(x) = L(x, y ). Optimization Group 15

Karush-Kuhn-Tucker (KKT) points Definition Let us assume that C = R n and the functions f, g 1,, g m are continuously differentiable functions. The vector (x, y ) R n+m is called a Karush Kuhn Tucker (KKT) point of (CO) if (i) g j (x ) 0, for all j = 1,..., m, (ii) 0 = f(x ) + y j g j(x ) (iii) y j g j(x ) = 0, for all j = 1,..., m (iv) y 0. NB: Saddle point of L KKT point of (CO) if the Slater condition holds. Optimization Group 16

Karush-Kuhn-Tucker (KKT) points (ctd.) Via Exercise 2.12 and 2.13 one can prove the following corollary of the Karush-Kuhn- Tucker theorem. Corollary 2.30 Let us assume that C = R n and the functions f, g 1,, g m are continuously differentiable convex functions and Slater regularity holds. There exists a KKT point (x, y ) if and only if x is an optimal solution of (CO). Optimization Group 17

Discussion Last week we proved that existence of a KKT point is a sufficient condition for optimality. Today we proved it is also a necessary condition if the Slater condition holds. Optimization Group 18

Exercise 3.3, parts i, ii, v instead of Exercise 2.15 We want to design a box with dimensions l b h such that the volume of the box is at least V, and the total surface area is minimal. min l,b,h Non-convex problem! Replace l by e x 1, etc: 2(lb + bh + lh), lbh V, l, b, h > 0. such that min 2(e x 1+x 2 + e x 2+x 3 + e x 1+x 3 ), x 1,x 2,x 3 x 1 + x 2 + x 3 ln(v ), x 1, x 2, x 3 R. Let f(x 1, x 2, x 3 ) = 2(e x 1+x 2 + e x 2+x 3 + e x 1+x 3 ) and g(x 1, x 2, x 3 ) = ln V (x 1 + x 2 + x 3 ). The transformed problem is of the form (CO) and satisfies Slater s regularity condition (why?). Optimization Group 19

Example (ctd.) KKT conditions: we are looking for some x F such that f(x) = y g(x) for some y 0, such that yg(x) = 0. 2 e x 1+x 2 + e x 1+x 3 e x 1+x 2 + e x 2+x 3 e x 1+x 3 + e x 2+x 3 = y 1 1 1 Solution to this system (KKT point): x 1 = x 2 = x 3 = 1 3 ln(v ), y = 4V 2/3 Original variables: l = e x 1 = V 1/3, b = e x 2 = V 1/3 and h = e x 3 = V 1/3. The optimal solution is the cube (l = b = h)! Optimization Group 20