Numerical Optimization

Similar documents
In view of (31), the second of these is equal to the identity I on E m, while this, in view of (30), implies that the first can be written

Quiz Discussion. IE417: Nonlinear Programming: Lecture 12. Motivation. Why do we care? Jeff Linderoth. 16th March 2006

Constrained Optimization Theory

ON LICQ AND THE UNIQUENESS OF LAGRANGE MULTIPLIERS

Optimality Conditions for Constrained Optimization

Introduction to Optimization Techniques. Nonlinear Optimization in Function Spaces

minimize x subject to (x 2)(x 4) u,

Second Order Optimality Conditions for Constrained Nonlinear Programming

Numerical Optimization

Constrained optimization

Numerical Optimization

Constrained Optimization and Lagrangian Duality

Lectures 9 and 10: Constrained optimization problems and their optimality conditions

Convex Optimization & Lagrange Duality

Chap 2. Optimality conditions

Constrained Optimization

CONSTRAINED NONLINEAR PROGRAMMING

Introduction to Machine Learning Lecture 7. Mehryar Mohri Courant Institute and Google Research

5. Duality. Lagrangian

Constraint qualifications for nonlinear programming

Nonlinear Optimization

Generalization to inequality constrained problem. Maximize

SECTION C: CONTINUOUS OPTIMISATION LECTURE 9: FIRST ORDER OPTIMALITY CONDITIONS FOR CONSTRAINED NONLINEAR PROGRAMMING

The Karush-Kuhn-Tucker (KKT) conditions

Sharpening the Karush-John optimality conditions

UNDERGROUND LECTURE NOTES 1: Optimality Conditions for Constrained Optimization Problems

Convex Optimization M2

ISM206 Lecture Optimization of Nonlinear Objective with Linear Constraints

I.3. LMI DUALITY. Didier HENRION EECI Graduate School on Control Supélec - Spring 2010

Lecture 18: Optimization Programming

SECTION C: CONTINUOUS OPTIMISATION LECTURE 11: THE METHOD OF LAGRANGE MULTIPLIERS

Convex Optimization Boyd & Vandenberghe. 5. Duality

Lecture 3. Optimization Problems and Iterative Algorithms

Solving generalized semi-infinite programs by reduction to simpler problems.

Extreme Abridgment of Boyd and Vandenberghe s Convex Optimization

MATH2070 Optimisation

Optimization Problems with Constraints - introduction to theory, numerical Methods and applications

Numerical Optimization

IE 5531: Engineering Optimization I

Constrained optimization: direct methods (cont.)

5 Handling Constraints

A FRITZ JOHN APPROACH TO FIRST ORDER OPTIMALITY CONDITIONS FOR MATHEMATICAL PROGRAMS WITH EQUILIBRIUM CONSTRAINTS

Lecture: Duality.

Duality. Lagrange dual problem weak and strong duality optimality conditions perturbation and sensitivity analysis generalized inequalities

Nonlinear Programming and the Kuhn-Tucker Conditions

Lecture 1: Introduction. Outline. B9824 Foundations of Optimization. Fall Administrative matters. 2. Introduction. 3. Existence of optima

CONVERGENCE ANALYSIS OF AN INTERIOR-POINT METHOD FOR NONCONVEX NONLINEAR PROGRAMMING

Lagrange Duality. Daniel P. Palomar. Hong Kong University of Science and Technology (HKUST)

Optimality, Duality, Complementarity for Constrained Optimization

Lecture 1: Introduction. Outline. B9824 Foundations of Optimization. Fall Administrative matters. 2. Introduction. 3. Existence of optima

Computational Optimization. Constrained Optimization Part 2

4TE3/6TE3. Algorithms for. Continuous Optimization

Date: July 5, Contents

Enhanced Fritz John Optimality Conditions and Sensitivity Analysis

Convex Optimization and Modeling

Lecture: Duality of LP, SOCP and SDP

We are now going to move on to a discussion of Inequality constraints. Our canonical problem now looks as ( ) = 0 ( ) 0

Lagrange duality. The Lagrangian. We consider an optimization program of the form

INTERIOR-POINT METHODS FOR NONCONVEX NONLINEAR PROGRAMMING: CONVERGENCE ANALYSIS AND COMPUTATIONAL PERFORMANCE

Lectures on Parametric Optimization: An Introduction

AN ABADIE-TYPE CONSTRAINT QUALIFICATION FOR MATHEMATICAL PROGRAMS WITH EQUILIBRIUM CONSTRAINTS. Michael L. Flegel and Christian Kanzow

Math 273a: Optimization Subgradients of convex functions

Mathematical Economics. Lecture Notes (in extracts)

Structural and Multidisciplinary Optimization. P. Duysinx and P. Tossings

Nonlinear Programming (Hillier, Lieberman Chapter 13) CHEM-E7155 Production Planning and Control

Inequality Constraints

Shiqian Ma, MAT-258A: Numerical Optimization 1. Chapter 4. Subgradient

Optimization. A first course on mathematics for economists

Numerical Optimization

Solution Methods. Richard Lusby. Department of Management Engineering Technical University of Denmark

Numerical optimization

CO 250 Final Exam Guide

TMA 4180 Optimeringsteori KARUSH-KUHN-TUCKER THEOREM

Chapter 3: Constrained Extrema

CONSTRAINT QUALIFICATIONS, LAGRANGIAN DUALITY & SADDLE POINT OPTIMALITY CONDITIONS

Numerical optimization. Numerical optimization. Longest Shortest where Maximal Minimal. Fastest. Largest. Optimization problems

CSCI : Optimization and Control of Networks. Review on Convex Optimization

The Relation Between Pseudonormality and Quasiregularity in Constrained Optimization 1

Characterizations of Solution Sets of Fréchet Differentiable Problems with Quasiconvex Objective Function

Examination paper for TMA4180 Optimization I

Optimality conditions for problems over symmetric cones and a simple augmented Lagrangian method

A Brief Review on Convex Optimization

The Karush-Kuhn-Tucker conditions

CE 191: Civil & Environmental Engineering Systems Analysis. LEC 17 : Final Review

Optimization. Escuela de Ingeniería Informática de Oviedo. (Dpto. de Matemáticas-UniOvi) Numerical Computation Optimization 1 / 30

ICS-E4030 Kernel Methods in Machine Learning

EE/AA 578, Univ of Washington, Fall Duality

A smoothing augmented Lagrangian method for solving simple bilevel programs

Support Vector Machines: Maximum Margin Classifiers

Nonlinear Programming, Elastic Mode, SQP, MPEC, MPCC, complementarity

CS-E4830 Kernel Methods in Machine Learning

Karush-Kuhn-Tucker Conditions. Lecturer: Ryan Tibshirani Convex Optimization /36-725

Subgradient. Acknowledgement: this slides is based on Prof. Lieven Vandenberghes lecture notes. definition. subgradient calculus

Tutorial on Convex Optimization: Part II

Computational Finance

Optimization for Machine Learning

4TE3/6TE3. Algorithms for. Continuous Optimization

Convex Optimization. Dani Yogatama. School of Computer Science, Carnegie Mellon University, Pittsburgh, PA, USA. February 12, 2014

University of California, Davis Department of Agricultural and Resource Economics ARE 252 Lecture Notes 2 Quirino Paris

The Kuhn-Tucker Problem

Transcription:

Constrained Optimization Computer Science and Automation Indian Institute of Science Bangalore 560 012, India. NPTEL Course on

Constrained Optimization Constrained Optimization Problem: min h j (x) 0, j = 1,..., l e i (x) = 0, i = 1,..., m x S Inequality constraint functions: h j : R n R Equality constraint functions: e i : R n R Assume all functions (f, h j s and e i s) are sufficiently smooth Feasible set: X = {x S : h j (x) 0, e i (x) = 0, j = 1,..., l, i = 1,..., m} Given problem: Minimize subject to x X Assume X to be nonempty set in R n

Local and Global Minimum Definition A point x X is said to be a global minimum point of f over X if f (x ) for all x X. If > f (x ) for all x X, x x, then x is said to be a strict global minimum point of f over X. Definition A point x X is said to be a local minimum point of f over Xif there exists ɛ > 0 such that f (x ) for all x X B(x, ɛ). x X is said to be a strict local minimum point of f over Xif there exists ɛ > 0 such that > f (x ) for all x X B(x, ɛ), x x.

Convex Programming Problem min h j (x) 0, j = 1,..., l e i (x) = 0, i = 1,..., m x S is a convex function e i (x) is affine (e i (x) = a T i x + b i, i = 1,..., m) h j (x) is a convex function for j = 1,..., l S is a convex set Any local minimum is a global minimum The set of global minima form a convex set

Consider the problem: min x X Different ways of solving this problem: Reformulation to an unconstrained problem needs to be done with care Solve the constrained problem directly

min x X An iterative optimization algorithm generates a sequence {x k } k 0, which converges to a local minimum. Constrained Minimization Algorithm (1) Initialize x 0 X, k := 0. (2) while stopping condition is not satisfied at x k (a) Find x k+1 X such that f (x k+1 ) < f (x k ). (b) k := k + 1 endwhile Output : x = x k, a local minimum of over the set X.

min x X Strict Local Minimum: There exists ɛ > 0 such that f (x ) < x X B(x, ɛ), x x At a local minimum of a constrained minimization problem: the function does not decrease locally by moving along directions which contain feasible points How to convert this statement to an algebraic condition?

min x X Definition A vector d R n, d 0 is said to be a feasible direction at x X if there exists δ 1 > 0 such that x + αd X for all α (0, δ 1 ). Let F(x) = Set of feasible directions at x X (w.r.t. X) Definition A vector d R n, d 0 is said to be a descent direction at x X if there exists δ 2 > 0 such that f (x + αd) < for all α (0, δ 2 ). Let D(x) = Set of descent directions at x X (w.r.t. f )

min h j (x) 0, j = 1,..., l e i (x) = 0, i = 1,..., m x R n X = {x R n : h j (x) 0, e i (x) = 0, j = 1,..., l, i = 1,..., m} At a local minimum x X, the function does not decrease by moving along feasible directions

min x X Theorem Let X be a nonempty set in R n and x X be a local minimum of f over X. Then, F(x ) D(x ) = φ. Proof. Let x X be a local minimum. By contradiction, assume that a nonzero d F(x ) D(x ). δ 1 > 0 x + αd X α (0, δ 1 ) and δ 2 > 0 f (x + αd) < f (x ) α (0, δ 2 ). Hence, x B(x, α) X < f (x ), for every α (0, min(δ 1, δ 2 )). This contradicts the assumption that x is a local minimum.

min x X x X is a local minimum F(x ) D(x ) = φ Consider any x X and assume f C 2 lim α 0 + f (x+αd) α = T d T d < 0 f (x + αd) < d is a descent direction d D(x) Let D(x) = {d : T d < 0} D(x) x X is a local minimum F(x ) D(x ) = φ If F(x ) = R n (every direction in R n is locally feasible), x X is a local minimum {d : f (x ) T d < 0} = φ f (x ) = 0 Can we characterize F(x ) algebraically for a constrained optimization problem?

Consider the problem: min h j (x) 0, j = 1,..., l x R n Assume f, h j C 2, j = 1,..., l X = {x R n : h j (x) 0, j = 1,..., l} Active constraints: A(x) = {j : h j (x) = 0} Lemma For any x X, F(x) = {d : h j (x) T d < 0, j A(x)} F(x)

Lemma For any x X, Proof. F(x) = {d : h j (x) T d < 0, j A(x)} F(x) Suppose F(x) is nonempty and let d F(x). Since h j (x) T d < 0 j A(x), d is a descent direction for h j, j A(x) at x. That is, δ 1 > 0 h j (x + αd) < h j (x) = 0 j A(x). Further, h j (x) < 0 j / A(x). Therefore, δ 3 > 0 h j (x + αd) < 0 α (0, δ 3 ), j / A(x) Thus, x + αd X α (0, min(δ 1, δ 3 )), and d F(x).

min h j (x) 0, j = 1,..., l x R n Let X = {x R n : h j (x) 0, j = 1,..., l}. For any x X, F(x) = {d : h j (x) T d < 0, j A(x)} F(x) and D(x) = {d : T d < 0} D(x). x X is a local minimum F(x ) D(x ) = φ F(x ) D(x ) = φ x X is a local minimum F(x ) D(x ) = φ This is only a necessary condition for a local minimum Utility of this condition depends on the constraint representation Cannot be directly used for equality constrained problems

min h j (x) 0, j = 1,..., l x R n Let X = {x R n : h j (x) 0, j = 1,..., l} x X is a local minimum F(x ) D(x ) = φ {d : h j (x ) T d < 0, j A(x )} {d : f (x ) T d < 0} = φ f (x ) T Let A = h j (x ) T, j A(x ) (1+ A(x ) ) n x X is a local minimum {d : Ad < 0} = φ

Farkas Lemma Let A R m n and c R n. Then, exactly one of the following two systems has a solution: (I) Ax 0, c T x > 0 for some x R n (II) A T y = c, y 0 for some y R m. Corollary Let A R m n. Then exactly one of the following systems has a solution: (I) Ax < 0 for some x R n (II) A T y = 0, y 0 for some nonzero y R m. x X is a local minimum {d : Ad < 0} = φ λ 0 0 and λ j 0, j A(x ) (not all λ s 0), such that λ 0 f (x ) + j A(x ) λ j h j (x ) = 0.

x X is a local minimum {d : Ad < 0} = φ λ 0 0 and λ j 0, j A(x ) (not all λ s 0), such that λ 0 f (x ) + j A(x ) λ j h j (x ) = 0. Easy to satisfy these conditions if h j (x ) = 0 for some j A(x ) or f (x ) = 0 Regular point: A point x X is said to be a regular point if the gradient vectors, h j (x ), j A(x ), are linearly independent. x X is a regular point λ 0 0

Letting λ j = 0 j / A(x ), we get the following conditions: λ 0 f (x ) + l λ j h j (x ) = 0 j=1 λ j h j (x ) = 0 j = 1,..., l λ j 0 j = 0,..., l (λ 0, λ) (0, 0) where λ T = (λ 1,..., λ l ).

Consider the problem: min h j (x) 0, j = 1,..., l x R n Assume x X to be a regular point. x is a local minimum λ j, j = 1,..., l such that f (x ) + l λ j h j (x ) = 0 j=1 λ j h j (x ) = 0 j = 1,..., l λ j 0 j = 1,..., l

Consider the problem: Karush-Kuhn-Tucker (KKT) Conditions min h j (x) 0, j = 1,..., l x R n X = {x R n : h j (x) 0, j = 1,..., l} x X, A(x ) = {j : h j (x ) = 0} KKT necessary conditions (First Order) : If x X is a local minimum and a regular point, then there exists a unique vector λ (= (λ 1,..., λ l ) T ) such that l f (x ) + λ j h j (x ) = 0 j=1 λ j h j (x ) = 0 j = 1,..., l λ j 0 j = 1,..., l

KKT necessary conditions (First Order) : If x X is a local minimum and a regular point, then there exists a unique vector λ (= (λ 1,..., λ l ) T ) such that l f (x ) + λ j h j (x ) = 0 j=1 λ j h j (x ) = 0 j = 1,..., l λ j 0 j = 1,..., l KKT point : (x, λ ), x X, λ 0 Lagrangian function : L(x, λ) = + l j=1 λ jh j (x) Lx(x, λ ) = 0 λ j : Lagrange multipliers, λ j 0 λ j h j (x ) = 0 λ j = 0 j / A(x ) : Complementary Slackness Condition

min h j (x) 0, j = 1,..., l x R n At a local minimum, active set is unknown Need to investigate all possible active sets for finding KKT points Example: min x 2 1 + x 2 2 x 2 1 x 1 + x 2 1 A KKT point can be a local maximum Example: min x 2 x 0

Constraint Qualification Every local minimum need not be a KKT point Example [Kuhn and Tucker, 1951] 1 min x 1 x 2 (1 x 1 ) 3 0 x 2 0 Linear Independence Constraint Qualification (LICQ) : h j (x ), j A(x ) are linearly independent Mangasarian-Fromovitz Constraint Qualification (MFCQ) {d : h j (x ) T d < 0, j A(x )} φ 1 H.W. Kuhn and A.W. Tucker, Nonlinear Programming, in Proceedings of the Second Berkeley Symposium on Mathematical Statistics and Probability, J. Neyman, ed., Berkeley, CA, 1951, University of California Press, pp. 481 492.

Consider the problem (CP): min h j (x) 0, j = 1,..., l x R n Assumption: f, h j, j = 1,..., l are differentiable convex functions CP is a convex program X = {x R n : h j (x) 0, j = 1,..., l} Every local minimum of a convex program is a global minimum The set of all optimal solutions to a convex program is convex If x X is a regular point, then for x to be a global minimum of CP, first order KKT conditions are necessary and sufficient.

Proof. Let (x, λ ) be a KKT point. We need to show that x is a global minimum of CP. We use the convexity of f and h j to prove this. Consider any x X. For a convex function f, f (x ) + f (x ) T (x x ). + λ j h j (x) j f (x ) + f (x ) T (x x ) + j λ j (h j (x ) + h j (x ) T (x x )) = (f (x ) + j λ j h j (x ) +( f (x ) + j λ j h j (x )) T (x x ) = f (x ) x X x is a global minimum of CP

Consider the problem (CP): min h j (x) 0, j = 1,..., l x R n Assumption: f, h j, j = 1,..., l are convex functions X = {x R n : h j (x) 0, j = 1,..., l} Slater s Constraint Qualification: There exists y X such that h j (y) < 0, j = 1,..., l Useful when the constraint functions h j are convex For example, the following program does not satisfy Slater s constraint qualification: min x 1 + x 2 (x 1 + 1) 2 + x 2 2 1 (x 1 1) 2 + x 2 2 1 (0, 0) T is the global minimum; but it is not a KKT point.

Consider the problem: min e i (x) = 0, i = 1,..., m x R n Assumption: f, e i, i = 1,..., m are smooth functions X = {x R n : e i (x) 0, i = 1,..., m} Let x X, A(x) = {i : e i (x) = 0} = {1,..., m} Definition A vector d R n is said to be a tangent of X at x if either d = 0 or there exists a sequence {x k } X, x k x k such that x k x, x k x x k x d d. The collection of all tangents of X at x is called the tangent set at x and is denoted by T (x).

min e i (x) = 0, i = 1,..., m x R n X = {x R n : e i (x) = 0, i = 1,..., m} Regular Point: A point x X is said to be a regular point if e i ( x), i = 1,..., m are linearly independent. At a regular point x X, T ( x) = {d : e i ( x) T d = 0, i = 1,..., m}

Let x X be a regular point and local extremum (minimum or maximum) of the problem Consider any d T (x ). Let x(t) be any smooth curve such that x(t) X x(0) = x, ẋ(0) = d a > 0 such that e(x(t)) = 0 t [ a, a] x is a regular point T (x ) = {d : e i (x ) T d = 0, i = 1,..., m} x is a constrained local extremum d dt f (x(t)) t=0 = 0 f (x ) T d = 0. If x is a regular point w.r.t. the constraints e i (x) = 0, i = 1,..., m and x is a local extremum point (a minimum or maximum) of f subject to these constraints, then f (x ) is orthogonal to the tangent set, T (x ).

Theorem Let x X be a regular point and be a local minimum. Then µ R m such that f (x ) + m µ i e i (x ) = 0. i=1 Proof. Let e(x) = (e 1 (x),..., e m (x)). x X is a local minimum. {d : f (x ) T d < 0, e(x ) T d = 0} = φ. Let C 1 = {(y 1, y 2 ) : y 1 = f (x ) T d, y 2 = e(x ) T d} and C 2 = {(y 1, y 2 ) : y 1 < 0, y 2 = 0} Note that C 1 and C 2 are convex and C 1 C 2 = φ. If C 1 and C 2 are nonempty convex sets in R n and C 1 C 2 = φ, µ R n (µ 0) such that µ T x 1 µ T x 2 x 1 C 1, x 2 C 2.

Proof. (continued) Therefore, (µ 0, µ) R m+1 such that µ 0 f (x ) T d+µ T ( e(x ) T d) µ 0 y 1 +µ T y 2 d R n, (y 1, y 2 ) C 2 Letting y 2 = 0, we get µ 0 0. Letting (y 1, y 2 ) = (0, 0), we get µ 0 f (x ) T d + µ T ( e(x ) T d) 0 d R n If we take d = (µ 0 f (x ) + µ T e(x )), we get (µ 0 f (x ) + µ T e(x )) 2 0. Therefore, µ 0 f (x ) + µ T e(x ) = 0 where (µ 0, µ) (0, 0) Note that, µ 0 > 0 since x is a regular point. Hence, f (x ) + µ T e(x ) = 0

Examples: 1 min x 1 3x 2 (x 1 1) 2 + x2 2 = 1 (x 1 + 1) 2 + x2 2 = 1 (0, 0) T is the only feasible point; (0, 0) T is not a regular point. 2 min x 1 + x 2 x1 2 + x2 2 = 1 local maximum : ( 1 1 2, 2 ) T local minimum : ( 1 2, 1 2 ) T

General Nonlinear Programming Problems min h j (x) 0, j = 1,..., l e i (x) = 0, i = 1,..., m f, h j (j = 1,..., l), e i (i = 1,..., m) are sufficiently smooth X = {x : h j (x) 0, e i (x) = 0, j = 1,..., l; i = 1,..., m} x X Active set of X at x : I = {j : h j (x ) = 0} All the equality constraints, E = {1,..., m} A(x ) = I E Assumption: x is a regular point. That is, { h j (x ) : j I} { e i (x ) : i E} is a set of linearly independent vectors

min h j (x) 0, j = 1,..., l e i (x) = 0, i = 1,..., m X = {x : h j (x) 0, e i (x) = 0, j = 1,..., l; i = 1,..., m} KKT necessary conditions (First Order) : If x X is a local minimum and a regular point, then there exist unique vectors λ R l + and µ R m such that l m f (x ) + λ j h j (x ) + µ i e i (x ) = 0 j=1 i=1 λ j h j (x ) = 0 j = 1,..., l λ j 0 j = 1,..., l KKT Point: (x X, λ R l +, µ R m ) satisfying above conditions First order KKT conditions also satisfied at a local max

Consider the problem (CP): min h j (x) 0, j = 1,..., l e i (x) = 0, i = 1,..., m Assumption: f, h j, j = 1,..., l are smooth convex functions e i (x) = a T x i b i, i = 1,..., m CP is a convex programming problem X = {x : h j (x) 0, e i (x) = 0, j = 1,..., l; i = 1,..., m} Assumption: Slater s Constraint Qualification holds for X. There exists y X such that h j (y) < 0, j = 1,..., l If X satisfies Slater s Constraint Qualification, then the first order KKT conditions are necessary and sufficient for a global minimum of a convex programming problem CP

Consider the problem : Interpretation of Lagrange Multipliers min h j (x) 0, j = 1,..., l X = {x : h j (x) 0, j = 1,..., l; } Let x X be a regular point and a local minimum Let A(x ) = {j : h j (x ) = 0} f (x ) + j A(x ) λ j h j (x ) = 0 Suppose the constraint h j(x), j A(x ) is perturbed to h j(x) ɛ h j(x ) (ɛ > 0) New problem: min h j (x) 0, j = 1,..., l, j j h j(x) ɛ h j(x )

For the new problem, let x ɛ be the solution. Assumption: A(x ) = A(x ɛ) For the constraint h j(x), h j(x ɛ) h j(x ) = ɛ h j(x ) (x ɛ x ) T h j(x ) ɛ h j(x ) For other constraints, h j (x), j j, h j (x ɛ) h j (x ) = 0 (x ɛ x ) T h j (x ) = 0 Change in the objective function, f (x ɛ) f (x ) (x ɛ x ) T f (x ) = (x ɛ x ) T (λ j h j (x ) j A(x ) = λ j ɛ h j(x ) df dɛ λ j x=x

Consider the problem (NLP): min h j (x) 0, j = 1,..., l e i (x) = 0, i = 1,..., m Let f, h j, e i C 2 for every j and i. X = {x : h j (x) 0, e i (x) = 0, j = 1,..., l; i = 1,..., m} x X Active set of X at x : I = {j : h j (x ) = 0} All the equality constraints, E = {1,..., m} A(x ) = I E Assumption: x is a regular point. That is, { h j (x ) : j I} { e i (x ) : i E} is a set of linearly independent vectors

Consider the problem (NLP): min h j (x) 0, j = 1,..., l e i (x) = 0, i = 1,..., m Define the Lagrangian function, L(x, λ, µ) = + l j=1 λ jh j (x) + m i=1 µ ie i (x) KKT necessary conditions (Second Order) : If x X is a local minimum of NLP and a regular point, then there exist unique vectors λ R l + and µ R m such that xl(x, λ, µ ) = 0 and λ j h j (x ) = 0 j = 1,..., l λ j 0 j = 1,..., l d T 2 xl(x, λ, µ )d 0 for all d h j (x ) T d 0, j I and e i (x ) T d = 0, i E.

KKT sufficient conditions (Second Order) : If there exist x X, λ R l + and µ R m such that such that xl(x, λ, µ ) = 0 and for all d 0 such that λ j h j (x ) = 0 j = 1,..., l λ j 0 j = 1,..., l d T 2 xl(x, λ, µ )d > 0 h j (x ) T d = 0, j I and λ j > 0 h j (x ) T d 0, j I and λ j = 0 e i (x ) T d = 0, i E, then x is a strict local minimum of NLP.

Existence and Uniqueness of Lagrange Multipliers Example: min x 1 x 2 (1 x 1 ) 3 0 x 1 0 x 2 0 x = (1, 0) T is the strict local minimum Cannot find a KKT point, (x, λ ) Linear Independence Constraint Qualification does not hold at (1, 0) T Add an extra constraint 2x 1 + x 2 2 Lagrange multipliers are not unique

Importance of Constraint Set Representation min (x 1 9 4 )2 + (x 2 2) 2 x 2 1 x 2 0 x 1 + x 2 6 x 1 0, x 2 0 Convex Programming Problem Slater s Constraint Qualification holds First order KKT conditions are necessary and sufficient at a global minimum KKT point does not have x = (2, 4) T Solution : x = ( 3 2, 9 4 )T Replace the first inequality in the constraints by (x 2 1 x 2 ) 3 0 ( 3 2, 9 4 )T is not regular for the new constraint representation!

Example: Find the point on the parabola x 2 = 1 5 (x 1 1) 2 that is closest to (1, 2) T, in the Euclidean norm sense. min (x 1 1) 2 + (x 2 2) 2 (x 1 1) 2 = 5x 2 x, µ is a KKT point : x = (1, 0) T and µ = 4 5 Satisfies second order sufficiency conditions x = (1, 0) T is a strict local minimum Reformulation to an unconstrained optimization problem

Unbounded problem Example: min 0.2(x 1 3) 2 + x2 2 x1 2 + x2 2 1 Unbounded objective function (1, 0) T is a strict local minimum

Example: min x 2 1 + x 2 2 + 1 4 x2 3 x 1 + x 3 = 1 x 2 1 + x 2 2 2x 1 = 1 (1 2, 0, 2 2) T is a strict local minimum. (1 + 2, 0, 2 + 2) T is a strict local maximum.