Second Order Optimality Conditions for Constrained Nonlinear Programming

Similar documents
SECTION C: CONTINUOUS OPTIMISATION LECTURE 9: FIRST ORDER OPTIMALITY CONDITIONS FOR CONSTRAINED NONLINEAR PROGRAMMING

SECTION C: CONTINUOUS OPTIMISATION LECTURE 11: THE METHOD OF LAGRANGE MULTIPLIERS

Numerical Optimization

The Fundamental Theorem of Linear Inequalities

Constrained Optimization Theory

Examination paper for TMA4180 Optimization I

Quiz Discussion. IE417: Nonlinear Programming: Lecture 12. Motivation. Why do we care? Jeff Linderoth. 16th March 2006

Lectures 9 and 10: Constrained optimization problems and their optimality conditions

Optimality Conditions for Constrained Optimization

TMA 4180 Optimeringsteori KARUSH-KUHN-TUCKER THEOREM

ISM206 Lecture Optimization of Nonlinear Objective with Linear Constraints

ON LICQ AND THE UNIQUENESS OF LAGRANGE MULTIPLIERS

minimize x subject to (x 2)(x 4) u,

Constraint qualifications for nonlinear programming

Date: July 5, Contents

Determination of Feasible Directions by Successive Quadratic Programming and Zoutendijk Algorithms: A Comparative Study

Continuous Optimisation, Chpt 6: Solution methods for Constrained Optimisation

8. Constrained Optimization

Karush-Kuhn-Tucker Conditions. Lecturer: Ryan Tibshirani Convex Optimization /36-725

Constrained Optimization

In view of (31), the second of these is equal to the identity I on E m, while this, in view of (30), implies that the first can be written

Computational Optimization. Constrained Optimization Part 2

Algorithms for constrained local optimization

Lecture 3 Introduction to optimality conditions

FIRST- AND SECOND-ORDER OPTIMALITY CONDITIONS FOR MATHEMATICAL PROGRAMS WITH VANISHING CONSTRAINTS 1. Tim Hoheisel and Christian Kanzow

Unconstrained optimization

Constrained optimization: direct methods (cont.)

The Karush-Kuhn-Tucker (KKT) conditions

OPTIMISATION /09 EXAM PREPARATION GUIDELINES

Lecture 1: Introduction. Outline. B9824 Foundations of Optimization. Fall Administrative matters. 2. Introduction. 3. Existence of optima

UNDERGROUND LECTURE NOTES 1: Optimality Conditions for Constrained Optimization Problems

Mathematical Economics. Lecture Notes (in extracts)

Lecture 1: Introduction. Outline. B9824 Foundations of Optimization. Fall Administrative matters. 2. Introduction. 3. Existence of optima

5 Handling Constraints

Written Examination

Lagrange Multipliers

Generalization to inequality constrained problem. Maximize

5.6 Penalty method and augmented Lagrangian method

Math 164-1: Optimization Instructor: Alpár R. Mészáros

MATH 4211/6211 Optimization Constrained Optimization

OPTIMISATION 2007/8 EXAM PREPARATION GUIDELINES

Exam: Continuous Optimisation 2015

Math 273a: Optimization Basic concepts

The Conjugate Gradient Method

Constrained Optimization and Lagrangian Duality

Taylor Series and stationary points

MATH2070 Optimisation

Lecture 3: Basics of set-constrained and unconstrained optimization

Assignment 1: From the Definition of Convexity to Helley Theorem

The Relation Between Pseudonormality and Quasiregularity in Constrained Optimization 1

Microeconomics I. September, c Leopold Sögner

Existence of minimizers

Chapter 1. Optimality Conditions: Unconstrained Optimization. 1.1 Differentiable Problems

Numerisches Rechnen. (für Informatiker) M. Grepl P. Esser & G. Welper & L. Zhang. Institut für Geometrie und Praktische Mathematik RWTH Aachen

Lecture 2 - Unconstrained Optimization Definition[Global Minimum and Maximum]Let f : S R be defined on a set S R n. Then

AM 205: lecture 18. Last time: optimization methods Today: conditions for optimality

Structural and Multidisciplinary Optimization. P. Duysinx and P. Tossings

CS-E4830 Kernel Methods in Machine Learning

Introduction to Optimization Techniques. Nonlinear Optimization in Function Spaces

Lecture 13: Constrained optimization

On sequential optimality conditions for constrained optimization. José Mario Martínez martinez

Optimization. A first course on mathematics for economists

There is a more global concept that is related to this circle of ideas that we discuss somewhat informally. Namely, a region R R n with a (smooth)

Numerical Optimization

Nonlinear Programming (NLP)

EC9A0: Pre-sessional Advanced Mathematics Course. Lecture Notes: Unconstrained Optimisation By Pablo F. Beker 1

Optimisation in Higher Dimensions

Optimization Theory. Lectures 4-6

Computational Optimization. Augmented Lagrangian NW 17.3

Lecture 3. Optimization Problems and Iterative Algorithms

Algorithms for nonlinear programming problems II

Lagrange Relaxation and Duality

Optimization Problems with Constraints - introduction to theory, numerical Methods and applications

Convex Optimization & Lagrange Duality

Implications of the Constant Rank Constraint Qualification

CS / ISyE 730 Spring 2015 Steve Wright

Introduction to Machine Learning Lecture 7. Mehryar Mohri Courant Institute and Google Research

5. Duality. Lagrangian

Optimality conditions for unconstrained optimization. Outline

Introduction to Nonlinear Stochastic Programming

4TE3/6TE3. Algorithms for. Continuous Optimization

SF2822 Applied Nonlinear Optimization. Preparatory question. Lecture 9: Sequential quadratic programming. Anders Forsgren

Algorithms for nonlinear programming problems II

March 8, 2010 MATH 408 FINAL EXAM SAMPLE

Convex Optimization and SVM

Lecture 15 Newton Method and Self-Concordance. October 23, 2008

Link lecture - Lagrange Multipliers

OPTIMIZATION THEORY IN A NUTSHELL Daniel McFadden, 1990, 2003

Support Vector Machines for Regression

Solving generalized semi-infinite programs by reduction to simpler problems.

1. Introduction. We consider mathematical programs with equilibrium constraints in the form of complementarity constraints:

Finite Dimensional Optimization Part III: Convex Optimization 1

Lyapunov Stability Theory

Optimization: Problem Set Solutions

1 Computing with constraints

1 Overview. 2 A Characterization of Convex Functions. 2.1 First-order Taylor approximation. AM 221: Advanced Optimization Spring 2016

Optimization Tutorial 1. Basic Gradient Descent

MATHEMATICAL ECONOMICS: OPTIMIZATION. Contents

MVE165/MMG631 Linear and integer optimization with applications Lecture 13 Overview of nonlinear programming. Ann-Brith Strömberg

Convex Optimization. Newton s method. ENSAE: Optimisation 1/44

Transcription:

Second Order Optimality Conditions for Constrained Nonlinear Programming Lecture 10, Continuous Optimisation Oxford University Computing Laboratory, HT 2006 Notes by Dr Raphael Hauser (hauser@comlab.ox.ac.uk)

We again consider the general nonlinear optimisation problem (NLP) min f(x) x R n s.t. g i (x) = 0 g i (x) 0 (i E), (i I). We will now derive second order optimality conditions for (NLP). For that purpouse, we assume that f and the g i (i E I) are twice continuously differentiable functions.

Definition 1: Let x R n be a feasible point for (NLP) and let x C 2( ( ɛ, ɛ), R n) be a path such that x(0) = x, d := d x(0) 0, dt g i (x(t)) = 0 (i E, t ( ɛ, ɛ)), g i (x(t)) 0 (i I, t [0, ɛ)). (1) Thus, we can imagine that x(t) is a smooth piece of trajectory of a point particle that passes through x at time t = 0 with nonzero speed d and moves into the feasible domain. We call x(t) a feasible exit path from x and the tangent vector d = d dt x(0) a feasible exit direction from x.

F d xt x*

The second order optimality analysis is based on the following observation: If x is a local minimiser of (NLP) and x(t) is a feasible exit path from x then x must also be a local minimiser for the univariate constrained optimisation problem min f(x(t)) s.t. t 0 Before we start looking at such problems more closely, we develop an alternative characterisation of feasible exit directions from x.

Definition 1 implies d T g i (x ) = d dt g i(x(t)) t=0 = d dt 0 = 0 g lim i (x(t)) 0 t 0+ (i E), t 0 (i A(x )). Therefore, the following are necessary conditions for d R n to be a feasible exit direction from x : d 0, d T g i (x ) = 0 (i E), d T g j (x ) 0 (j A(x )). (2)

On the other hand, if the LICQ holds at x then Lemma 1 of Lecture 9 shows that (2) implies the existence of a feasible exit path from x such that d x(0) = d, dt (3) g i (x(t)) = td T g i (x ) (i E A(x ). (4) Thus, when the LICQ holds then (2) is also a sufficient condition and hence an exact characterisation for d to be a feasible exit path from x.

Second Order Necessary Optimality Conditions Let x be a local minimiser of (NLP) where the LICQ holds. The KKT conditions say that there exists a vector λ of Lagrange multipliers such that D x L(x, λ ) = 0, λ j 0 λ i g i(x ) = 0 g j (x ) 0 g i (x ) = 0 (j I), (i E I), (j I), (i E), where L(x, λ) = f(x) i λ i g i is the Lagrangian associated with (NLP). (5)

Now let x(t) be a feasible exit path from x with exit direction d, and let us consider the restricted problem min f(x(t)) s.t. t 0 (6) Since x is a local minimiser of (NLP), t = 0 must be a local minimiser of (6). By Taylor s theorem and the KKT conditions, f(x(t)) = f(x ) + td T f(x ) + O(t 2 ) =f(x ) + t m i=1 λ i dt g i (x ) + O(t 2 ).

We thus wish to show that for small t 0, t m i=1 λ i dt g i (x ) + O(t 2 ) 0. (7) Note that λ i dt g i (x ) = 0 (i E I \ A(x )), so that these terms can be omitted from (7). But what about indices j A(x )? We have to distinguish two different cases:

Case 1: there exists an index j A(x ) such that d T g j (x ) > 0. Then for all 0 < t 1, f(x(t)) =f(x ) + t m i=1 λ i dt g i (x ) + O(t 2 ) f(x ) + tλ j dt g j (x ) + O(t 2 ) > f(x ). Thus, in this case f strictly increases along the path x(t) for small positive t even if d2 dt2f(x(0)) was negative. Because of the constraint g j, nothing can be said about the Dxxf(x 2 )d.

Case 2: λ i dt g i (x ) = 0 (i I E). (8) In this case the above argument fails to guarantee that f locally increases along path x(t). We only know that d/dt f(x(0)) = 0, that is, x is a stationary point of (6). But this might very well be a local maximiser of the restricted problem. Second order derivatives d2 dt2f(x(0)) now decide whether t = 0 is a local minimiser of the restricted problem (6), yielding additional necessary information in this case!

f G F x2 x* xt d x1

Theorem 1: 2nd Order Necessary Optimality Conditions. Let x be a local minimiser of (NLP) where the LICQ holds. Let λ R m be a Lagrange multiplier vector such that (x, λ ) satisfy the KKT conditions. Then we have d T D xx L(x, λ )d 0 (9) for all feasible exit directions d from x that satisfy (8).

Proof: Let d 0 satisfy (2) and (8), and let x C 2( ( ɛ, ɛ), R n) be a feasible exit path from x corresponding to d. Then L ( x(t), λ ) (4) = f(x(t)) m i=1 λ i tdt g i (x ) (8) = f(x(t)).

Therefore, Taylor s theorem implies f(x(t)) = L(x, λ ) + td x L(x, λ )d + t2 ( d T D xx L(x, λ )d + D x L(x, λ ) d2 2 dt 2x(0)) + O(t 3 ) KKT = f(x ) + t2 2 dt D xx L(x, λ )d + O(t 3 ). If it were the case that d T D xx L(x, λ )d < 0 then f(x(t)) < f(x ) for all t sufficiently small, contradicting the assumption that x is a local minimiser. Therefore, it must be the case that d T D xx L(x, λ )d 0.

Sufficient Optimality Conditions: In unconstrained minimisation we found that strengthening the second order condition D 2 f(x) 0 to D 2 f(x) 0 led to sufficient optimality conditions. Does the same happen when we change the inequality in (9) to a strict inequality? Our next result shows that this is indeed the case.

There are two issues that need to be addressed in the proof: The first is that x is a strict local minimiser for the restricted problem (6). This is easy to prove using Taylor expansions. The second, more delicate issue is to show that it suffices to look at the univariate problems (6) for all possible feasible exit paths from x.

Theorem: Sufficient Optimality Conditions. Let (x, λ ) R n R m be such that the KKT conditions (5) hold, the LICQ holds, and d T D xx L(x, λ )d > 0 for all feasible exit directions d R n from x that satisfy λ i dt g i (x ) = 0 (i I E). Then x is a strict local minimiser.

Proof: Let us assume to the contrary of our claim that x is not a local minimiser. Then there exists a sequence of feasible points (x k ) N such that lim k x k = x and f(x k ) f(x ) k N. (10) x k x The sequence x k x lies on the unit sphere which is a compact set. The Bolzano Weierstrass theorem therefore im- plies that we can extract a subsequence (x ki ) i N, k i < k j

(i < j), such that the limiting direction d := lim k d ki exists, where d ki = x k i x x ki x. Since d lies on the unit sphere we have d 0. Replacing the old sequence by the new one we may assume without loss of generality that k i i. Let us check that d satisfies the conditions d 0, d T g i (x ) = 0 (i E), d T g j (x ) 0 (j A(x )). (11)

and hence is a feasible exit direction: d T g j (x g j (x i ) g j (x ) ) = lim i x i x = lim i 0 = 0 g lim j (x i ) 0 i (j E), x i x 0 (j A(x )). By Taylor s theorem, f(x ) f(x k ) = f(x ) + x k x f(x ) T d k + O( x k x 2 ). Therefore, f(x ) T d = lim k f(x ) T d k 0. (12)

On the other hand, the KKT conditions and (11) imply d T f(x ) = m i=1 λ i dt g i (x ) 0. (13) But (12) and (13) can be jointly true only if λ i dt g i (x ) = 0 (i I E). The assumption of the theorem therefore implies that d T D xx L(x, λ )d > 0. (14)

On the other hand, f(x ) f(x k ) KKT f(xk ) = L(x k, λ ) m i=1 λ i g i(x k ) (since λ i 0 for i I and x k is feasible) or = L(x, λ ) + x k x D x L(x, λ )d T k + x k x 2 d T k 2 D xxl(x, λ )d k + O( x k x 3 ) KKT = f(x ) + x k x 2 d T k 2 D xxl(x, λ )d k + O( x k x 3 ), d T k D xxl(x, λ )d k O( x k x ).

Taking limits, we obtain d T D xx L(x, λ )d = lim k d T k D xxl(x, λ )d k 0. Since this contradicts (14), our assumption about the existence of the sequence (x k ) N must have been wrong.

Reading Assignment: Lecture-Note 10.