On sequential optimality conditions for constrained optimization. José Mario Martínez martinez

Similar documents
Some Remarks on Constrained Optimization. José Mario Martínez

ON AUGMENTED LAGRANGIAN METHODS WITH GENERAL LOWER-LEVEL CONSTRAINTS. 1. Introduction. Many practical optimization problems have the form (1.

On Augmented Lagrangian methods with general lower-level constraints

On the minimization of possibly discontinuous functions by means of pointwise approximations

On the complexity of an Inexact Restoration method for constrained optimization

Augmented Lagrangian methods under the Constant Positive Linear Dependence constraint qualification


Derivative-free methods for nonlinear programming with general lower-level constraints*

Sequential equality-constrained optimization for nonlinear programming

Priority Programme 1962

5 Handling Constraints

Sequential equality-constrained optimization for nonlinear programming


College of William & Mary Department of Computer Science

Constrained Optimization

Optimization Problems with Constraints - introduction to theory, numerical Methods and applications

Second-order negative-curvature methods for box-constrained and general constrained optimization

Algorithms for constrained local optimization

minimize x subject to (x 2)(x 4) u,

Infeasibility Detection and an Inexact Active-Set Method for Large-Scale Nonlinear Optimization

On Augmented Lagrangian Methods with General Lower-Level Constraints

arxiv: v2 [math.oc] 13 Apr 2018

AN AUGMENTED LAGRANGIAN AFFINE SCALING METHOD FOR NONLINEAR PROGRAMMING

Constrained Optimization Theory

Quiz Discussion. IE417: Nonlinear Programming: Lecture 12. Motivation. Why do we care? Jeff Linderoth. 16th March 2006

A New Penalty-SQP Method

Continuous Optimisation, Chpt 6: Solution methods for Constrained Optimisation

Constrained optimization: direct methods (cont.)

Some new facts about sequential quadratic programming methods employing second derivatives

Inexact-Restoration Method with Lagrangian Tangent Decrease and New Merit Function for Nonlinear Programming 1 2

ON LICQ AND THE UNIQUENESS OF LAGRANGE MULTIPLIERS

Second Order Optimality Conditions for Constrained Nonlinear Programming

Some Notes on Approximate Optimality Conditions in Scalar and Vector Optimization Problems

Sufficient Conditions for Finite-variable Constrained Minimization

A New Sequential Optimality Condition for Constrained Nonsmooth Optimization

Penalty and Barrier Methods. So we again build on our unconstrained algorithms, but in a different way.

Examples of dual behaviour of Newton-type methods on optimization problems with degenerate constraints

Algorithms for Constrained Optimization

Primal-dual relationship between Levenberg-Marquardt and central trajectories for linearly constrained convex optimization

Numerical Optimization

An Inexact Sequential Quadratic Optimization Method for Nonlinear Optimization

A GLOBALLY CONVERGENT STABILIZED SQP METHOD: SUPERLINEAR CONVERGENCE

2.3 Linear Programming

10 Numerical methods for constrained problems

AM 205: lecture 19. Last time: Conditions for optimality Today: Newton s method for optimization, survey of optimization methods

Numerical Optimization

Combining stabilized SQP with the augmented Lagrangian algorithm

1 Computing with constraints

Nonlinear Programming, Elastic Mode, SQP, MPEC, MPCC, complementarity

5.6 Penalty method and augmented Lagrangian method

SF2822 Applied Nonlinear Optimization. Preparatory question. Lecture 9: Sequential quadratic programming. Anders Forsgren

A SHIFTED PRIMAL-DUAL INTERIOR METHOD FOR NONLINEAR OPTIMIZATION

Constraint qualifications for nonlinear programming

4TE3/6TE3. Algorithms for. Continuous Optimization

Constrained Derivative-Free Optimization on Thin Domains

SECTION C: CONTINUOUS OPTIMISATION LECTURE 9: FIRST ORDER OPTIMALITY CONDITIONS FOR CONSTRAINED NONLINEAR PROGRAMMING

A STABILIZED SQP METHOD: SUPERLINEAR CONVERGENCE

Some Inexact Hybrid Proximal Augmented Lagrangian Algorithms

Outline. Roadmap for the NPP segment: 1 Preliminaries: role of convexity. 2 Existence of a solution

ALADIN An Algorithm for Distributed Non-Convex Optimization and Control

AM 205: lecture 19. Last time: Conditions for optimality, Newton s method for optimization Today: survey of optimization methods

Rejoinder on: Critical Lagrange multipliers: what we currently know about them, how they spoil our lives, and what we can do about it

Differentiable exact penalty functions for nonlinear optimization with easy constraints. Takuma NISHIMURA

Simple Iteration, cont d

Computational Optimization. Constrained Optimization Part 2

Optimality Conditions for Constrained Optimization

Global convergence of trust-region algorithms for constrained minimization without derivatives

Lecture 13: Constrained optimization

On the behavior of Lagrange multipliers in convex and non-convex infeasible interior point methods

An Active Set Strategy for Solving Optimization Problems with up to 200,000,000 Nonlinear Constraints

N. L. P. NONLINEAR PROGRAMMING (NLP) deals with optimization models with at least one nonlinear function. NLP. Optimization. Models of following form:

c 2012 Society for Industrial and Applied Mathematics

Constraint Identification and Algorithm Stabilization for Degenerate Nonlinear Programs

CONSTRAINED NONLINEAR PROGRAMMING

Algorithms for nonlinear programming problems II

Multidisciplinary System Design Optimization (MSDO)

Machine Learning. Support Vector Machines. Manfred Huber

TMA 4180 Optimeringsteori KARUSH-KUHN-TUCKER THEOREM

Lectures 9 and 10: Constrained optimization problems and their optimality conditions

MS&E 318 (CME 338) Large-Scale Numerical Optimization

Applications of Linear Programming

A SHIFTED PRIMAL-DUAL PENALTY-BARRIER METHOD FOR NONLINEAR OPTIMIZATION

Examination paper for TMA4180 Optimization I

CE 191: Civil and Environmental Engineering Systems Analysis. LEC 05 : Optimality Conditions

Computational Optimization. Augmented Lagrangian NW 17.3

Scientific Computing: An Introductory Survey

Scientific Computing: An Introductory Survey

Algorithms for nonlinear programming problems II

8. Constrained Optimization

Optimization. Escuela de Ingeniería Informática de Oviedo. (Dpto. de Matemáticas-UniOvi) Numerical Computation Optimization 1 / 30

CSCI 1951-G Optimization Methods in Finance Part 09: Interior Point Methods

CS-E4830 Kernel Methods in Machine Learning

Lecture 3: Lagrangian duality and algorithms for the Lagrangian dual problem

Primal-Dual Interior-Point Methods

A smoothing augmented Lagrangian method for solving simple bilevel programs

SECTION C: CONTINUOUS OPTIMISATION LECTURE 11: THE METHOD OF LAGRANGE MULTIPLIERS

Optimization. Charles J. Geyer School of Statistics University of Minnesota. Stat 8054 Lecture Notes

Lecture 11 and 12: Penalty methods and augmented Lagrangian methods for nonlinear programming

A STABILIZED SQP METHOD: GLOBAL CONVERGENCE

Convex Optimization. Dani Yogatama. School of Computer Science, Carnegie Mellon University, Pittsburgh, PA, USA. February 12, 2014

Transcription:

On sequential optimality conditions for constrained optimization José Mario Martínez www.ime.unicamp.br/ martinez UNICAMP, Brazil 2011

Collaborators This talk is based in joint papers with Roberto Andreani (Campinas), Benar F. Svaiter (IMPA), Lucio T. Santos (Campinas) and Gabriel Haeser (U. F. São Paulo). 1 R. Andreani, J. M. Martínez and B. F. Svaiter. A new sequential optimality condition for constrained optimization and algorithmic consequences. SIAM Journal on Optimization 20, pp. 3533-3554 (2010). 2 R. Andreani, G. Haeser, J. M. Martínez. On sequential optimality conditions for smooth constrained optimization. To appear in Optimization 2011. 3 J. M. Martínez and B. F. Svaiter. A practical optimality condition without constraint qualifications for nonlinear programming. Journal of Optimization Theory and Applications 118, pp. 117-133 (2003).

Nonlinear Programming Problem NLP Minimize f (x) subject to h(x) = 0, g(x) 0, where x R n, h(x) R m, g(x) R p and all the functions are smooth.

Finding global minimizers is difficult. Affordable algorithms for solving large-scale problems usually guarantee convergence only to stationary points. Affordable algorithms are useful tools in the process of finding global solutions (Multistart and so on).

Pointwise Optimality conditions Stationary points are points that satisfy some necessary optimality condition. In Nonlinear Programming, (pointwise) Necessary Optimality conditions take the form: The point x is feasible and fulfills the KKT conditions OR fails to satisfy the XXX constraint qualification. The strength of a pointwise necessary optimality condition is associated with the weakness of the constraint qualification. (Weak constraint qualifications are good.)

Sequential Optimality conditions Nonlinear Programming algorithms are iterative. They generate sequences {x k } that, presumably, converge to stationary points which, in fact, are never reached exactly. So, any practical algorithm needs to decide, at each iterate x k whether x k is an approximate solution (approximate stationary point) or not. Stopping criterion!. This motivates to study Sequential Optimality conditions. A feasible point x satisfies a Sequential Optimality condition if there exists a sequence that converges to x and fulfills some property PPP. Local minimizers satisfy Sequential Optimality conditions. The fulfillment of PPP may be verified at each iterate of a practical algorithm.

Example: Approximate Gradient Projection AGP A feasible point x satisfies AGP (Martínez-Svaiter 2003) if there exists a sequence {x k } x such that the projected gradient P(x k f (x k ) x k on a linear approximation of the constraints tends to zero. AGP is a strong optimality condition: Every local minimizer satisfies AGP and AGP implies the pointwise optimality condition KKT or not-cpld. (CPLD is a weak constraint qualification that says that positively linear dependent gradients of active constraints at x remain linearly dependent in a neighborhood of x ). AGP generates the natural stopping criterion for Inexact Restoration Methods and other algorithms for Nonlinear Programming.

Approximate KKT condition We say that a feasible point x satisfies AKKT if there exist sequences {x k } R n, {λ k } R m, {µ k } R p such that lim x k = x, k and lim f (x k ) + h(x k )λ k + g(x k )µ k = 0 k lim min{ g i(x k ), µ k i } = 0, for all i = 1,..., p. k

Recall that a feasible point x satisfies KKT if there exist λ R m, µ R p such that f (x ) + h(x )λ + g(x )µ = 0 and min{ g i (x ), µ i } = 0, for all i = 1,..., p. Consequence: KKT implies AKKT. (Take x k = x, λ k = λ, µ k = µ for all k).

Properties of the Approximate KKT condition Every local minimizer satisfies AKKT (Even if it does not satisfy KKT). (No constraint qualification is needed.) AKKT is a strong optimality condition. (It implies KKT or not-cpld.) AKKT generates the natural stopping criterion h(x k ) ε, g(x k ) + ε, and f (x k ) + h(x k )λ k + g(x k )µ k ε min{ g i (x k ), µ k i } ε for all i = 1,..., p. Sequences generated by well-established Nonlinear Programming algorithms satisfy the Approximate KKT condition.

Pointwise and Sequential Optimality Conditions L AGP AGP KKT or not CQ AKKT KKT or not CPLD C AGP KKT or not MFCQ

Complementary Approximate KKT condition (CAKKT) CAKKT is a new sequential optimality condition (Andreani, Martínez, Svaiter, 2010). We say that a feasible point x satisfies AKKT if there exist sequences {x k } R n, {λ k } R m, {µ k } R p + such that lim x k = x, k and lim f (x k ) + h(x k )λ k + g(x k )µ k = 0, k lim g i(x k )µ k i = 0, for all i = 1,..., p k lim h i(x k )λ k i = 0, for all i = 1,..., m. k

Properties of the Complementary Approximate KKT condition Every local minimizer x satisfies CAKKT. (Even if x is not KKT.) (Independently of constraint qualifications.) CAKKT is strictly stronger than AKKT and strictly stronger than the Approximate Gradient Projection (AGP) Condition. CAKKT is strictly stronger than the pointwise optimality condition KKT or not-cpld.

CAKKT generates the natural? stopping criterion h(x k ) ε, g(x k ) + ε, f (x k ) + h(x k )λ k + g(x k )µ k ε, and g i (x k )µ k i ε for all i = 1,..., p h i (x k )λ k i ε for all i = 1,..., m. Sequences generated by well-established Nonlinear Programming algorithms satisfy CAKKT Really?.

Example: Augmented Lagrangian (Algencan) At each outer iteration we compute x k, an approximate stationary point of the augmented Lagrangian L(x, λ k, µ k ) = f (x) + ρ k 2 ) 2 [(h(x) + λk + (g(x) + µk ρ k ρ k ) 2 + ]. We update λ k+1 = λ k + ρ k h(x k ) and µ k+1 = (µ k + ρ k g(x k )) +. We update ρ k+1 = γρ k+1 (γ > 1) if max{ h(x k, max(g(x k ), µ k+1 ) > r max{ h(x k 1, max(g(x k 1 ), µ k ). Else, we maintain ρ k+1 = ρ k. We project λ k+1 and µ k+1 on safeguarding boxes.

Generalized Lojasiewicz Inequality (GLI) f : R n R satisfies GLI at x if f (z) f (x) ϕ(z) f (z) for all z in a neighborhood of x, where ϕ is continuous and ϕ(x) = 0 Every reasonable (for example, analytic) function f : R n R satisfies GLI. We will assume that the constraint functions h i, g j of our NLP problem satisfy GLI at the feasible points

Augmented Lagrangian Algorithm and CAKKT Theorem Assume that x is a feasible limit point of a sequence generated by the Augmented Lagrangian method and that the GLI assumption holds at x. Then, the associated subsequence (x k, λ k+1, µ k+1 ) fulfills the Complementary Approximate KKT condition. Consequence: At least one reasonable NLP algorithm converges to CAKKT points.

GLI Assumption cannot be eliminated Sequences generated by the Augmented Lagrangian method generate Approximate KKT sequences without the GLI condition. However, for convergence to Complementary Approximate KKT points, the GLI condition cannot be eliminated. Example: Consider the problem where Minimize x subject to h(x) = 0, h(x) = x 4 sin ( ) 1 x if x 0, and h(0) = 0. h does not fulfill GLI at x = 0. We are able to define an instance of the algorithm in which the sequence x k converges to x = 0 and the condition h(x k )λ k+1 0 does not hold.

What happens with other reasonable NLP algorithms? Essentially, we don t know. We performed a good number of numerical experiments with SQP (Sequential Quadratic Programming) in cases in which the solution of the problem is not KKT. We observed that the approximate complementarity condition λ k h(x k ) 0 generally holds but the Lagrangian condition f (x k ) + h(x k )λ k 0 does not! So, contrary to expected, SQP seems to fail to satisfy CAKKT not because of the C but because of the KKT.

Should CAKKT be used as stopping criterion? (Perhaps Naive) Answer: If one is using an algorithm that provably generates CAKKT sequences, CAKKT should be used as stopping criterion since, eventually, it will be satisfied. If your algorithm does not generate CAKKT sequences, it seems better not to use CAKKT at all.