A STABILIZED SQP METHOD: GLOBAL CONVERGENCE

Similar documents
A STABILIZED SQP METHOD: SUPERLINEAR CONVERGENCE

A GLOBALLY CONVERGENT STABILIZED SQP METHOD: SUPERLINEAR CONVERGENCE

A stabilized SQP method: superlinear convergence

A GLOBALLY CONVERGENT STABILIZED SQP METHOD

REGULARIZED SEQUENTIAL QUADRATIC PROGRAMMING METHODS

A SHIFTED PRIMAL-DUAL INTERIOR METHOD FOR NONLINEAR OPTIMIZATION

A SHIFTED PRIMAL-DUAL PENALTY-BARRIER METHOD FOR NONLINEAR OPTIMIZATION

An Inexact Sequential Quadratic Optimization Method for Nonlinear Optimization

A PRIMAL-DUAL TRUST REGION ALGORITHM FOR NONLINEAR OPTIMIZATION

CONVEXIFICATION SCHEMES FOR SQP METHODS

What s New in Active-Set Methods for Nonlinear Optimization?

Infeasibility Detection and an Inexact Active-Set Method for Large-Scale Nonlinear Optimization

5 Handling Constraints

A Primal-Dual Augmented Lagrangian Penalty-Interior-Point Filter Line Search Algorithm

Part 5: Penalty and augmented Lagrangian methods for equality constrained optimization. Nick Gould (RAL)

Algorithms for Constrained Optimization

AN AUGMENTED LAGRANGIAN AFFINE SCALING METHOD FOR NONLINEAR PROGRAMMING

Algorithms for constrained local optimization

Constrained optimization: direct methods (cont.)

A PRIMAL-DUAL AUGMENTED LAGRANGIAN

A Primal-Dual Interior-Point Method for Nonlinear Programming with Strong Global and Local Convergence Properties

MS&E 318 (CME 338) Large-Scale Numerical Optimization

Scientific Computing: An Introductory Survey

Scientific Computing: An Introductory Survey

Some new facts about sequential quadratic programming methods employing second derivatives

Penalty and Barrier Methods. So we again build on our unconstrained algorithms, but in a different way.

Numerical Methods for PDE-Constrained Optimization

CONVERGENCE ANALYSIS OF AN INTERIOR-POINT METHOD FOR NONCONVEX NONLINEAR PROGRAMMING

minimize x subject to (x 2)(x 4) u,

An Inexact Newton Method for Nonlinear Constrained Optimization

1. Introduction. We consider nonlinear optimization problems of the form. f(x) ce (x) = 0 c I (x) 0,

Outline. Scientific Computing: An Introductory Survey. Optimization. Optimization Problems. Examples: Optimization Problems

Preprint ANL/MCS-P , Dec 2002 (Revised Nov 2003, Mar 2004) Mathematics and Computer Science Division Argonne National Laboratory

An Active Set Strategy for Solving Optimization Problems with up to 200,000,000 Nonlinear Constraints


CONSTRAINED NONLINEAR PROGRAMMING

Written Examination

arxiv: v1 [math.oc] 20 Aug 2014

An Inexact Newton Method for Optimization

Adaptive augmented Lagrangian methods: algorithms and practical numerical experience

4TE3/6TE3. Algorithms for. Continuous Optimization

SF2822 Applied Nonlinear Optimization. Preparatory question. Lecture 9: Sequential quadratic programming. Anders Forsgren

1 Computing with constraints

Inexact Newton Methods and Nonlinear Constrained Optimization

N. L. P. NONLINEAR PROGRAMMING (NLP) deals with optimization models with at least one nonlinear function. NLP. Optimization. Models of following form:

Optimality Conditions for Constrained Optimization

A globally and quadratically convergent primal dual augmented Lagrangian algorithm for equality constrained optimization

MODIFYING SQP FOR DEGENERATE PROBLEMS

INTERIOR-POINT METHODS FOR NONCONVEX NONLINEAR PROGRAMMING: CONVERGENCE ANALYSIS AND COMPUTATIONAL PERFORMANCE

Key words. Nonconvex quadratic programming, active-set methods, Schur complement, Karush- Kuhn-Tucker system, primal-feasible methods.

Constrained Optimization Theory

UNDERGROUND LECTURE NOTES 1: Optimality Conditions for Constrained Optimization Problems

Part 4: Active-set methods for linearly constrained optimization. Nick Gould (RAL)

A New Penalty-SQP Method

E5295/5B5749 Convex optimization with engineering applications. Lecture 8. Smooth convex unconstrained and equality-constrained minimization

10 Numerical methods for constrained problems

Constrained Optimization

PDE-Constrained and Nonsmooth Optimization

Examination paper for TMA4180 Optimization I

2.3 Linear Programming

Lagrange Duality. Daniel P. Palomar. Hong Kong University of Science and Technology (HKUST)

Convex Optimization M2

On the Local Quadratic Convergence of the Primal-Dual Augmented Lagrangian Method

Additional Exercises for Introduction to Nonlinear Optimization Amir Beck March 16, 2017

Lecture 13: Constrained optimization

E5295/5B5749 Convex optimization with engineering applications. Lecture 5. Convex programming and semidefinite programming

Lectures 9 and 10: Constrained optimization problems and their optimality conditions

A globally convergent Levenberg Marquardt method for equality-constrained optimization

A SECOND DERIVATIVE SQP METHOD : LOCAL CONVERGENCE

SEQUENTIAL QUADRATIC PROGAMMING METHODS FOR PARAMETRIC NONLINEAR OPTIMIZATION

Numerical Optimization

Optimality, Duality, Complementarity for Constrained Optimization

5.6 Penalty method and augmented Lagrangian method

A null-space primal-dual interior-point algorithm for nonlinear optimization with nice convergence properties

AN INTERIOR-POINT METHOD FOR NONLINEAR OPTIMIZATION PROBLEMS WITH LOCATABLE AND SEPARABLE NONSMOOTHNESS

Generalization to inequality constrained problem. Maximize

Structural and Multidisciplinary Optimization. P. Duysinx and P. Tossings

A Trust Funnel Algorithm for Nonconvex Equality Constrained Optimization with O(ɛ 3/2 ) Complexity

Part 3: Trust-region methods for unconstrained optimization. Nick Gould (RAL)

Interior Methods for Mathematical Programs with Complementarity Constraints

arxiv: v1 [math.na] 8 Jun 2018

Lecture 3. Optimization Problems and Iterative Algorithms

Lecture Note 5: Semidefinite Programming for Stability Analysis

Sequential Quadratic Programming Method for Nonlinear Second-Order Cone Programming Problems. Hirokazu KATO

Motivation. Lecture 2 Topics from Optimization and Duality. network utility maximization (NUM) problem:

Computational Finance

5. Duality. Lagrangian

Unconstrained optimization

Optimization Problems with Constraints - introduction to theory, numerical Methods and applications

The Squared Slacks Transformation in Nonlinear Programming

Numerisches Rechnen. (für Informatiker) M. Grepl P. Esser & G. Welper & L. Zhang. Institut für Geometrie und Praktische Mathematik RWTH Aachen

On the use of piecewise linear models in nonlinear programming

HYBRID FILTER METHODS FOR NONLINEAR OPTIMIZATION. Yueling Loh

Lecture 3: Lagrangian duality and algorithms for the Lagrangian dual problem

MATH 5720: Unconstrained Optimization Hung Phan, UMass Lowell September 13, 2018

Extreme Abridgment of Boyd and Vandenberghe s Convex Optimization

Constrained Optimization and Lagrangian Duality

Lecture 11 and 12: Penalty methods and augmented Lagrangian methods for nonlinear programming

Iterative Reweighted Minimization Methods for l p Regularized Unconstrained Nonlinear Programming

Nonlinear Programming, Elastic Mode, SQP, MPEC, MPCC, complementarity

Transcription:

A STABILIZED SQP METHOD: GLOBAL CONVERGENCE Philip E. Gill Vyacheslav Kungurtsev Daniel P. Robinson UCSD Center for Computational Mathematics Technical Report CCoM-13-4 Revised July 18, 2014, June 23, 2015 Abstract Stabilized sequential quadratic programming (SQP) methods for nonlinear optimization are designed to provide a sequence of iterates with fast local convergence regardless of whether or not the active-constraint gradients are linearly dependent. This paper concerns the global convergence properties of a stabilized SQP method with a primal-dual augmented Lagrangian merit function. The proposed method incorporates several novel features. (i) A flexible line search is used based on a direction formed from an approximate solution of a strictly convex QP subproblem and, when one exists, a direction of negative curvature for the primaldual merit function. (ii) When certain conditions hold, an approximate QP solution is computed by solving a single linear system defined in terms of an estimate of the optimal active set. The conditions exploit the formal equivalence between the conventional stabilized SQP subproblem and a bound-constrained QP associated with minimizing a quadratic model of the merit function. (iii) It is shown that with an appropriate choice of termination condition, the method terminates in a finite number of iterations without the assumption of a constraint qualification. The method may be interpreted as an SQP method with an augmented Lagrangian safeguarding strategy. This safeguarding becomes relevant only when the iterates are converging to an infeasible stationary point of the norm of the constraint violations. Otherwise, the method terminates with a point that approximately satisfies certain second-order necessary conditions for optimality. In this situation, if all termination conditions are removed, then the limit points either satisfy the same second-order necessary conditions exactly or fail to satisfy a wea second-order constraint qualification. (iv) The global convergence analysis concerns a specific algorithm that estimates the least curvature of the merit function at each step. If negative curvature directions are omitted, the analysis still applies and establishes convergence to either firstorder solutions or infeasible stationary points. The superlinear convergence of the iterates and the formal local equivalence to stabilized SQP is established in a companion paper (Report CCoM 14-01, Center for Computational Mathematics, University of California, San Diego, 2014). Key words. Nonlinear programming, augmented Lagrangian, sequential quadratic programming, SQP methods, stabilized SQP, primal-dual methods, second-order optimality. Department of Mathematics, University of California, San Diego, La Jolla, CA 92093-0112 (pgill@ucsd.edu). Research supported by National Science Foundation grant DMS-1318480. Agent Technology Center, Department of Computer Science, Faculty of Electrical Engineering, Czech Technical University in Prague. (vyacheslav.ungurtsev@fel.cvut.cz) Research supported by the European social fund within the framewor of realizing the project Support of inter-sectoral mobility and quality enhancement of research teams at the Czech Technical University in Prague, CZ.1.07/2.3.00/30.0034. Department of Applied Mathematics and Statistics, Johns Hopins University, Baltimore, MD 21218-2682 (daniel.p.robinson@jhu.edu). Research supported in part by National Science Foundation grant DMS- 1217153. 1

A Stabilized SQP Method: Global Convergence 2 1. Introduction The nonlinear optimization problem under consideration has the form (NP) minimize f(x) subject to c(x) = 0, x 0, x R n where c: R n R m and f : R n R are twice-continuously differentiable. This problem format assumes that all general inequality constraints have been converted to equalities by the use of slac variables. Methods for solving problem (NP) are easily extended to the more general setting with l x u. For problem (NP), the vector g(x) is used to denote f(x), the gradient of f at x. The matrix J(x) denotes the m n constraint Jacobian, which has ith row c i (x) T, the gradient of the ith constraint function c i at x. The Lagrangian associated with (NP) is L(x, y, z) = f(x) c(x) T y z T x, where y and z are m- and n-vectors of dual variables associated with the equality constraints and nonnegativity constraints, respectively. The Hessian of the Lagrangian with respect to x is denoted by H(x, y) = 2 f(x) m i=1 y i 2 c i (x). Sequential quadratic programming (SQP) methods are an important class of methods for nonlinearly constrained optimization (for a survey, see, e.g., [8, 21]). The th iteration of a conventional SQP method involves the solution of a quadratic programming (QP) subproblem in which a local quadratic model of the Lagrangian is minimized subject to the linearized constraints. In this paper we focus on the properties of an SQP method that uses a merit function to ensure global convergence. In this case, a scalar-valued merit function is used to provide a measure of the quality of a given point as an estimate of a solution. Stabilized SQP methods are designed to resolve some of the numerical and theoretical difficulties associated with SQP methods when they are applied to ill-posed or degenerate nonlinear problems (see, e.g., Hager [22], Wright [35 37], Oberlin and Wright [31], Fernández and Solodov [15], Izmailov and Solodov [23]). Broadly speaing, stabilized SQP methods are designed to provide a sequence of iterates with fast local convergence regardless of whether or not the active-constraint gradients are linearly dependent. Given an estimate (x, y ) of a primal-dual solution (x, y ) of problem (NP), stabilized SQP methods compute a new primal-dual estimate by solving a QP subproblem of the form minimize g(x ) T (x x ) + 1 x,y 2 (x x ) T H(x, y )(x x ) + 1 2 µ y 2 subject to c(x ) + J(x )(x x ) + µ(y y ) = 0, x 0, (1.1) where µ (µ > 0) is a small scalar parameter. There are two important issues associated with the formulation of a practical stabilized SQP method. The first is that stabilized SQP methods have no global convergence theory. This implies that stabilized methods must start by solving the QP subproblem associated with a conventional globally convergent SQP method and switch to the stabilized QP subproblem (1.1) when it is determined that the iterates are in the proximity of a solution. This strategy may require several switches between a conventional and a stabilized SQP method before the neighborhood of a solution is identified correctly. The second issue concerns the assumptions needed to guarantee a fast local convergence rate. In general, the Hessian of the Lagrangian is not positive definite in the neighborhood of a solution, which implies that the QP problem (1.1) may be nonconvex and may have local or unbounded solutions. Nonconvex QP is NP-hard even for the calculation of a local minimizer [10, 17]. When establishing the local convergence rate of stabilized SQP methods, the potential nonconvexity of the QP subproblem implies that assumptions must be made regarding which solution of (1.1) is found. For example, some analyses require a global

A Stabilized SQP Method: Global Convergence 3 solution of (1.1) (see, e.g., [24,25]), or require that the same local solution is found at each step. This paper concerns the formulation and global convergence analysis of a stabilized SQP method that uses a primal-dual augmented Lagrangian merit function to ensure convergence from an arbitrary starting point. The analysis indicates that the method has the same strong first- and second-order convergence properties that have been established for augmented Lagrangian methods, while being able to transition seamlessly to stabilized SQP with fast local convergence in the neighborhood of a solution. The principal contributions of the paper are the following. 1. The method uses a flexible line search along a direction formed from an approximate solution of a strictly convex QP subproblem and, when one exists, a direction of negative curvature for the primal-dual merit function. The superlinear convergence of the iterates and the formal local equivalence to stabilized SQP is established in a companion paper (see Gill, Kungurtsev and Robinson [18]). It is not necessary to solve an indefinite QP subproblem, and no assumptions are necessary about the quality of each subproblem solution. 2. When certain conditions hold, an approximate QP solution is computed by solving a single linear system defined in terms of an estimate of the optimal active set. These conditions may be satisfied at any iterate, but are most liely to be satisfied in the neighborhood of a solution. The conditions exploit the formal equivalence between the conventional stabilized SQP subproblem and a bound-constrained QP associated with minimizing a quadratic model of the merit function. 3. Convergence to first-order KKT points is established under weaer conditions than those assumed in [20]. It is shown that with an appropriate choice of termination condition, the method terminates in a finite number of iterations without the assumption of a constraint qualification. The method may be interpreted as a SQP method with an augmented Lagrangian safeguarding strategy. This safeguarding becomes relevant only when the iterates are converging to an infeasible stationary point of the norm of the constraint violations. Otherwise, the method terminates with a point that approximately satisfies certain second-order necessary conditions for optimality. In this situation, if all termination conditions are removed, then limit points either satisfy the same second-order necessary conditions exactly or fail to satisfy a wea second-order constraint qualification. For the main algorithm proposed in Section 2, the cost of each iteration is dominated by the cost of factoring two matrices with so-called regularized KKT structure (see Section 2). However, the algorithm is easily modified so that the negative curvature direction is omitted, in which case the line search is defined in terms of the approximate QP solution only. The resulting algorithm requires only one factorization per iteration and constitutes a globally convergent stabilized SQP method for finding first-order points. The main algorithm is intended for applications in which some guarantee of convergence to second-order solutions is needed (see, e.g., [1, 9, 12]). For other applications, the negative curvature direction may be omitted completely as described above, or may be computed in the final stages of the optimization as a chec that the iterates are converging to a point

A Stabilized SQP Method: Global Convergence 4 satisfying second-order optimality conditions. In the latter case the results of this paper and its companion paper [18] imply that a direction of negative curvature, when it is computed, may be used without impeding the overall convergence rate or preventing global convergence. The remainder of the paper is organized as follows. This section concludes with a review of the first-order optimality conditions for (NP) and the properties of the primal-dual augmented Lagrangian function. Section 2 provides the details of the second-order primal-dual stabilized SQP method. The global convergence of the method is established in Section 3. Section 4 provides final comments and conclusions. 1.1. Notation Unless explicitly indicated otherwise, denotes the vector two-norm or its induced matrix norm. The least eigenvalue of a symmetric matrix A will be denoted by λ min (A). Given vectors a and b with the same dimension, the vector with ith component a i b i is denoted by a b. Similarly, min(a, b) is a vector with components min(a i, b i ). The vectors e and e j denote, respectively, the column vector of ones and the jth column of the identity matrix I; the dimensions of e, e j and I are defined by the context. Given vectors x and y, the long vector consisting of the elements of x augmented by elements of y is denoted by (x, y). The ith component of a vector will be denoted by [ ] i, e.g., [v] i is the ith component of the vector v. For a given l-vector u and index set S, the quantity [u] S denotes the subvector of components u j such that j S {1, 2,..., l }. Similarly, if M is a symmetric l l matrix, then [M ] S denotes the symmetric matrix with elements m ij for i, j S {1, 2,..., l }. Given a sequence {α j } j 0 of scalars, vectors, or matrices, we write α j = O ( β j ) if there exists a positive constant γ such that α j γβ j for some positive scalar sequence {β j } j 0. 1.2. Bacground At the basic level, the method proceeds to compute a first-order KKT point of (NP), which is defined formally as follows. Definition 1.1. (First-order KKT point of (NP)) The vector x is called a first-order KKT point for problem (NP) if there exists a dual vector y such that r(x, y ) = 0, where r(x, y) = ( c(x), min ( x, g(x) J(x) T y )). (1.2) Any (x, y ) satisfying r(x, y ) = 0, is called a first-order KKT pair. If a suitable constraint qualification holds, then a local minimizer x for problem (NP) is a first-order KKT point. In this case, any y associated with a KKT pair (x, y ) is a vector of Lagrange multipliers for the constraints c(x) = 0. Given a first-order KKT point x, the nonempty set of dual vectors Y(x ) is defined so that Y(x ) = {y R m : (x, y) satisfies r(x, y) = 0}. (1.3) For any x 0, the active set at x is given by A(x) = {i : [x] i = 0}. (1.4)

A Stabilized SQP Method: Global Convergence 5 At a primal-dual first-order solution (x, y ), second-order conditions for (NP) may be defined in terms of a partition of A(x ). The index set of strongly active variables is given by A + (x, y ) = {i A(x ) : [g(x ) J(x ) T y ] i > 0} (1.5) and, similarly, the set of wealy active variables is A 0 (x, y ) = {i A(x ) : [g(x ) J(x ) T y ] i = 0}. (1.6) A set of second-order conditions is given in the following definition. Definition 1.2. (Second-order sufficient conditions) A primal-dual pair (x, y ) satisfies the second-order sufficient optimality conditions for problem (NP) if it satisfies the firstorder conditions r(x, y ) = 0 (cf. (1.2)), and where C(x, y ) is the critical cone p T H(x, y )p > 0 for all p C(x, y ) \ {0}, (1.7) C(x, y ) = null ( J(x ) ) {p : p i = 0 for i A + (x, y ), p i 0 for i A 0 (x, y ) }. The proposed method is based on the primal-dual augmented Lagrangian function of Gill and Robinson [20], which was first used to define a regularized SQP method in which the linear system of equations solved at each iteration of the QP subproblem are guaranteed to be nonsingular. The function is given by M(x, y ; y E, µ) = f(x) c(x) T y E + 1 2µ c(x) 2 + 1 2µ c(x) + µ(y ye ) 2, (1.8) where µ is a positive penalty parameter and y E is a Lagrange multiplier estimate. (The function M is a member of a one-parameter family of functions that includes the conventional augmented Lagrangian. For more details, see Robinson [33], and Gill and Robinson [19].) The next result, which is proved in Theorem A.1 of the Appendix, motivates the use of (1.8) as an SQP merit function. Theorem 1.1. If (x, y ) is a solution of problem (NP) that satisfies the second-order sufficient optimality conditions given by Definition 1.2, then for the choice y E = y, there exists a positive µ such that for all 0 < µ < µ, the point (x, y ) satisfies the second-order sufficient optimality conditions for the problem of minimizing the primal-dual function M(x, y ; y E, µ) of (1.8) subject to the nonnegativity constraints x 0. The result of Theorem 1.1 implies that problem (NP) may be replaced by a sequence of boundconstrained problems with objective function M(x, y ; y E, µ) defined in terms of a sequence of multiplier estimates {y E }. This approach defines an inner/outer iteration structure, with the inner iteration being those of the active-set method used to minimize a quadratic model of the merit function subject to the bound constraints. Each quadratic model is defined in terms of the gradient and Hessian of M. For given values y E and µ, the gradient and Hessian of M at (x, y) may be written in the form ( g(x) J(x) T ( π(x) + (π(x) y) ) ) M(x, y ; y E, µ) = µ(y π(x)), (1.9)

A Stabilized SQP Method: Global Convergence 6 and ( ( ) ) H x, π(x) + (π(x) y) + 2 2 M(x, y ; y E µ, µ) = J(x)T J(x) J(x) T, (1.10) J(x) µi where π(x) denotes the vector-valued function π(x ; y E, µ) = y E c(x)/µ. If v = (x, y ) denotes the th estimate of a primal-dual solution of problem (NP), one possible local quadratic model of the change in M(x, y ; y E, µ) is given by Q (v ; µ) = (v v ) T M(v ; y E, µ) + 1 2 (v v ) T B(v ; µ)(v v ), (1.11) where y E estimates a vector of Lagrange multipliers, and B(v ; µ) is defined by replacing π(x ) by y in the leading bloc of 2 M, i.e., ( H(x, y ) + 2 µ B(x, y ; µ) = J(x ) T J(x ) J(x ) T ). (1.12) J(x ) µi Both this approximation and the exact Hessian 2 M have the same characteristic doublyaugmented structure involving the term (2/µ)J(x ) T J(x ) in the leading diagonal bloc. However, unlie 2 M, the matrix B(x, y ; µ) is independent of the multiplier estimate y E (cf. (1.10)). The benefit of using B(x, y ; µ) to define the quadratic model (1.11) is that the QP subproblem minimize v Q (v ; µ) subject to [v] i 0, i = 1, 2,..., n, (1.13) is formally equivalent to the stabilized SQP subproblem minimize g T x,y (x x ) + 1 2 (x x ) T H(x, y )(x x ) + 1 2 µ y 2 subject to c(x ) + J(x )(x x ) + µ(y y E) = 0, x 0, (1.14) (see Gill and Robinson [20]). This equivalence suggests an algorithm based on minimizing (1.13) with Q defined with a small µ analogous to the stabilized SQP perturbation. However, if M(x, y ; y E, µ) is to serve as a line-search merit function, it should be defined with a value of µ that is not too small. Accordingly, different penalty parameters µ R and µ are defined, with µ R used for the definition of the local quadratic model, and µ used for the line-search merit function. In the neighborhood of a solution, µ R and µ are computed so that µr µ, which implies that µ R plays the role of the stabilization parameter µ in (1.1). 2. A Second-Order Primal-Dual Stabilized SQP Algorithm In this section we describe a method designed to compute a point satisfying certain first- and second-order conditions for a solution of (NP). At the th iterate (x, y ) a flexible line search is performed along a direction formed from two primal-dual directions d and s, either of which may be zero. A nonzero d is a descent direction for the merit function; a nonzero s is a direction of negative curvature for the quadratic model Q (v ; µ R 1 ). The descent direction d is either a local descent step designed to facilitate a fast rate of local convergence, or a global descent step designed to encourage global convergence. In each case, d is an approximate solution of a QP subproblem with bound constraints.

A Stabilized SQP Method: Global Convergence 7 Overview of the step computation. If a solution of the subproblem (1.13) is written in terms of the step d from the base point v, then the QP optimality conditions are given by [ Q (v +d ; µ R )] F = 0, [ Q (v +d ; µ R )] A 0, and [v +d ] i 0 for i = 1, 2,..., n, (2.1) where A is the active set (1.4) and F is the set of free variables associated with the primal-dual QP subproblem, i.e., F(x) = {1, 2,..., n + m} \ A(x). (2.2) The local descent step is the solution of an equality-constrained QP subproblem defined by relaxing the optimality conditions (2.1). (See Section 2.3 for more details.) The global descent step is the solution of a convex inequality constrained QP subproblem. If v = (x, y ) is not close to a solution, the matrix B(v ; µ R ) may not be positive definite and (1.13) is not an appropriate QP subproblem for the global descent step. Instead, a strictly convex boundconstrained QP subproblem is defined that preserves the positive curvature of the nonconvex quadratic model (1.11) on the reduced space associated with the free variables. The strictly convex QP is defined in terms of the matrix B(v ; µ) such that ) (Ĥ(x, y ) + B(x 2, y ; µ) = µ J(x ) T J(x ) J(x ) T (2.3) J(x ) µi with the symmetric matrix Ĥ(x, y ) defined so that Ĥ(x, y )+(1/µ)J(x ) T J(x ) is positive definite. With this definition, B(x, y ; µ) is equal to B(x, y ; µ) when B(x, y ; µ) is sufficiently positive definite. A method for computing a suitable symmetric (but not necessarily positive definite) Ĥ(x, y ) is given by Gill and Robinson [20, Section 4]. Once B(x, y ; µ) has been computed, the global descent step is computed by solving the QP subproblem minimize Q (v) = (v v ) T M(v v ; y E, µr ) + 1 2 (v v ) T B(v ; µ R )(v v ) subject to [v] i 0, i = 1, 2,..., n. (2.4) Throughout what follows, the unique primal-dual solution of this QP subproblem is denoted by v, which has primal and dual components v = ( x, ŷ ). The global descent step is then computed as d = v v = ( x x, ŷ y ). For numerical stability, it is important that all computations be performed without the explicit calculation of the matrices 2 M or B. The relevant properties of B may be determined by exploiting the relationship between the three matrices ( H(x, y) + 2 µ J(x)T J(x) J(x) T J(x) µi ), H(x, y) + 1 µ J(x)T J(x) and ( H(x, y) J(x) T which are said to have doubly-augmented form, augmented Lagrangian form and regularized KKT form, respectively. These matrices are related via the matrix identities ( H + 1+ν µ J T J νj T ) ( I 1 = µ J T ) ( H + 1 µ J T ) ( ) J 0 I 0 1 νj νµi 0 I 0 νµi µ J I (2.5) ( I 1+ν = νµ J T ) ( ) ( ) H J T I 1 0 ν I, (2.6) J µi I J(x) µi ),

A Stabilized SQP Method: Global Convergence 8 which hold for all positive ν (see, e.g., Forsgren and Gill [16], Robinson [33], and Gill and Robinson [20]). For simplicity of exposition, the formulation and analysis of the main algorithm is given in terms of matrices in augmented Lagrangian form. However, in practice, all computations are performed by factoring matrices in regularized KKT form. At no point is it necessary to form a matrix in doubly-augmented or augmented Lagrangian form. The term regularized KKT form stems from the role of µ as a regularization parameter for the matrix in (2.6). (For other examples of this type of regularization, see [2, 20, 34].) In practice, the global descent step is found by solving the stabilized QP form (1.14) of the bound-constrained subproblem with H(x, y ) = Ĥ(x, y ). Gill and Robinson [20, Theorem 5.2] show that the application of a conventional active-set method to (1.14) involves the solution of a sequence of linear equations in regularized KKT form. At the start of an outer iteration, the optimal active set of problem (NP) is estimated by an ɛ-active set A ɛ defined in terms of a positive scalar ɛ that reflects the distance of (x, y) to a first-order optimal point of problem (NP). The ɛ-active set is defined as A ɛ (x, y, µ) = { i : x i ɛ, with ɛ min ( ɛ a, max ( µ, r(x, y) γ) )}, (2.7) where γ and ɛ a are fixed scalars satisfying 0 < γ < 1 and 0 < ɛ a < 1, and the function r(x, y) is defined by (1.2). Similarly, the ɛ-free set is the complement of A ɛ in {1, 2,..., n + m}, i.e., F ɛ (x, y, µ) = {1, 2,..., n + m} \ A ɛ (x, y, µ), (2.8) (cf. (2.2)). The ɛ-free set defines the composition of certain matrices used to compute a direction of negative curvature for Q (v ; µ R 1 ), and the estimate of the active set for the convex QP subproblem. The following simple argument shows that F(x ) F ɛ (x, y, µ) when µ is sufficiently small and (x, y) is close to a first-order KKT pair (x, y ). If i A(x ), then x i > 0. Assume that (x, y) is sufficiently close to (x, y ) and µ is sufficiently small that: (i) r(x, y) γ < µ x i ; (ii) x i x i < x i µ; and (iii) µ ɛ a. Then A ɛ (x, y, µ) = { i : x i min ( ɛ a, max ( µ, r(x, y) γ) )} = { i : x i µ }, and the inequality x i x i < x i µ implies x i > µ, in which case i A ɛ(x, y, µ). We have shown that if i A(x ) then i A ɛ (x, y, µ), which implies that F(x ) F ɛ (x, y, µ), as required. Summary of the main algorithm. The computation associated with the th iteration of the main algorithm (Algorithm 5 of Section 2.5) is arranged into five principal steps. 1. Given (x, y ) and the regularization parameter µ R 1 from the previous iteration, compute F ɛ (x, y, µ R 1 ) and B(x, y ; µr 1 ). Compute the nonnegative scalar ξ and vector s (1) such that, if ξ > 0, then ( ξ, s (1) ) approximates the most negative or least eigenpair of B(x, y ; µ R 1 ) (see Section 2.1). If this computation gives ξ = 0, then s (1) = 0. If B(x, y ; µ R 1 ) is sufficiently positive definite then (ξ, s(1) ) = 0. (See Algorithm 1.) 2. Use ξ and r(x, y ) to compute y E and µr for the th iteration (see Section 2.2).

A Stabilized SQP Method: Global Convergence 9 3. Compute the positive-definite matrix B(x, y ; µ R ) such that B(x, y ; µ R ) = B(x, y ; µ R ) if B(x, y ; µ R ) is positive definite. Compute d = (p, q ) as either a local or global descent step. The vector p of primal components of d satisfies x + p 0. (See Section 2.3 and Algorithm 2.) 4. Compute a direction of negative curvature s = (u, w ) by rescaling the direction s (1). The vector u of primal components of s satisfies x + p + u 0. (See Section 2.3 and Algorithm 3.) 5. Perform a flexible line search along the vector v = s + d = (u + p, w + q ). (See Section 2.4 and Algorithm 4.) Update the line-search penalty parameter µ. 2.1. Computing a direction of negative curvature for Q The values of the regularization parameter µ R and multiplier estimate ye for the th iteration depend on an estimate of the smallest eigenvalue of B Fɛ (x, y ; µ R 1 ) computed from the ɛ- free rows and columns of the approximate Hessian B(x, y ; µ R 1 ) (1.12). This estimate may ( be computed in terms of an estimate of λ min HFɛ + (1/µ R 1 )J T J Fɛ Fɛ), where HFɛ is the matrix of ɛ-free rows and columns of H(x, y ), and J Fɛ is the matrix of ɛ-free columns of J(x ). Algorithm 1 summarizes the principal calculations associated with finding a nonnegative estimate ξ of max{0, λ min }, where λ min denotes the least eigenvalue of H Fɛ +(1/µ R 1 )J T J. Fɛ Fɛ If the computed ξ is positive, then a by-product of the computation is a feasible approximate direction of negative curvature for the merit function M(x, y ; y E, µr 1 ). This direction is combined with a descent direction to form the direction of search for the line search (see Section 2.3). Algorithm 1 Least curvature estimate of Q. 1: procedure LEAST CURVATURE ESTIMATE(x, y, µ R 1, J, H ) 2: Compute H Fɛ and J Fɛ as submatrices of H and J associated with F ɛ (x, y, µ R 1 ); 3: if H Fɛ + (1/µ R 1 )J T J is positive semidefinite then Fɛ Fɛ 4: ξ = 0; u (1) = 0; w (1) = 0; 5: else 6: Compute u Fɛ 0 such that (2.9) is satisfied; 7: ξ = u T Fɛ( HFɛ + (1/µ R 1 )J T J Fɛ Fɛ) ufɛ / u Fɛ 2 > 0; 8: u (1) = 0; [u (1) ] Fɛ = u Fɛ ; 9: w (1) = (1/µ R 1 )J u(1) ; 10: end if 11: s (1) = (u (1), w(1) ); 12: return (s (1), ξ ); 13: end procedure Algorithm 1 assumes that a procedure is available for determining whether or not H Fɛ + (1/µ R 1 )J T J is positive semidefinite. If H Fɛ Fɛ Fɛ + (1/µ R 1 )J T J is not positive semidefinite, Fɛ Fɛ the procedure must provide a direction u Fɛ satisfying u T Fɛ ( H Fɛ + 1 ( J )u T µ J R Fɛ Fɛ Fɛ θλ min H Fɛ + 1 ) J T 1 µ J u R Fɛ Fɛ Fɛ 2 < 0, (2.9) 1

A Stabilized SQP Method: Global Convergence 10 where θ is a positive scalar that is independent of x and y. Several alternative methods may be used to find the vector u Fɛ, but the details of the computation are not relevant for the proof of global convergence. A specific procedure appropriate for a matrix in regularized KKT form is given by Forsgren and Gill [16, Section 4.3]. The inequality (2.9) allows the definition of a direction of negative curvature for the quadratic model Q (v ; µ R 1 ) when H Fɛ + (1/µ R 1 )J T J is not positive semidefinite (see Lemma 2.1 below). Fɛ Fɛ The vector w (1) computed in Step 9 of Algorithm 1 is the change in the dual variables as a function of the change u (1) in the primal variables. The definition of w (1) ensures that J u (1) + µ R 1 w(1) = 0, which implies that the step (x + u (1), y + w(1) ) does not change the residuals of the constraints of the stabilized QP subproblem (1.14). It also facilitates a proof that the resulting direction s (1) is a direction of negative curvature for the quadratic model Q (v ; µ R 1 ) when ξ > 0 (see Lemma 2.1). The next lemma gives the properties of the quantities computed by Algorithm 1. Lemma 2.1. Suppose that ξ and s (1) = (u (1), w(1) ) are computed by Algorithm 1, and that the matrix H Fɛ + (1/µ R 1 )J T J is not positive semidefinite. Then Fɛ Fɛ (1) u (1) 0, s (1) 0, ξ > 0; and (2) s (1) is a direction of negative curvature for the quadratic model Q (v ; µ R 1 ) in (1.11). In particular, s (1) satisfies s (1)T (1/µ R 1 )J T J Fɛ Fɛ) with θ the value associated with the bound (2.9). B(v ; µ R 1 )s(1) θ u (1) 2 < 0, where θ = θλ min ( HFɛ + Proof. To simplify the notation, the suffix is omitted and the quantities µ R 1, θ, s (1) u (1), w(1), ξ, J(x ), H(x, y ), and B(v ; µ R 1 ) are denoted by µ, θ, s, u, w, ξ, J, H and B, respectively. As H Fɛ + (1/µ)J T J is not positive semidefinite by assumption, Algorithm 1 computes a Fɛ Fɛ nonzero direction u Fɛ that satisfies (2.9). In this case, u 0, s 0, λ min (H Fɛ + 1 ) µ J T J < 0, and ξ > 0, (2.10) Fɛ Fɛ which proves part (1). The definition of B implies that s T Bs = ( ) T ( u H + 2 w µ J T J J T J µi ) ( ) u w = u T Hu + 2 µ ut J T Ju + 2u T J T w + µ w 2 = u T( H + 1 µ J T J, ) u. (2.11) The definition of u in Step 8 of Algorithm 1, and the requirement that u Fɛ satisfies (2.9) yield the inequality u T( H + 1 ( µ J T J )u θλ min H Fɛ + 1 ) µ J T J u 2. (2.12) Fɛ Fɛ Given the definition θ = θλ min (H Fɛ + 1 µ J T Fɛ J Fɛ ), it must hold that θ < 0, where the strict inequality follows from (2.10). The result now follows directly from (2.11), (2.12), and the definition u = u (1) as a nonzero vector in Algorithm 1.

A Stabilized SQP Method: Global Convergence 11 2.2. Updating the multiplier estimate and regularization parameter The multiplier estimate y E is updated to be the current dual vector y as long as there is improvement in some measure of the distance to a primal-dual second-order solution (x, y ). The main algorithm (Algorithm 5 of Section 2.5) uses the feasibility and optimality measures η(x ) and ω(x, y, ξ ) such that η(x ) = c(x ) and ω(x, y, ξ ) = max ( min(x, g(x ) J(x ) T y ), ξ ), (2.13) where the quantity ξ is computed by Algorithm 1. Given η(x ) and ω(x, y, ξ ), weighted combinations of the feasibility and optimality measures are computed as φ V (x, y ) = η(x ) + βω(x, y, ξ ) and φ O (x, y, ξ ) = βη(x ) + ω(x, y, ξ ), where β is a fixed scalar such that 0 < β 1. The subscripts on φ V and φ O reflect the preference given to reducing the of violation measure or the optimality measure, as defined by the placement of the parameter β. The update y E = y is performed if φ V (v ) 1 2 φmax V or φ O (v, ξ ) 1 2 φmax O, (2.14) in which case the th iterate is called a V-iterate or an O-iterate, respectively. A V-Oiterate is any point for which one or both of these conditions holds. The associated iteration (or iteration index) is called a V-O iteration. The quantities φ max V and φ max O are positive bounds that are reduced during the solution process. At a V-O-iterate, the regularization parameter is updated as { ( min µ R = µ R 0, max ( ) γ r, ξ ), if max ( ) r, ξ > 0; 1 2 µr 1, otherwise, (2.15) for some fixed γ (0, 1), and where r r(x, y ) is the first-order optimality measure (1.2). Note that the definition (2.15) implies that the sequence {µ R } is not necessarily monotonic. If the conditions for a V-O-iterate do not hold, (x, y ) is checed to determine if it is an approximate second-order solution of the bound-constrained problem minimize x,y M(x, y ; y 1 E, µr 1 ) subject to x 0. (2.16) Specifically, (x, y ) is tested using the conditions min(x, x M(x, y ; y E 1, µr 1 )) τ 1, y M(x, y ; y E 1, µr 1 ) τ 1 µr 1, and ξ τ 1, (2.17a) (2.17b) (2.17c) where τ 1 is a positive tolerance. If these conditions are satisfied, then (x, y ) is called an M-iterate and the parameters are updated as in a conventional augmented Lagrangian method; i.e., the multiplier estimate y 1 E is replaced by the safeguarded value y E = max ( y max e, min( y, y max e ) ) (2.18)

A Stabilized SQP Method: Global Convergence 12 for some large positive constant y max, and the new regularization parameter is given by { µ R = min ( 1 2 µr 1, max ( ) γ ) r, ξ, if max(r, ξ ) > 0; 1 2 µr 1, otherwise. (2.19) In addition, a new tolerance τ is computed such that τ = 1 2 τ 1. Numerical results given by Gill, Kungurtsev and Robinson [18] indicate that M-iterates occur infrequently relative to the total number of iterations. Finally, if neither (2.14) nor (2.17) are satisfied, then y E = ye 1 and µr = µr 1. As the multiplier estimates and regularization parameter are fixed at their current values in this case, the th iterate is called an F-iterate. 2.3. Definition of the line-search direction Once µ R and ye have been computed, the line-search directions s and d are computed. This section provides more details of the computation of the local and global descent directions, and direction of negative curvature introduced in the overview of the step computation. The computations are summarized in Algorithms 2 and 3 below. The local descent direction. A local descent direction is computed if B Fɛ (x, y ; µ R ) is positive definite and v is a V-O-iterate (in which case y E = y ). Under these conditions, d = v v, where v is the unique solution of the QP subproblem minimize v Q (v ; µ R ) subject to [v] Aɛ = 0, (2.20) where Q (v ; µ R ) is the quadratic model (1.11). Given a feasible point v(0) such that [ v (0) ] = 0, and [ v(0) Aɛ ] = [v Fɛ ], (2.21) Fɛ the vector v is computed in the form v (0) + v (0), where [ v(0) ] = 0 and [ v(0) Aɛ ] satisfy Fɛ the equations B Fɛ [ v (0) ] = [ Q Fɛ ( v(0) ; µr )]. This direction is used in the line search if Fɛ it satisfies the conditions v + d feasible, [ Q (v + d ; µ R )] t Aɛ e, and M T d < 0, (2.22) where t is a small positive scalar computed in Algorithm 2. The condition on the ɛ-active components of Q relaxes the analogous gradient condition on the active components of Q in (2.1) associated with solving (1.13), and serves to reduce the possibility of unnecessary QP iterations in the neighborhood of a solution. In the companion paper [18, Theorem 3.14] it is shown that, in the limit, a local descent direction is used at every iteration, and that the iterates are equivalent to those of a conventional stabilized SQP method. These properties facilitate a proof of superlinear convergence under mild assumptions. If the conditions for computing the local descent step are not satisfied, or the local descent step does not satisfy the conditions in (2.22), then a global descent direction is found by solving the convex QP problem (2.4). If the local descent step has been computed, then the QP solver for the global descent step can reuse the factorization needed to solve the equations for [ v (0) ] Fɛ.

A Stabilized SQP Method: Global Convergence 13 The global descent direction. The global descent direction is d = v v, where v = ( x, ŷ ) is the unique solution of the strictly convex QP problem (2.4). This subproblem may be solved using a conventional active-set algorithm in which each iteration requires the solution of a linear system of equations with matrix in regularized KKT form. Given an initial feasible point v (0) for problem (2.4), active-set methods generate a feasible sequence { v (j) } j>0 such that Q ( v (j) ) Q ( v (j 1) ) and v (j) minimizes Q (v) on a woring set W j of the nonnegativity constraints. The first woring set W 0 is the ɛ-active set A ɛ (x, y, µ R ), which defines the initial feasible point v (0) as in (2.21). Algorithm 2 Computation of the primal-dual descent direction d. 1: procedure DESCENT DIRECTION(x, y, y E, µr, J, H ) 2: Constants: 0 < λ < min{γ, 1 γ} < 1; 3: B = B(x, y ; µ R ); Compute a positive-definite B from B; 4: M = M(x, y ; y E, µr ); t = r(x, y ) λ ; = 0; [ v(0) ] = [v Fɛ ] ; Fɛ 6: if (B Fɛ is positive definite and v is a V-O-iterate) then 5: [ v (0) ] Aɛ 7: [ v (0) ] = 0; Solve B Aɛ Fɛ [ v(0) ] = [ Q Fɛ ( v(0) ; µr )] ; v Fɛ = v(0) + v (0) ; 8: d = v v ; 9: if (v + d is feasible and M Td < 0 and [ Q (v + d ; µ R )] t Aɛ e) then 10: return d ; [local descent direction] 11: end if 12: end if 13: [ v (0) ] Aɛ = 0; Solve B Fɛ [ v (0) ] Fɛ = [ Q ( v (0) )] Fɛ ; 14: Compute α 0 0 and v (1) such that v (1) = v (0) + α 0 v (0) is feasible for the bounds; 15: Solve the convex QP (2.4) for v, starting at v (1) ; 16: d = v v ; [global descent direction] 17: return d ; 18: end procedure The direction of negative curvature. The nonzero vector s (1) = (u (1), w(1) ) computed as a by-product of the computation of ξ in Algorithm 1 is a direction of negative curvature for the quadratic model Q (v ; µ R 1 ). In Algorithm 3, s(1) is rescaled to provide the direction s used in the line search. First, an intermediate direction s (2) is computed such that { s (2) (u (2), w(2) ) = (u (1), w(1) ), if M(v ; ye, µr )T s (1) > 0; (u (1), w(1) ), otherwise, (2.23) which implies that s (2) is a non-ascent direction for M(v ; y E, µr ) at v = (x, y ). The direction s (2) is then scaled again by a positive σ to give a direction s such that: (i) v + s + d is feasible for the nonnegativity constraints, and (ii) the primal portions of s and d have comparable norms. (Only the primal part of s is influential in the scaling because the dual part w (1) of the base vector s (1) = (u (1), w(1) ) is a function of the primal part u(1), see Section 2.1.) The choice of σ not only ensures that s and d have approximately equal weight in the definition of v + s + d, but also gives a sensibly scaled s when s (1) is arbitrarily

A Stabilized SQP Method: Global Convergence 14 large or small. As only the primal variables are subject to nonnegativity constraints, the scalar σ must be chosen to satisfy [x + σ u (2) + p ] i = [x + σ u (2) + x x ] i = [ x + σ u (2) ] i 0, where x is the primal solution of the subproblem (2.4). The value of σ is chosen as large as possible subject to the restriction that the resulting primal step σ u (2) is no greater than the primal step x x (= p ) from the QP subproblem. The curvature estimate ξ computed by Algorithm 1 is included in the bound to give an appropriately scaled direction when p is small or zero. These considerations lead to the definition { } σ = argmax σ : x + σu (2) 0, σu (2) max(ξ, x x ). (2.24) σ 0 The direction of negative curvature used in the line search is then s (u, w ) = σ (u (2) The computation of the vector s is summarized in Algorithm 3., w(2) ) = σ s(2). (2.25) Algorithm 3 Feasible direction of negative curvature for the merit function. 1: procedure NEGATIVE CURVATURE DIRECTION(s (1) {, ξ, x, d, J, H ) 2: s (2) s (1) =, if M(v ; ye, µr )T s (1) > 0; s (1), otherwise; 3: Let p and u (2) be the first n components of d and s (2) { ; 4: σ = argmax σ 0 σ : x + p + σu (2) 0, σu (2) max(ξ, p )} ; 5: s = σ s (2) ; [scaled curvature direction] 6: return s ; 7: end procedure The direction s is zero if no direction of negative curvature exists for Q, in which case the value of ξ computed by Algorithm 1 is zero. However, the next result considers one important situation in which σ is positive and s is nonzero. Lemma 2.2. If d = 0 and ξ > 0, then σ > 0 and s 0. Proof. A positive ξ in Algorithm 1 implies that the vectors u Fɛ, u (1) and u (2) are nonzero. This gives a nonzero s (1) regardless of the value of w (1). The definition of s(2) in (2.23) can only change the sign of u (1), and hence s(2) must be nonzero also. It remains to show that σ is positive. The assumption that d is zero implies that x, the solution of the QP subproblem (1.14), is identical to x and x x is zero. The positive value of ξ provides a strictly positive upper bound on σ that will be achieved provided there is no index i F ɛ (x, y, µ R 1 ) such that [x ] i = 0. This is assured by the definition of the ɛ-active set from (2.7), and the fact that µ R 1 > 0.

A Stabilized SQP Method: Global Convergence 15 2.4. The flexible line search and penalty-parameter update The flexible line search defined in Algorithm 4 is a generalization of the line search used in the first-order primal-dual method of Gill and Robinson [20]. (The idea of a flexible line search was proposed by Curtis and Nocedal [13] in the context of minimizing an l 1 penalty function.) Given a primal-dual search direction v = d + s, and a line-search penalty parameter µ, consider the univariate function Ψ (α ; µ) = M(v +α v ; y E, µ). (The multiplier estimate y E remains fixed during the line search and is omitted as a parameter of Ψ.) The line search is based on approximating Ψ (α ; µ) by the line-search model function ψ (α ; µ, l ) = Ψ (0 ; µ) + αψ (0 ; µ) + 1 2 (l 1)α 2 min ( 0, v T B(v ; µ R 1 ) v ), (2.26) where Ψ denotes the derivative with respect to α. The scalar l is either 1 or 2, depending on the required order of the line-search model. The value l = 1 implies that Ψ is an affine function, which gives a first-order line-search model. The value l = 2 defines a quadratic Ψ and second-order line-search model. The choice of line-search model for Algorithm 4 is motivated by superlinear convergence considerations. In particular, it allows the acceptance of a unit step in the neighborhood of a solution, which is a ey result in establishing the asymptotic equivalence between Algorithm 5 and a conventional stabilized SQP method (see [18, Theorem 3.12]). A conventional line search requires that the step α produces a ratio of the actual and predicted reduction in Ψ that is at least γ s, where γ s is a scalar satisfying 0 < γ s < 1. The reduction requirement may be written in terms of the Armijo condition Ψ (0 ; µ) Ψ (α ; µ) γ s ( ψ (0 ; µ, l ) ψ (α ; µ, l ) ). The flexible line search defined in Algorithm 4 requires that α satisfies the modified Armijo condition Ψ (0 ; µ F ) Ψ (α ; µ F ) γ s( ψ (0 ; µ R, l ) ψ (α ; µr, l ) ) (2.27) for some value of µ F such that µf [µr, µ ]. In practice, the step may be found by reducing α by a constant factor until either ρ (α ; µ, l ) γ s or ρ (α ; µ R, l ) γ s, where ρ (α ; µ, l ) = ( Ψ (0 ; µ) Ψ (α ; µ) ) / ( ψ (0 ; µ R, l ) ψ (α ; µr, l ) ). The Armijo procedure is not executed if d = s = 0, or d = 0 and the curvature of the merit function M(v ; y E, µr ) along the vector s is not sufficiently large compared to the curvature of the quadratic model. The condition for negligible curvature involves the calculation of the scalar ξ R = st 2 M(v ; y E, µr )s / u 2 = u T ( H(x, ŷ ) + (1/µ R )J(x ) T J(x ) ) u / u 2, (2.28) with ŷ = π + (π y ) (cf. (2.11)). If the Armijo procedure is not executed (i.e., α = 0, or α = 1 with d = s = 0) then v +1 = v. In this case, it must hold that µ R < µr 1 (see Lemma 2.3(2) and Lemma 2.4(3)). The analysis will show that the property lim µ R = 0 together with the definitions of V-, O-, and M-iterates are ey in establishing the convergence results for Algorithm 5. On completion of the flexible line search, the iteration ends with the computation of the line-search penalty parameter µ +1 for the next iteration. The strategy described below is

A Stabilized SQP Method: Global Convergence 16 Algorithm 4 Flexible line search. 1: procedure FLEXIBLE LINE SEARCH(d, s, y E, µ, µr, µr 1, l, J, H ) 2: Constant: γ s (0, 1 2 ); 3: Compute M = M(x, y ; y E, µr ); 4: if s = 0 and d = 0 then 5: α = 1; 6: else if (d 0 or M T s < 0 or µ R = µr 1 ) then 7: α = 1; 8: while ρ (α ; µ R, l ) < γ s and ρ (α ; µ, l ) < γ s do 9: α = 1 2 α ; 10: end while 11: else [d = 0, s 0, ξ > 0] 12: ξ R = st 2 M(v ; y E, µr )s / u 2 ; [by definition, s = (u, w )] 13: if ξ R > γ sξ then 14: α = 1; 15: while ρ (α ; µ R, l ) < γ s and ρ (α ; µ, l ) < γ s do 16: α = 1 2 α ; 17: end while 18: else 19: α = 0; 20: end if 21: end if 22: return α 0 23: end procedure motivated by the goal of decreasing the penalty parameter only when the trial step indicates that the merit function computed with µ has not been sufficiently reduced. In particular, µ +1 is computed as { µ, if ρ (α ; µ, l ) γ s, or d = s = 0, or α = 0, µ +1 = max ( 1 2 µ, ) (2.29) µr, otherwise. 2.5. The main algorithm The stabilized second-order primal-dual SQP algorithm is formally stated as Algorithm 5. The algorithm is terminated if one of two sets of conditions holds. The first is r(x, y ) τ stop, ξ τ stop, and µ R 1 τ stop, (2.30) where τ stop is a given positive stopping tolerance, r(x, y) is defined in (1.2), and ξ is computed in Algorithm 1. When the conditions in (2.30) hold, then (x, y ) is an approximate secondorder KKT point for (NP). The motivation for the second set of termination conditions is to identify convergence to an infeasible stationary point (ISP), i.e., a minimizer of minimize x R n 1 2 c(x) 2 2 subject to x 0 (2.31)

A Stabilized SQP Method: Global Convergence 17 at which the constraint c(x) = 0 is violated. The first-order optimality conditions for (2.31) may be written in the form min ( x, J(x) T c(x) ) = 0, (2.32) which motivates the second set of termination conditions: min ( c(x ), τ stop ) > µ R, min ( x, J(x ) T c(x ) ) τ stop, with an M-iteration. (2.33) If these conditions hold, then the constraint violation is bounded below by µ R (µr > 0) and v = (x, y ) is an M-iterate that approximately satisfies (2.32). The requirements that is an M-iteration and µ R is small relative to min ( ) c(x ), τ stop have the practical benefit of reducing the lielihood of premature termination when progress towards an approximate second-order KKT point is still possible. The termination conditions (2.30) and (2.33) constitute the conditions used in Steps 10 and 25 of Algorithm 5. Discussion. Each iteration of Algorithm 5 requires the solution of equations involving the ɛ- free rows and columns of B Fɛ (v ; µ R 1 ) and B Fɛ(v ; µ R ). The solve with B Fɛ(v ; µ R 1 ) is used to compute the least-curvature estimate ξ in Step 8. Solves with B Fɛ (v ; µ R ) are required to compute the local descent step and the solution of the QP subproblem. This implies that if µ R µr 1 it is necessary to compute the factorizations of two matrices in regularized KKT form at each iteration. If the computation of ξ is omitted, and the line search is defined in terms of the descent step d only, then one KKT factorization is required at each iteration. Moreover, the analysis of [18] shows that this variant of Algorithm 5 is a globally convergent stabilized SQP method for finding first-order solutions. 2.6. Properties of the main algorithm The next lemma establishes some properties associated with the ey computational steps of Algorithm 5. Lemma 2.3. Let d and s be vectors computed by Algorithm 5. (1) If d = 0, then (a) min(x, g(x ) J(x ) T y ) = 0, and π(x, y E, µr ) = y ; (b) if the th iterate is a V-O-iterate, then r(x, y ) = 0; (c) if the th iterate is an M-iterate and y y max, then r(x, y ) = 0. (2) If d = s = 0, then ξ = 0, the th iterate is not an F-iterate, and µ R < µr 1. Proof. For all parts of this lemma it is assumed that d = 0, in which case it must hold that M(v ; y E, µr )T d = 0. It follows from Step 9 of Algorithm 2 that d is a global descent step. For part (1a), as d is zero, the optimality conditions for the bound-constrained QP subproblem (2.4) give ( ( min x, x M(v 0 = ; y E, µr )) ) ( min(x, g y M(v ; y E, µr ) = J T ( π + (π y ) ) ) µ R (y π ), where π = π(x, y E, µr ). As µr is positive, it follows that y = π and min(x, g J T y ) = 0. (2.34)

A Stabilized SQP Method: Global Convergence 18 Algorithm 5 Second-order primal-dual SQP algorithm. 1: procedure PDSQP2(x 1, y 1 ) 2: Constants: {τ stop, γ s } (0, 1 2 ), 0 < γ < 1, and 0 < ɛ a 1; 3: Choose y0 E Rm, {τ 0, φ max V, φ max O } (0, ), and 0 < µ R 0 µ 1 < ; 4: = 1; 5: loop 6: Compute the ɛ-free set F ɛ (x, y, µ R 1 ) from (2.8); 7: J = J(x ); H = H(x, y ); ( (1) 8: s, ξ ) = LEAST CURVATURE ESTIMATE(x, y, µ R 1, J, H ); 9: Compute r(x, y ) from (1.2); [Algorithm 1] 10: if (termination criterion (2.30) holds) then 11: return the approximate second-order KKT point (x, y ); 12: end if 13: if (φ V (x, y ) 1 2 φmax V ) then [V-iterate] 14: φ max V = 1 2 φmax V ; y E = y ; τ = τ 1 ; 15: Set µ R as in (2.15); 16: else if (φ O (x, y, ξ ) 1 2 φmax O ) then [O-iterate] 17: φ max O = 1 2 φmax O ; y E = y ; τ = τ 1 ; 18: Set µ R as in (2.15); 19: else if ((x, y ) satisfies (2.17a) (2.17c)) then [M-iterate] 20: Set y E as in (2.18); τ = 1 2 τ 1; 21: Set µ R as in (2.19); 22: else [F-iterate] 23: y E = ye 1 ; 24: end if τ = τ 1 ; µr = µr 1 ; 25: if (termination condition (2.33) holds) then 26: exit with the approximate infeasible stationary point x. 27: end if 28: Compute the ɛ-free set F ɛ (x, y, µ R ) from (2.8); 29: d = DESCENT DIRECTION(x, y, y E, µr, J, H ); [Algorithm 2] 30: s = CURVATURE DIRECTION(s (1), ξ, x, d, J, H ); [Algorithm 3] 31: if (d 0 and s = 0 and (x, y ) is a V-O-iterate) then 32: l = 1; 33: else 34: l = 2; 35: end if 36: µ = max(µ R, µ ); 37: α = FLEXIBLE LINE SEARCH(d, s, y E, µ, µr, µr 1, l, J, H ); [Algorithm 4] 38: Set µ +1 as in (2.29); 39: v +1 = (x +1, y +1 ) = v + α d + α s ; 40: = + 1; 41: end loop 42: end procedure

A Stabilized SQP Method: Global Convergence 19 This completes the proof of part (1a). As d is zero and the th iterate is a V-O-iterate by assumption, it follows from the update for y E given by Algorithm 5, the definition of π, and part (1a) that y E = y = π = ye c(x )/µr, (2.35) which implies that c(x ) = 0. Combining this with (1.2) and (2.34) gives r(x, y ) = 0, which proves part (1b). If the th iterate is an M-iterate and y y max, it follows from the update for y E in (2.18), and the definition of π that (2.35) holds. As in part (1b), this proves that c(x ) = 0. Combining this result with (1.2) and (2.34) yields r(x, y ) = 0, which establishes part (1c). The assumption for part (2) to hold is that both d and s are zero. Suppose that the curvature result does not hold, i.e., assume that ξ > 0. As d is zero, x is optimal for the QP subproblem (2.4) and x = x. In this case, the assumption that ξ > 0 in the definitions of u (2) and its associated scale factor σ in (2.24) imply that σ > 0. However, if σ > 0 and s = 0 in the definition of s (2) (2.25), then not only is s (2) = 0, but also s (1) = 0 from (2.23). This gives the required contradiction because s (1) must be nonzero for ξ to be positive in Algorithm 1. It follows that ξ = 0, as required. The proof that the th iterate cannot be an F-iterate when both d and s are zero is also by contradiction. For an F-iterate, the parameters µ R and ye used in the definition of the QP subproblem are µ R = µr 1 and y E = ye 1. (2.36) As the solution of the subproblem is d = 0 by assumption, part (1a) and the result that ξ = 0 imply that (x, y ) satisfies the conditions (2.17) for an M-iterate. This is a contradiction because a point is classified as an F-iterate if it is not a V-O-iterate and the conditions for an M-iterate fail to hold. It remains to show that the regularization parameter is decreased. Assume to the contrary that µ R = µr 1. It has already been shown that x cannot be an F-iterate. Moreover, if x were an M-iterate then the update (cf. (2.19)) would imply that µ R < µr 1. It follows that x must be a V-O-iterate. The properties of the update (2.15) associated with a V-O-iterate, and the assumption that µ R = µr 1 imply that max ( r, ξ ) > 0. This is a contradiction because r = 0 follows from part (1b), and it has been shown above that ξ = 0. As all possible cases have been exhausted, it holds that µ R < µr 1. The next lemma summarizes the principal properties of the flexible line-search computation and the penalty parameter update scheme. In particular, the result of part (1) implies that the line search is guaranteed to terminate in a finite number of steps, and as a consequence, the algorithm is well-defined. Lemma 2.4. If f and c are twice continuously differentiable, then the following properties hold. (1) The while-loops given by Steps 8 and 15 of Algorithm 4 terminate with α > 0. (2) If µ < µ 1 for some 1, then either the while-loop given by Step 8 or the while-loop given by Step 15 of Algorithm 4 was executed. (3) If α = 0, then the th iterate is not an F-iterate and µ R < µr 1.