A REDUCED HESSIAN METHOD FOR LARGE-SCALE CONSTRAINED OPTIMIZATION by Lorenz T. Biegler, Jorge Nocedal, and Claudia Schmid ABSTRACT We propose a quasi-

Similar documents
NUMERICAL EXPERIENCE WITH A REDUCED HESSIAN METHOD FOR LARGE SCALE CONSTRAINED OPTIMIZATION by Lorenz T. Biegler, Jorge Nocedal, Claudia Schmid and Da

A new ane scaling interior point algorithm for nonlinear optimization subject to linear equality and inequality constraints

5 Handling Constraints

1 Introduction Sequential Quadratic Programming (SQP) methods have proved to be very ecient for solving medium-size nonlinear programming problems [12

154 ADVANCES IN NONLINEAR PROGRAMMING Abstract: We propose an algorithm for nonlinear optimization that employs both trust region techniques and line

2 Chapter 1 rely on approximating (x) by using progressively ner discretizations of [0; 1] (see, e.g. [5, 7, 8, 16, 18, 19, 20, 23]). Specically, such

National Institute of Standards and Technology USA. Jon W. Tolle. Departments of Mathematics and Operations Research. University of North Carolina USA

Algorithms for Constrained Optimization


AM 205: lecture 19. Last time: Conditions for optimality Today: Newton s method for optimization, survey of optimization methods

Unconstrained optimization

1. Introduction In this paper we study a new type of trust region method for solving the unconstrained optimization problem, min f(x): (1.1) x2< n Our

Search Directions for Unconstrained Optimization

2. Quasi-Newton methods

Higher-Order Methods

Nonmonotonic back-tracking trust region interior point algorithm for linear constrained optimization

Newton s Method. Ryan Tibshirani Convex Optimization /36-725

Constrained optimization: direct methods (cont.)

A Piecewise Line-Search Technique for Maintaining the Positive Definiteness of the Matrices in the SQP Method

2 JOSE HERSKOVITS The rst stage to get an optimal design is to dene the Optimization Model. That is, to select appropriate design variables, an object

On the Local Quadratic Convergence of the Primal-Dual Augmented Lagrangian Method

17 Solution of Nonlinear Systems

Constrained Optimization Theory

Methods for Unconstrained Optimization Numerical Optimization Lectures 1-2

A globally and R-linearly convergent hybrid HS and PRP method and its inexact version with applications

SF2822 Applied Nonlinear Optimization. Preparatory question. Lecture 9: Sequential quadratic programming. Anders Forsgren

Survey of NLP Algorithms. L. T. Biegler Chemical Engineering Department Carnegie Mellon University Pittsburgh, PA

f = 2 x* x g 2 = 20 g 2 = 4

AM 205: lecture 19. Last time: Conditions for optimality, Newton s method for optimization Today: survey of optimization methods

Quasi-Newton Methods

Newton s Method. Javier Peña Convex Optimization /36-725

SECTION: CONTINUOUS OPTIMISATION LECTURE 4: QUASI-NEWTON METHODS

University of Maryland at College Park. limited amount of computer memory, thereby allowing problems with a very large number

Lecture 14: October 17

5 Quasi-Newton Methods

QUADRATICALLY AND SUPERLINEARLY CONVERGENT ALGORITHMS FOR THE SOLUTION OF INEQUALITY CONSTRAINED MINIMIZATION PROBLEMS 1

Written Examination

Extra-Updates Criterion for the Limited Memory BFGS Algorithm for Large Scale Nonlinear Optimization M. Al-Baali y December 7, 2000 Abstract This pape

Linear Algebra (part 1) : Vector Spaces (by Evan Dummit, 2017, v. 1.07) 1.1 The Formal Denition of a Vector Space

IBM Research Report. Line Search Filter Methods for Nonlinear Programming: Motivation and Global Convergence

Structural and Multidisciplinary Optimization. P. Duysinx and P. Tossings

A trust region method based on interior point techniques for nonlinear programming

Optimality Conditions for Constrained Optimization

Preconditioned conjugate gradient algorithms with column scaling

IE 5531: Engineering Optimization I

Infeasibility Detection and an Inexact Active-Set Method for Large-Scale Nonlinear Optimization

1 Computing with constraints

Constrained Nonlinear Optimization Algorithms

min f(x). (2.1) Objectives consisting of a smooth convex term plus a nonconvex regularization term;

Jorge Nocedal. select the best optimization methods known to date { those methods that deserve to be

Maria Cameron. f(x) = 1 n

An Active Set Strategy for Solving Optimization Problems with up to 200,000,000 Nonlinear Constraints

2 jian l. zhou and andre l. tits The diculties in solving (SI), and in particular (CMM), stem mostly from the facts that (i) the accurate evaluation o

Scientific Computing: An Introductory Survey

Scientific Computing: An Introductory Survey

system of equations. In particular, we give a complete characterization of the Q-superlinear

E5295/5B5749 Convex optimization with engineering applications. Lecture 8. Smooth convex unconstrained and equality-constrained minimization

Improved Damped Quasi-Newton Methods for Unconstrained Optimization

Constrained Optimization

1. Introduction Let the least value of an objective function F (x), x2r n, be required, where F (x) can be calculated for any vector of variables x2r

Nonlinear Optimization: What s important?

An Inexact Newton Method for Nonlinear Constrained Optimization

September Math Course: First Order Derivative

Algorithms for constrained local optimization

Unconstrained Multivariate Optimization

MS&E 318 (CME 338) Large-Scale Numerical Optimization

On fast trust region methods for quadratic models with linear constraints. M.J.D. Powell

Flexible Penalty Functions for Nonlinear Constrained Optimization

Arc Search Algorithms

IPAM Summer School Optimization methods for machine learning. Jorge Nocedal

A SHIFTED PRIMAL-DUAL PENALTY-BARRIER METHOD FOR NONLINEAR OPTIMIZATION

Lecture 15: SQP methods for equality constrained optimization

March 8, 2010 MATH 408 FINAL EXAM SAMPLE

Step lengths in BFGS method for monotone gradients

Determination of Feasible Directions by Successive Quadratic Programming and Zoutendijk Algorithms: A Comparative Study

Merging Optimization and Control

arxiv: v1 [math.oc] 10 Apr 2017

1. Introduction. In this paper we discuss an algorithm for equality constrained optimization problems of the form. f(x) s.t.

1. Introduction The nonlinear complementarity problem (NCP) is to nd a point x 2 IR n such that hx; F (x)i = ; x 2 IR n + ; F (x) 2 IRn + ; where F is

A PROJECTED HESSIAN GAUSS-NEWTON ALGORITHM FOR SOLVING SYSTEMS OF NONLINEAR EQUATIONS AND INEQUALITIES

An Inexact Newton Method for Optimization

118 Cores, D. and Tapia, R. A. A Robust Choice of the Lagrange Multiplier... For example: chemical equilibrium and process control oil extraction, ble

Optimization Tutorial 1. Basic Gradient Descent

x 2. x 1

1. Nonlinear Equations. This lecture note excerpted parts from Michael Heath and Max Gunzburger. f(x) = 0

On the use of piecewise linear models in nonlinear programming

Optimal Newton-type methods for nonconvex smooth optimization problems

Introduction to Nonlinear Optimization Paul J. Atzberger

An Inexact Sequential Quadratic Optimization Method for Nonlinear Optimization

ON THE CONNECTION BETWEEN THE CONJUGATE GRADIENT METHOD AND QUASI-NEWTON METHODS ON QUADRATIC PROBLEMS

8 Numerical methods for unconstrained problems

Part 5: Penalty and augmented Lagrangian methods for equality constrained optimization. Nick Gould (RAL)

Optimization Methods. Lecture 18: Optimality Conditions and. Gradient Methods. for Unconstrained Optimization

A STABILIZED SQP METHOD: SUPERLINEAR CONVERGENCE

Stochastic Optimization Algorithms Beyond SG

A COMBINED CLASS OF SELF-SCALING AND MODIFIED QUASI-NEWTON METHODS

ISM206 Lecture Optimization of Nonlinear Objective with Linear Constraints

CS 542G: Robustifying Newton, Constraints, Nonlinear Least Squares

1. Introduction. We consider the general smooth constrained optimization problem:

Optimization II: Unconstrained Multivariable

Transcription:

A REDUCED HESSIAN METHOD FOR LARGE-SCALE CONSTRAINED OPTIMIZATION by Lorenz T. Biegler, 1 Jorge Nocedal, 2 and Claudia Schmid 1 1 Chemical Engineering Department, Carnegie Mellon University, Pittsburgh, PA 15213. Partial support for these authors is acnowledged from the Engineering Design Research Center, an NSF-sponsored Engineering Research Center at Carnegie Mellon University. 2 Department ofelectrical Engineering and Computer Science, Northwestern University, Evanston IL 60208, nocedal@eecs.nwu.edu also, Mathematics and Computer Science Division, Argonne National Laboratory, Argonne, IL 60439. This author was supported by National Science Foundation grants CCR- 9101359 and ASC-9213149, by U.S. Department of Energy grant DE-FG02-87ER25047-A004, and by the Oce of Scientic Computing, U.S. Department of Energy, under Contract W-31-109-Eng-38. 1

A REDUCED HESSIAN METHOD FOR LARGE-SCALE CONSTRAINED OPTIMIZATION by Lorenz T. Biegler, Jorge Nocedal, and Claudia Schmid ABSTRACT We propose a quasi-newton algorithm for solving large optimization problems with nonlinear equality constraints. It is designed for problems with few degrees of freedom and is motivated by the need to use sparse matrix factorizations. The algorithm incorporates a correction vector that approximates the cross term Z T WYp Y in order to estimate the curvature in both the range and null spaces of the constraints. The algorithm can be considered to be, in some sense, a practical implementation of an algorithm of Coleman and Conn. We give conditions under which local and superlinear convergence is obtained. Key words: successive quadratic programming, reduced Hessian methods, constrained optimization, quasi-newton method, large-scale optimization. Abbreviated title: A Reduced Hessian Method 1. Introduction We consider the nonlinear optimization problem min f(x) (1:1) x2r n subject to c(x) =0 (1:2) where f : R n! R and c : R n! R m are smooth functions. We are particularly interested in the case when the number of variables n is large, and the algorithm we propose, which is a variation of the successive quadratic programming method, is designed to be ecient in this case. We assume that the rst derivatives of f and c are available, but our algorithm does not require second derivatives. The successive quadratic programming (SQP) method for solving (1.1){(1.2) generates, at an iterate x, a search direction d by solving min d2r n g(x ) T d + 1 2 dt W (x )d (1:3) subject to c(x )+A(x ) T d =0 (1:4) 1

where g denotes the gradient off, W denotes the Hessian of the Lagrangian function L(x ) = f(x)+ T c(x), and A denotes the n m matrix of constraint gradients A new iterate is then computed as A(x) =[rc 1 (x) ::: rc m (x)]: (1:5) x +1 = x + d (1:6) where is a steplength parameter chosen so as to reduce the value of the merit function. In this study we will use the `1 merit function (x) =f(x)+c(x) 1 (1:7) where is a penalty parameter see, for example, Conn (1973), Han (1977) or Fletcher (1987). We could have used other merit functions, but the essential points we wish to convey in this article are not dependent upon the particular choice of the merit function. The solution of the quadratic program (1.3){(1.4) can be written in a simple form if we choose a suitable basis of R n to represent the search direction d. For this purpose, we introduce a nonsingular matrix of dimension n, which we write as where Y 2 R nm and Z 2 R n(n;m),andwe assume that [Y Z ] (1:8) A T Z =0: (1:9) (From now onwe abbreviate A(x )asa, g(x )asg, etc.) Thus Z is a basis for the tangent space of the constraints. We can now express d, the solution to (1.3){(1.4), as d = Y p Y + Z p Z (1:10) for some vectors p Y 2 R m and p Z 2 R n;m. Due to (1.9) the linear constraints (1.4) become c + A T Y p Y =0: (1:11) If we assume that A has full column ran, then the nonsingularity of[y Z ] and equation (1.9) imply that the matrix A T Y is nonsingular, so that p Y is determined by (1.11): Substituting this in (1.10), we have Note that p Y = ;[A T Y ] ;1 c : (1:12) d = ;Y [A T Y ] ;1 c + Z p Z : (1:13) Y [A T Y ] ;1 (1:14) is a right inverse of A T and that the rst term in (1.13) represents a particular solution of the linear equations (1.4). We havethus reduced the size of the SQP subproblem, which can now be expressed exclusively in terms of the variables p Z. Indeed, substituting (1.10) into (1.3), considering Y p Y as constant, and ignoring constant terms,we obtain the unconstrained quadratic problem min (Z T p Z 2R n;m g + Z T W Y p Y ) T p Z + 1 2 p T Z (Z T W Z )p Z : (1:15) 2

If we assume that Z T W Z is positive denite, the solution of (1.15) is p Z = ;(Z T W Z ) ;1 [Z T g + Z T W Y p Y ]: (1:16) This determines the search direction of the SQP method. We are particularly interested in the class of problems in which thenumber ofvariables n is large, but n ; m is small. In this case it is practical to approximate Z T W Z using a variable metric formula such asbfgs. On the other hand, the matrix Z T W Y, of dimension (n;m)m may be too expensive to compute directly when m is large. For this reason several authors simply ignore the \cross term" Z T W Y p Y in (1.16) and compute only an approximation to the reduced Hessian Z T W Z see Coleman and Conn (1984), Nocedal and Overton (1985), and Xie (1991). This approach is quite adequate when the basis matrices Y and Z in (1.8) are chosen to be orthonormal (Gurwitz and Overton (1989)). For large problems, however, computing orthogonal bases can be expensive, and it is more ecient to obtain Y and Z by simple elimination of variables (cf. Fletcher (1987)). Unfortunately, in this case ignoring the cross term Z T W Y p Y can mae the algorithm inecient, as is illustrated by an example given in a companion paper (Biegler, Nocedal, and Schmid (1993)). The central point is that the range space component Y p Y may bevery large, and ignoring the contribution from the cross term in (1.16) can result in a poor step. Therefore, in this paper we suggest ways of approximating the cross term Z T W Y p Y bya vector w, [Z T W Y ]p Y w (1:17) without computing the matrix Z T W Y. We consider two approaches for calculating w the rst involves an approximation to the matrix [Z T W Y ] using Broyden's update, and the second generates w using nite dierences. We will show that the rate of convergence of the new algorithm is 1-step Q-superlinear, as opposed to the 2-step superlinear rate for methods that ignore the cross term (Byrd (1985) and Yuan (1985)). The null space step (1.16) of our algorithm will be given by p Z = ;(Z T W Z ) ;1 [Z T g + w ] (1:18) where 0 < 1 is a damping factor to be discussed later on. To describe our rst strategy for computing the vector w,we consider a quasi-newton method in which the rectangular matrix Z T W is approximated by a matrix S, using Broyden's method. We then obtain w bymultiplying this matrix by Y p Y, that is, w = S Y p Y : How should S be updated? Since W +1 = r 2 xxl(x +1 +1 ), we have that Z T W +1(x +1 ; x ) Z T [r x L(x +1 +1 ) ;r x L(x +1 )] (1:19) when x +1 is close to x. We use this relation to establish the following secant equation: we demand that S +1 satisfy S +1 (x +1 ; x )=Z T [r x L(x +1 +1 ) ;r x L(x +1 )]: (1:20) One point in this derivation requires clarication. In the left-hand side of (1.19) wehave Z T W +1, and not Z T W +1 +1. We could have used Z +1 in (1.19), avoiding an inconsistency of indices, but this is not necessary since we will show that using Z instead of Z +1 in (1.20) results in algorithms with all the desirable properties. This fact will not be surprising to readers familiar 3

with the analysis of SQP methods see, for example, Coleman and Conn (1984) or Nocedal and Overton (1985). In addition, using Z allows updating of S +1 and B +1 prior to creating Z +1 at the new point. Let us now consider how to approximate the reduced Hessian matrix Z T W Z. Using (1.6) and (1.10) in (1.20), we obtain [S +1 Z ] p Z = ; S +1 (Y p Y )+Z T [r x L(x +1 +1 ) ;r x L(x +1 )]: Since S +1 approximates Z T W, this suggests the following secant equation for B +1, the quasi- Newton approximation to the reduced Hessian Z T W Z : where s is dened by and y by with B +1 s = y (1:21) s = p Z y = Z T [r x L(x +1 +1 ) ;r x L(x +1 )] ; w (1:22) We willupdateb by the BFGS formula (cf. Fletcher (1987)) w = S +1 (Y p Y ): (1:23) B +1 = B ; B s s T B s T B s + y y T y T s (1:24) provided s T y is suciently positive. We highlight a subtle but important point. We have dened two correction terms, w and w. Both are approximations to the cross term (Z T WY)p Y. The rst term, w,which is needed to dene the null-space step (1.18) and thus the new iterate x +1 maes use of the matrix S. The second term, w,which is used in (1.22) to dene the BFGS update of B, is computed by using the new Broyden matrix S +1 and taes into account the steplength. We will see below that it is useful to incorporate the most recent information in w. Note that this requires the Broyden update to be applied before the vector y for the BFGS update can be calculated from (1.22). The Lagrange multiplier estimates needed in the denition (1.22) of y are dened by = ;[Y T A ] ;1 Y T g : (1:25) This formula is motivated by the fact that, at a solution x of (1.1){(1.2), we have ;g = A, and since Y [A T Y ] ;1 is a right inverse of A T, = ;[Y T A ] ;1 Y T g : Using the same right inverse (1.14) in the denitions of p Y and will allow us a convenient simplication in the formulae presented in the following sections. We stress, however, that other Lagrange multiplier estimates can be used and that the best choice in practice might be the one that involves the least computation or storage. We can now outline the sequential quadratic programming method analyzed in this paper. Algorithm I 4

1. Choose constants 2 (0 1=2) and 0 with 0 << 0 < 1. Set :=1,and choose a starting point x 1 and an (n ; m) (n ; m) symmetric and positive denite starting matrix B 1. 2. Evaluate f g c and A, and compute Y and Z. 3. Compute p Y by solving the system (A T Y )p Y = ;c : (range space step) (1.26) 4. Compute an approximation w to (Z T W Y )p Y. 5. Choose the damping parameter 2 (0 1] and compute p Z from Dene the search direction by B p Z = ;[Z T g + w ]: (null space step) (1.27) d = Y p Y + Z p Z : (1.28) 6. Set =1,andchoose the weight of the merit function (1.7). 7. Test the line search condition (x + d ) (x )+ D (x d ) (1.29) where D (x d ) is the directional derivative of the merit function in the direction d. 8. If (1.29) is not satised, choose a new 2 [ 0 ] and go to (7) otherwise set x +1 = x + d : (1.30) 9. Evaluate f +1 g +1 c +1, and A +1, and compute Y +1 and Z +1. 10. Compute the Lagrange multiplier estimate Dene w (as will be discussed in x3), and compute and +1 = ;[Y T +1 A +1] ;1 Y T +1 g +1: (1:31) s = p Z (1:32) y = Z T [r x L(x +1 +1 ) ;r x L(x +1 )] ; w : (1:33) If the update criterion (to be discussed in x3.3) is satised, compute B +1 by the BFGS formula (1.24) else set B +1 = B. 11. Set := + 1, and go to (3). The algorithm has been left in a very general form, but in the next sections we discuss all its aspects in detail. In x2 we consider the choice of the basis matrices Y and Z. In x3 we describe the calculation of the correction terms w and w, the conditions under which BFGS updating taes place, the choice of the damping parameter, and the procedure for updating the weight in the merit function. In x4 and x5 we analyze of the local behavior of the algorithm and 5

show that the rate of convergence is at least R-linear. In x6 we present a superlinear convergence result, and some nal remars in x7 conclude the paper. We now mae a few comments about our notation. Throughout the paper, the vectors p Y and p Z are computed at x and could be denoted by p Y () and p Z (), but we will normally omit the superscript for simplicity. The symbol denotes the l 2 vector norm or the corresponding induced matrix norm. When using the l 1 or l 1 norms we willindicate it explicitly by writing 1 or 1. A solution of problem (1.1) is denoted by x, and we dene e = x ; x and =maxfe e +1 g: (1:34) Here, and for the rest of the paper, rl(x ) indicates the gradient of the Lagrangian with respect to x only. 2. The Basis Matrices As long as Z spans the null space of A T, and [Y Z ] is nonsingular, the choice of Y and Z is arbitrary. However, from the viewpoint of numerical stability and robustness of the algorithm it is desirable to dene Y and Z to be orthonormal, that is, Z(x) T Z(x) = I n;m Y (x) T Y (x) = I m Y (x) T Z(x) = 0: One way of obtaining these matrices is by forming the QR factorization of A. For large problems, however, computing this QR factorization is often too expensive. Therefore many researchers, including Gabay (1982), Gilbert (1991), Fletcher (1987), Murray and Prieto (1992), and Xie (1991), consider other, nonorthogonal choices of Y and Z. For example, if we partition x into m basic or dependent variables (which without loss of generality are assumed to be the rst m variables) and n ; m nonbasic or control variables, we induce the partition A(x) T =[C(x) N(x)] (2:1) where the m m basis matrix C(x) is assumed to be nonsingular. We now dene Z(x) andy (x) to be ;C(x) ;1 N(x) I Z(x) = Y (x) = : (2:2) I 0 When A(x) is large and sparse, a sparse LU decomposition of C(x) can often be computed eciently, and this approach will be considerably less expensive than the QR factorization of A. Note that from the assumed nonsingularity ofc(x) both Y (x) and Z(x) vary smoothly with x, provided the same partition of the variables is maintained. In our implementation of the new algorithm (Biegler, Nocedal, and Schmid (1993)) we choose Y and Z by (2.2). There is a price to pay for using nonorthogonal bases. If the matrix C is ill conditioned (and this can be dicult to detect), the step computation may beinaccurate. Moreover, even if the basis is well conditioned, the range space step Y p Y can be large, and ignoring the cross term can cause serious diculties. This phenomenon is illustrated in a two-dimensional example given by Biegler, Nocedal, and Schmid (1993). It is shown in that example that if the cross term Z T W Y p Y is ignored, the ratio x + d =x can be arbitrarily large, even close to the solution. It is also shown that these ineciencies disappear if the cross term is approximated as suggested in the following sections. 6

In the rest of the paper we allow much freedom in the choice of the basis matrices. They canbegiven by (2.2), can be orthonormal, or can be chosen in other ways. The only restrictions we impose are that A T Z = 0 is satised, that the n n matrix [Y Z ] is nonsingular and well conditioned, and that this matrix varies smoothly in a neighborhood of the solution. 3. Further Details of the Algorithm In this section we consider how to calculate approximations w and w to (Z T W Y )p Y to be used in the determination of the search direction p Z and in updating B, respectively. We also discuss when to sip the BFGS update of the reduced Hessian approximation, as well as the selection of the damping factor and the penalty parameter. To calculate approximations to (Z T WY)p Y,we propose two approaches. First, we consider a nite dierence approximation to Z T W along the direction Y p Y. While this approach requires additional evaluations of reduced gradients at each iteration, it gives rise to a very good step. The second, more economical approach denes w and w in terms of a Broyden approximation to Z T W, as discussed in x1, and requires no additional function or gradient evaluations. Our algorithm will normally use this second approach, but as we will later see, it is sometimes necessary to use nite dierences. 3.1. Calculating w and w through Finite Dierences We rst calculate the range space step p Y at x through Equation (1.26). Next we compute the reduced gradient of the Lagrangian at x + Y p Y and dene w = Z T [rl(x + Y p Y ) ;rl(x )]: (3:1) After the step to the new iterate x +1 has been taen, we dene w = Z T [rl(x + Y p Y +1 ) ;rl(x +1 )] (3:2) which requires a new evaluation of gradients if 6= 1. Thus, up to three evaluations of the objective function gradient may be required at each iteration. We note that this nite-dierence approach isvery similar to the algorithm of Coleman and Conn (1982, 1984). Starting at a point z, the Coleman-Conn algorithm (with steplength =1) is given by Z p Z = ;Z(z )B ;1 Z(z ) T g(z ) (3:3) Y p Y = ;Y (z )[A(z ) T Y (z )] ;1 c(z + Z p Z ) (3:4) z +1 = z + Z p Z + Y p Y : (3:5) Let us now consider Algorithm I, and to better illustrate its similarity with the Coleman and Conn method, let us assume that instead of (3.1), w is dened by w = Z(x + Y p Y ) T g(x + Y p Y ) ; Z(x ) T g(x ) which diers from (3.1) by terms of order O(p Y ). Then Algorithm I with = 1 and =1is given by Y p Y = ;Y (x )[A(x ) T Y (x )] ;1 c(x ): (3:6) 7

Z p Z = ;Z(x )B ;1 [Z(x ) T g(x )+w ] = ;Z(x )B ;1 [Z(x + Y p Y ) T g(x + Y p Y )]: (3.7) x +1 = x + Y p Y + Z p Z : (3:8) The similarity between the two approaches is apparent in Figure 1, especially if we consider the intermediate points in the Coleman-Conn iteration to be the starting and nal points, respectively. z z + Z p z x z +1 x + Y p y x +1 Coleman-Conn step The step of Algorithm I Figure 1 Comparison of Coleman-Conn method and Algorithm I In the Coleman-Conn algorithm, the approximation B to the reduced Hessian Z T W Z is obtained bymoving along the null space direction Z p Z, and maing a new evaluation of the function and constraint gradients. To be more precise, Coleman and Conn dene y = Z T [rl(x + Z p Z ) ;rl(x )] and s = Z T [x +1 ; x ]andapply a quasi-newton formula to update B. Algorithm I, using nite dierences, amounts essentially to the same thing. To see this, note that if Formula (3.2) is used in (1.33), then y = Z T [rl(x +1 +1 ) ;rl(x + Y p Y +1 )] which represents a dierence in reduced gradients of the Lagrangian along the null space direction Z p Z. Byrd (1990) and Gilbert (1989) showed that the sequence fz + Z p Z g (but not the sequence fz g) generated by the Coleman-Conn method converges one-step Q-superlinearly. If Algorithm I always computed the correction terms w and w by nite dierences, its cost and convergence behavior would be similar to those of the Coleman-Conn method (except when 6=1,which requires one extra gradient evaluation for Algorithm I). However, we will often be able to avoid the use of nite dierences and instead use the more economical approach discussed next. 3.2. Using Broyden's Method to Compute w and w We can approximate the rectangular matrix Z T W by a matrix S updated by Broyden's method, and then compute w and w by post-multiplying this matrix by Y p Y or by amultiple 8

of this vector. As discussed in x1, it is reasonable to impose the secant equation (1.20) on this Broyden approximation, which can therefore be updated by the formula (cf. Fletcher (1987)) where and We now dene S +1 = S + (y ; S s )s T s T s (3:9) y = Z T [rl(x +1 +1 ) ;rl(x +1 )] (3:10) s = x +1 ; x : (3:11) w = S Y p Y and w = S +1 Y p Y : (3:12) It should be noted that this approach requires the storage of the (n ; m) n matrix S, in addition to the reduced Hessian approximation, B. For problems where n ; m is small, this expense is far less than the storage of a full Hessian approximation to W. On the other hand, if n ; m is not very small, it may be preferable to use a limited-memory implementation of Broyden's method. Here the matrices S are represented implicitly, using, for example, the compact representation described in Byrd, Nocedal, and Schnabel (1992). The advantage of the limited memory implementation is that it requires the storage of only a few n;vectors to represent S. Since is no guarantee that the Broyden approximations S will remain bounded, we need to safeguard them. At the beginning of the algorithm we choose a positive constant ;and dene ( ; w if w p w := Y p 1=2 Y w ;p Y 1=2 (3:13) w otherwise. The correction w will be safeguarded in a dierent way. We choose a sequence of positive numbers f g such that 1 =1 < 1, and we set w := ( w if w p Y = w p Y w otherwise. (3:14) As the iterates converge to the solution, p Y! 0, so that from (3.12) and from the boundedness of Y we see that these safeguards allow the Broyden updates S to become unbounded, but in a controlled manner. We will show in x4 and x5 that with the safeguards (3.13) and (3.14) Algorithm I is locally and R-linearly convergent and that this implies that the Broyden updates S do, in fact, remain bounded, so that the safeguards become inactive asymptotically. Our Broyden approximation to the correction terms w and w was motivated by recent wor of Gurwitz (1993). She approximates Z T W Z by the BFGS formula with and s = Z T [x +1 ; x ] y = Z T [rl(x +1 +1 ) ;rl(x +1 )] and approximates Z T W Y by a matrix D using Broyden's formula (3.9) with s = Y T [x +1 ; x ] 9

y = Z T [rl(x +1 +1 ) ;rl(x +1 )] ; B p Z : Since the updates may not always be dened, Gurwitz proposes to sometimes sip the update of B or D. She shows 1-step Q-superlinear convergence if and only if one of the updates is taen at each iteration, but this cannot be guaranteed. The analysis of this paper will show that it is preferable to update an approximation to Z T W, as in Algorithm I, instead of an approximation to Z T W Y, as proposed by Gurwitz, since our approach leads to 1-step superlinear convergence in all cases. A related method was derived by Coleman and Fenyes (1992). Their lower partition BFGS formula (LPB) simultaneously updates approximations to Z T W Z and Z T W Y,by means of a new variational problem. The resulting updating formula requires the solution of a cubic equation, and its roots can correspond to cases where updates should be avoided (e.g., s T y 0). The drawbac of this approach is that choosing the correct root is not always easy. An earlier proposal bytagliaferro (1989) consists of approximating the matrices Z T W Z and Z T W Y using the PSB update formula and Broyden's method, respectively. One disadvantage of this approach is that the matrices generated by this updating procedure may become very ill conditioned. 3.3. Update Criterion It is well nown that the BFGS update (1.24) is well dened only if the curvature condition s T y > 0 is satised. This condition can always be enforced in the unconstrained case by performing an appropriate line search see, for example, Fletcher (1987). When constraints are present, however, the curvature condition s T y > 0 can be dicult to obtain, even near the solution. To show this, we rst note from (1.33), (1.28), and (1.32) and from the Mean Value theorem that Z 1 y = Z T r 2 xx L(x + d +1 )d d ; w where we have dened Thus 0 Z T W ~ d ; w = Z T W ~ Z s + Z T W ~ Y p Y ; w (3.15) s T y = s T ~W = Z 1 0 r 2 xx L(x + d +1 )d: (3:16) Z T ~ W Z s + s T Z T ~ W Y p Y ; s T w : (3:17) Near the solution, the rst term on the right-hand side will be positive, since Z T W ~ Z can be assumed positive denite. Nevertheless, the last two terms are of uncertain sign and can mae s T y negative. Several reduced Hessian methods in the literature set w equal to zero for all, and update B only if p Y is small enough compared with s that the rst term in the right-hand side of (3.17) dominates the second term (see Nocedal and Overton (1985), Gurwitz and Overton (1989), and Xie (1991)). Sipping the BFGS update may appear to be a crude heuristic, but we argue that it gives rise to a sound algorithm. First of all, the last two terms in (3.17) normally converge to zero faster than the rst term, so that the right-hand side of (3.17) will often be positive near the solution and BFGS updating will tae place frequently. Furthermore, if the right-hand side of (3.17) is 10

negative, the range space step Y p Y is relatively large, resulting in sucient progress towards the solution. These arguments will be made more precise in x5. We therefore opt for sipping the BFGS update, when necessary, and we now present a strategy for deciding when to do so. Recall that, dened by (1.34), converges to zero if the iterates converge to x. Update Criterion I Choose aconstant fd > 0 and a sequence ofpositive numbers f g such that 1 =1 < 1 (this is the same sequence f g that was used in (3.14)). If w is computed bybroyden's method, and if both s T y > 0 and p Y 2 p Z (3:18) hold at iteration, thenupdate the matrix B by means of the BFGS formula (1.24) with s and y given by (1.32) and (1.33). Otherwise, set B +1 = B. If w is computed by nite dierences, and if both s T y > 0 and p Y fd p Z = 1=2 (3:19) hold at iteration, thenupdate the matrix B by means of the BFGS formula (1.24) with s and y given by (1.32) and (1.33). Otherwise, set B +1 = B. Note that requires nowledge of the solution vector x and is therefore not computable. However, we will later see that can be replaced by any quantity that is of the same order as the error e, for example, the optimality conditions (Z T g +c ). Nevertheless, for convenience we willleave in (3.19). We now closely consider the properties of the BFGS matrices B when Update Criterion I is used. Let us dene cos = st B s s B s (3:20) which, as we will see, is a measure of the goodness of the null space step Z p Z. We begin by restating a theorem from Byrd and Nocedal (1989) regarding the behavior of cos when the matrix B is updated by the BFGS formula. Theorem 3.1 Let fb g be generated by the BFGS formula (1.24) where, for all 1, s 6=0 and y T s s T s y 2 y T s m>0 (3.21) M: (3.22) Then, there exist constants 1 2 3 > 0 such that, for any 1, the relations hold for at least d 1 e values of j 2 [1 ]. 2 cos j 1 (3.23) 2 B js j s j 3 (3.24) 11

This theorem refers to the iterates for which BFGS updating taes place but since, for the other iterates, B +1 = B, the theorem characterizes the whole sequence of matrices fb g. Theorem 3.1 states that, if s T y is always suciently positive, in the sense that Conditions (3.21) and (3.22) are satised, then at least half of the iterates at which updating taes place are such that cos j is bounded away from zero and B j s j = O(s j ). Since it will be useful to refer easily to these iterates, we mae the following denition. Denition 3.1 We dene J to be the set of iterates for which (3.23) and (3.24) hold. We call J the set of \good iterates" and dene J = J \f1 2 ::: g. Note that if the matrices B are updated only a nite number of times, their condition number is bounded, and (3.23){(3.24) are satised for all. Thus in this case all iterates are good iterates. We now study the case when BFGS updating taes place an innite number of times. Let us assume that all functions under consideration are smooth and bounded. If at a solution point x the reduced Hessian Z T W Z is positive denite, then for all x in a neighborhood of x the smallest eigenvalue of Z T W ~ Z is bounded away fromzero ( W ~ is dened in (3.16)). We now show that in such a neighborhood Update Criterion I implies (3.21){(3.22). Let us rst consider the case when w is computed bybroyden's method. Using (3.17), (3.18), and (3.14), and since converges to zero, we have s T y Cs 2 ; O( 2 s 2 ) ; O( s 2 ) ms 2 (3.25) for some positive constants C m. Also, from (3.15), (3.18), and (3.14) we have that y O(s )+O( 2 s )+O( s ) O(s ): (3.26) We thus see from (3.25){(3.26) that there is a constant M such that for all for which updating taes place, y 2 y T s M which together with (3.25) shows that (3.21){(3.22) hold when Broyden's method is used. If w is computed by the nite-dierence formula (3.2), we see from (1.33) and the Mean Value theorem that there is a matrix ^W such that y = Z T [rl(x +1 +1 ) ;rl(x + Y p Y +1 )] Z T ^W Z s : Reasoning as before we see that (3.25) and (3.26) also hold in this case, and that (3.21){(3.22) are satised in the case when nite dierences are used. We have therefore established the following result. Lemma 3.1 In a neighborhood of a solution point x, and whenever BFGS updating taes place as stipulated by Update Criterion I, s T y is suciently positive in the sense that (3.21){(3.22) hold. 12

3.4. Choosing and We will now see that by appropriately choosing the penalty parameter and the damping parameter for w, the search direction generated by Algorithm I is always a descent direction for the merit function. Moreover, for the good iterates J, it is a direction of strong descent. Since d satises the linearized constraint (1.11), it is easy to show (see Eq. (2.24) of Byrd and Nocedal (1991)) that the directional derivative ofthe`1 merit function in the direction d is given by D (x d )=g T d ; c 1 : (3:27) The fact that the same right inverse of A T Recalling the decomposition (1.28) and using (3.28), we obtain is used in (1.26) and (1.31) implies that g T Y p Y = T c : (3:28) D (x d ) = g T Z p Z ; c 1 + T c = (Z T g + w ) T p Z ; w T p Z ; c 1 + T c : (3.29) Now from (1.32) and (1.27) we have that Substituting this in (3.20), we obtain B s = ; (Z T g + w ): (3:30) cos = ;(ZT g + w ) T p Z Z T g + w p Z : (3:31) Recalling the inequality T c 1 c 1, and using (3.31) in (3.29), we obtain, for all, D (x d ) ;Z T g + w p Z cos ; w T p Z ; ( ; 1 )c 1 : (3:32) Note also from (3.30) and (1.32) that s B s = p Z Z T g + w : (3:33) We now concentrate on the good iterates J, as given in Denition 3.1. If j 2 J, we have from (3.33) and (3.24) that 1 3 Z T j g j + j w j p (j) Z 1 2 Z T j g j + j w j : (3:34) Using this and (3.23) in (3.32), we obtain, for j 2 J, D j (x j d j ) ; 1 3 Z T j g j + j w j 2 cos j ; w T j p(j) Z ; ( j ; j 1 )c j 1 ; 1 Zj T g j 2 ; 2 j cos j 3 3 (g T j Z jw j ) ; j w T j p(j) Z where we have dropped the nonpositive term ;j 2 cos jw j 2 = 3. 3 > 1 (it is dened as an upper bound in (3.24)), we have D j (x j d j ) ; 1 3 Z T j g j 2 + h 2 j cos j jg T j Z jw j j; j w T j p(j) Z 13 ; ( j ; j 1 )c j 1 Since we can assume that i ; ( j ; j 1 )c j 1 :

It is now clear that if for some constant, and if then for all j 2 J, 2 j cos j jg T j Z j w j j; j w T j p (j) Z c j 1 (3:35) j j 1 +2 (3:36) D j (x j d j ) ; 1 3 Z T j g j 2 ; c j 1 : (3:37) This means that if (3.35) and (3.36) hold, then for the good iterates j 2 J, the search direction d j is a strong direction of descent for the `1 merit function in the sense that the rst-order reduction is proportional to the KKT error. We will choose so that (3.35) holds for all iterations. To show how to do this, we note from (1.27) that p Z = ;B ;1 ZT g ; B ;1 w so that, for j =, (3.35) can be written as [2 cos jg T Z w j + w T B;1 ZT g + w T B;1 w ] c 1 : (3:38) Clearly this condition is satised for a suciently small and positive value of. Specically, at the beginning of the algorithm we choose a constant >0 and, at every iteration, dene = minf1 ^ g (3:39) where ^ is the largest value that satises (3.38) as an equality. The penalty parameter must satisfy (3.36), so we dene it at every iteration of the algorithm by ;1 if = ;1 1 +2 (3:40) 1 +3 otherwise. The damping factor and the updating formula for the penalty parameter have been dened so as to give strong descent for the good iterates J. We nowshow that they ensure that the search direction is also a direction of descent (but not necessarily of strong descent) for the other iterates, 62 J. Since (3.35) holds for all iterations by ourchoice of,wehave in particular Using this and (3.40) in (3.32), we have ; w T p Z c 1 : D (x d ) ;Z T g + w p Z cos ; c 1 : (3:41) The directional derivative is thus nonpositive. Furthermore, since w = 0 whenever c = 0 (regardless of whether w is obtained by nite dierences or through Broyden's method), it is easy to show that this directional derivative can be zero only at a stationary point of problem (1.1){(1.2). 3.5. The Algorithm We can now give a complete description of the algorithm that incorporates all the ideas discussed so far and that species the only remaining question, namely, when to apply nite 14

dierences and when to use Broyden's method to approximate the cross term. The idea is to consider the relative sizes of p Y and p Z. Update Criterion I generates the three regions R 1 R 2, and R 3 illustrated in Figure 2. The algorithm starts by computing w through Broyden's method and by calculating p Y and p Z. If the search direction is in R 1 or R 3,we proceed. Otherwise we recompute w by nite dierences, use this value to recompute p Z, and proceed. The reason for applying nite dierences in this fashion is that in the middle region R 2 Broyden's method is not good enough, nor is the convergence suciently tangential, to give a superlinear step. Therefore we must resort to nite dierences to obtain a good estimate of w. The motivation behind this strategy will become clearer when we study the rate of convergence of the algorithm in x6. p Y 1 2 p Y = fd p Z = R 3 p Y = 2 p Z R 2 R 1 p Z Figure 2 Three regions generated by Update Criterion I Note from Updating Criterion I that the BFGS update of B is sipped if the search direction is in R 3. A precise description of the algorithm follows. Algorithm II 1. Choose constants 2 (0 1=2), >0 and 0 with 0 << 0 < 1, and positive constants ;and fd for Conditions (3.13) and (3.19), respectively. For Conditions (3.14) and (3.18), select a summable sequence of positive numbers f g. Set := 1, and choose a starting point x 1, an initial value 1 for the penalty parameter, an (n ; m) (n ; m) symmetric and positive denite starting matrix B 1 and an (n ; m) n starting matrix S 1. 2. Evaluate f g c, and A, and compute Y and Z. 3. Set ndi = false and compute p Y by solving the system (A T Y )p Y = ;c : (range space step) (3:42) 4. Calculate w using Broyden's method, from Equations (3.12) and (3.13). 15

5. Choose the damping parameter from Equations (3.38) and (3.39), and compute p Z from B p Z = ;[Z T g + w ]: (null space step) (3:43) 6. If (3.19) is satised and (3.18) is not satised, set ndi = true and recompute w from Equation (3.1). (In practice we replace by Z T g + c in (3.19).) 7. If ndi = true, use this new value of w to choose the damping parameter from Equations (3.38) and (3.39), and recompute p Z from Equation (3.43). 8. Dene the search direction by and set =1. 9. Test the line search condition d = Y p Y + Z p Z (3:44) (x + d ) (x )+ D (x d ): (3:45) 10. If (3.45) is not satised, choose a new 2 [ 0 ] and go to 9 otherwise set 11. Evaluate f +1 g +1 c +1 A +1, and compute Y +1 and Z +1. 12. Compute the Lagrange multiplier estimate and update so as to satisfy (3.40). x +1 = x + d : (3:46) +1 = ;[Y T +1 A +1] ;1 Y T +1 g +1 (3:47) 13. Update S +1 using Equations (3.9) to (3.11). If ndi = false, calculate w bybroyden's method through Equations (3.12) and (3.14) otherwise calculate w by (3.2). 14. If (s T y 0) or if (findiff = true and (3.19) is not satised) or if (findiff = false and (3.18) is not satised), set B +1 = B. Else, compute s = p Z (3:48) y = Z T [rl(x +1 +1 ) ;rl(x +1 )] ; w (3:49) and compute B +1 bythebfgs formula (1.24). 15. Set := + 1, and go to 3. We mentioned in x3.1 that, when using nite dierences, there are various ways of dening w and w, but for concreteness we now assume in steps 6 and 13 that they are computed by (3.1) and (3.2), respectively. We should also point out that the curves in Figure 2 may intersect, creating a fourth region, and in practice we should stipulate a new set of conditions in this region. We discuss these conditions in another paper that considers the implementation of the algorithm (Biegler, Nocedal, and Schmid (1993)). In the next sections we present several convergence results for Algorithm II. The analysis, which does not assume that the BFGS matrices B or the Broyden matrices S are bounded, is based on the results of Byrd and Nocedal (1991), who have studied the convergence of the Coleman-Conn updating algorithm. We also mae use of some results of Xie (1991), who has 16

analyzed the algorithm proposed by NocedalandOverton (1985) using nonorthogonal bases Y and Z. The main dierence between this paper and that of Xie stems from our use of the correction terms w and w,which are not employed in his method. 4. Semi-Local Behavior of the Algorithm We rst show that the merit function decreases signicantly at the good iterates J and that this gives the algorithm a wea convergence property. To establish the results of this section, we mae the following assumptions. Assumptions 4.1 The sequence fx g generated by Algorithm II is contained in a convex set D with the following properties: (I) The functions f : R n! R and c : R n! R m and their rst and second derivatives are uniformly bounded in norm over D. (II) The matrix A(x) has full column ran for all x 2 D, and there exist constants 0 and 0 such that Y (x)[a(x) T Y (x)] ;1 0 Z(x) 0 (4:1) for all x 2 D. (III) For all 1 for which B is updated, (3.21) and (3.22) hold. (IV) The correction term w is chosen so that there is a constant >0such that for all, w c 1=2 : (4:2) Note that Condition (I) is rather strong, since it would often be satised only if D is bounded, and it is far from certain that the iterates will remain in a bounded set. Nevertheless, the convergence result of this section can be combined with the local analysis of x5 to give a satisfactory semi-global result. Condition (II) requires that the basis matrices Y and Z be chosen carefully, and is important to obtain good behavior in practice. Note that (4.1) and (3.42) imply that Y p Y 0 c : (4:3) Condition (III) is justied by Lemma 3.1. Condition (III) and Theorem 3.1 ensure that at least half of the iterates at which BFGS updating taes place are good iterates. Wehave left some freedom in the choice of w since (4.2) suces for the analysis of this section. Relation (4.2) holds for the nite-dierence approach, since (3.1) implies that w = O(Y p Y )and since (I) ensures that fc g is uniformly bounded (see (5.21)). Furthermore, the safeguard (3.13) and (4.3) immediately imply that (4.2) is satised when the Broyden approximation is used. The following result concerns the good iterates J, as given in Denition 3.1. Lemma 4.1 If Assumptions 4.1 hold and if j = is constant for all suciently large j, then there isapositive constant such that for all large j 2 J, (x j ) ; (x j+1 ) Z T j g j 2 + c j 1 : (4:4) 17

Proof. Using (3.37), we have for all j 2 J D j (x j d j ) ;b 2 Z T j g j 2 + c j 1 (4:5) where b 2 =min( 1 = 3 ). Note that the line search enforces the Armijo condition (3.45), j (x j ) ; j (x j+1 ) ; j D j (x j d j ): (4:6) It is then clear from (4.5) that (4.4) holds, provided the j, j 2 J, can be bounded from below. Suppose that j < 1, which means that (4.6) failed for a steplength ~: where j (x j +~d j ) ; j (x j ) >~D j (x j d j ) (4:7) ~ j (4:8) (see Step 10 of Algorithm II). On the other hand, expanding to second order, we have j (x j +~d j ) ; j (x j ) ~D j (x j d j )+~ 2 b 1 d j 2 (4:9) where b 1 depends on j. Combining (4.7) and (4.9), we have Next we show that, for j 2 J, ( ; 1)~D j (x j d j ) < ~ 2 b 1 d j 2 : (4:10) d j 2 b 3 [Z T j g j 2 + c j 1 ] (4:11) for some constant b 3. To dothis,we mae repeated use of the following elementary result: Using (3.44), (4.12), (4.1), and (4.3), we have a b 0 ) a 2 +2ab + b 2 3a 2 +3b 2 : (4:12) d j 2 Z j p (j) Z 2 +2Z j p (j) Z Y j p (j) Y + Y j p (j) Y h i 2 3 Z j p (j) Z 2 + Y j p (j) Y 2 h i 3 2 0 p(j) Z 2 + 2 c 0 j 2 : (4.13) Also by (3.34), (4.12), and (4.2) and noting that 1,wehave that for j 2 J p (j) Z 2 1 2 2 3 2 2 3 2 2 Z T j g j 2 +2 j Z T j g jw j + 2 j w j 2 Z T j g j 2 + 2 j w j 2 Z T j g j 2 + 2 c j 1 since j 1. Since c j 1 is uniformly bounded on D, we see from this relation and (4.13) that (4.11) holds, where b 3 =maxf9 2 0 =2 2 3(32 2 0 =2 2 + 2 0 sup c(x))g: x2d 18

Combining (4.10), (4.5), and (4.11), and recalling that <1, we obtain ~ > (1 ; )b 2 b 1 b 3 : (4:14) This relation and (4.8) imply that the steplengths j are bounded away from zero for all j 2 J. Since by assumption j = for all large j, we conclude that (4.4) holds with = b 2 minf1 (1; )b 2 =(b 1 b 3 )g. 2 It is now easytoshow that the penalty parameter settles down and that the set of iterates is not bounded away from stationary points of the problem. Theorem 4.2 If Assumptions 4.1 hold, then the weights f g are constant for all suciently large and lim inf!1 (ZT g + c )=0: Proof. First note that by Assumptions 4.1 (I){(II) and (3.47) that f g is bounded. Therefore, since the procedure (3.40) increases by at least whenever it changes the penalty parameter, it follows that there are an index 0 and a value such that for all > 0, = +2. If BFGS updating is performed an innite number of times, by Assumptions 4.1-(III) and Theorem 3.1 there is an innite set J of good iterates, and by Lemma 4.1 and the fact that the Armijo condition (3.45) forces (x ) to decrease at each iterate, we have that for > 0, (x 0 ) ; (x +1 ) = X X j= 0 ( (x j ) ; (x j+1 )) j2j\[ X 0 ] j2j\[ 0 ] ( (x j ) ; (x j+1 )) [Z T j g j 2 + c j 1 ]: By Assumption 4.1(I) (x) is bounded below for all x 2 D, so the last sum is nite, and thus the term inside the square bracets converges to zero. Therefore lim (Zj T g j + c j 1 )=0: (4:15) j2j j!1 If BFGS updating is performed a nite number of times, then, as discussed after Denition 3.1, all iterates are good iterates, and in this case we obtain the stronger result lim!1 (ZT g + c 1 )=0: 2 5. Local Convergence In this section we show thatifx is a local minimizer that satises the second-order optimality conditions, and if the penalty parameter is chosen large enough, then x is a point of attraction 19

for the sequence of iterates fx g generated by Algorithm II. To prove this result, we will mae the following assumptions. In what follows, G denotes the reduced Hessian of the Lagrangian function, namely, G = Z T r2 xx L(x )Z : (5:1) Assumptions 5.1 The point x is a local minimizer for problem (1.1){(1.2), at which the following conditions hold. (1) The functions f : R n! R and c : R n! R m are twice continuously dierentiable in a neighborhood of x, and their Hessians are Lipschitz continuous in a neighborhood of x. (2) The matrix A(x ) has full column ran. This implies that there exists a vector 2 R m such that rl(x )=g(x )+A(x ) =0: (3) For all q 2 R n;m, q 6= 0,wehave q T G q>0: (4) There exist constants 0, 0, and c such that, for all x in a neighborhood of x, Y (x)[a(x) T Y (x)] ;1 0 Z(x) 0 (5:2) and [Y (x) Z(x)] ;1 c : (5:3) (5) Z(x) and (x) are Lipschitz continuous in a neighborhood of x. That is, there exist constants Z and such that for all x z near x. (x) ; (z) x ; z (5.4) Z(x) ; Z(z) Z x ; z (5.5) Note that (1), (3), and (5) imply that for all (x ) suciently near (x ), and for all q 2 R n;m, mq 2 q T G(x )q Mq 2 (5:6) for some positive constants m M. We also note that Assumptions 5.1 ensure that the conditions (3.21){(3.22) required by Theorem 3.1 hold whenever BFGS updating taes place in a neighborhood of x, as shown in Lemma 3.1. Therefore Theorem 3.1 can be applied in the convergence analysis. The following two lemmas are proved by Xie (1991) for very general choices of Y and Z. His result generalizes Lemmas 4.1 and 4.2 of Byrd and Nocedal (1991) see also Powell (1978). Lemma 5.1 If Assumptions 5.1 hold, then for all x suciently near x for some positive constants 1 2. 1 x ; x c(x) + Z(x) T g(x) 2 x ; x (5:7) This result states that, near x, the quantities c(x) and Z(x) T g(x) may be regarded as a measure of the error at x. The next lemma states that, for a large enough weight, the merit function may also be regarded as a measure of the error. 20

Lemma 5.2 Suppose that Assumptions 5.1 hold at x. Then for any > 1 there exist constants 3 > 0 and 4 > 0, such that for all x suciently near x 3 x ; x 2 (x) ; (x ) 4 Z(x) T g(x) 2 + c(x) 1 : (5:8) Note that the left inequality in (5.8) implies that, for a suciently large value of the penalty parameter, the merit function will have a strong local minimizer at x. We will now use the descent property of Algorithm II to show convergence of the algorithm. However, because of the nonconvexity of the problem, the line search could generate a step that decreases the merit function but that taes us away from the neighborhood of x. To rule this out, we mae the following assumption. Assumption 5.2 The line search has the property that, for all large, ((1 ; )x + x +1 ) (x ) for all 2 [0 1]. In other words, x +1 is in the connected component of the level set fx : (x) (x )g that contains x. There is no practical line search algorithm that can guarantee this condition, but it is liely to hold close to x. Assumption 5.2 is made by Byrd, Nocedal, and Yuan (1987) when analyzing the convergence of variable metric methods for unconstrained problems, as well as by Byrd and Nocedal (1991) in the analysis of Coleman-Conn updates for equality constrained optimization. Lemma 5.3 Suppose that the iterates generated by Algorithm II (with a line search satisfying Assumptions 5.2) are contained in a convex region D satisfying Assumptions 4.1. If an iterate x 0 is suciently close to a solution point x that satises Assumptions 5.1, and if the weight 0 is large enough, then the sequence ofiterates converges to x. Proof. By Assumptions 4.1 (I){(II) and (3.47) we now that f g is bounded. Therefore the procedure (3.40) ensures that the weights are constant, say = for all large. Moreover, if an iterate gets suciently close to x,wenowby (3.40) and by the continuityofthat >. For such value of, Lemma 5.2 implies that the merit function has a strict local minimizer at x. Now suppose that once the penalty parameter has settled, and for a given >0, there is an iterate x 0 such that jjx 0 ; x jj 3 2 2 4^ 0 where ^ 0 is such that 1 ^ 0. Assumption 5.2 shows that for any 0, x is in the connected component of the level set of x 0 that contains x 0,andwe can assume that is small enough that Lemmas 5.1 and 5.2 hold in this level set. Thus since (x ) (x 0 )for 0, and since we can assume that Z T 0 g 0 1, we have from Lemmas 5.1 and 5.2, for any 0 jjx ; x jj ; 1 2 3 ( (x ) ; (x )) 1 2 ; 1 2 3 ( (x 0 ) ; (x )) 1 2 4 3 1 2 Z T 0 g 0 2 + c 0 1 1 2 1 4 2 Z T 1 0 g 0 2 +^ 0 c 0 2 3 21

: 2 4^ 0 3 1 2 x 0 ; x jj This implies that the whole sequence of iterates remains in a neighborhood of radius of x. If is small enough, we concludeby(5.8),by the monotonicity off (x )g, and by Theorem 4.2 that the iterates converge to x. 2 The assumptions of this lemma, which is modeled after a result in Xie (1991), are restrictive especially the assumption on the penalty parameter. One can relax these assumptions and obtain a stronger result, such as Theorem 4.3 in Byrd and Nocedal (1991), but the proof would be more complex and is not particularly relevant to Algorithm II since it is based only on the properties of the merit function. Therefore, instead of further analyzing the local convergence properties of the new algorithm, we will study its rate of convergence. 5.1. R-Linear Convergence For the rest of the paper we assume that the line search strategy satises Assumption 5.2. We also assume that the iterates generated by Algorithm II converge to a point x at which Assumptions 5.1 hold, which implies that for all large, = >. The analysis that follows depends on how often BFGS updating is applied. To mae this concept precise, we dene U to be the set of iterates at which BFGS updating taes place, and let The number of elements in U will be denoted by ju j. U = f : B +1 = BFGS(B s y )g (5:9) U = U \f1 2 ::: g: (5:10) Theorem 5.4 Suppose that the iterates fx g generated by Algorithm II converge to a point x that satises Assumptions 5.1. Then for any 2 U and any j for some constants C>0 and 0 r<1. Proof. Using (4.4) and (5.8), we have for i 2 J, Let us dene r =(1; = 4 ) 1 4. Then for i 2 J x j ; x Cr ju j (5:11) (x i ) ; (x i+1 ) 4 [ (x i ) ; (x )] : (5:12) (x i+1 ) ; (x ) r 4 [ (x i ) ; (x )] : (5:13) We now that the merit function decreases at each step, and by (5.8) we have, for j and 2 U, x j ; x ; 1 2 3 ( (x j ) ; (x )) 1 2 ; 1 2 3 ( (x ) ; (x )) 1 2 : 22

We continue in this fashion, bounding the right-hand side by terms involving earlier iterates, but using now (5.13) for all good iterates. Since by Theorem 3.1 at least half of the iterates at which updating taes place are good iterates (i.e., jj j 1 2 ju j), we have x j ; x ; 1 2 3 ; 1 2 3 h r 4jJj ( (x 1 ) ; (x ))i 1 2 h r 2jUj ( (x 1 ) ; (x ))i 1 2 [ ; 1 2 3 ( (x 1 ) ; (x )) 1 2 ]r ju j Cr ju j : This result implies that if fju j=g is bounded away from zero, then Algorithm II is R-linearly convergent. However, BFGS updating could tae place only a nite number of times, in which case this ratio would converge to zero. It is also possible for BFGS updating to tae place an innite number of times, but every time less often, in such away thatju j=! 0. We therefore need to examine the iteration more closely. We mae use of the matrix function dened by (B) =tr(b) ; ln(det(b)) (5:14) where tr denotes the trace, and det the determinant. It can be shown that ln cond(b) < (B) (5:15) for any positive denite matrix B (Byrd and Nocedal (1989)). We alsomae use of the weighted quantities ~y = G ;1=2 y ~s = G 1=2 s (5:16) and ~B = G ;1=2 B G ;1=2 (5:17) cos ~ = ~st ~ B ~s ~ B ~s ~s (5:18) ~q = ~st ~ B ~s ~s T ~s : (5:19) One can show (see Eq. (3.22) of Byrd and Nocedal (1989)) that if B is updated by the BFGS formula, then ( ~ B+1 ) = ( ~ B )+ ~y 2 + ~y T ~s 1 ; ~q cos 2 ~ +ln ; 1 ; ln ~yt ~s ~s T ~s + ln cos 2 ~ ~q cos 2 ~ : (5.20) This expression characterizes the behavior of the BFGS matrices B and will be crucial to the analysis of this section. Before we can mae use of this relation, however, we need to consider the accuracy of the correction terms. We begin by showing that when nite dierences are used to estimate w and w, these are accurate to second order. 23 2

Lemma 5.5 If at the iterate x,thecorrections w and w are computed by the nite-dierence formulae (3.1){(3.2), and if x is suciently close to a solution point x that satises Assumptions 5.1, then w = O(p Y ) (5:21) and w ; Z T W Y p Y = O( p Y ) (5:22) w ; Z T W Y p Y = O( p Y ): (5:23) Proof. Recalling that rl(x ) =g(x)+a(x), wehave from (3.1) that w = Z T [rl(x + Y p Y ) ;rl(x )] = Z T [rl(x + Y p Y ) ;rl(x )] + Z T [(A(x + Y p Y ) ; A )( ; )] Z 1 = Z T r 2 xxl(x + Y p Y )d Y p Y + Z T [(A(x + Y p Y ) ; A )( ; )] 0 Z T W Y p Y + Z T [(A(x + Y p Y ) ; A )( ; )]: (5.24) Let us assume that x is in the neighborhood of x where (5.2){(5.5) hold. Then ; = O(e )=O( ), where is dened by (1.34).Therefore the last term in (5.24) is O(p Y ), which proves (5.21). Also, a simple computation shows that [Z T W ; Z T W ]Y p Y = O( p Y ): (5:25) Using these facts in (5.24) yields the desired result (5.22). To prove (5.23), we note only that 1 and reason in the same manner. 2 Next we show that the condition number of the matrices B is bounded and that, in the limit, at the iterates U at which BFGS updating taes place, the matrices B are accurate approximations of the reduced Hessian of the Lagrangian. Theorem 5.6 Suppose that the iterates fx g generated by Algorithm II converge to a solution point x that satises Assumptions 5.1. Then fb g and fb ;1 g are bounded, and for all 2 U (B ; Z T W Z )p Z = o(d ): (5:26) Proof. We will only consider iterates for which BFGS updating of B taes place. We have from (3.49), (3.46), (3.44), (3.16), and (3.48) y = Z T [rl(x +1 +1 ) ;rl(x +1 )] ; w Z 1 = Z T r 2 xxl(x + d +1 )d d ; w 0 = Z T W ~ (Z p Z + Y p Y ) ; w = Z T W ~ Z s + (Z T W ~ ; Z T W )Y p Y +( Z T W Y p Y ; w ): (5.27) Since w can be computed by Broyden's method or by nite dierences, we need to consider these two cases separately. 24