Pacific Journal of Optimization (Vol. 2, No. 3, September 2006) ABSTRACT

Similar documents
1. Introduction. We consider the mathematical programming problem

Examples of dual behaviour of Newton-type methods on optimization problems with degenerate constraints

Some new facts about sequential quadratic programming methods employing second derivatives

On attraction of linearly constrained Lagrangian methods and of stabilized and quasi-newton SQP methods to critical multipliers

Rejoinder on: Critical Lagrange multipliers: what we currently know about them, how they spoil our lives, and what we can do about it

A globally convergent Levenberg Marquardt method for equality-constrained optimization

Journal of Convex Analysis Vol. 14, No. 2, March 2007 AN EXPLICIT DESCENT METHOD FOR BILEVEL CONVEX OPTIMIZATION. Mikhail Solodov. September 12, 2005

A note on upper Lipschitz stability, error bounds, and critical multipliers for Lipschitz-continuous KKT systems

Journal of Convex Analysis (accepted for publication) A HYBRID PROJECTION PROXIMAL POINT ALGORITHM. M. V. Solodov and B. F.

Inexact Josephy Newton framework for generalized equations and its applications to local analysis of Newtonian methods for constrained optimization

CONVERGENCE ANALYSIS OF AN INTERIOR-POINT METHOD FOR NONCONVEX NONLINEAR PROGRAMMING

c 2012 Society for Industrial and Applied Mathematics

Merit functions and error bounds for generalized variational inequalities

On attraction of linearly constrained Lagrangian methods and of stabilized and quasi-newton SQP methods to critical multipliers

Some Inexact Hybrid Proximal Augmented Lagrangian Algorithms

A convergence result for an Outer Approximation Scheme

Newton-Type Methods: A Broader View

On the acceleration of augmented Lagrangian method for linearly constrained optimization

Local convergence of the method of multipliers for variational and optimization problems under the noncriticality assumption

Lecture 3: Lagrangian duality and algorithms for the Lagrangian dual problem

ON REGULARITY CONDITIONS FOR COMPLEMENTARITY PROBLEMS

Some Properties of the Augmented Lagrangian in Cone Constrained Optimization

On well definedness of the Central Path

Damián Fernández 1 and Mikhail Solodov 2*

Priority Programme 1962

Globalizing Stabilized Sequential Quadratic Programming Method by Smooth Primal-Dual Exact Penalty Function

Critical solutions of nonlinear equations: local attraction for Newton-type methods

Combining stabilized SQP with the augmented Lagrangian algorithm

On the Local Quadratic Convergence of the Primal-Dual Augmented Lagrangian Method

Lecture 3. Optimization Problems and Iterative Algorithms

Constrained Optimization

ON A CLASS OF NONSMOOTH COMPOSITE FUNCTIONS

The effect of calmness on the solution set of systems of nonlinear equations

Constrained Optimization and Lagrangian Duality

Critical solutions of nonlinear equations: Stability issues

A FRITZ JOHN APPROACH TO FIRST ORDER OPTIMALITY CONDITIONS FOR MATHEMATICAL PROGRAMS WITH EQUILIBRIUM CONSTRAINTS

MODIFYING SQP FOR DEGENERATE PROBLEMS

MS&E 318 (CME 338) Large-Scale Numerical Optimization

Technische Universität Dresden Institut für Numerische Mathematik

FIXED POINTS IN THE FAMILY OF CONVEX REPRESENTATIONS OF A MAXIMAL MONOTONE OPERATOR

A doubly stabilized bundle method for nonsmooth convex optimization

FIRST- AND SECOND-ORDER OPTIMALITY CONDITIONS FOR MATHEMATICAL PROGRAMS WITH VANISHING CONSTRAINTS 1. Tim Hoheisel and Christian Kanzow

ON LICQ AND THE UNIQUENESS OF LAGRANGE MULTIPLIERS

Lagrange Relaxation and Duality

A GLOBALLY CONVERGENT STABILIZED SQP METHOD: SUPERLINEAR CONVERGENCE

On the convergence properties of the projected gradient method for convex optimization

Nonlinear Programming, Elastic Mode, SQP, MPEC, MPCC, complementarity

SOME STABILITY RESULTS FOR THE SEMI-AFFINE VARIATIONAL INEQUALITY PROBLEM. 1. Introduction

5 Handling Constraints

Convergence Analysis of Perturbed Feasible Descent Methods 1

A STABILIZED SQP METHOD: SUPERLINEAR CONVERGENCE

Generalization to inequality constrained problem. Maximize

Introduction to Nonlinear Stochastic Programming

Subgradient Methods in Network Resource Allocation: Rate Analysis

Primal/Dual Decomposition Methods

A Proximal Method for Identifying Active Manifolds

Introduction to Optimization Techniques. Nonlinear Optimization in Function Spaces

INTERIOR-POINT METHODS FOR NONCONVEX NONLINEAR PROGRAMMING: CONVERGENCE ANALYSIS AND COMPUTATIONAL PERFORMANCE

A Continuation Method for the Solution of Monotone Variational Inequality Problems

Iteration-complexity of a Rockafellar s proximal method of multipliers for convex programming based on second-order approximations

Numerical Optimization

A globally and R-linearly convergent hybrid HS and PRP method and its inexact version with applications

ON AUGMENTED LAGRANGIAN METHODS WITH GENERAL LOWER-LEVEL CONSTRAINTS. 1. Introduction. Many practical optimization problems have the form (1.

Solving Dual Problems

Extreme Abridgment of Boyd and Vandenberghe s Convex Optimization

A Primal-Dual Interior-Point Method for Nonlinear Programming with Strong Global and Local Convergence Properties

The Squared Slacks Transformation in Nonlinear Programming

Solving generalized semi-infinite programs by reduction to simpler problems.

1 Introduction We consider the problem nd x 2 H such that 0 2 T (x); (1.1) where H is a real Hilbert space, and T () is a maximal monotone operator (o

Convex Optimization & Lagrange Duality

Critical Lagrange multipliers: what we currently know about them, how they spoil our lives, and what we can do about it

AN AUGMENTED LAGRANGIAN AFFINE SCALING METHOD FOR NONLINEAR PROGRAMMING

c 2002 Society for Industrial and Applied Mathematics

Finite Convergence for Feasible Solution Sequence of Variational Inequality Problems

A QP-FREE CONSTRAINED NEWTON-TYPE METHOD FOR VARIATIONAL INEQUALITY PROBLEMS. Christian Kanzow 1 and Hou-Duo Qi 2

WHEN ARE THE (UN)CONSTRAINED STATIONARY POINTS OF THE IMPLICIT LAGRANGIAN GLOBAL SOLUTIONS?

2.098/6.255/ Optimization Methods Practice True/False Questions

McMaster University. Advanced Optimization Laboratory. Title: A Proximal Method for Identifying Active Manifolds. Authors: Warren L.

Convex Optimization and Modeling

Using exact penalties to derive a new equation reformulation of KKT systems associated to variational inequalities

Quiz Discussion. IE417: Nonlinear Programming: Lecture 12. Motivation. Why do we care? Jeff Linderoth. 16th March 2006

Constrained Optimization Theory

Decision Science Letters

Constraint Identification and Algorithm Stabilization for Degenerate Nonlinear Programs

The Relation Between Pseudonormality and Quasiregularity in Constrained Optimization 1

Nonlinear Programming and the Kuhn-Tucker Conditions

AN INEXACT HYBRID GENERALIZED PROXIMAL POINT ALGORITHM AND SOME NEW RESULTS ON THE THEORY OF BREGMAN FUNCTIONS. M. V. Solodov and B. F.

A sensitivity result for quadratic semidefinite programs with an application to a sequential quadratic semidefinite programming algorithm

AN INEXACT HYBRID GENERALIZED PROXIMAL POINT ALGORITHM AND SOME NEW RESULTS ON THE THEORY OF BREGMAN FUNCTIONS. May 14, 1998 (Revised March 12, 1999)

On proximal-like methods for equilibrium programming

An inexact subgradient algorithm for Equilibrium Problems

An Accelerated Hybrid Proximal Extragradient Method for Convex Optimization and its Implications to Second-Order Methods

Convex Optimization M2

Stationary Points of Bound Constrained Minimization Reformulations of Complementarity Problems1,2

5. Duality. Lagrangian

A Primal-Dual Augmented Lagrangian Penalty-Interior-Point Filter Line Search Algorithm

Convex Optimization. Dani Yogatama. School of Computer Science, Carnegie Mellon University, Pittsburgh, PA, USA. February 12, 2014

AN ABADIE-TYPE CONSTRAINT QUALIFICATION FOR MATHEMATICAL PROGRAMS WITH EQUILIBRIUM CONSTRAINTS. Michael L. Flegel and Christian Kanzow

Math 273a: Optimization Subgradients of convex functions

Optimality, Duality, Complementarity for Constrained Optimization

Transcription:

Pacific Journal of Optimization Vol., No. 3, September 006) PRIMAL ERROR BOUNDS BASED ON THE AUGMENTED LAGRANGIAN AND LAGRANGIAN RELAXATION ALGORITHMS A. F. Izmailov and M. V. Solodov ABSTRACT For a given iterate generated by the augmented Lagrangian or the Lagrangian relaxation based method, we derive estimates for the distance to the primal solution of the underlying optimization problem. The estimates are obtained using some recent contributions to the sensitivity theory, under appropriate first or second order sufficient optimality conditions. The given estimates hold in situations where nown algorithm-independent) error bounds may not apply. Examples are provided which show that the estimates are sharp. Key words. Error bound, augmented Lagrangian, Lagrangian relaxation, sensitivity. AMS subject classifications. 90C30, 90C33, 90C55, 65K05 Research of the first author is supported by the Russian Foundation for Basic Research Grant 04-01- 00341 and the RF President s Grant NS-1815.003.1 for the support of leading scientific schools. The second author is supported in part by CNPq Grants 300734/95-6 and 471780/003-0, by PRONEX Optimization, and by FAPERJ. Moscow State University, Faculty of Computational Mathematics and Cybernetics, Department of Operations Research, Leninsiye Gori, GSP-, 11999 Moscow, Russia. Email: izmaf@ccas.ru Instituto de Matemática Pura e Aplicada, Estrada Dona Castorina 110, Jardim Botânico, Rio de Janeiro, RJ 460-30, Brazil. Email: solodov@impa.br

1 Introduction We consider the optimization problem minimize fx) subject to F x) = 0, Gx) 0, 1.1) where f : R n R is a smooth function, F : R n R l and G : R n R m are smooth mappings. Specifically, for a local solution x of 1.1), we assume that f, F and G are twice differentiable at x. The stationary points of problem 1.1) and the associated Lagrange multipliers are characterized by the Karush Kuhn Tucer KKT) optimality system where L x, λ, µ) = 0, F x) = 0, µ 0, Gx) 0, µ, Gx) = 0, 1.) x L : R n R l R m R, Lx, λ, µ) = fx) + λ, F x) + µ, Gx), is the standard Lagrangian function of problem 1.1). Let M x) be the set of Lagrange multipliers associated with x. In this paper, we do not invoe any specific constraint qualifications, but the main line of our discussion presumes that the multiplier set M x) is nonempty. This setting has been lately receiving much attention in the literature, e.g., [7, 1, 11, 6, 17], with one of the important motivations coming from optimization problems with complementarity constraints. We shall further assume a first or second order) sufficient condition for optimality, so that x is a strict local solution of problem 1.1). In this setting, a primal) local error bound is the following property: x x = Oδx, λ, µ)), 1.3) where δ : R n R l R m R +. The function δ should be easily) computable and should provide a reasonable upper bound for the distance from x to x. For example, it should tend to zero when x, λ, µ) tends to { x} M x), or to a given point of this set. Moreover, it is desirable that the bound should be sharp, i.e., not improvable under the given assumptions. The point x, λ, µ) in 1.3) can either be arbitrary or can be generated by some primal-dual) algorithm for solving 1.1) so that there is a certain relation between x and λ, µ), in which case δ may also depend on parameters involved in the algorithm). In the former case, we shall call the error bound algorithm-independent, and in the latter case algorithm-based. We refer the reader to [14] for a survey of error bounds and their applications. The setting of an arbitrary x, λ, µ) is somewhat more traditional in the study of error bounds. Nevertheless, the case when the error bound holds along trajectories generated by some specific algorithm is also of interest and importance. The significance of an algorithm-based error bound in the context of the algorithm being considered, is essentially the same as that of an algorithmindependent bound. In particular, both provide quantitative information about convergence and, as a consequence, a reliable stopping test for the algorithm without a valid error bound, a stopping test based on the corresponding residual says nothing about how close we are to a solution of the problem). Furthermore, it can be possible to obtain an algorithm-based 1

error bound under assumptions weaer or different from the ones required for an algorithmindependent bound, as discussed next. Let C x) be the critical cone of problem 1.1) at x, that is, C x) = {ξ R n F x)ξ = 0, G i x), ξ 0, i A x), f x), ξ 0}, where A x) = {i = 1,..., m G i x) = 0} is the set of indices of inequality constraints active at x. It is nown [8, Lemma ], [5, Theorem ]) that an algorithm-independent bound holds in a neighbourhood of x, λ, µ) satisfying the second-order sufficient condition L x x, λ, µ)ξ, ξ > 0 ξ C x) \ {0}. 1.4) In that case, 1.3) holds with δ being the norm of the natural residual of the violation of the KKT conditions 1.). We note that the cited result assumes only the above second-order sufficient condition, and in particular does not require any constraint qualifications. There exist also other error bounds under assumptions which subsume some constraint qualifications, see [1] for a detailed discussion and comparisons. In the sequel, we shall derive algorithm-based error bounds related to the classical augmented Lagrangian and Lagrangian relaxation algorithms. We shall further provide examples showing that the given estimates are sharp, i.e., that they cannot be improved. Our analysis will assume either the first-order sufficient condition FOSC): C x) = {0}, 1.5) or the second-order sufficient condition SOSC): ξ C x) \ {0} λ, µ) M x) s.t. L x, λ, µ)ξ, ξ > 0. 1.6) x Obviously, 1.6) is a weaer assumption than 1.4), as the latter needs existence of the universal multiplier λ, µ) M x), suitable for all ξ C x) \ {0}. But perhaps even more importantly, even if the universal multipliers exist, we shall mae no assumption of the dual sequence generated by the the given algorithm converging to the set of universal multipliers. In other words, the dual sequence will be allowed to approach the multipliers which do not satisfy 1.4), as long as 1.6) holds. In this situation, the algorithm-independent bound based on 1.4) is not applicable along the trajectory of the algorithm. Of course, FOSC 1.5) implies both 1.4) and 1.6). But we shall consider FOSC separately, because it allows to obtain a better estimate. Note also that if M x) =, then FOSC and SOSC are formally equivalent. Some Sensitivity Results In this section, we state some recent sensitivity results [10], adapted for our purposes.

Let x R n be a strict local solution of 1.1), and let f, F and G be twice differentiable at x. Fix some ε > 0 such that x is the unique global) solution of the problem minimize fx) subject to F x) = 0, Gx) 0, x B ε x),.1) where B ε x) = {x R n x x ε}. For each pair y, z) R l R m, consider the following perturbation of problem.1): minimize fx) subject to F x) = y, Gx) z, x B ε x)..) Let ωy, z) and Sy, z) stand for the optimal value and the solution set of problem.), respectively. In Theorem.1 below, we assume that for y, z) R l R m, the following upper bound on the optimal value ωy, z) holds: ωy, z) f x) + O y, z) )..3) If the point x satisfies the Mangasarian Fromovitz constraint qualification, then.3) holds for arbitrary perturbations. However, in the absence of constraint qualifications, this property does not hold for arbitrary perturbations. But, as we show below,.3) is satisfied in the context of this paper. This is precisely the advantage of algorithm-based setting, which induces rather specific perturbations. The following theorem is a direct consequence of [10, Theorems, 3]. Theorem.1 Assume that.3) is satisfied. Then for y, z) R l R m, the following assertions hold: i) If FOSC 1.5) holds, then ii) If SOSC 1.6) holds, then and.4) holds as well. sup x x = O y, z) ), x Sy, z) ωy, z) = f x) + O y, z) )..4) sup x x = O y, z) 1/ ), x Sy, z) 3

3 Augmented Lagrangian In this section, all norms are -norms. Let c > 0. The augmented Lagrangian for problem 1.1) is given by L c : R n R l R m R, L c x, λ, µ) = fx) + λ, F x) + c F x) + 1 max{0, cg i x) + µ i }) µ i ). c Given some c > 0 and λ, µ) R n R m +, consider the associated subproblem minimize L c x, λ, µ) subject to x R n. 3.1) We next establish the relationship between 3.1) and perturbations of the original problem 1.1). Proposition 3.1 For any λ, µ) R n R m + and c > 0, the point x λ, µ, c which solves 3.1) is a solution of minimize fx) subject to F x) = F x λ, µ, c ), G i x) max{ µ i /c, G i x λ, µ, c )}, i = 1,..., m. 3.) Proof. Clearly, x λ, µ, c is feasible in 3.). Assume that there exists ˆx which is also feasible in 3.) and it holds that fˆx) < fx λ, µ, c ). We first show that max{0, cg i ˆx) + µ i } max{0, cg i x λ, µ, c ) + µ i }, i = 1,..., m. 3.3) The relation obviously holds if cg i ˆx) + µ i 0. Suppose cg i ˆx) + µ i > 0. Then G i ˆx) > µ i /c, and the fact that ˆx is feasible in 3.) implies that G i ˆx) G i x λ, µ, c ). Therefore, 3.3) holds also in that case. Using 3.3) and the fact that F ˆx) = F x λ, µ, c ), we further obtain L c ˆx, λ, µ) = fˆx) + λ, F ˆx) + c F ˆx) + 1 max{0, cg i ˆx) + µ i }) µ i ) c < fx λ, µ, c ) + λ, F x λ, µ, c ) + c F x λ, µ, c) + 1 max{0, cg i x λ, µ, c ) + µ i }) µ i ) c = L c x λ, µ, c, λ, µ), which contradicts the definition of x λ, µ, c. 4

The augmented Lagrangian algorithm or the method of multipliers), see [9, 15, 16,, 4], is the following procedure. Given some λ, µ ) R n R m + and c > 0, the primal iterate x is generated by solving 3.1) with λ, µ) = λ, µ ), c = c. After this, the multiplier estimates are updated by λ +1 = λ + c F x ), µ +1 i = max{0, µ i + c G i x )}, i = 1,..., m, 3.4) the parameter c is possibly adjusted, and the process is repeated. Assuming that the iterative process described above generates a primal sequence {x } converging to the strict local solution x, and a bounded dual sequence {λ, µ )}, we are interested in quantifying the convergence, i.e., obtaining an estimate of the distance from x to x in terms of a nown quantity. Recalling 3.), for each define ) 1/ δ = F x ) + max{ µ i /c, G i x )}). 3.5) Theorem 3.1 Let c c for all, where c > 0 is arbitrary. Suppose that the sequence {λ, µ )} generated according to 3.4) is bounded. For each, let x be a solution of 3.1) with λ, µ) = λ, µ ), c = c. Suppose that the sequence {x } converges to x, which is a solution of 1.1). Then the following assertions hold: i) If FOSC 1.5) holds, then ii) If SOSC 1.6) holds, then and 3.7) holds as well. x x = Oδ ), 3.6) fx ) = f x) + Oδ ). 3.7) x x = Oδ 1/ ), 3.8) Proof. Since {x } converges to x, we can assume that x B ε x) for all, where ε is the same as in.1). We first show that under the given assumptions, δ 0 as. Clearly, {F x )} 0. For i A x), since G i x ) 0 while µ i /c 0, it holds that max{ µ i /c, G i x )} 0 as. Let i A x). Then for all large enough, say, it holds that G i x ) β, where β > 0. Suppose that there exists an infinite subsequence of indices { l } such that µ l i > 0 for all l. For l >, we obtain that 0 < µ l i = µ l 1 i + c l 1G i x l 1 l 1 ) = µ i + c j G i x j ) µ i cβ l 1 ), j= which results in a contradiction for l sufficiently large so that l is sufficiently large). We conclude that µ i = 0 for all i A x) and sufficiently large. Therefore, max{ µ i /c, G i x )} = 0 for all i A x) and sufficiently large. This concludes the proof that δ 0. 5

Next, note that for any a, b R and c > 0, it holds that 1 max{0, ca + b}) b ) = c c max{ b/c, a}) + b max{ b/c, a}. Indeed, if a b/c, then the right-hand side of the relation above becomes b /c) b /c = b /c), which is the same as the left-hand side. If a > b/c, then the right-hand side is equal to c a + cab)/c) = ca / + ab, which is the same as the left-hand side. Using the above relation in the definition of L c, we obtain Observe that fx ) = L c x, λ, µ ) λ, F x ) c F x ) c max{ µ i /c, G i x )}) µ i max{ µ i /c, G i x )} L c x, λ, µ ) λ, F x ) c F x ) c max{ µ i /c, G i x )}) µ i max{ µ i /c, G i x )} 3.9) L c x, λ, µ ) λ, F x ) µ i max{ µ i /c, G i x )}. 3.10) L c x, λ, µ ) = f x) + 1 c max{0, c G i x) + µ i }) µ i ) ). For i A x), we have max{0, c G i x) + µ i }) = max{0, µ i } = µ i, because µ i 0. For i A x), we have max{0, c G i x) + µ i } = max{0, c G i x)} = 0 = µ i, which holds for all sufficiently large, because for such, as shown above, we have that µ i = 0. It follows that for all sufficiently large, Hence, from 3.10) we obtain that L c x, λ, µ ) = f x). fx ) f x) λ, F x ) µ i max{ µ i /c, G i x )} = f x) + Oδ ), where we have taen into account the boundedness of {λ, µ )}, and where δ is defined by 3.5). Taing into account Proposition 3.1 and applying Theorem.1 to the perturbation 3.) of problem 1.1), we obtain the announced results. We note that in the given context, a sufficient condition for the dual sequence {λ, µ )} generated according to 3.4) to be bounded is the Mangasarian Fromovitz constraint qualification at x. Indeed, by the necessary optimality condition for problem 3.1), we have 6

that 0 = L c x x, λ, µ ) = f x ) + F x )) T λ + c F x )) T F x ) + max{0, c G i x ) + µ i }G ix ) = f x ) + F x )) T λ +1 + i A x) µ +1 i G ix ), 3.11) where we have taen into account that µ i = 0 for all large enough and i A x), as shown above. If {x } x while {λ, µ )} is unbounded, dividing the above relation by λ +1, µ +1 ) and passing onto the limit as possibly along an appropriate subsequence), we obtain the existence of some λ, µ) R l R A x) + ) \ {0} such that 0 = F x)) T λ + µ i G i x), i A x) which contradicts the dual form of) Mangasarian Fromovitz constraint qualification at the point x. In Theorem 3.1, we do not assume that M x), but in fact, this must be the case under the stated assumptions. Indeed, from 3.11) we obtain that for every limit point λ, µ) of {λ, µ )}, it holds that f x) + F x)) T λ + µ i G i x) = 0. i A x) Moreover, as mentioned above, for all large enough we have µ i = 0 for all i A x), and µ 0 according to 3.4). Hence, λ, µ) M x). The following observation can also be useful. By the direct computation, from 3.4) it can be seen that δ = 1 λ +1 λ, µ +1 µ ). c From this representation of δ, it is evident that δ 0 provided {λ, µ )} is bounded while c, or provided {λ, µ )} is convergent while c is bounded away from zero, without any assumptions on {x }. Note, however, that the assumptions of Theorem 3.1 above are different, and so δ 0 had to be established by other considerations. The next result is more in the spirit of penalty methods. We do not consider any specific rule for updating the dual variables, and in particular, they can even be fixed the classical penalty method is formally obtained by setting λ, µ ) = 0 for all in the definition of L c ). In that setting, to ensure convergence it in general must hold that c + as. Theorem 3. Suppose that the sequence {λ, µ )} is bounded and that c + as. For each, let x be a solution of 3.1) with λ, µ) = λ, µ ), c = c. Suppose that the sequence {x } converges to x, which is a solution of 1.1). Then both assertions of Theorem 3.1 hold. Moreover, under 3.7) in particular, if SOSC 1.6) holds), we have that δ = O1/c ). 3.1) 7

Proof. Again, we can assume that x B ε x) for all. Because {µ } is bounded and c +, it holds that µ i /c 0. Since {Gx )} G x) 0, we conclude that max{ µ i /c, G i x )} 0 for all i = 1,..., m. Taing also into account that {F x )} 0, it follows that δ 0 as alternatively, the argument preceding the theorem could be used here). As in the proof of Theorem 3.1, we have the relation 3.10). However, it now holds that L c x, λ, µ ) = f x) + 1 c = f x) 1 c f x), max{0, c G i x) + µ i }) µ i ) ) i A x) µ i ) 3.13) where the second equality follows from the fact that c G i x) + µ i < 0 for all i A x) and all large enough recall that c + while {µ } is bounded). Combining this relation with 3.10), we again obtain that fx ) f x) λ, F x ) µ i max{ µ i /c, G i x )} = f x) + Oδ ), and the assertions of Theorem 3.1 follow. Suppose now that 3.7) holds. Then combining 3.9) and 3.13), we obtain that Hence, f x) + Oδ ) = fx ) f x) 1 µ i ) λ, F x ) c c F x ) i A x) c max{ µ i /c, G i x )}) µ i max{ µ i /c, G i x )} = f x) 1 µ i ) c c F x ) c max{ µ i /c, G i x )}) + Oδ ). i A x) 1 c i A x) µ i ) + c F x ) + c Evidently, for all i A x) and each large enough, it holds that max{ µ i /c, G i x )}) = Oδ ). 3.14) max{ µ i /c, G i x )} = µ i /c. Hence, 1 c i A x) µ i ) = c max{ µ i /c, G i x )}), i A x) 8

and recalling the definition of δ, it now follows from 3.14) that Therefore, 3.1) holds. c δ = Oδ ). Note that while M x) is not assumed in Theorem 3., this is again the case in the setting of 3.7) the second part of Theorem 3.). Indeed, defining the auxiliary sequences λ = λ + c F x ), µ i = max{0, µ i + c G i x )}, i = 1,..., m, 3.15) it can be seen by direct computation that δ = 1 c λ λ, µ µ ). 3.16) The last assertion of Theorem 3. implies that the sequence {c δ } is bounded. Hence, 3.16) combined with boundedness of {λ, µ )} implies boundedness of { λ, µ )}. The fact that every limit point of the latter sequence belongs to M x) can be established the same way as its counterpart for the method of multipliers; see the discussion following the proof of Theorem 3.1. The following observation is also worth mentioning. Suppose that under the assumptions of Theorem 3., M x) is a singleton, i.e., M x) = { λ, µ)}. Then { λ, µ } must converge to this λ, µ), whatever one taes as {λ, µ )}. Moreover, if good multiplier approximations are used, i.e., {λ, µ )} λ, µ), then estimate 3.1) can be sharpened. Specifically, from 3.16) it follows that δ = o1/c ). 3.17) We complete this section with two examples demonstrating that the estimates obtained above are sharp. The first example below satisfies FOSC, while the second satisfies SOSC. It is interesting to note that the second example does not satisfy constraint qualifications the multiplier set is unbounded). Example 3.1 Let n = l = 1, m = 0, fx) = x + x /, F x) = x. The only feasible point and hence, the only solution) of 1.1) is x = 0. Furthermore, M x) = { 1}, and C x) = er F x) = {0}, that is, FOSC 1.5) holds. It can be easily seen that for each λ R and c > 0, the only solution of 3.1) is given by x λ, c = λ + 1)/c + 1). We first consider the case when {λ } R is generated according to the first equality in 3.4), while c > 0 is fixed. Let x = x λ, c for each. Then λ +1 + 1 = λ + cx + 1 = λ c λ + 1 c + 1 + 1 = λ + 1 c + 1, 9

and since c > 0, it is now evident that {λ } λ = 1, which is the only multiplier associated with x. Hence, {x } x, δ = x 0, fx ) = x + o x ), and the estimates 3.6), 3.7) are exact for the latter, note that for each, x = δ if λ 0 < 1, and x = δ if λ 0 > 1). We next consider the case when λ R is fixed, while c +. Let x = x λ, c for each. All the conclusions for the previous case remain valid note that for each, x = δ if λ < 1, and x = δ if λ > 1). Moreover, for each, λ = λ + 1 c + 1, where λ is defined according to the first equality in 3.15). Clearly, { λ } λ. Finally, estimate 3.1) obviously holds and is in general sharp it can be improved if we let λ tend to λ, in which case 3.17) is valid). Example 3. Let n = l =, m = 0, fx) = x 1 + x 1 / + x4 /, F x) = x 1, x ). The only feasible point and hence, the only solution) of 1.1) is x = 0. Furthermore, M x) = { 1} R, and it can be easily verified that SOSC 1.6) holds. It can be easily seen that for each λ R and c > 0, the only solution of 3.1) is given by ) λ 1+1 c+1, 0 x λ, c = ) ) 1/ λ 1+1 c+1, ± λ c+1 if λ 0, if λ < 0. Consider the case when {λ } R is generated according to the first equality in 3.4), while c > 0 is fixed. Let x = x λ, c for each. Let λ = 1, 0) M x). Then, assuming that λ < 0, we obtain λ +1 λ = λ 1 + cx 1, λ + cx ) ) λ = λ 1 c λ 1 + 1 ) c + 1, λ c λ c + 1 λ λ = 1 + 1 c + 1, λ ) c + 1 = λ λ c + 1. In particular, if λ 0 < 0, then λ remains negative for each, and since c > 0, it is evident that {λ } λ. Hence, {x } x and δ = λ 1 + 1) + λ ) ) 1/ c + 1 0. Fix an arbitrary θ > 0 and tae λ 0 = 1 ± θ, θ). Then λ = 1 + ±θ ) c + 1), θ c + 1), 10

) ) x θ = c + 1) +1, θ 1/ c + 1) +1, ) fx θ ) = c + 1) +1 + o 1 c + 1) +1, θ δ = c + 1) +1, and the estimates 3.7), 3.8) are exact. Consider now the case when λ R is fixed, while c +. Fix an arbitrary θ > 0 and tae λ = 1 ± θ, θ). Let x = x λ, c for each. All the conclusions for the previous case remain valid. Moreover, for each, λ = 1 + ±θ c + 1, θ c + 1 where λ is defined according to the first equality in 3.15). Clearly, { λ } λ. Finally, estimate 3.1) obviously holds and is in general sharp it can be improved if we let λ tend to λ, that is, let θ to tend to 0, in which case 3.17) is valid). Finally, consider the case when λ = 1, 1). For each, set x = x λ, c, and note that x 1 = 0 while x 0. At x, λ), the residual of the optimality system 1.) is of order x ). Hence, the algorithm-independent error bound of [8] does not hold along the sequence {x, λ)}. At the same time, δ = x ), and the estimate 3.8) holds and is exact. The same conclusions apply if λ is not fixed at λ = 1, 1), but the generated dual sequence tends to this λ. 4 Lagrangian Relaxation The approach of Lagrangian relaxation [3, Chapter 7] is a useful tool for solving various classes of optimization problems [13]. It consists of solving 1.1) via solving its dual ), maximize ϕλ, µ) subject to λ, µ), 4.1) where = {λ, µ) R l R m + ϕλ, µ) > }, ϕ : R l R m + R, ϕλ, µ) = inf Lx, λ, µ). x Rn The dual problem 4.1) is a concave maximization problem, in general nonsmooth, which is solved by appropriate subgradient or bundle methods [3, Chapter 7]. Bundle methods are in general more reliable and practical, although the subgradient methods are also sometimes useful, thans to their simplicity of implementation. Below we shall consider explicitly the subgradient method only, because introducing the more sophisticated bundle methods would have required an extensive discussion which is secondary to the subject of this paper. However, we believe that bundle methods can also be treated within our framewor. 11

We shall assume that the problem minimize Lx, λ, µ) subject to x R n 4.) has a solution x λ, µ for every λ, µ) R l R m + of interest artificial bounds on x are often introduced to guarantee this, so that the dual function ϕ is finite everywhere). The following relation between 4.) and perturbations of the original problem 1.1) is well nown we include its short proof for completeness). Proposition 4.1 [3, Theorem 10.1] For any λ, µ) R n R m +, a point x λ, µ which solves 4.), is a solution of minimize fx) subject to F x) = F { x λ, µ ), Gi x G i x) λ, µ ), if µ i > 0, max{0, G i x λ, µ )}, if µ i = 0, i = 1,..., m. 4.3) Proof. For any x R n which is feasible in 4.3), we obtain that fx λ, µ ) fx) fx λ, µ ) fx) + µ i >0 µ i G i x λ, µ ) G i x)) = fx λ, µ ) fx) + λ, F x λ, µ ) F x) + µ, Gx λ, µ ) Gx) = Lx λ, µ, λ, µ) Lx, λ, µ). Since x λ, µ is feasible in 4.3), the assertion follows. As is well nown [3, Chapter 7] and easy to see, solving 4.) provides not only the value of ϕ at the point λ, µ), but also at no additional cost) of one of its subgradients. In particular, it holds that F x λ, µ ), Gx λ, µ )) ϕλ, µ). 4.4) The subgradient Lagragian relaxation based method is the following iterative procedure. Given some λ, µ ) R l R m +, the first part of the iteration consists of solving the minimization problem 4.) with λ, µ) = λ, µ ). This generates a primal point, which we shall denote x. After this, the dual variables are updated by the projected subgradient step λ +1, µ +1 ) = P R l R m λ, µ ) + α + g ), g ϕλ, µ ), α > 0, where P R l R m denotes the orthogonal projection onto the set Rl R m + +, and α is the stepsize. In particular, the implementation based on 4.4) gives λ +1 = λ + α F x ), µ +1 i = max{0, µ i + α G i x )}, i = 1,..., m. 4.5) 1

Theoretically, for convergence of the subgradient projection method, the stepsize sequence {α } must satisfy α g = + 4.6) and =0 α g < +. 4.7) =0 More precisely, if the dual problem 4.1) is solvable, and conditions 4.6) and 4.7) hold, then the method generates the trajectory which converges to a solution of 4.1). If, in addition, the solution set of problem 4.1) is bounded, then trajectory converges to this set even if 4.7) is replaced by a weaer condition lim α g = 0. In our setting, we assume that the sequence {λ, µ )} is bounded and the corresponding sequence {x } converges to x, the strict local solution of 1.1) under consideration. The latter is not automatic, but can be expected to happen in some situations for example, if convexity and constraint qualifications are assumed). Under these assumptions, the sequence {g } is automatically bounded, and 4.6) implies α = +, 4.8) =0 which is the condition to be used below. In this setting, we are again interested in obtaining an estimate of the distance from x to x. Recalling 4.3), define δ = F x ) + G i x )) + max{0, G i x )}) µ i >0 µ i =0 1/. 4.9) Theorem 4.1 Suppose that the solution set of the dual problem 4.1) is nonempty and that the sequence {λ, µ )} is generated according to 4.5) and 4.8), where for each, x is a solution of 4.) for λ, µ) = λ, µ ). Suppose that the sequence {x } converges to x, which is a solution of 1.1). Then all the assertions of Theorem 3.1 hold, with δ defined by 4.9). Proof. As before, we can assume that x B ε x) for all. To prove that δ 0 as, we first show that µ i = 0 for all i A x) and all sufficiently large. Let i A x) and suppose the opposite, i.e., that there exists an infinite subsequence of indices { l } such 13

that µ l i > 0 for all l. For all large enough, say, it holds that G i x ) β, where β > 0. Then for l >, we obtain that 0 < µ l i = µ l 1 i + α l 1G i x l 1 l 1 ) = µ i + α j G i x j l 1 ) µ i β α j, j= j= which results in a contradiction when l and then l ), because of the first condition in 4.8). It follows that if µ i > 0 for some large enough, then i A x), so that G ix ) 0. It now easily follows that δ 0. We have that fx ) + λ, F x ) + µ, Gx ) = Lx, λ, µ ) L x, λ, µ ) = f x) + λ, F x) + µ, G x) f x), where the first inequality is by the definition of x, and the last follows from F x) = 0, G x) 0 and µ 0. Hence, fx ) f x) λ, F x ) µ, Gx ). In particular, taing into account the boundedness of {λ, µ )} this sequence converges, by our assumptions and the properties of the subgradient projection method) and the definition 4.9) of δ, we conclude that fx ) f x) + Oδ ). The assertions follow from Proposition 4.1 and Theorem.1. It can be verified that Examples 3.1 and 3. also show that the corresponding estimates obtained for the subgradient Lagrangian relaxation based method are sharp. We shall not provide any specifics, as conceptually the situation is very similar to the case of the augmented Lagrangian, considered in full detail in Section 3. References [1] M. Anitescu. Nonlinear programs with unbounded Lagrange multiplier sets, 00. Preprint ANL/MCS-P796-000, Mathematics and Computer Science Devision, Argonne National Laboratory. [] D.P. Bertseas. Constrained Optimization and Lagrange Multiplier Methods. Academic Press, New Yor, 198. [3] J.F. Bonnans, J.Ch. Gilbert, C. Lemaréchal, and C. Sagastizábal. Numerical Optimization: Theoretical and Practical Aspects. Springer Verlag, Berlin, Germany, 003. 14

[4] A.R. Conn, N.I.M. Gould, and Ph.L. Toint. A globally convergent augmented Lagrangian algorithm for optimization with general constraints and simple bounds. SIAM Journal of Numerical Analysis, 8:545 57, 1991. [5] A. Fischer. Local behavior of an iterative framewor for generalized equations with nonisolated solutions. Mathematical Programming, 94:91 14, 00. [6] R. Fletcher, S. Leyffer, D. Ralph, and S. Scholtes. Local convergence of SQP methods for mathematical programs with equilibrium constraints, 00. Numerical Analysis Report NA/09, Department of Mathematics, University of Dundee. [7] W.W. Hager. Stabilized sequential quadratic programming. Computational Optimization and Applications, 1:53 73, 1999. [8] W.W. Hager and M.S. Gowda. Stability in the presence of degeneracy and error estimation. Mathematical Programming, 85:181 19, 1999. [9] M.R. Hestenes. Multiplier and gradient methods. Journal of Optimization Theory and Applications, 4:303 330, 1969. [10] A.F. Izmailov. Mathematical programs with complementarity constraints: regularity, optimality comditions, and sensitivity. Computational Mathematics and Mathematical Physics, 44:1145 1164, 004. [11] A.F. Izmailov and M.V. Solodov. Newton-type methods for optimization problems without constraint qualifications. SIAM Journal on Optimization, 15:10 8, 004. [1] A.F. Izmailov and M.V. Solodov. Karush-Kuhn-Tucer systems: regularity conditions, error bounds and a class of Newton-type methods. Mathematical Programming, 95:631 650, 003. [13] C. Lemaréchal. Lagrangian relaxation. In Computational Combinatorial Optimization, Lecture Notes in Computer Science, 41, pages 11 156. Springer, Berlin, 001. [14] J.-S. Pang. Error bounds in mathematical programming. Mathematical Programming, 79:99 33, 1997. [15] M.J.D. Powell. A method for nonlinear constraints in minimization problems. In R. Fletcher, editor, Optimization, pages 83 98. Academic Press, London, 1969. [16] R.T. Rocafellar. Augmented Lagrangians and applications of the proximal point algorithm in convex programming. Mathematics of Operations Research, 1:97 116, 1976. [17] S.J. Wright. An algorithm for degenerate nonlinear programming with rapid local convergence. Optimization Technical Report 03-0, Computer Sciences Department, University of Wisconsin-Madison, October 003. 15