Critical solutions of nonlinear equations: local attraction for Newton-type methods

Similar documents
Critical solutions of nonlinear equations: Stability issues

Rejoinder on: Critical Lagrange multipliers: what we currently know about them, how they spoil our lives, and what we can do about it

A globally convergent Levenberg Marquardt method for equality-constrained optimization

A note on upper Lipschitz stability, error bounds, and critical multipliers for Lipschitz-continuous KKT systems

Newton-Type Methods: A Broader View

Some new facts about sequential quadratic programming methods employing second derivatives

Pacific Journal of Optimization (Vol. 2, No. 3, September 2006) ABSTRACT

On attraction of linearly constrained Lagrangian methods and of stabilized and quasi-newton SQP methods to critical multipliers

Local convergence of the method of multipliers for variational and optimization problems under the noncriticality assumption

ON REGULARITY CONDITIONS FOR COMPLEMENTARITY PROBLEMS

The effect of calmness on the solution set of systems of nonlinear equations

1. Introduction. We consider the mathematical programming problem

Globalizing Stabilized Sequential Quadratic Programming Method by Smooth Primal-Dual Exact Penalty Function

Critical Lagrange multipliers: what we currently know about them, how they spoil our lives, and what we can do about it

Examples of dual behaviour of Newton-type methods on optimization problems with degenerate constraints

c 2002 Society for Industrial and Applied Mathematics

On attraction of linearly constrained Lagrangian methods and of stabilized and quasi-newton SQP methods to critical multipliers

Technische Universität Dresden Institut für Numerische Mathematik

Merit functions and error bounds for generalized variational inequalities

Combining stabilized SQP with the augmented Lagrangian algorithm

Journal of Convex Analysis (accepted for publication) A HYBRID PROJECTION PROXIMAL POINT ALGORITHM. M. V. Solodov and B. F.

Inexact Josephy Newton framework for generalized equations and its applications to local analysis of Newtonian methods for constrained optimization

Journal of Convex Analysis Vol. 14, No. 2, March 2007 AN EXPLICIT DESCENT METHOD FOR BILEVEL CONVEX OPTIMIZATION. Mikhail Solodov. September 12, 2005

ON A CLASS OF NONSMOOTH COMPOSITE FUNCTIONS

c 2012 Society for Industrial and Applied Mathematics

Some Properties of the Augmented Lagrangian in Cone Constrained Optimization

A convergence result for an Outer Approximation Scheme

1. Introduction. In this paper, we consider the constrained equation

Spectral gradient projection method for solving nonlinear monotone equations

Damián Fernández 1 and Mikhail Solodov 2*

Technische Universität Dresden Herausgeber: Der Rektor

Introduction to Optimization Techniques. Nonlinear Optimization in Function Spaces

1. Introduction. Consider the following parameterized optimization problem:

On well definedness of the Central Path

FIRST- AND SECOND-ORDER OPTIMALITY CONDITIONS FOR MATHEMATICAL PROGRAMS WITH VANISHING CONSTRAINTS 1. Tim Hoheisel and Christian Kanzow

Finite Convergence for Feasible Solution Sequence of Variational Inequality Problems

A derivative-free nonmonotone line search and its application to the spectral residual method

CONVERGENCE PROPERTIES OF COMBINED RELAXATION METHODS

A SIMPLY CONSTRAINED OPTIMIZATION REFORMULATION OF KKT SYSTEMS ARISING FROM VARIATIONAL INEQUALITIES

On the acceleration of augmented Lagrangian method for linearly constrained optimization

Constrained Optimization

Maria Cameron. f(x) = 1 n

5 Handling Constraints

1 Introduction We consider the problem nd x 2 H such that 0 2 T (x); (1.1) where H is a real Hilbert space, and T () is a maximal monotone operator (o

Primal-dual relationship between Levenberg-Marquardt and central trajectories for linearly constrained convex optimization

1 Computing with constraints

FIXED POINTS IN THE FAMILY OF CONVEX REPRESENTATIONS OF A MAXIMAL MONOTONE OPERATOR

An Accelerated Hybrid Proximal Extragradient Method for Convex Optimization and its Implications to Second-Order Methods

A STABILIZED SQP METHOD: SUPERLINEAR CONVERGENCE

The Relation Between Pseudonormality and Quasiregularity in Constrained Optimization 1

Kantorovich s Majorants Principle for Newton s Method

Contraction Methods for Convex Optimization and Monotone Variational Inequalities No.16

Stationary Points of Bound Constrained Minimization Reformulations of Complementarity Problems1,2

Convergence Analysis of Perturbed Feasible Descent Methods 1

A doubly stabilized bundle method for nonsmooth convex optimization

MS&E 318 (CME 338) Large-Scale Numerical Optimization

Affine covariant Semi-smooth Newton in function space

On the complexity of the hybrid proximal extragradient method for the iterates and the ergodic mean

An inexact subgradient algorithm for Equilibrium Problems

20 J.-S. CHEN, C.-H. KO AND X.-R. WU. : R 2 R is given by. Recently, the generalized Fischer-Burmeister function ϕ p : R2 R, which includes

arxiv: v1 [math.fa] 16 Jun 2011

Algorithms for nonlinear programming problems II

On the Local Quadratic Convergence of the Primal-Dual Augmented Lagrangian Method

Hyperbolicity singularities in rarefaction waves

A GLOBALLY CONVERGENT STABILIZED SQP METHOD: SUPERLINEAR CONVERGENCE

Multipoint secant and interpolation methods with nonmonotone line search for solving systems of nonlinear equations

Implications of the Constant Rank Constraint Qualification

Relationships between upper exhausters and the basic subdifferential in variational analysis

On Total Convexity, Bregman Projections and Stability in Banach Spaces

Absolute Value Programming

A CHARACTERIZATION OF STRICT LOCAL MINIMIZERS OF ORDER ONE FOR STATIC MINMAX PROBLEMS IN THE PARAMETRIC CONSTRAINT CASE

A FRITZ JOHN APPROACH TO FIRST ORDER OPTIMALITY CONDITIONS FOR MATHEMATICAL PROGRAMS WITH EQUILIBRIUM CONSTRAINTS

Bulletin of the. Iranian Mathematical Society

Constrained Optimization Theory

Newton-type Methods for Solving the Nonsmooth Equations with Finitely Many Maximum Functions

On convergence rate of the Douglas-Rachford operator splitting method

On the convergence properties of the projected gradient method for convex optimization

Some Inexact Hybrid Proximal Augmented Lagrangian Algorithms

A Regularized Directional Derivative-Based Newton Method for Inverse Singular Value Problems

Structural and Multidisciplinary Optimization. P. Duysinx and P. Tossings

Stationarity and Regularity of Infinite Collections of Sets. Applications to

1. Introduction The nonlinear complementarity problem (NCP) is to nd a point x 2 IR n such that hx; F (x)i = ; x 2 IR n + ; F (x) 2 IRn + ; where F is

AN INEXACT HYBRID GENERALIZED PROXIMAL POINT ALGORITHM AND SOME NEW RESULTS ON THE THEORY OF BREGMAN FUNCTIONS. M. V. Solodov and B. F.

A Proximal Method for Identifying Active Manifolds

AN INEXACT HYBRID GENERALIZED PROXIMAL POINT ALGORITHM AND SOME NEW RESULTS ON THE THEORY OF BREGMAN FUNCTIONS. May 14, 1998 (Revised March 12, 1999)

A Continuation Method for the Solution of Monotone Variational Inequality Problems

A double projection method for solving variational inequalities without monotonicity

Differentiable exact penalty functions for nonlinear optimization with easy constraints. Takuma NISHIMURA

An Accelerated Newton Method for Equations with Semismooth Jacobians and Nonlinear Complementarity Problems

1. Introduction. We consider the classical variational inequality problem [1, 3, 7] VI(F, C), which is to find a point x such that

First-order optimality conditions for mathematical programs with second-order cone complementarity constraints

Radius Theorems for Monotone Mappings

CHARACTERIZATIONS OF LIPSCHITZIAN STABILITY

WHEN ARE THE (UN)CONSTRAINED STATIONARY POINTS OF THE IMPLICIT LAGRANGIAN GLOBAL SOLUTIONS?

MODIFYING SQP FOR DEGENERATE PROBLEMS

Algorithms for Constrained Optimization

TMA 4180 Optimeringsteori KARUSH-KUHN-TUCKER THEOREM

University of Houston, Department of Mathematics Numerical Analysis, Fall 2005

Obstacle problems and isotonicity

A Smoothing Newton Method for Solving Absolute Value Equations

Transcription:

Mathematical Programming manuscript No. (will be inserted by the editor) Critical solutions of nonlinear equations: local attraction for Newton-type methods A. F. Izmailov A. S. Kurennoy M. V. Solodov Received: March 06 / Revised: November 06 Abstract We show that if the equation mapping is -regular at a solution in some nonzero direction in the null space of its Jacobian (in which case this solution is critical; in particular, the local Lipschitzian error bound does not hold), then this direction defines a star-like domain with nonempty interior from which the iterates generated by a certain class of Newtontype methods necessarily converge to the solution in question. This is despite the solution being degenerate, and possibly non-isolated (so that there are other solutions nearby). In this sense, Newtonian iterates are attracted to the specific (critical) solution. Those results are related to the ones due to A. Griewank for the basic Newton method but are also applicable, for example, to some methods developed specially for tackling the case of potentially non-isolated solutions, including the Levenberg Marquardt and the LP-Newton methods for equations, and the stabilized sequential quadratic programming for optimization. Keywords Newton method critical solutions -regularity Levenberg Marquardt method linear-programming-newton method stabilized sequential quadratic programming. Mathematics Subject Classification (000) 90C33 65K0 49J53. This research is supported in part by the Russian Foundation for Basic Research Grant 7-0-005, by the Russian Science Foundation Grant 5--00, by the Ministry of Education and Science of the Russian Federation (the Agreement number 0.a03..0008), by VolkswagenStiftung Grant 5540, by CNPq Grants PVE 409/04-9 and 30374/05-3, and by FAPERJ. A. F. Izmailov VMK Faculty, OR Department, Lomonosov Moscow State University (MSU), Uchebniy Korpus, Leninskiye Gory, 999 Moscow, Russia; and Peoples Friendship University of Russia, Miklukho-Maklaya Str. 6, 798 Moscow, Russia. E-mail: izmaf@ccas.ru A. S. Kurennoy Department of Mathematics, Physics and Computer Sciences, Derzhavin Tambov State University, TSU, Internationalnaya 33, 39000 Tambov, Russia. E-mail: akurennoy@cs.msu.ru M. V. Solodov IMPA Instituto de Matemática Pura e Aplicada, Estrada Dona Castorina 0, Jardim Botânico, Rio de Janeiro, RJ 460-30, Brazil. E-mail: solodov@impa.br

Izmailov, Kurennoy, and Solodov Introduction We consider a nonlinear equation Φ(u) = 0, () where the mapping Φ : R p R p is smooth enough (precise smoothness assumptions will be stated later as needed). This paper is concerned with convergence properties of Newtontype methods for solving the equation () when it has a singular solution ū (i.e., the matrix Φ (ū) is singular). Of particular interest is the difficult case when ū may be a non-isolated solution of (). Note that if ū is a non-isolated solution, it is necessarily singular. To describe the class of methods in question, we define the perturbed Newton method (pnm) framework for equation () as follows. For the given iterate u k R p, the next iterate is u k+ = u k + v k, with v k satisfying the following linear equation in v: Φ(u k ) + (Φ (u k ) + Ω(u k ))v = ω(u k ). () In (), the mappings Ω : R p R p p and ω : R p R p are certain perturbation terms which may have different roles, and individually or collectively define specific methods within the general pnm framework. In particular, if Ω 0 and ω 0, then () reduces to the iteration system of the basic Newton method. The most common setting would be for Ω to characterize a perturbation of the iteration matrix of the basic Newton method (i.e., the difference between the matrix a given method actually employs, compared to the exact Jacobian Φ (u k )), and for ω to account for possible inexactness in solving the corresponding linear system of equations. An example of this setting within this paper, is the stabilized Newton- Lagrange method (stabilized sequential quadratic programming for equality-constrained optimization), considered in Section 3.3 below. However, we emphasize that our framework is not restricted to this situation. In particular, subproblems of a given method need not even be systems of linear equations, as long as they can be related to () a posteriori. One example is the linear-programming-newton (LP-Newton) method discussed in Section 3. below, which solves linear programs, and for which the perturbation term ω is implicit (i.e., it does not have an explicit analytical formula, but its properties are known). In this respect, we also comment that the way we shall employ the perturbation mappings Ω and ω is somewhat unusual, in the following sense. They may describe a given method not necessarily on the whole neighborhood of a solution of interest, but possibly only in some relevant star-like domain of convergence; see the discussion of the Levenberg Marquardt method in Section 3. and of the LP-Newton method in Section 3.. This, however, is exactly what is needed in the presented convergence analysis, as it is shown that the generated iterates do in fact stay within the domain in question and, within this set, Ω and ω that we construct do adequately represent the given algorithms. Finally, we note that the specific methods that we consider in this paper have been designed to tackle the difficult cases when () has degenerate/nonisolated solutions, and in this sense the perturbation terms in () that describe these methods can be regarded as structural, i.e., introduced intentionally for improving convergence properties in the degenerate cases. The assumptions imposed on Ω and ω are only related to their size, which allows ω to cover naturally precision control when the subproblems are solved approximately. However, as already commented, the use of ω can also be quite different. In the analysis of the LP-Newton method in Section 3., ω is implicit and is not related to solving subporoblems approximately. Our convergence results assume a certain -regularity property of the solution of (), which implies that this solution is critical in the sense of [5]. We next state the relevant definitions, and discuss the relations between those concepts.

Critical solutions of nonlinear equations: Newton methods 3 Recall first that if Φ is differentiable at a solution ū of the equation (), then it holds that T Φ (0) (ū) kerφ (ū), where Φ (0) is the solution set of (), and by T S (u) we denote the contingent cone to a set S R p at u S, i.e. the tangent cone as defined in [6, Definition 6.]: T S (u) = {v R p {t k } R + : {t k } 0+, dist(u +t k v, S) = o(t k )}. Recall that according to [6, Corollary 6.9] (see also [6, Definition 6.4] and the original definition in [3, Definition.4.6]), assuming that a set S R p is closed near u S, this set is Clarke-regular at u if the multifunction T S ( ) is inner semicontinuous at u relative to S (the latter meaning that for any v T S (u) and any sequence {u k } S convergent to u, there exists a sequence {v k } convergent to v such that v k T S (u k ) for all k). The following notion was introduced in [5]. Definition Assuming that Φ is differentiable at a solution ū of the equation (), this solution is referred to as noncritical if the set Φ (0) is Clarke-regular at ū, and Otherwise, the solution ū is referred to as critical. T Φ (0) (ū) = kerφ (ū). (3) Observe that according to the above definition, any isolated singular solution is necessarily critical. As demonstrated in [5], if Φ satisfies some mild and natural smoothness assumptions, then noncriticality of ū is equivalent to the local Lipschitzian error bound on the distance to the solution set in terms of the natural residual of the equation (): dist(u, Φ (0)) = O( Φ(u) ) holds as u R p tends to ū. Moreover, it is also equivalent to the upper-lipschitzian stability with respect to right-hand side perturbations of (): any solution u(w) of the perturbed equation Φ(u) = w, close enough to ū, satisfies dist(u(w), Φ (0)) = O( w ) as w R p tends to 0. Accordingly, criticality of ū means the absence of the properties above. The interest in critical/noncritical solutions of nonlinear equations originated from the study of special Lagrange multipliers in equality-constrained optimization, also called critical [3,8,9,4,], [, Chapter 7]. For the relations between critical solutions of equations and critical multipliers in optimization, see [5]. It had been demonstrated that critical Lagrange multipliers tend to attract dual sequences generated by a number of Newton-type methods for optimization [8,9], [, Chapter 7]. In this paper, we show that critical solutions of nonlinear equations also serve as attractors, in this case for methods described by the pnm framework (). For a symmetric bilinear mapping B : R p R p R p and an element v R p, we define the linear operator B[v] : R p R p by B[v]u = B[v, u]. The notion of -regularity is a useful tool in nonlinear analysis and optimization theory; see, e.g., the book [], as well as [,6, 7,9,0] for some applications. The essence of the construction is that when a mapping Φ is irregular at ū (Φ (ū) is singular), first-order information is insufficient to adequately represent Φ around ū, and so second-order information has to come into play. To this end, we have the following.

4 Izmailov, Kurennoy, and Solodov Definition Assuming that Φ is twice differentiable at ū R p, Φ is said to be -regular at the point ū in the direction v R p if the p p-matrix Ψ(ū; v) = Φ (ū) + ΠΦ (ū)[v] (4) is nonsingular, where Π is the projector in R p onto an arbitrary fixed complementary subspace of imφ (ū), along this subspace. In our convergence analysis of Newton-type methods described by (), we shall assume that Φ is -regular at a solution ū of () in some direction v kerφ (ū) \ {0}. It turns out that this implies that the solution ū is necessarily critical in the sense of Definition. We show this next. Proposition Let Φ be twice differentiable at a solution ū of the equation () and -regular at this solution in some direction v kerφ (ū) \ {0}. Then (3) does not hold and, in particular, ū is a critical solution of (). Proof Let v kerφ (ū)\{0}. Suppose that (3) holds. Then v T Φ (0)(ū). Thus, there exist a sequence {t k } of positive reals and a sequence {r k } R p such that {t k } 0, r k = o(t k ), and for all k it holds that Therefore, which implies that 0 = Φ(ū +t k v + r k ) = Φ (ū)r k + t k Φ (ū)[v, v] + o(t k ). t k Φ (ū)[v, v] + o(t k ) imφ (ū), Φ (ū)[v, v] imφ (ū). Hence, ΠΦ (ū)[v, v] = 0, where Π is specified in Definition. Since also Φ (ū)v = 0, from (4) we conclude that v kerψ(ū; v). As v 0, this contradicts -regularity (specifically, the nonsigularity of Ψ(ū; v)). Thus, (3) cannot hold, and ū must be a critical solution. In Section, we shall prove that if Φ is -regular at a (critical) solution ū of () in some direction v kerφ (ū) \ {0}, then v defines a domain star-like with respect to ū, with nonempty interior, from which the iterates that satisfy the pnm framework () necessarily converge to ū. In this sense, the iterates are attracted specifically to ū, even though there may be other nearby solutions. These results are related to [], where the pure Newton method was considered (i.e., () with Ω 0 and ω 0). An interesting extension of the results in [] to the case when Φ is not necessarily twice differentiable, but has a Lipschitzcontinuous first derivative, has been proposed in [5]. The latter reference also gives an application to smooth equation reformulations of complementarity problems. In Section 3, we demonstrate how the general results for the pnm framework () apply to some specific Newton-type methods. These include the classical Levenberg Marquardt method [4, Chapter 0.] and the LP-Newton method [6] for nonlinear equations, and the stabilized Newton Lagrange method for optimization (or stabilized sequential quadratic programming) [8,,7,0]; see also [, Chapter 7]. We finish this section with some words about our notation. Throughout,, is the Euclidian inner product, and unless specified otherwise, is the Euclidian norm, where the space is always clear from the context. Then the unit sphere is S = {u u = }. For a linear operator (a matrix) A, we denote by kera its null space, and by ima its image (range) space. The notation I stands for the identity matrix. A set U is called star-like with respect to u U if tû + ( t)u U for all û U and all t [0, ].

Critical solutions of nonlinear equations: Newton methods 5 Local convergence of pnm iterates to a critical solution Results of this section are related to those in [], where the basic Newton method was considered (i.e, pnm () with Ω 0 and ω 0). To the best of our knowledge, this is the only directly related reference, as it also allows for non-isolated solutions. Other literature on the basic Newton method for singular equations uses assumptions that imply that a solution, though possibly singular, must be isolated. Let ū be a solution of the equation (). Then every u R p is uniquely decomposed into the sum u = u +u, with u (kerφ (ū)) and u kerφ (ū). The corresponding notation will be used throughout the rest of the paper. We start with the following counterpart of [, Lemma 4.], which establishes that, under appropriate assumptions, the pnm subproblem () has the unique solution for all u k close enough to ū, in a certain star-like domain. Lemma Let Φ : R p R p be twice differentiable near ū R p, with its second derivative Lipschitz-continuous with respect to ū, that is, Φ (u) Φ (ū) = O( u ū ) as u ū. Let ū be a solution of the equation (), and assume that Φ is -regular at ū in a direction v R p S. Let Π stand for the orthogonal projector onto (imφ (ū)). Let Ω : R p R p p satisfy the following properties: Ω(u) = O( u ū ) (5) as u ū, and for every > 0 there exist ε > 0 and δ > 0 such that for every u R p \ {ū} satisfying u ū ε, u ū u ū v δ, (6) it holds that Let ω : R p R p satisfy ΠΩ(u) u ū. (7) ω(u) = O( u ū ) (8) as u ū. Then there exist ε = ε( v) > 0 and δ = δ( v) > 0 such that for every u R p \ {ū} satisfying u ū ε, u ū u ū v δ, (9) the equation () with u k = u has the unique solution v, satisfying u + v ū = O( u ū ), (0) as u ū. u + v ū = (u ū ) + O( u ū ) +O( u ū Πω(u)) + O( ΠΩ(u) ) + O( u ū ) ()

6 Izmailov, Kurennoy, and Solodov Proof Under the stated smoothness assumptions, without loss of generality we can assume that ū = 0 and Φ(u) = Au + B[u, u] + R(u), () where A = Φ (0) R p p, B = Φ (0) is a symmetric bilinear mapping from R p R p to R p, and the mapping R : R p R p is differentiable near 0, with R(u) = O( u 3 ), R (u) = O( u ) (3) as u 0. Substituting the form of Φ stated in () into (), and multiplying () be (I Π) and by Π, we decompose () into the following two equations: ( ) (A + (I Π)(B[u] + R (u) + Ω(u)))v = Au (I Π) B[u, u] + R(u) ω(u) (I Π)(B[u] + R (u) + Ω(u))v, (4) and ( ) Π(B[u] + R (u) + Ω(u))(v + v ) = Π B[u, u] + R(u) ω(u). (5) Let ε > 0 and δ > 0 be arbitrary and fixed for now. From this point on, we consider only those u R p \ {0} that satisfy (9). Define the family of linear operators A (u) : (kera) ima as the restriction of (A + (I Π)(B[u] + R (u) + Ω(u))) to (kera). Let A : (kera) ima be the restriction of A to (kera). Then, taking into account (8) and (3), the equality (4) can be written as A (u)v = A u (I Π)(B[u] + R (u) + Ω(u))v + O( u ) (6) as u 0. Evidently, A is invertible, and according to (5) and (3), A (u) = A + O( u ). The latter implies that A (u) is invertible, provided ε > 0 is small enough, and (A (u)) = ( A ) + O( u ) as u 0 (see, e.g., [, Lemma A.6]). Therefore, (6) can be written as where M (u) : kera (kera), v = u + M (u)v + O( u ), (7) M (u) = (A (u)) (I Π)(B[u] + R (u) + Ω(u)) = O( u ) (8) as u 0. Substituting (7) into (5), and taking into account (3), we obtain that ( ) Π(B[u] + R (u) + Ω(u))(I + M (u))v = Π B[u, u] ω(u) + Π(B[u] + Ω(u))u +O( u 3 ). (9)

Critical solutions of nonlinear equations: Newton methods 7 Define the family of linear operators B(u) : kera (ima) as the restriction of Π(B[u] + R (u) + Ω(u))(I + M (u)) to kera. Let B(u) : kera (ima) be the restriction of ΠB[u] to kera. Then (9) can be written as B(u)v = (( ) ) B(u)u + Π B[u] + Ω(u) u + ω(u) + O( u 3 ) (0) as u 0. The -regularity of Φ at 0 in the direction v means precisely that B( v) is nonsingular. Then, possibly after reducing δ > 0, by [, Lemma A.6] we obtain the existence of C > 0 such that for every u R p \{0} satisfying the second relation in (9), B(u/ u ) is invertible, and ( B ( u u )) C. () According to (3), (5), (8), it holds that B ( u u ) = B ( u u ) + u ΠΩ(u) + O( u ). Choosing > 0 small enough, and further reducing ε > 0 and δ > 0 if necessary, by (7) and (), and by [, Lemma A.6], we now obtain that B ( u u ) is invertible, and ( B ( u u )) = ( B ( u u )) + O ( u ΠΩ(u) ) + O( u ). Employing again (5), we further conclude that (B(u)) = ( B(u)) + O ( u ΠΩ(u) ) + O() = O ( u ). Therefore, (0) is uniquely solvable, and its unique solution has the form v = ( ) u +( B(u)) Π B[u, u ] + ω(u) +O( ΠΩ(u) )+O( u ) = O( u ) () as u 0, where the last estimate employs (5) and (8). Substituting () into (7), and employing (8) again, we finally obtain that v = u + O( u ) (3) as u 0. From () and (3), we have the needed estimates (0) and (). Remark From the proof of Lemma it can be seen that under the assumptions of this lemma (removing the assumptions on ω which are not needed for the following), the values ε = ε( v) > 0 and δ = δ( v) > 0 can be chosen in such a way that for every u R p \ {ū} satisfying (9), the matrix Φ (u)+ω(u) is invertible, and (Φ (u)+ω(u)) = O( u ū ) as u ū. When Ω 0, this result is a particular case of [, Lemma 3.]. We proceed to establish convergence of the iterates satisfying the pnm framework (), from any starting point in the relevant domain. This result is a generalization of [, Lemma 5.].

8 Izmailov, Kurennoy, and Solodov Theorem Let Φ : R p R p be twice differentiable near ū R p, with its second derivative being Lipschitz-continuous with respect to ū, that is, Φ (u) Φ (ū) = O( u ū ) as u ū. Let ū be a solution of the equation (), and assume that Φ is -regular at ū in a direction v kerφ (ū) S. Let Ω : R p R p p and ω : R p R p satisfy the estimates (5), (8), as well as ΠΩ(u) = O( u ū ) + O( u ū ) (4) and Πω(u) = O( u ū u ū ) + O( u ū 3 ) (5) as u ū. Then for every ε > 0 and δ > 0, there exist ε = ε( v) > 0 and δ = δ( v) > 0 such that any starting point u 0 R p \ {ū} satisfying u 0 ū ε, u 0 ū u 0 ū v δ (6) uniquely defines the sequence {u k } R p such that for each k it holds that v k = u k+ u k solves (), u k ū, the point u = u k satisfies (9), the sequence {u k } converges to ū, the sequence { u k ū } converges to zero monotonically, as k, and u k+ ū u k+ ū = O( uk ū ) (7) u k+ ū lim k u k ū =. (8) Proof We again assume that ū = 0, and Φ is given by () with R satisfying (3). Considering that v = 0, observe first that if u R p \ {0} satisfies the second condition in (6) with some δ (0, ), then u u = u u v u u v δ. This implies that and hence, so that Then u δ u, (9) u u + u δ u + u, ( δ) u u. (30) u u v u u v + u u u u u u v + u u u δ + u u δ, (3)

Critical solutions of nonlinear equations: Newton methods 9 where the second inequality employs the fact that u / u is the metric projection of u/ u onto kera. From (4) and (9) it evidently follows that Ω satisfies the corresponding assumptions in Lemma. Therefore, according to this lemma, there exist ε > 0 and δ > 0 such that for every u R p \{ū} satisfying (9), the equation () with u k = u has the unique solution v, and and u + v = O( u ) u + v = u + O( u ) + O( u ) as u 0, where (4) and (5) were again taken into account. In what follows, we use these values ε and δ, with the understanding that if we prove the assertion of the theorem for these specific values of ε and δ, it will also be valid for any larger values of those constants. At the same time, since Lemma allows for these ε and δ, it certainly allows for any smaller values. Therefore, there exists C > 0 such that and hence, u + v C u, (3) u + v u u + v u + u ), (33) u C( u + u ) u + v u + v u +C( u + u ). (34) From (9), (9) and (30) with δ = δ, and from (34), we further derive that ( ) ( ) δ C( δ + ε) u u + v u + v +C( δ + ε) u. Reducing ε > 0 and δ > 0 if necessary, so that and setting we then obtain that where q = δ δ +C( δ + ε) <, C( δ + ε), q + = +C( δ + ε), q u u + v u + v q + u, (35) 0 < q < q + <. (36) By (33), the right inequality in (34), and the left inequality in (35), we have that u + v u + v u u = (u + v) u u u + v u u + v C( u + u ) q u = C ( ) u q u + u, (37)

0 Izmailov, Kurennoy, and Solodov u + v u + v u u = (u + v ) u u u + v u u + v Now choose ε (0, ε], δ (0, δ] satisfying ( δ + δ + C q C( u + u ) q u = C ( ) u q u + u. (38) ( C q + ) ε q + ) δ, (39) and assume that (6) holds for u 0 R p \ {0}. Suppose that u j R p \ {0}, satisfy () and (9) with u = u j for all j =,..., k. Then by the choice of ε > 0 and δ > 0, there exists the unique u k+ satisfying (), and by (9), (3), (3), (35), (36), (37), (38), it holds that 0 < u k+ q + u k q + u k... q k+ + u 0 q k+ + ε ε ε, u k+ u k+ v u k u k v + u k+ u k+ uk u k u k u k v + u k u k uk u k + u k+ u k+ uk u k... u 0 u 0 v k + u j j= u j u j u j + u k+ u k+ uk u k ( ) k C u δ + j j= q u j + u j + C ( u k ) q u k + uk ( ) δ + C k u q j u j + u j δ + C q δ + C q δ + C q δ + C q δ, j=0 ( u 0 k u 0 + C u j + j= q ( δ + C q ( δ + C q ( δ + k j= q j + ε + k j=0 k j=0 q j +ε ε + ε ) q + q + ) ) ε ( C q + q + u j where the last inequality is by (39). Therefore, (9) holds with u = u k+. We have thus established that there exists the unique sequence {u k } R p such that for each k the point u k satisfies () and (9) with u = u k. By (35) and (36), it then follows that u k ū for all k, and {u k } converges to 0. ) )

Critical solutions of nonlinear equations: Newton methods According to (3) and (35), it holds that for all k. This yields (7). Furthermore, according to (34), C uk + uk u k u k+ u k+ C q u k uk+ u k +C uk + uk u k, where by (7) both sides tend to / as k. This gives (8). Remark In Theorem, by the monotonicity of the sequence { u k ū }, for every k large enough it holds that u k+ ū u k+ ū u k ū. Therefore, (7) implies the estimates u k+ ū u k+ ū = O( uk ū ) and as k. Furthermore, u k+ ū = O( u k ū ) (40) u k+ ū u k ū + u k ū uk+ ū u k ū uk+ ū u k ū, where by (7) and (8) both sides tend to / as k. Therefore, Finally, lim k u k+ ū u k = ū. (4) u k+ ū u k+ ū u k uk+ ū ū u k ū uk+ ū u k ū where by (40) and (4) both sides tend to / as k. Therefore, u k+ ū lim k u k ū =. + uk+ ū u k ū Theorem establishes the existence of a set with nonempty interior, which is star-like with respect to ū, and such that any sequence satisfying the pnm relation () and initialized from any point of this set, converges linearly to ū. Moreover, if Φ is -regular at ū in at least one direction v kerφ (ū), then the set of such v is open and dense in kerφ (ū) S: its complement is the null set of the nontrivial homogeneous polynomial detb( ) considered on kerφ (ū) S. The union of convergence domains coming with all such v is also a star-like convergence domain with nonempty interior. In the case when Φ (ū) = 0 (full degeneracy), this domain is quite large. In particular, it is asymptotically dense : the only excluded directions are those in which Φ is not -regular at ū, which is the null set of a nontrivial homogeneous polynomial. Beyond the case of full degeneracy, the convergence

Izmailov, Kurennoy, and Solodov domain given by Theorem is at least not asymptotically thin. Though it is also not asymptotically dense. For the (unperturbed) Newton method, the existence of asymptotically dense star-like domain of convergence was established in [, Theorem 6.]. Specifically, it was demonstrated that one Newton step from a point u 0 in this domain leads to the convergence domain coming with the appropriate v = π(u 0 )/ π(u 0 ), where π(u 0 ) = (u0 ū ) + ( B(u 0 ū)) ΠB[u 0 ū, u 0 ū ]; see () above. Deriving a result like this for the pnm scheme () is hardly possible in general, at least without rather restrictive assumptions on perturbation terms. Perhaps results along those lines can be derived for specific methods, rather than for the general pnm framework, but such developments are not known at this time. 3 Applications to some specific algorithms In this section we show how our general results for the pnm framework () can be applied to some specific methods. In particular, we consider the following algorithms, all developed for tackling the difficult case of non-isolated solutions: the Levenberg Marquardt method and the LP-Newton method for equations, and the stabilized Newton Lagrange method for optimization. We start with observing that the assumptions (5), (8), (4) and (5) on the perturbations terms in Theorem hold automatically if Ω(u) = O( Φ(u) ), ω(u) = O( u ū Φ(u) ) (4) as u ū. Indeed, this readily follows from the relation Φ(u) = Φ (ū)(u ū ) + O( u ū ) as u ū. Thus, to apply Theorem, in the part of perturbations properties it is sufficient to verify (4). 3. Levenberg Marquardt method An iteration of the classical Levenberg Marquardt method [4, Chapter 0.] consists in solving the following subproblem: minimize Φ(uk ) + Φ (u k )v + σ(uk ) v, v R p, (43) where u k R p is the current iterate and σ : R p R + defines the regularization parameter. In [9], it was established that under the Lipschitzian error bound condition (i.e., being initialized near a noncritical solution ū of ()), the method described by (43) with σ(u) = Φ(u) generates a sequence which is quadratically convergent to a (nearby) solution of (). For analysis under the Lipschitzian error bound condition of a rather general framework that includes the Levenberg Marquardt method, see [5, 8]. Our interest here is the case of critical solutions, when the error bound does not hold.

Critical solutions of nonlinear equations: Newton methods 3 First, note that the unique (if σ(u k ) > 0) minimizer of the convex quadratic function in (43) is characterized by the linear system (Φ (u k )) T Φ(u k ) + ((Φ (u k )) T Φ (u k ) + σ(u k )I)v = 0. (44) We next show how (44) can be embedded into the pnm framework (). In particular, we construct the perturbation terms for which the conditions (4) hold, and hence Theorem is applicable, and which correspond to (44) on the relevant domain of convergence. As a result, we obtain the following convergence assertions. Corollary Let Φ : R p R p be twice differentiable near ū R p, with its second derivative Lipschitz-continuous with respect to ū. Let ū be a solution of the equation (), and assume that Φ is -regular at ū in a direction v kerφ (ū) S. Then for any τ, there exist ε = ε( v) > 0 and δ = δ( v) > 0 such that any starting point u 0 R p \ {ū} satisfying (6) uniquely defines the sequence {u k } R p such that for each k it holds that v k = u k+ u k solves (43) with σ(u) = Φ(u) τ, u k ū, the sequence {u k } converges to ū, the sequence { u k ū } converges to zero monotonically, and (7) and (8) hold. Moreover, if Φ (ū) = 0, then the same assertion is valid with any τ 3/. Proof Define ε = ε( v) > 0 and δ = δ( v) > 0 according to Remark, where we set Ω 0. Define the set K = K( v) = {u R p \ {ū} (9) holds}. (45) Then Φ (u) is invertible for all u K, and (Φ (u)) = O( u ū ) as u ū (this can also be concluded directly from [, Lemma 3.], with an appropriate choice of ε > 0 and δ > 0). We next define the mappings Ω and ω. First, we set Ω(u) = 0 and ω(u) = 0 for all u R p \K. Of course, those vanishing perturbation terms for u K have no relation to (44). The point is that we next show that with the appropriate definitions for u K, we obtain Ω and ω satisfying (4). Then Theorem ensures that if the starting point satisfies (6) with appropriate ε > 0 and δ > 0, it follows that the subsequent iterates are well defined and remain in the set K. Finally, in K, the constructed Ω and ω do correspond to (44). Let u K. Considering (44) with u k = u, and multiplying both sides of this relations by the matrix (((Φ (u k )) T ) = (((Φ (u k )) ) T, we obtain that Φ(u k ) + (Φ (u k ) + σ(u k )((Φ (u k )) ) T )v = 0, which is the pnm iteration system () with the perturbation terms given by Then Ω(u) = σ(u)(((φ (u)) ) T, ω(u) = 0, u K. Ω(u) = O( u ū σ(u)). Therefore, the needed first estimate in (4) would hold if σ(u) = O( u ū Φ(u) ) (46) as u K tends to ū. Let σ(u) = Φ(u) τ with τ > 0. Then (46) takes the form Φ(u) τ = O( u ū ), and since Φ(u) = O( u ū ), the last estimate is satisfied as u ū if τ.

4 Izmailov, Kurennoy, and Solodov Moreover, in the case when Φ (ū) = 0 (full singularity) it holds that Φ(u) = O( u ū ) as u ū, and hence, in this case, the appropriate values are all τ 3/. The construction is complete. As the exhibited Ω and ω satisfy (4) as u ū (regardless of whether u stays in K or not), Theorem is applicable. In particular, it guarantees that for appropriate starting points all the iterates stay in K. In this set, the perturbation terms define the Levenberg Marquardt iterations (44). Thus, the assertions follow from Theorem. The following example is taken from DEGEN test collection [4]. 0.5 0 0.5 λ.5.5 3 3.5 4 3 0 3 x (a) Iterative sequences (b) Domain of attraction to the critical solution Fig. : Levenberg Marquardt method with τ = for Example. Example (DEGEN 00) Consider the equality-constrained optimization problem minimize x subject to x = 0. Stationary points and associated Lagrange multipliers of this problem are characterized by the Lagrange optimality system which has the form of the nonlinear equation () with Φ : R R, Φ(u) = (x( + λ), x ), u = (x, λ). The unique feasible point (hence, the unique solution, and the unique stationary point) of this problem is x = 0, and the set of associated Lagrange multipliers is the entire R. Therefore, the solution set of the Lagrange system (i.e., the primal-dual solution set) is { x} R. The unique critical solution is ū = ( x, λ) with λ =, the one for which Φ (ū) = 0 (full singularity). In Figures and, the vertical line corresponds to the primal-dual solution set. These figures show some iterative sequences generated by the Levenberg Marquardt method, and the domains from which convergence to the critical solution was detected. Using zoom in, and taking smaller areas for starting points, does not significantly change the picture in Figure, corresponding to τ =. At the same time, such manipulations with Figures a and b put in evidence that for τ = 3/, the domain of convergence is in fact asymptotically dense (see Figures c and d). These observations are in agreement with Theorem.

Critical solutions of nonlinear equations: Newton methods 5 0.5 0 0.5 λ.5.5 3 3.5 4 3 0 3 x (a) Iterative sequences (b) Domain of attraction to the critical solution 0.75 0.8 0.85 λ 0.9 0.95 0. 0.05 0 0.05 0. 0.5 0. x (c) Iterative sequences (d) Domain of attraction to the critical solution Fig. : Levenberg Marquardt method with τ = 3/ for Example. 3. LP-Newton method The LP-Newton method was introduced in [6]. For the equation (), the iteration subproblem of this method has the form minimize γ subject to Φ(u k ) + Φ (u k )v γ Φ(u k ), v γ Φ(u k ), (v, γ) R p R. (47) The subproblem (47) always has a solution if Φ(u k ) 0 (naturally, if Φ(u k ) = 0 the method stops). If the l -norm is used, this is a linear programming problem (hence the name). As demonstrated in [5, 6] (see also [8]), local convergence properties of the LP-Newton method (under the error bound condition, i.e., near noncritical solutions) are the same as for the Levenberg Marquardt algorithm. Again, our setting is rather that of critical solutions. The proof of the following result is again by placing (47) within the pnm framework (). It is interesting that in this case Ω 0, while ω( ) is defined implicitly: there is no analytic expression for it.

6 Izmailov, Kurennoy, and Solodov Corollary Under the assumptions of Corollary, there exist ε = ε( v) > 0 and δ = δ( v) > 0 such that for any starting point u 0 R p \ {ū} satisfying (6) the following assertions are valid: (a) There exists a sequence {u k } R p such that for each k the pair (v k, γ k+ ) with v k = u k+ u k and some γ k+ solves (47). (b) For any such sequence, u k ū for each k, the sequence {u k } converges to ū, the sequence { u k ū } converges to zero monotonically, and (7) and (8) hold. Proof By the second constraint in (47), the equality () holds for Ω 0 and some ω( ) satisfying ω(u) γ(u) Φ(u), (48) where γ(u) is the optimal value of the subproblem (47) with u k = u. Note that ω( ) would satisfy (4), and thus the assumptions of Theorem, if γ(u) = O( Φ(u) u ū ) (49) as u ū. Indeed, from the first constraint in (47) we then obtain that Φ(u) + Φ (u)v γ(u) Φ(u) = O( u ū Φ(u) ), which implies the second estimate in (4). We thus have to establish (49) on the relevant set. In this respect, the construction is similar to what had been done in Section 3. above. Define ε = ε( v) > 0 and δ = δ( v) > 0 according to Lemma applied with Ω 0 and ω 0, and define the set K according to (45). Then the step v(u) of the (unperturbed) Newton method from any point u K exists, it is uniquely defined, and by (0) and (), it holds that v(u) = O( u ū ). We now define the mappings Ω and ω. Similarly to the case of the Levenberg Marquardt method, we first set them identically equal to zero on R p \ K. Let u K. Then the point (v, γ) = (v(u), v(u) / Φ(u) ) is feasible in (47), and hence, γ(u) γ = Φ(u) v(u) = O( Φ(u) u ū ) as u K tends to ū. As in the proof of Corollary, we note that the constructed Ω and ω satisfy (4) as u ū. Therefore, Theorem ensures that for appropriate starting points all the iterates stay in K, and in this set the perturbation terms defined hereby correspond to (47). In particular, the assertions then follow from Theorem. Observe that for the LP-Newton method, the values of ω( ) are defined in a posteriori manner, after v k is computed. For this reason, Theorem cannot yield uniqueness of the iterative sequence: the next iterate can be defined by any ω( ) satisfying (48) for u = u k, and different choices of appropriate ω( ) may give rise to different next iterates. Figure 3 shows the same information for the LP-Newton method as Figure for the Levenberg Marquardt algorithm, with the same conclusions.

Critical solutions of nonlinear equations: Newton methods 7 0.5 0 0.5 λ.5.5 3 3.5 4 3 0 3 x (a) Iterative sequences (b) Domain of attraction to the critical solution 0.5 0.6 0.7 λ 0.8 0.9 0. 0. 0 0. 0. 0.3 0.4 0.5 x (c) Iterative sequences (d) Domain of attraction to the critical solution Fig. 3: LP-Newton method for Example. 3.3 Equality-constrained optimization and the stabilized Newton Lagrange method We next turn our attention to the origin of the critical solutions issues, namely, to the equality-constrained optimization problem minimize f (x) subject to h(x) = 0, (50) where f : R n R and h : R n R l are smooth. The Lagrangian L : R n R l R of this problem is given by L(x, λ) = f (x) + λ, h(x). Then stationary points and associated Lagrange multipliers of (50) are characterized by the Lagrange optimality system L (x, λ) = 0, h(x) = 0, x with respect to x R n and λ R l. The Lagrange optimality system is a special case of nonlinear equation (), corresponding to setting p = q = n + l, u = (x, λ), ( ) L Φ(u) = (x, λ), h(x). (5) x

8 Izmailov, Kurennoy, and Solodov The stabilized Newton Lagrange method (or stabilized sequential quadratic programming) was developed for solving the Lagrange optimality system (or optimization problem) when the multipliers associated to a stationary point need not be unique [8,,7,0]; see also [, Chapter 7]. For the current iterate u k = (x k, λ k ) R n R l, the iteration subproblem of this method is given by minimize f (x k ),ξ + L x (xk, λ k )ξ,ξ + σ(uk ) η subject to h(x k ) + h (x k )ξ σ(u k )η = 0, where the minimization is in the variables (ξ, η) R n R l, and σ : R n R l R + now defines the stabilization parameter. Equivalently, the following linear system (characterizing stationary points and associated Lagrange multipliers of this subproblem) is solved: L x (xk, λ k ) + L x (xk, λ k )ξ + (h (x k )) T η = 0, h(x k ) + h (x k )ξ σ(u k )η = 0. (5) With a solution (ξ k, η k ) of (5) at hand, the next iterate is given by u k+ = (x k + ξ k, λ k + η k ). Note that if σ 0, then (5) becomes the usual Newton Lagrange method, i.e., the basic Newton method applied to the Lagrange optimality system. For a given Lagrange multiplier λ associated with a stationary point x of problem (50), define the linear subspace { Q( x, λ) = ξ kerh ( x) L x ( x, λ)ξ im ( h ( x) ) } T. Recall that the multiplier λ is called critical if Q( x, λ) {0}; see [3,8]. Otherwise λ is noncritical. As demonstrated in [0], if initialized near a primal-dual solution with a noncritical dual part, the stabilized Newton Lagrange method with σ(u) = Φ(u), where Φ is given by (5), generates a sequence which is superlinearly convergent to a (nearby) solution. Again, of current interest is the critical case. Evidently, the iteration (5) fits the pnm framework () for Φ defined in (5), taking ( ) 0 0 Ω(u) =, ω 0. 0 σ(u)i These perturbations satisfy the assumptions in Theorem if, e.g., σ(u) = Φ(u) τ, τ. We thus conclude the following. Corollary 3 Let f : R n R and h : R n R l be three times differentiable near a stationary point x R n of problem (50), with their third derivatives Lipschitz-continuous with respect to x, and let λ R l be a Lagrange multiplier associated to x. Assume that the mapping Φ defined in (5) is -regular at ū = ( x, λ) in a direction v kerφ (ū) S. Then for any τ, there exist ε = ε( v) > 0 and δ = δ( v) > 0 such that any starting point u 0 = (x 0, λ 0 ) (R n R l ) \ {ū} satisfying (6) uniquely defines a sequence {(x k, λ k )} R n R l such that for each k it holds that (ξ k, η k ) = (x k+ x k, λ k+ λ k ) solves (5) with σ(u) = Φ(u) τ, u k ū, the sequence {u k } converges to ū, the sequence { u k ū } converges to zero monotonically, and (7) and (8) hold.

Critical solutions of nonlinear equations: Newton methods 9 Table : Cases of convergence to λ = in Example (%) ε L-M L-M L-M LP-N sn-l N-L τ = τ = 3/ τ = 38 44 56 54 4 83 0.5 34 5 68 55 57 86 0.5 36 57 79 58 73 86 0. 38 68 86 73 83 90 0.0 38 84 86 86 88 86 Table : Cases of convergence to λ = (, 3) in Example (%) ε L-M L-M L-M LP-N sn-l N-L τ = τ = 3/ τ = 6 3 4 4 65 0.5 7 37 45 7 75 0.5 3 9 47 48 5 8 0. 3 33 63 55 6 9 0.0 4 60 89 8 59 97 It is worth to mention that the above results and discussion can be extended to variational problems (of which optimization is a special case), see [7], [, Chapter 7]. We proceed with some further considerations. A multiplier is said to be critical of order one if dimq( x, λ) =. The following was established in [5]. Proposition Let f : R n R and h : R n R l be three times differentiable at a stationary point x R n of problem (50), and let λ R l be a Lagrange multiplier associated to x. Let Q( x, λ) be spanned by some ξ R n \ {0}, i.e., λ is a critical multiplier of order one. If rankh ( x) = l, then kerφ (ū) contains elements of the form v = ( ξ,η) with some η R l, and Φ is -regular at ū in every such direction if and only if h ( x)[ ξ, ξ ] imh ( x). If rankh ( x) l, then Φ cannot be -regular at ū in any direction v kerφ (ū). If h ( x) = 0, and l or h ( x)[ ξ, ξ ] = 0, then Φ cannot be -regular at ū in any direction v kerφ (ū). In the last two cases specified in the proposition above, Theorem is not applicable; these cases require special investigation. In the last case, when l but h ( x)[ ξ, ξ ] 0, allowing for non-isolated critical multipliers, the effect of attraction of the basic Newton- Lagrange method to such multipliers had been studied in [3] for fully quadratic problems. On the other hand, if, e.g., l =, h ( x) = 0, and h ( x)[ ξ, ξ ] 0, then Theorem is applicable with v = ( ξ, η) for every η R. Taking here η = 0 recovers the results in [7] for the basic Newton-Lagrange method and the fully quadratic case. Moreover, this is exactly the situation that we have in Example. Table reports on the percentage of detected cases of dual convergence to the unique critical multiplier λ = in Example, for the algorithms discussed above, depending on the size ε of the region for starting points around ( x, λ). In the table, L-M refers to the Levenberg Marquardt method (with different values for the power τ that defines the regularization parameter), LP-N refers to the LP-Newton method, N-L to the Newton-Lagrange method (i.e., (5) with σ 0), and sn-l to the stabilized Newton-Lagrange method. The case when dimq( x, λ) (i.e., when λ is critical of order higher than ) opens wide possibilities for -regularity in the needed directions, and such solutions are often specially attractive for Newton-type iterates.

0 Izmailov, Kurennoy, and Solodov 4 3 λ 0 3 4 0 4 6 8 λ Fig. 4: Critical multipliers in Example. Example (DEGEN 030) Consider the equality-constrained optimization problem minimize x x + x 3 subject to x + x x 3 = 0, x x 3 = 0. Here, x = 0 is the unique solution, h ( x) = 0, and the set of associated Lagrange multipliers is the entire R. Critical multipliers are those satisfying λ = or (λ 3) λ =. In Figure 4, critical multipliers are those forming the vertical straight line and two branches of the hyperbola. According to Proposition, the -regularity property cannot hold for Φ defined in (5), at ( x, λ) for any direction v kerφ (ū), for all critical multipliers λ, except for λ = (, ± 3), which are the two intersection points of the vertical line and the hyperbola. One can directly check that the mapping Φ is indeed -regular at ū = ( x, λ) in some directions v kerφ (ū). Table reports on the percentage of detected cases of dual convergence to λ = (, 3), for the algorithms discussed above, depending on the size ε of the region for starting points around ( x, λ). Acknowledgment. The authors thank the two anonymous referees for their comments on the original version of this article. References. A.V. Arutyunov. Optimality Conditions: Abnormal and Degenerate Problems. Kluwer Academic Publishers, Dordrecht, 000.

Critical solutions of nonlinear equations: Newton methods. E.R. Avakov. Extremum conditions for smooth problems with equality-type constraints. USSR Comput. Math. Math. Phys. 5 (985), 4 3. 3. F.H. Clarke. Optimization and Nonsmooth Analysis. John Wiley, New York, 983. 4. DEGEN. http://w3.impa.br/~optim/degen collection.zip. 5. F. Facchinei, A. Fischer, and M. Herrich. A family of Newton methods for nonsmooth constrained systems with nonisolated solutions. Math. Methods Oper. Res. 77 (03), 433 443. 6. F. Facchinei, A. Fischer, and M. Herrich. An LP-Newton method: Nonsmooth equations, KKT systems, and nonisolated solutions. Math. Program. 46 (04), 36. 7. D. Fernández and M. Solodov. Stabilized sequential quadratic programming for optimization and a stabilized Newton-type method for variational problems. Math. Program. 5 (00), 47 73. 8. A. Fischer, M. Herrich, A.F. Izmailov, and M.V. Solodov. Convergence conditions for Newton-type methods applied to complementarity systems with nonisolated solutions. Comput. Optim. Appl. 63 (06), 45 459. 9. H. Gfrerer and B.S. Mordukhovich. Complete characterizations of tilt stability in nonlinear programming under weakest qualification conditions. SIAM J. Optim. 5 (05), 08 9. 0. H. Gfrerer and J.V. Outrata. On computation of limiting coderivatives of the normal-cone mapping to inequality systems and their applications. Optim. 65 (06), 67 700.. A. Griewank. Starlike domains of convergence for Newton s method at singularities. Numer. Math. 35 (980), 95.. W.W. Hager. Stabilized sequential quadratic programming. Comput. Optim. Appl. (999), 53 73. 3. A.F. Izmailov. On the analytical and numerical stability of critical Lagrange multipliers. Comput. Math. Math. Phys. 45 (005), 930 946. 4. A.F. Izmailov, A.S. Kurennoy, and M.V. Solodov. A note on upper Lipschitz stability, error bounds, and critical multipliers for Lipschitz-continuous KKT systems. Math. Program. 4 (03), 59 604. 5. A.F. Izmailov, A.S. Kurennoy, and M.V. Solodov. Critical solutions of nonlinear equations: stability issues. Math. Program., 06. DOI 0.007/s007-06-047-x. 6. A.F. Izmailov and M.V. Solodov. Error bounds for -regular mappings with Lipschitzian derivatives and their applications. Math. Program. 89 (00), 43 435. 7. A.F. Izmailov and M.V. Solodov. The theory of -regularity for mappings with Lipschitzian derivatives and its applications to optimality conditions. Math. Oper. Res. 7 (00), 64 635. 8. A.F. Izmailov and M.V. Solodov. On attraction of Newton-type iterates to multipliers violating secondorder sufficiency conditions. Math. Program. 7 (009), 7 304. 9. A.F. Izmailov and M.V. Solodov. On attraction of linearly constrained Lagrangian methods and of stabilized and quasi-newton SQP methods to critical multipliers. Math. Program. 6 (0), 3 57. 0. A.F. Izmailov and M.V. Solodov. Stabilized SQP revisited. Math. Program. (0), 93 0.. A.F. Izmailov and M.V. Solodov. Newton-Type Methods for Optimization and Variational Problems. Springer Series in Operations Research and Financial Engineering. 04.. A.F. Izmailov and M.V. Solodov. Critical Lagrange multipliers: what we currently know about them, how they spoil our lives, and what we can do about it. TOP 3 (05), 6. Rejoinder of the discussion: TOP 3 (05), 48 5. 3. A.F. Izmailov and E.I. Uskov. Attraction of Newton method to critical Lagrange multipliers: fully quadratic case. Math. Program. 5 (05), 33 73. 4. J. Nocedal and S.J. Wright. Numerical Optimization. nd edition. Springer-Verlag, New York, Berlin, Heidelberg, 006. 5. C. Oberlin and S.J. Wright. An accelerated Newton method for equations with semismooth Jacobians and nonlinear complementarity problems. Math. Program. 7 (009), 355 386. 6. R.T. Rockafellar and R.J.-B. Wets. Variational Analysis. Springer-Verlag, Berlin, 998. 7. E.I. Uskov. On the attraction of Newton method to critical Lagrange multipliers. Comp. Math. Math. Phys. 53 (03), 099. 8. S.J. Wright. Superlinear convergence of a stabilized SQP method to a degenerate solution. Comput. Optim. Appl. (998), 53 75. 9. N. Yamashita and M. Fukushima. On the rate of convergence of the Levenberg Marquardt method. Computing. Suppl. 5 (00), 37 49.