Keywords: Nonlinear least-squares problems, regularized models, error bound condition, local convergence.

Similar documents
Università di Firenze, via C. Lombroso 6/17, Firenze, Italia,

Trust-region methods for rectangular systems of nonlinear equations

Spectral gradient projection method for solving nonlinear monotone equations

Higher-Order Methods

Worst Case Complexity of Direct Search

Optimal Newton-type methods for nonconvex smooth optimization problems

Worst Case Complexity of Direct Search

On affine scaling inexact dogleg methods for bound-constrained nonlinear systems

Trust Regions. Charles J. Geyer. March 27, 2013

Suppose that the approximate solutions of Eq. (1) satisfy the condition (3). Then (1) if η = 0 in the algorithm Trust Region, then lim inf.

This manuscript is for review purposes only.

Part 3: Trust-region methods for unconstrained optimization. Nick Gould (RAL)

A globally and R-linearly convergent hybrid HS and PRP method and its inexact version with applications

Cubic regularization of Newton s method for convex problems with constraints

Worst-Case Complexity Guarantees and Nonconvex Smooth Optimization

A derivative-free nonmonotone line search and its application to the spectral residual method

An example of slow convergence for Newton s method on a function with globally Lipschitz continuous Hessian

The effect of calmness on the solution set of systems of nonlinear equations

1. Introduction. We consider the numerical solution of the unconstrained (possibly nonconvex) optimization problem

Linear algebra issues in Interior Point methods for bound-constrained least-squares problems

Namur Center for Complex Systems University of Namur 8, rempart de la vierge, B5000 Namur (Belgium)

University of Houston, Department of Mathematics Numerical Analysis, Fall 2005

j=1 r 1 x 1 x n. r m r j (x) r j r j (x) r j (x). r j x k

Worst case complexity of direct search

Handout on Newton s Method for Systems

1. Nonlinear Equations. This lecture note excerpted parts from Michael Heath and Max Gunzburger. f(x) = 0

Technische Universität Dresden Herausgeber: Der Rektor


Local convergence analysis of the Levenberg-Marquardt framework for nonzero-residue nonlinear least-squares problems under an error bound condition

Scientific Computing: An Introductory Survey

Scientific Computing: An Introductory Survey

arxiv: v1 [math.oc] 1 Jul 2016

A PROJECTED HESSIAN GAUSS-NEWTON ALGORITHM FOR SOLVING SYSTEMS OF NONLINEAR EQUATIONS AND INEQUALITIES

Methods for Unconstrained Optimization Numerical Optimization Lectures 1-2

Unconstrained optimization

Newton Method with Adaptive Step-Size for Under-Determined Systems of Equations

MS&E 318 (CME 338) Large-Scale Numerical Optimization

Numerisches Rechnen. (für Informatiker) M. Grepl P. Esser & G. Welper & L. Zhang. Institut für Geometrie und Praktische Mathematik RWTH Aachen

Line Search Methods for Unconstrained Optimisation

1. Introduction. We analyze a trust region version of Newton s method for the optimization problem

Multipoint secant and interpolation methods with nonmonotone line search for solving systems of nonlinear equations

Complexity of gradient descent for multiobjective optimization

5 Handling Constraints

An Alternative Three-Term Conjugate Gradient Algorithm for Systems of Nonlinear Equations

CONVERGENCE BEHAVIOUR OF INEXACT NEWTON METHODS

A Regularized Directional Derivative-Based Newton Method for Inverse Singular Value Problems

Primal-dual relationship between Levenberg-Marquardt and central trajectories for linearly constrained convex optimization

system of equations. In particular, we give a complete characterization of the Q-superlinear

Outline. Scientific Computing: An Introductory Survey. Optimization. Optimization Problems. Examples: Optimization Problems

Cubic-regularization counterpart of a variable-norm trust-region method for unconstrained minimization

Newton-type Methods for Solving the Nonsmooth Equations with Finitely Many Maximum Functions

On the complexity of an Inexact Restoration method for constrained optimization

A Trust Funnel Algorithm for Nonconvex Equality Constrained Optimization with O(ɛ 3/2 ) Complexity

How to Characterize the Worst-Case Performance of Algorithms for Nonconvex Optimization

c 2007 Society for Industrial and Applied Mathematics

A SIMPLY CONSTRAINED OPTIMIZATION REFORMULATION OF KKT SYSTEMS ARISING FROM VARIATIONAL INEQUALITIES

Derivative-Free Trust-Region methods

x 2 x n r n J(x + t(x x ))(x x )dt. For warming-up we start with methods for solving a single equation of one variable.

A trust region algorithm with a worst-case iteration complexity of O(ɛ 3/2 ) for nonconvex optimization

A Simple Primal-Dual Feasible Interior-Point Method for Nonlinear Programming with Monotone Descent

Cubic regularization in symmetric rank-1 quasi-newton methods

8 Numerical methods for unconstrained problems

On Generalized Primal-Dual Interior-Point Methods with Non-uniform Complementarity Perturbations for Quadratic Programming

A new ane scaling interior point algorithm for nonlinear optimization subject to linear equality and inequality constraints

STOP, a i+ 1 is the desired root. )f(a i) > 0. Else If f(a i+ 1. Set a i+1 = a i+ 1 and b i+1 = b Else Set a i+1 = a i and b i+1 = a i+ 1

Global convergence of trust-region algorithms for constrained minimization without derivatives

WHEN ARE THE (UN)CONSTRAINED STATIONARY POINTS OF THE IMPLICIT LAGRANGIAN GLOBAL SOLUTIONS?

On the Local Convergence of an Iterative Approach for Inverse Singular Value Problems

Conditional Gradient (Frank-Wolfe) Method

1. Introduction. We consider the general smooth constrained optimization problem:

A null-space primal-dual interior-point algorithm for nonlinear optimization with nice convergence properties

GLOBALLY CONVERGENT GAUSS-NEWTON METHODS

Self-Concordant Barrier Functions for Convex Optimization

Numerical Experience with a Class of Trust-Region Algorithms May 23, for 2016 Unconstrained 1 / 30 S. Smooth Optimization

On the Local Quadratic Convergence of the Primal-Dual Augmented Lagrangian Method

Numerical Methods for Large-Scale Nonlinear Systems


E5295/5B5749 Convex optimization with engineering applications. Lecture 8. Smooth convex unconstrained and equality-constrained minimization

OPER 627: Nonlinear Optimization Lecture 14: Mid-term Review

An introduction to complexity analysis for nonconvex optimization

Parameter Optimization in the Nonlinear Stepsize Control Framework for Trust-Region July 12, 2017 Methods 1 / 39

A PENALIZED FISCHER-BURMEISTER NCP-FUNCTION. September 1997 (revised May 1998 and March 1999)

Step lengths in BFGS method for monotone gradients

On the complexity of the hybrid proximal extragradient method for the iterates and the ergodic mean

A Distributed Newton Method for Network Utility Maximization, II: Convergence

Applying a type of SOC-functions to solve a system of equalities and inequalities under the order induced by second-order cone

A QP-FREE CONSTRAINED NEWTON-TYPE METHOD FOR VARIATIONAL INEQUALITY PROBLEMS. Christian Kanzow 1 and Hou-Duo Qi 2

1. Introduction The nonlinear complementarity problem (NCP) is to nd a point x 2 IR n such that hx; F (x)i = ; x 2 IR n + ; F (x) 2 IRn + ; where F is

Affine scaling interior Levenberg-Marquardt method for KKT systems. C S:Levenberg-Marquardt{)KKTXÚ

A globally convergent Levenberg Marquardt method for equality-constrained optimization

Implications of the Constant Rank Constraint Qualification

On fast trust region methods for quadratic models with linear constraints. M.J.D. Powell

Notes on Numerical Optimization

Trust Region Methods. Lecturer: Pradeep Ravikumar Co-instructor: Aarti Singh. Convex Optimization /36-725

An Accelerated Hybrid Proximal Extragradient Method for Convex Optimization and its Implications to Second-Order Methods

Complexity analysis of second-order algorithms based on line search for smooth nonconvex optimization

Some new developements in nonlinear programming

Search Directions for Unconstrained Optimization

A PRIMAL-DUAL TRUST REGION ALGORITHM FOR NONLINEAR OPTIMIZATION

Convex Optimization. Newton s method. ENSAE: Optimisation 1/44

Transcription:

STRONG LOCAL CONVERGENCE PROPERTIES OF ADAPTIVE REGULARIZED METHODS FOR NONLINEAR LEAST-SQUARES S. BELLAVIA AND B. MORINI Abstract. This paper studies adaptive regularized methods for nonlinear least-squares problems where the model of the objective function used at each iteration is either the Euclidean residual regularized by a quadratic term or the Gauss-Newton model regularized by a cubic term. For suitable choices of the regularization parameter the role of the regularization term is to provide global convergence. In this paper we investigate the impact of the regularization term on the local convergence rate of the methods and establish that, under the well-known error bound condition, quadratic convergence to zero-residual solutions is enforced. This result extends the existing analysis on the local convergence properties of adaptive regularized methods. In fact, the known results were derived under the standard full rank condition on the Jacobian at a zero-residual solution while the error bound condition is weaker than the full rank condition and allows the solution set to be locally nonunique. Keywords: Nonlinear least-squares problems, regularized models, error bound condition, local convergence.. Introduction. In this paper we discuss the local convergence behaviour of adaptive regularized methods for solving the nonlinear least-squares problem (.) min f(x) = x IR n 2 F (x) 2 2, where F : IR n IR m is a given vector-valued continuously-differentiable function. Adaptive regularization approaches for unconstrained optimization base each iteration upon a quadratic or a cubic regularization of standard models for f. Their construction follows from observing that, for suitable choices of the regularization parameter, the regularized model overestimates the objective function. The role of the adaptive regularization is to control the distance between successive iterates and to provide global convergence of the procedures. Interesting connections among these approaches, trust-region and linesearch methods are established in [23]. First ideas on adaptive cubic regularization of the Newton s quadratic model for f can be found in [5] in the context of affine invariant Newton methods; further results were successively obtained in [25]. In [2] it was shown that the use of a local cubic overestimator for f yields an algorithm with global and fast local convergence properties and a good global iteration complexity. Elaborating on these ideas, an adaptive cubic regularization method for unconstrained minimization problems was proposed in [4]; it employs a cubic regularization of the Newton s model and allows approximate minimization of the model and/or a symmetric approximation to the Hessian matrix of f. This new approach enjoys good global and local convergence properties as well as the same (in order) worst-case iteration complexity bound as in [2]. Adaptive regularized methods have been also studied for the specific case of nonlinear least-square problems. Complexity bounds for the method proposed in [4] applied to potentially rank-deficient nonlinear least-squares problems were given in [8]. Moreover, such an approach was specialized to the solution of (.) along with Dipartimento di Ingegneria Industriale, Università di Firenze, viale G.B. Morgagni 40, 5034 Firenze, Italia, stefania.bellavia@unifi.it, benedetta.morini@unifi.it Work partially supported by INdAM-GNCS, under the 203 Projects Numerical Methods and Software for Large-Scale Optimization with Applications to Image Processing.

the updating rules for the regularization parameter [4]. Regarding quadratic regularization, a model consisting of the Euclidean residual regularized by a quadratic term was proposed in [20] for nonlinear systems and then extended to general nonlinear least-squares problems allowing the use of approximate minimizers of the model [2]. Further recent works on adaptive cubic regularization concern its extension to constrained nonlinear programming problems [7, 8] and its application in a barrier method for solving a nonlinear programming problem [3]. In this paper we focus on two adaptive regularized methods for nonlinear leastsquare problems introduced in [2, 4] and investigate their local convergence properties. The model used in [2] is a Euclidean residual regularized by a quadratic term, whereas the model used in [4] is the Gauss-Newton model regularized by a cubic term. These methods are especially suited for computing zero-residual solutions of (.) which is a case of interest for example when F is the map of a square nonlinear system of equations or it models the detection of feasible points in nonlinear programming [, 8] The two procedures considered are known to be quadratically convergent to zeroresidual solutions where the Jacobian of F is full rank, see [2, 4]. Here we go a step further and show that the presence of the regularization term provides fast local convergence under weaker assumptions. Thus we can conclude that the regularization has a double role in enhancing the properties of the underlying unregularized methods: besides guaranteeing global convergence, it enforces strong local convergence properties. Our local convergence analysis concerns zero-residual solutions of (.) satisfying an error bound condition and covers square problems (m = n), overdetermined problems (m > n) and underdetermined problems (m < n). Error bounds were introduced in mathematical programming in order to bound, in terms of a computable residual function, the distance of a point from the typically unknown solution set [22]. These conditions have been widely used for studying the local convergence of various optimization methods: proximal methods for minimization problems [6], Newton-type methods for complementarity problems [24], derivative-free methods for nonlinear least-squares [27], Levenberg-Marquardt methods for constrained and unconstrained nonlinear least-squares [, 0, 2, 3, 7, 8, 26, 28]. Our study is motivated by these latter results and the connection established here between the Levenberg-Marquadt methods and the adaptive regularized methods. A similar insight on such a connection between the two classes of procedures is given in [3]. Following [26] we consider the case where the norm of F provides a local error bound for (.) and prove Q-quadratic convergence of ARQ and ARC methods. The error bound condition considered can be valid at locally nonunique zero-residual solutions of (.). In fact, it may hold, irrespective of the dimensions m and n of F, at solutions where the Jacobian J of F is not full rank. Thus our study establishes novel local convergence properties of ARQ and ARC under milder conditions than the standard full rank condition on J assumed in [2, 4]. In the context of Levenberg- Marquardt methods, such fast local convergence properties were defined as strong properties, see [7]. The paper is organized as follows: in 2 we describe the two methods under study, discuss their connection with the Levenberg-Marquardt methods and the error bound condition used. In 3 we provide a convergence result that paves the way for the local convergence analysis carried out in 4. In this latter section we also show how to compute an approximate minimizer of the model and retain fast local convergence 2

properties of the studied approaches. In 5 we discuss the case where the approximate step is computed by minimizing the model in a subspace and its implication on the local convergence properties. In 6 we make some concluding remarks. Notations. For the differentiable mapping F : R n R m, the Jacobian matrix of F at x is denoted by J(x). The gradient and the Hessian matrix of the smooth function f(x) = F (x) 2 /2 are denoted by g(x) = J(x) T F (x) and H(x) respectively. When clear from the context, the argument of a mapping is omitted and, for any function h, the notation h k is used to denote h(x k ). The 2-norm is denoted by x. For any vector y R n, the ball with center y and radius ρ is indicated by B ρ (y), i.e. B ρ (y) = {x : x y ρ}. The identity matrix n n is indicated by I. 2. Adaptive regularized quadratic and cubic algorithms. In this section we discuss the adaptive quadratic and cubic regularized methods proposed in [2, 4, 4] and summarize some of their properties. We first introduce the models proposed for problem (.). Given some iterate x k, in [2] a model consisting of the euclidean residual F k + J k p regularized by a quadratic term is introduced. Letting σ k be a dynamic strictly positive parameter, the model takes the form (2.) m Q k (p) = F k + J k p + σ k p 2. If the Jacobian J of F is globally Lipschitz continuous (with constant 2L) and σ k = L, then m Q k reduces to the modified Gauss-Newton model proposed in [20]. Whenever σ k L, F is overestimated around x k by means of m Q k, i.e. F (x k + p) m Q k (p). Alternatively, in [4] the cubic regularization of the Gauss-Newton model (2.2) m C k (p) = 2 F k + J k p 2 + 3 σ k p 3, is used with a dynamic strictly positive parameter σ k. This model is motivated by the cubic overestimation of f devised in [4, 5, 2]. In fact, if the Hessian H of f is Lipschitz continuous (with constant 2L), then f(x k + p) f k + p T g k + 2 pt H k p + 3 L p 3. Thus, (2.2) is obtained replacing L by σ k and considering the first order approximation Jk T J k to H k ; it is well-known that the latter approximation is reasonable in a neighborhood of a zero-residual solution of problem (.) []. Before addressing the use of the models m Q k and mc k in the solution of (.), we review their properties and the form of the minimizer. Lemma 2.. Suppose that σ k > 0. Then the models m Q k and mc k are strictly convex. Proof. The strict convexity of m Q k is proved in [2, Lemma 2.]. Regarding mc k, the function F k + J k p 2 is convex and the function p 3 is strictly convex. Lemma 2.2. Let F : R n R m be continuously differentiable and suppose g k 0. Then, 3

i) If p k is the minimizer of mq k, then there is a nonnegative λ k such that (p k, λ k ) solves (2.3) (2.4) (J T k J k + λi)p = g k, λ = 2σ k J k p + F k. Moreover, λ k is such that (2.5) λ k [0, 2σ k F k ]. ii) If there exists a solution (p k, λ k ) of (2.3) and (2.4) with λ k > 0, then p k is the minimizer of m Q k. Otherwise, the minimizer of mq k is given by the minimum norm solution of the linear system Jk T J kp = g k. iii) The vector p k is the unique minimizer of mc k if and only if there exists a positive λ k such that (p k, λ k ) solves (2.3) and (2.6) λ k = σ k p k. Proof. The results for m Q k are given in [2, Lemma 4., Lemma 4.3]. The hypothesis for m C k follows from the application of [4, Theorem 3.] to such a model. The above lemma shows that the minimizer of the quadratic and cubic regularized models solves the shifted linear system (2.3) for a specific value λ = λ k which depends on the model and is given in (2.4) and (2.6) respectively. In the rest of the paper, the notation m k will be used to indicate the model, irrespective of its specific form, in all the expressions that are valid for both m Q k and m C k. Further, for a given λ 0, we let p(λ) be the minimum-norm solution of (2.3), and p k = p(λ k ) be the minimizer in Lemma 2.2 without distinguishing between the models; it will be inferred from the context whether p k minimizes either mq k or mc k. The vector p(λ) can be characterized in terms of the singular values of J k as follows. Lemma 2.3. [2, Lemma 4.2] Assume g k = 0 and let p(λ) be the minimum norm solution of (2.3) with λ 0. Assume furthermore that J k is of rank l and its singular-value decomposition is given by U k Σ k Vk T where Σ k = diag(ς k,..., ςν k ), with ν = min(m, n). Then, denoting r k = ((r k ), (r k ) 2,..., (r k ) ν ) T = Uk T F k, we have that (2.7) p(λ) 2 = l i= (ς k i (r k) i ) 2 ((ς k i )2 + λ) 2. In the literature, adaptive regularized methods for (.) compute the step from one iterate to the next as an approximate minimizer of either m Q k or mc k and test its progress toward a solution. In [2] a procedure based on the use of the model m Q k for all k 0 is proposed; here it is named Adaptive Quadratic Regularization (ARQ). On the other hand, in [4] a procedure based on the use of the model m C k for all k 0 is given and denoted Adaptive Cubic Regularization (ARC). The description of kth iteration of ARQ and ARC is summarized in Algorithm 2. where the string Method denotes the name of the method, i.e. it is either ARQ or ARC. The trial step selection consists in finding an approximate minimizer p k of the model which produces a value of m k smaller than that achieved by the Cauchy point (2.8). This step is accepted and the new iterate x k+ is set to x k + p k if a sufficient decrease in the objective is achieved; otherwise, the step is rejected and x k+ is set to 4

x k. We note that the denominator in (2.0) and (2.) is strictly positive whenever the current iterate is not a first-order critical point. As a consequence, the algorithm is well defined and the sequence {f k } is non-increasing. The rules for updating the parameter σ k parallel those for updating the trust-region size in trust-region methods [2, 4]. Algorithm 2. and standard trust-region methods [9] belong to the same unifying framework, see [23]. In a trust-region method with Gauss-Newton model the step is an approximate minimizer of the model subject to the explicit constraint p k k for some adaptive trust-region radius k. In the adaptive regularized methods, p k is an approximate minimizer of a regularized Gauss-Newton model and the stepsize is implicitly controlled. Algorithm 2.: kth iteration of ARQ and ARC Given x k and the constants σ k > 0, > η 2 > η > 0, γ 2 γ >, γ 3 > 0. If Method= ARQ let m k be m Q k, else let m k be m C k. Step : Set (2.8) p c k = α k g k, α k = argmin m k ( αg k ). α 0 Compute an approximate minimizer p k of m k (p) such that (2.9) m k (p k ) m k (p c k). Step 2: If Method= ARQ compute (2.0) ρ k = F (x k) F (x k + p k ) F (x k ) m Q k (p, k) else compute (2.) Step 3: Set ρ k = 2 F (x k) 2 2 F (x k + p k ) 2 2 F (x k) 2 m C k (p k). { xk + p x k+ = k if ρ k η, otherwise. Step 4: Set (0, σ k ] if ρ k η 2 (very successful), (2.2) σ k+ [σ k, γ σ k ) if η ρ k < η 2 (successful), [γ σ k, γ 2 σ k ) otherwise (unsuccessful). x k The convergence properties of ARQ and ARC methods have been studied in [2] and [4] under standard assumptions for unconstrained nonlinear least-squares problems and optimization problems respectively. Both ARQ and ARC show global convergence to first-order critical points of (.), see [2, Theorem 3.8], [4, Corollary 2.6]. 5

Moreover, imposing a certain level of accuracy in the computation of the approximate minimizer p k, quadratic asymptotic convergence to a zero-residual solution is achieved. Specifically, the sequence generated {x k } is Q-quadratically convergent to a zero-residual solution x if J(x ) is full rank; this result is valid for ARQ under any relation between the dimensions m and n of F ([2, Theorems 4.9, 4.0, 4.]) while it is proved for ARC method whenever m n [4, Corollary 4.0]. The purpose of this paper is to show that, even if J is not of full-rank at the solution, ARQ and ARC are Q-quadratically convergent to a zero-residual solution provided that the so-called error bound condition holds. This property is suggested by two issues discussed in the next section: the connection between adaptive regularization methods and the Levenberg-Marquardt method and the local convergence properties of the latter under the error bound condition. 2.. Connection of the steps in ARQ, ARC and Levenberg-Marquardt methods. In a Levenberg-Marquardt method [9], given x k and a positive scalar µ k, the quadratic model for f around x k takes the form (2.3) m LM k (s) = 2 F k + J k s 2 2 + 2 µ k s 2. Letting s k be the minimizer, the new iterate x k+ is set to x k + s k if it provides a sufficient decrease of f; otherwise x k+ is set to x k. Clearly, the minimizer s k of m LM k is the solution of the shifted linear system (2.3) where λ = µ k. The difference between s k and the step p(λ) in ARQ and ARC methods lies in the shift parameter used. In the Levenberg-Marquardt approach the regularization parameter can be chosen as proposed by Moré in the renowned paper [9]. In the adaptive regularized methods, the optimal value of λ k for the minimizer p k = p(λ k ) depends on the regularization parameter σ k and satisfies (2.4) in ARQ and (2.6) in ARC. Moreover, for an approximate minimizer p k = p(λ k ) the value of λ k will be close to λ k on the base of specified accuracy requirements. In [26], Yamashita and Fukushima showed that Levenberg-Marquardt methods may converge locally quadratically to a zero-residual solutions of (.) satisfying a certain error bound condition. Letting S denote the nonempty set of zero-residual solutions of (.) and d(x, S) denote the distance between the point x and the set S, such a condition is defined as follows. Assumption 2.. A point x S satisfies the error bound condition if there exist positive constants χ and α such that (2.4) α d(x, S) F (x), for all x B χ(x ). By extending the theory in [26], Fan and Yuan [3] and Behling and Fischer [] showed that, under the same condition, Levenberg-Marquardt methods converge locally Q- quadratically provided that (2.5) µ k = O( F k δ ), for some δ [, 2]. In [7] this property was defined as a strong local convergence property since it is weaker than the standard full rank condition on the Jacobian J of F. Inequality (2.4) bounds the distance of vectors in a neighbourhood of x to the set S in terms of the computable residual F and depends on the solution x. Remarkably, it allows the solution set S to be locally nonunique [7]. Specifically, in case of overdetermined 6

or square residual functions F, i.e. m n, the condition J(x ) is full rank implies that (2.4) holds, see e.g. [8, Lemma 4.2]. On the other hand, the converse is not true and (2.4) may hold even in the case where J(x ) is not full rank, irrespective to the relationship between m and n. To see this, consider the example given in [0, p. 608] where F : R 2 R 2 has the form F (x, x 2 ) = ( e x x2, (x x 2 )(x x 2 2) ) T, We have that S = {x R 2 : x = x 2 }, J is singular at any point in S. As d(x, S) = ( 2/2) x x 2, the error bound condition is satisfied with α = in a proper neighbourhood of any point x S. Slight modifications of the previous example show that (2.4) is weaker than the full rank condition in the overdetermined and underdetermined cases too. In the first case, an example is given by F : R 2 R 3 such that F (x, x 2 ) = ( e x x2, (x x 2 )(x x 2 2), sin(x x 2 ) ) T, The error bound condition can be showed proceeding as before, by noting that S = {x R 2 : x = x 2 }, J is not full rank at any point in S, d(x, S) = ( 2/2) x x 2 and the error bound condition is satisfied with α = in a proper neighbourhood of any point x S. For the underdetermined case, let us consider the problem where F : R 3 R 2 is given by F (x, x 2, x 3 ) = ( e x x2 x3, (x x 2 x 3 )(x x 2 x 3 2) ) T. Then S = {x R 3 : x x 2 x 3 = 0} and J is rank deficient everywhere in S. As d(x, S) = ( 3/3) x x 2 x 3, again the error bound condition is satisfied with α = in a proper neighbourhood of any point x S. In this paper we show that, under Assumption 2., ARQ and ARC exhibit the same strong convergence properties as the Levenberg-Marquardt methods. Since the existing local results in literature are valid as long as J(x ) is of full-rank, our new results offer a further insight into the effects of the regularizations employed in ARQ and ARC. 3. Local convergence of a sequence. In this section we analyze the local behaviour of a sequence {x k } admitting a limit point x in the set S of the zeroresidual solution of (.). For x k sufficiently close to x and under suitable conditions on the step taken and the behaviour of d(x k, S), we show that the sequence {x k } converges to x Q-quadratically. The theorem proved below uses technicalities from [3] but it does not involve the error bound condition. It will be used in 4 because we will show that, under Assumption 2., the sequences generated by the two Adaptive Regularized methods satisfy its assumptions. Theorem 3.. Suppose that x S and that {x k } is a sequence with limit point x. Let q k R n be such that, (3.) q k Ψd(x k, S), if x k B ɛ (x ), for some positive Ψ and ɛ, and (3.2) x k+ = x k + q k, if x k B ψ (x ), 7

for some ψ (0, ɛ). Then, {x k } converges to x Q-quadratically whenever (3.3) d(x k + q k, S) Γd(x k, S) 2, if x k B ψ (x ), for some positive Γ. Proof. In order to prove that the sequence {x k } is convergent, we show that it is a Cauchy sequence. Consider a fixed positive scalar ζ such that { } ψ ζ min + 4Ψ, (3.4). 2Γ We first prove that if x k B ζ (x ), then (3.5) x k+l B ψ (x ), for all l. Consider the case l = first. Since x k B ζ (x ) and ζ < ψ, it follows from (3.2) that x k+ = x k +q k and by (3.) we get x k +q k x x k x + q k ζ( + Ψ). Then, (3.4) gives x k + q k B ψ (x ) and (3.5) holds for l =. Assume now that (3.5) holds for iterations k + j, j = 0,..., l. By (3.) and (3.2) x k+l x x k+l x k+l +... + x k x (3.6) l ζ + q k+j j=0 l ζ + Ψ d(x k+j, S). j=0 To provide an upper bound for (3.6), we use (3.3) and obtain (3.7) d(x k+j, S) Γd(x k+j, S) 2... Γ (2j ) d(x k, S) 2j Γ (2j ) ζ 2j, for j = 0,..., l. Moreover, since (3.4) implies Γζ 2, it follows (3.8) d(x k+j, S) 2ζ ( ) 2 j, 2 and l l ( ) 2 j d(x k+j, S) 2ζ. 2 j=0 j=0 Thus, since 2 j > j for j 0, we get l l ( ) j d(x k+j, S) 2ζ 4ζ, 2 j=0 j=0 and by (3.6) x k+l x ζ( + 4Ψ) ψ, 8

where in the last inequality we have used the definition of ζ in (3.4). Then, we have proved (3.5) and by (3.2) we can conclude that x k+j+ = x k+j + q k+j, for j 0. Using (3.8) and proceeding as above we have t t x k+r x k+t q k+j Ψ d(x k+j, S) 4Ψζ. j=r Then {x k } is a Cauchy sequence and it is convergent. Since x is a limit point of the sequence we deduce that x k x. We finally show the convergence rate of the sequence. Let k sufficiently large so that x k+j B ζ (x ) for j 0. Then, conditions (3.) (3.4) and (3.7) give j=r x k+ x q k+j+ j=0 ΨΓ d(x k, S) 2 + (Γd(x k, S)) 2j+ 2 d(x k, S) 2 j= ΨΓ + (Γζ) j d(x k, S) 2 3ΨΓ d(x k, S) 2 3ΨΓ x k x 2. This shows the local Q-quadratic convergence of the sequence {x k }. j= 4. Strong local convergence of ARQ and ARC. In this section we provide new results on the local convergence rate of ARQ and ARC methods to zero-residual solutions of (.). Under appropriate assumptions including the error bound condition on a limit point x in the set S, we show that the sequence {x k } generated satisfies the conditions stated in Theorem 3.. The results obtained are valid irrespective of the relation between the dimensions m and n of F. In the following we make the following assumptions: Assumption 4.. F : R n R m is continuously differentiable and for some solution x S there exists a constant ɛ > 0 such that J is Lipschitz continuous with constant 2k in a neighbourhood B 2ɛ (x ) of x. By Assumption 4. some technical results follow. The continuity of J implies that (4.) J(x) κ J, for all x B 2ɛ (x ) and some κ J > 0. Moreover, for any point x, let [x] S S be a vector such that x [x] s = d(x, S). Then, with x S, we have that [x k ] S B 2ɛ (x ) whenever x k B ɛ (x ), as [x k ] S x [x k ] S x k + x k x 2 x k x 2ɛ. As a consequence, by [, Lemma 4..9] (4.2) F k κ J [x k ] S x k, if x k B ɛ (x ), and (4.3) F k κ J x x k, if x k B 2ɛ (x ). 9

Further, (4.4) F (x + p) F (x) J(x)p k p 2, for any x and x + p in B 2ɛ (x ), see [, Lemma 4..2], and consequently (4.5) F k + J k ([x k ] S x k ) k d(x k, S) 2, if x k B ɛ (x ). 4.. Analysis of the step. We prove that the step p k generated by Algorithm 2. satisfies condition (3.), i.e. (4.6) p k Ψd(x k, S), if x k B ɛ (x ), for some strictly positive Ψ and ɛ. We start showing that the minimizer p k of the models satisfies (4.7) p k Θd(x k, S), for some positive scalar Θ, if x k is sufficiently close to x. Lemma 4.. Let Assumption 4. hold and x S be a limit point of the sequence {x k } generated by Algorithm 2.. Suppose that σ k > σ min > 0 for all k 0. Then, if x k B ɛ (x ), then p k satisfies (4.7). Proof. Let x k B ɛ (x ). Consider the ARQ method. Since p k minimizes the model m Q k we have (4.8) m Q k (p k) m Q k ([x k] S x k ), and by (4.5) (4.9) m Q k (p k) F k + J k ([x k ] S x k ) + σ k [x k ] S x k 2 (k + σ k )d(x k, S) 2. Using (2.) we get p k 2 m Q k (p k )/σ k, and the desired result follows from (4.9) and the assumption σ k > σ min > 0. For ARC method we proceed as above and get m C k (p k) 2 F k + J k ([x k ] S x k ) 2 + 3 σ k [x k ] S x k 3 ( 2 k2 ɛ + ) 3 σ k d(x k, S) 3. Since (2.2) implies p k 3 3m C k (p k )/σ k, the proof is completed. We now consider the case of practical interest where the minimizer of the model is approximately computed and characterize such an approximation. We proceed supposing that the couple (p k, λ k ) satisfies the following two assumptions Assumption 4.2. The step p k has the form p k = p(λ k ), i.e. it solves (2.3) with λ = λ k. Assumption 4.3. The scalar λ k is such that λ k + τ k λ k λ k( + τ k ), 0

for a given τ k [0, τ max ]. Trivially, these assumptions are satisfied when p k = p k. The upper bound τ max on τ k ensures that λ k goes to zero as fast as λ k does. In Section 4.3 we will show that, letting τ k be a threshold chosen by the user, practical implementations of ARQ and ARC provide a step p k satisfying both the above conditions. The bound on the norm of p k is now derived. Lemma 4.2. Suppose that Assumption 4. holds and that x S is a limit point of the sequence {x k } generated by Algorithm 2.. Let σ k σ min > 0 for all k 0. Then there exists a positive Ψ such that if x k B ɛ (x ) and (p k, λ k ) satisfies Assumptions 4.2 and 4.3, then (4.6) holds. Proof. From Lemma 2.3 the function p(λ) 2 is monotonic decreasing for λ 0. Then, when λ k λ k we have p(λ k) p(λ k ) and by (4.7) the hypothesis follows. More generally, Assumption 4.3 yields (4.0) Moreover, from (2.7) ( ) λ p k 2 = + τ k ( ) p(λ k ) 2 λ p k 2. + τ k l (ςi k(r k) i ) 2 ( (ςi k)2 + λ k + τ k i= = ( + τ k ) 2 l ( + τ k ) 2 i= l i= = ( + τ k ) 2 p(λ k) 2. ) 2 (ς k i (r k) i ) 2 ( (ς k i ) 2 ( + τ k ) + λ k) 2 (ς k i (r k) i ) 2 ( (ς k i ) 2 + λ k) 2 Therefore, from (4.7) and τ k τ max we get the required result. 4.2. Successful iterations and convergence of the sequence {x k }. In order to apply Theorem 3., the second step is to prove that iteration k of ARQ and ARC is successful, i.e. (4.) x k+ = x k + p k, if x k B ψ (x ), for some ψ (0, ɛ). In the rest of the section the error bound Assumption 2. is supposed on the limit point x and the scalar ɛ in Assumption 4. is possibly reduced to be such that ɛ χ, where χ is as in (2.4). Moreover we require that (4.2) for some positive σ max. σ k σ max for all k 0,

In the results below we will make use of the interpretation of p k as the minimizer of the model m LM k given in (2.3) with µ k = λ k. Thus, by (4.5) m LM k (p k ) m LM k ([x k ] S x k ) = 2 F k + J k ([x k ] S x k ) 2 + 2 λ k [x k ] S x k 2 (4.3) 2 k2 [x k ] S x k 4 + 2 λ k [x k ] S x k 2, whenever x k B ɛ (x ). Lemma 4.3. Let Assumptions 4., 4.2 and 4.3 hold and x S be a limit point of the sequence {x k } generated by the ARQ method satisfying Assumption 2.. Suppose that σ max σ k σ min > 0 for all k 0. Then there exist positive scalars ψ and Λ such that, if x k B ψ (x ), (4.4) m Q k (p k) Λd(x k, S) 3/2, and iteration k is very successful. Proof. Let ψ = ɛ/( + Ψ) where ɛ and Ψ are the scalars in Lemma 4.2. Assume that x k B ψ (x ). Using (4.3), Assumption 4.3, (2.5) and (4.2) ( ) m LM k (p k ) 2 k2 ψ + σ k κ J ( + τ max ) d(x k, S) 3. Then, using (2.3) J k p k + F k and by (2.) and (4.6) 2m LM k (p k ) k 2 ψ + 2σ k κ J ( + τ max ) d(x k, S) 3/2, m Q k (p k) k 2 ψ + 2σ k κ J ( + τ max ) d(x k, S) 3/2 + σ k Ψ 2 d(x k, S) 2. This latter inequality and (4.2) yield (4.4). Finally, we show that iteration k is very successful. Since (4.6) gives (4.5) x k + p k x x k x + p k ψ + Ψd(x k, S) ψ( + Ψ), from the definition of ψ, it follows that x k + p k B ɛ (x ). Using (4.4) the quantity F (x k + p k ) can be bounded as (4.6) (4.7) F (x k + p k ) F (x k + p k ) F k J k p k + F k + J k p k k p k 2 + m Q k (p k) Thus, condition (2.0) can be bounded below as ρ k = F (x k + p k ) m Q k (p k) F k m Q k (p k) Moreover, (4.4) and (2.4) yield κ * p k 2 F k m Q k (p k). F k m Q k (p k) F k Λd(x k, S) 3/2 ( α 3/2 Λ F k /2 ) F k, 2

and the last expression is positive if F k is small enough, i.e. if x k is sufficiently close to x. Thus, by (4.6) and (2.4) we obtain ρ k k α 2 Ψ 2 F k α 3/2 Λ F k /2, and reducing ψ, if necessary, we get ρ k η 2 for the fixed value > η 2 > 0. Regarding the ARC algorithm, we have an analogous result shown in the next lemma. Lemma 4.4. Let Assumptions 4., 4.2 and 4.3 hold and x S is a limit point of the sequence {x k } generated by the ARC method satisfying Assumption 2.. Suppose that σ max σ k σ min > 0 for all k 0. Then there exist positive scalars ψ and Λ such that, if x k B ψ (x ), (4.8) m C k (p k ) Λd(x k, S) 3, and iteration k is very successful. Proof. Let ψ = ɛ/( + Ψ) where ɛ and Ψ are the scalars in Lemma 4.2. Assume that x k B ψ (x ). By (4.3), (2.6), (4.7) and Assumption 4.3 m LM k (p k ) ( k 2 2 ψ + σ k Θ( + τ max ) ) d(x k, S) 3. Consequently, by the definition of m C k, inequality 2 J kp k + F k 2 m LM k (p k ) and (4.6) we get m C k (p k ) 2 ( k 2 ψ + σ k Θ( + τ max ) ) d(x k, S) 3 + 3 σ kψ 3 d(x k, S) 3 and (4.8) follows from (4.2). Let now focus on very successful iterations. Proceeding as in Lemma 4.3 we can derive (4.5), i.e. x k + p k B ɛ (x ). Thus, by using (4.6), (4.4) and (4.8) we have F (x k + p k ) 2 F (x k + p k ) F k J k p k 2 + F k + J k p k 2 + 2(F k + J k p k ) T (F (x k + p k ) F k J k p k ) k 2 p k 4 + 2m C k (p k ) + 2 F (x k + p k ) F k J k p k F k + J k p k k 2 Ψ 4 d(x k, S) 4 + 2m C k (p k ) + 2k Ψ 2 2Λd(x k, S) 7/2, i.e., there exists a constant Φ such that 2 F (x k + p k ) 2 m C k (p k ) Φd(x k, S) 7/2. Consequently, ρ k in (2.) can be bounded below as ρ k = 2 F (x k + p k ) 2 m C k (p k) 2 F k 2 m C k (p k) Φd(x k, S) 7/2 2 F k 2 m C k (p k), and (4.8) and (2.4) give 2 F k 2 m C k (p k ) ( ) 2 F k 2 Λd(x k, S) 3 F k 2 2 α3 Λ F k, 3

where the last expression is positive if x k is close enough to x. Thus, by (2.4) ρ k α7/2 Φ F k 3/2 2 α3 Λ F k, and reducing ψ, if necessary, we get ρ k η 2 for the fixed value > η 2 > 0. The next lemma establishes the dependence of d(x k + p k, S) upon d(x k, S) whenever x k is sufficiently close to x. Lemma 4.5. Let Assumptions 2., 4. and 4.2 hold. Suppose that p k satisfies (4.6) and λ k d(x k, S) ξ, with ξ (0, 2]. Then, if x k is close enough to x (4.9) d(x k + p k, S) Γd(x k, S) min{ξ+, 2}, for some positive Γ. Proof. The proof is given in [, Lemma 4]. With the previous results at hand, exploiting Theorem 3., we are ready to show the local convergence behaviour of both adaptive regularized approaches. Corollary 4.6. Let Assumptions 4., 4.2 and 4.3 hold and x S be a limit point of the sequence {x k } generated by the ARQ method satisfying Assumption 2.. Suppose that σ max σ k σ min > 0 for all k 0. Then, {x k } converges to x Q-quadratically. Proof. Lemmas 4.2 and 4.3 guarantee conditions (3.) and (3.2). Moreover, by Assumption 4.3, (2.5) and (4.2) λ k λ k( + τ max ) 2( + τ max )σ k F k 2( + τ max )σ max κ J d(x k, S). Therefore, by Lemma 4.5 we have that (4.9) holds with ξ = and such an inequality coincides with (3.3). Then, the proof is completed by using Theorem 3.. Corollary 4.7. Let Assumptions 4., 4.2 and 4.3 hold and x S be a limit point of the sequence {x k } generated by the ARC method satisfying Assumption 2.. Suppose that σ max σ k σ min > 0 for all k 0. Then, {x k } converges to x Q-quadratically. Proof. Lemmas 4.2 and 4.4 guarantee conditions (3.) and (3.2). Further, by Assumption 4.3, (2.6) and (4.7) λ k λ k( + τ max ) ( + τ max )σ k p k ( + τ max )σ max Θ d(x k, S). Therefore, by Lemma 4.5 the inequality (4.9) holds with ξ =, i.e. (3.3) is met. Then, the proof is completed by using Theorem 3.. The previous results have been obtained supposing that σ k [σ min, σ max ]. The lower bound σ k σ min can be straightforwardly enforced in the algorithm for some small specified threshold σ min. Concerning condition σ k σ max, we now discuss when it is satisfied. Focusing on the ARQ method, in [2, Lemma 4.7] it has been proved that (4.2) holds under the following two assumptions on J(x): i) J(x) is uniformly 4

bounded above for all k 0 and all x [x k, x k + p k ]; ii) there exist positive constants κ L, κ S such that, if x x k κ S and x [x k, x k +p k ], then J(x) J k κ L x x k for all k 0. For the ARC method, suppose that F is twice continuously differentiable and the Hessian matrix H of f is globally Lipschitz continuous in IR n, with Lipschitz constant κ H. We now show two occurrences where (4.2) holds. By [4, Lemma 5.2], (4.2) is guaranteed if (4.20) (H k J T k J k )p k C p k 2, for all k 0 and some constant C > 0. Since H k J T k J k = m i= F i(x k ) 2 F i (x k ), where F i is the i-th component of F, i m, then (4.20) is satisfied provided that F k κ F p k for some κ F > 0 and all k 0. Alternatively, (4.2) holds if x k x. In fact, f(x k + p k ) m k (p k ) = 2 pt k ( H(ζk ) J T k J k ) pk σ k 3 p k 3, for some ζ k on the line segment (x k, x k + p k ), see [4, Equation (4.2)]. Then, f(x k + p k ) m k (p k ) 2 H(ζ k) H(x k ) p k 2 + 2 (H(x k) J T k J k )p k p k σ k 3 p k 3 2 κ H p k 3 + κ H F k p k 2 σ k 3 p k 3, for some positive κ H. Thus, for all k sufficiently large (4.6) and (2.4) yield ( f(x k + p k ) m k (p k ) αψκ H + κ H αψ σ ) k α 2 Ψ 2 F k 3, 3 and for σ k > 3(αΨκ H + κ H )/(αψ) the iteration is very successful. Consequently, the updating rule (2.2) gives σ k+ σ k and (4.2) follows. Summarizing, we have shown convergence results for ARQ and ARC that are analogous to results known in literature for the Levenberg-Marquardt methods when the parameter µ k has the form (2.5). Concerning the choice of σ k and µ k, we underline that the rule for fixing σ k is simpler to implement than the rule for choosing µ k. In fact, σ k is fixed on the base of the adaptive choice in Algorithm 2. while (2.5) leaves the choice of both δ and the constant multiplying F k open and this may have an impact on the practical behaviour of the Levenberg-Marquardt methods [7, 8]. Finally, in [5] the model m C k (p) has been generalized to (4.2) m 2,β k (p) = 2 F k + J k p 2 + β σ k p β, with β [2, 3]. Trivially, m 2,β k (p) reduces to mc k (p) when β = 3. The adaptive regularized procedure based on the use of m 2,β k (p) can be analyzed using the same arguments as above. Assume the same assumptions as in Corollary 4.7. Then the adaptive procedure converges to x S superlinearly when β (2, 3). In fact, (4.6) holds and a slight modification of Lemma 4.4 shows that eventually all iterations are very successful. By [5], λ k = σ k p k β 2 and proceeding as in Corollary 4.7 λ k ( + τ max )σ max Θ β 2 d(x k, S) β 2. Thus, Lemma 4.5 implies d(x k + p k, S) Γd(x k, S) β, and a straightforward adaptation of Theorem 3. yields Q-superlinear convergence with rate β. 5

4.3. Computing the trial step. In this section we consider a viable way devised in [2, 4] for computing an approximate minimizer of m k and enforcing Assumptions 4.2 and 4.3. In such an approach a couple (p k, λ k ) satisfying (4.22) p k = p(λ k ), (J T k J k + λ k I)p k = g k, with λ k being an approximation to λ k, is sought. The scalar λ k can be obtained applying a root-finding solver to the so-called secular equation. In fact, from Lemma 2.2, the optimal scalar λ k for ARQ solves the scalar nonlinear equation (4.23) ρ(λ) = λ 2σ k J k p(λ) + F k = 0. The function ρ (λ) may change sign in (0, + ), while the reformulation (4.24) ψ Q (λ) = ρ(λ) λ = 0, is such that ψ Q (λ) is convex and strictly decreasing in (0, + ) [2]. Analogously, in ARC the scalar λ solves the scalar nonlinear equation (4.25) which can be reformulated as (4.26) ρ(λ) = λ σ k p(λ) = 0, ψ C (λ) = ρ(λ) λ p(λ) = 0. The function ψ C (λ) is convex and strictly decreasing in (0, + ) [4]. In what follows we will use the notation ψ(λ) in all the expressions that holds for both functions. Due to the monotonicity and convexity properties of ψ(λ), either the Newton or the secant method applied to (4.24) and (4.26) converges globally and monotonically to the positive root λ k for any initial guess in (0, λ k ). We refer to [2] and [4] for details on the evaluation of ψ(λ) and its first derivatives. Clearly p k satisfies Assumption 4.2, while Assumption 4.3 is met if a suitable stopping criterion is imposed to the root-finding solver. Let the initial guess be λ 0 k (0, λ k ). Then, the sequence {λl k } generated is such that λl k > λ0 k for any l > 0 and by Taylor expansion there exists λ (λ l k, λ k ) such that ψ(λ l k) = ψ ( λ)(λ l k λ k), i.e. λ k λ l k = ψ(λl k ) ψ ( λ). In principle, if the iterative process is stopped when (4.27) ψ(λ l k) < τ k λ l kψ ( λ), then the couple (p(λ l k ), λl k ) satisfies Assumptions 4.2, 4.3. A practical implementation of this stopping criterion can be carried out by using an upper bound λ U on λ. Since ψ(λ) is convex it follows that ψ (λ) is strictly increasing, ψ ( λ) ψ ( λ U ) and (4.27) is guaranteed by enforcing ψ(λ l k) < τ k λ l kψ ( λ U ). Possible choices for λ U follows from using the condition λ λ k. In particular, in ARQ method Lemma 2.2 suggests λ U = 2σ k F k. In ARC method, Lemma 2.2 and the monotonic decrease of p(λ) in (0, ), shown in Lemma 2.3, yield λ U = σ k p(λ l k ). Finally we note that if the bisection process is used to get an initial guess for the Newton process, then λ U can be taken as the right extreme of the last bracketing interval computed. 6

5. Computing the trial step in a subspace. The strategy for computing the trial step devised in the previous section requires the solution of a sequence of linear systems of the form (4.22). Namely, for each value of λ l k generated by the root-finding solver applied to the secular equation, the computation of ψ(λ l k ) is needed and this calls for the solution of (4.22). A different approach can be used when large scale problems are solved and the factorization of coefficient matrix of (4.22) is unavailable due to cost or memory limitations. In such an approach the model m k is minimized over a sequence of nested subspaces. The Golub and Kahan bi-diagonalization process [5] is used to generate such subspaces and minimizing the model in the subspaces is quite inexpensive. The minimization process is carried out until a step p s k satisfying (5.) p m k (p s k) ω k, is computed for some positive tolerance ω k. We now study the effect of using the step p k = p s k in Algorithm 2.. At termination of the iterative process the couple (λ k, p s k ) satisfies: (5.2) (5.3) (Jk T J k + λ k I)p s k + g k = r k λ k φ(p s k) = 0 where φ(p s k) = { 2σk F k + J k p s k when m k = m Q k σ k p s k when m k = m C k. From condition (5.) and the form of p m k it follows that (5.4) r k ω k, where (5.5) ω k = { ωk F k + J k p s k when m k = m Q k ω k when m k = m C k. Since p s k does not satisfy Assumption 4.2 some properties of the approximate minimizer p k discussed in Section 4.3 are not shared by p s k and we cannot rely on the convergence theory developed in the previous section. The relation between p s k and p(λ k) follows from noting that (5.2) yields i.e., (J T k J k + λ k I)p s k = (J T k J k + λ k I)p(λ k ) + r k, (5.6) p s k p(λ k ) = (J T k J k + λ k I) r k. On the other hand, Assumption 4.3 can be satisfied choosing the tolerance ω k in (5.) small enough. In this section we show when the strong local convergence behaviour of the ARQ and ARC procedures is retained if p s k is used. We start studying the quadratic model and the step generated by the ARQ method. 7

Lemma 5.. Let {x k } be the sequence generated by the ARQ method with steps p k = p s k satisfying (5.) (5.3). Suppose that Assumptions 4. and 4.3 hold and x S is a limit point of {x k }. If σ k σ min > 0 for all k 0, and (5.7) ω k θ F k 3/2, for a positive θ, then there exists a positive Ψ such that, (5.8) p s k Ψd(x k, S), whenever x k is sufficiently close to x. Moreover, if (4.2) holds then there exists a positive Λ > 0 such that (5.9) m Q k (ps k) Λd(x k, S) 3/2, whenever x k is sufficiently close to x. Proof. Let x k B ɛ (x ). By (5.4), (5.5) and (5.3) we get and using (5.7) and (4.2) (J T k J k + λ k I) r k λ k ω k 2σ min ω k, (5.0) (Jk T J k + λ k I) r k θκ3/2 J d(x k, S) 3/2. 2σ min Since p(λ k ) satisfies (4.6) by Lemma 4.2, and (5.6) gives p s k p(λ k) + (J T k J k + λ k I) r k, inequality (5.8) holds. Let us now focus on inequality (5.9). Assuming x k B ψ (x ) where ψ is the scalar in Lemma 4.3, and using (4.), (4.4) and (4.2) we obtain and by (5.6) and (5.0) m Q k (ps k) F k + J k p(λ k ) + J k (p s k p(λ k )) + σ k p s k 2 Λd(x k, S) 3/2 + κ J p s k p(λ k ) + σ max Ψ2 d(x k, S) 2, m Q k (ps k) Λd(x k, S) 3/2 + θκ5/2 J d(x k, S) 3/2 + σ 2σ Ψ2 max d(x k, S) 2, min which completes the proof. The following Lemma shows the corresponding result for the ARC method. Lemma 5.2. Let {x k } be the sequence generated by the ARC method with steps p k = p s k satisfying (5.) (5.3). Suppose that Assumptions 4. and 4.3 hold and x S is a limit point of the sequence {x k }. If σ k σ min > 0 for all k 0, and (5.) ω k θ k F k 3/2, θ k = κ θ min(, p s k ), for a positive scalar κ θ, then there exists a positive Ψ such that (5.8) holds for x k sufficiently close to x. Moreover, if (4.2) holds then there exists Λ > 0 such that (5.2) m C k (p s k) Λd(x k, S) 3, whenever x k is sufficiently close to x. 8

(5.3) Proof. Let x k B ɛ (x ). First, note that by (4.2), (5.3), (5.4), (5.5) and (5.) (Jk T J k + λ k I) r k ω k λ k σ k p s k ω k, κ θκ 3/2 J d(x k, S) 3/2. σ min Since p(λ k ) satisfies (4.6) by Lemma 4.2, and (5.6) gives p s k p(λ k) + (J T k J k + λ k I) r k, (5.8) holds. Moreover, letting x k S ψ (x ), and using (4.), (4.8), (5.6) and (5.3) we get m C k (p s k) 2 F k + J k p(λ k ) 2 + 2 J k(p s k p(λ k )) 2 +(F k + J k p(λ k )) T J k (p s k p(λ k )) + 3 σ k p s k 3 Λd(x k, S) 3 + κ2 θ κ5 J 2σmin 2 d(x k, S) 3 + κ θκ 5/2 J 2Λ d(x k, S) 3 + σ min 3 σ Ψ max 3 d(x k, S) 3, and this yields (5.2) We are now able to state the local convergence results for ARQ and ARC methods in the case where the step is computed in a subspace. Corollary 5.3. Let {x k } be the sequence generated by the ARQ method with steps p k = p s k satisfying (5.) (5.3). Suppose Assumptions 4. and 4.3 hold and that x S is a limit point of {x k } satisfying Assumption 2.. If 0 < σ min σ k σ max for all k 0 and (5.7) holds, then {x k } converges to x Q-quadratically. Proof. Using (5.8) and (5.9) and proceeding as in the proof of Lemma 4.3, we can prove that iteration k is very successful whenever x k is sufficiently close to x. Moreover, as (5.) and (5.7) are satisfied, Lemma 4 in [] guarantees that inequality (3.3) holds and Theorem 3. yields the hypothesis. Corollary 5.4. Let {x k } be the sequence generated by the ARC method with steps p k = p s k satisfying (5.) (5.3). Assume that Assumptions 4. and 4.3 hold and x S is a limit point of {x k } satisfying Assumption 2.. If 0 < σ min σ k σ max for all k 0 and (5.) holds, then {x k } converges to x Q-quadratically. Proof. Using (5.8) and (5.2) and proceeding as in the proof of Lemma 4.4, we can prove that iteration k is very successful whenever x k is sufficiently close to x. Moreover, as (5.) and (5.) are satisfied, Lemma 4 in [] guarantees that inequality (3.3) holds and Theorem 3. yields the hypothesis. 6. Conclusion. In this paper, we have studied the local convergence behaviour of two adaptive regularized methods for solving nonlinear least-squares problems and we have established local quadratic convergence to zero-residual solutions under an error bound assumption. Interestingly, this condition is considerably weaker than the standard assumptions used in literature and the results obtained are valid for under and over-determined problems as well as for square problems. The theoretical analysis carried out shows that the regularizations enhance the properties of the underlying unregularized methods. The focus on zero-residual solutions is a straightforward consequence of models considered which are regularized Gauss-Newton models. Further results on potentially rank-deficient nonliner leastsquares have been given in [8] for adaptive cubic regularized models employing suitable approximations of the Hessian of the objective function. 9

REFERENCES [] R. Behling and A. Fischer. A unified local convergence analysis of inexact constrained Levenberg-Marquardt methods. Optimization Letters, 6, 927 940, 202. [2] S. Bellavia, C. Cartis, N. I. M. Gould, B. Morini, and Ph. L. Toint. Convergence of a Regularized Euclidean Residual Algorithm for Nonlinear Least-Squares. SIAM Journal on Numerical Analysis, 48, 29, 200. [3] H. Y. Benson and D. F. Shanno. Interior-Point methods for nonconvex nonlinear programming: cubic regularization. Technical report, 202, http://www.optimizationonline.org/db FILE/202/02/3362.pdf. [4] C. Cartis, N. I. M. Gould, and Ph. L. Toint. Adaptive cubic overestimation methods for unconstrained optimization. Part I: motivation, convergence and numerical results. Mathematical Programming A, 27, 245 295, 20. [5] C. Cartis, N. I. M. Gould, and Ph. L. Toint. Trust-region and other regularisations of linear least-squares problems. BIT Numerical Mathematics, 49, 2-53, 2009 [6] C. Cartis, N. Gould and Ph. L. Toint. Adaptive cubic overestimation methods for unconstrained optimization. Part II: worst-case function-evaluation complexity. Mathematical Programming A, 30, 295-39, 20. [7] C. Cartis, N. Gould and Ph. L. Toint. An adaptive cubic regularization algorithm for nonconvex optimization with convex constraints and its function-evaluation complexity. IMA Journal on Numerical Analysis, 4, 662 695, 202. [8] C. Cartis, N. I. M. Gould, and Ph. L. Toint. On the evaluation complexity of cubic regularization methods for potentially rank-deficient nonlinear least-squares problems and its relevance to constrained nonlinear optimization. SIAM Journal on Optimization, 23, pp. 553 574, 203. [9] A. R. Conn, N. I. M. Gould, and Ph. L. Toint. Trust-Region Methods. No. in the MPS SIAM series on optimization. SIAM, Philadelphia, USA, 2000. [0] H. Dan, N. Yamashita and M. Fukushima. Convergence properties of the inexact Levenberg- Marquardt method under local error bound conditions. Optimization Methods and Software, 7, 605-626, 2002. [] J.E. Dennis, R.B. Schnabel, Numerical methods for unconstrained optimization and nonlinear equations Prentice Hall, Englewood Cliffs, NJ, 983. [2] J. Fan and J. Pan. Inexact Levenberg-Marquardt method for nonlinear equations. Discrete and Continuous Dynamical Systems, Series B, 4, 223 232, 2004. [3] J. Fan, Y. Yuan. On the quadratic convergence of the Levenberg-Marquardt method without nonsingularity assumption Computing, 74, 23 39, 2005. [4] N. I. M. Gould, M. Porcelli, and Ph. L. Toint. Updating the regularization parameter in the adaptive cubic regularization algorithm. Computational Optimization and Applications, 53, 22, 202. [5] A. Griewank. The modification of Newton s method for unconstrained optimization by bounding cubic terms. Technical Report NA/2 (98), Department of Applied Mathematics and Theoretical Physics, University of Cambridge, United Kingdom, 98. [6] W. W. Hager and H. Zhang, Self-adaptive inexact proximal point methods. Computational Optimization and Applications, 39, 6 8, 2008. [7] C. Kanzow, N. Yamashita, and M. Fukushima. Levenberg-Marquardt methods with strong local convergence properties for solving nonlinear equations with convex constraints, Journal of Computational and Applied Mathematics, 72, 375 397, 2004. [8] M. Macconi, B. Morini and M. Porcelli, Trust-region quadratic methods for nonlinear systems of mixed equalities and inequalities. Applied Numerical Mathematics, 59, 859 876, 2009. [9] J.J. Moré. The Levenberg-Marquardt algorithm: implementation and theory. Lecture Notes in Mathematics, 630, 05-6, Springer, Berlin, 978. [20] Yu. Nesterov. Modified Gauss-Newton scheme with worst-case guarantees for global performance. Optimization Methods and Software, 22, 469 483, 2007. [2] Yu. Nesterov and B. T. Polyak. Cubic regularization of Newton s method and its global performance. Mathematical Programming, 08, 77 205, 2006. [22] J.S. Pang, Error bounds in mathematical programming Mathematical Programming, 79, 299 332, 997. [23] Ph. L. Toint. Nonlinear Stepsize Control, Trust Regions and Regularizations for Unconstrained Optimization. Optimization Methods and Software, 28, 82 95, 203. [24] P. Tseng. Error bounds and superlinear convergence analysis of some Newton-type methods in optimization. in Nonlinear Optimization and Applications, 2, eds. G. Di Pillo and F. Giannessi, Kluwer Academic Publishers, Dordrecht Appl. Optim. 36, 445 462, 2000. 20

[25] M. Weiser, P. Deuflhard, and B. Erdmann. Affine conjugate adaptive Newton methods for nonlinear elastomechanics. Optimization Methods and Software, 22, 43 43, 2007. [26] N. Yamashita and M. Fukushima. On the rate of convergence of the Levenberg-Marquardt method. Computing Supplementa, 5, 237 249, 200. [27] H. Zhang and A. R. Conn. On the local convergence of a derivative-free algorithm for leastsquares minimization. Computational Optimization and Applications, 5, 48 507,202. [28] D. Zhu. Affine scaling interior Levenberg-Marquardt method for bound-constrained semismooth equations under local error bound conditions. Journal of Computational and Applied Mathematics, 29, 98 26, 2008. 2