PRIMAL DUAL METHODS FOR NONLINEAR CONSTRAINED OPTIMIZATION

Similar documents
On the Local Quadratic Convergence of the Primal-Dual Augmented Lagrangian Method

Primal-Dual Exterior Point Method for Convex Optimization

Primal-Dual Lagrangian Transformation method for Convex

update (PDNRD) for convex optimization. We proved the global convergence, established 1.5-Q-superlinear rate of convergence

LAGRANGIAN TRANSFORMATION IN CONVEX OPTIMIZATION

Support Vector Machine via Nonlinear Rescaling Method


10 Numerical methods for constrained problems

Primal-Dual Nonlinear Rescaling Method for Convex Optimization 1

Optimisation in Higher Dimensions

12. Interior-point methods

Lagrangian Transformation and Interior Ellipsoid Methods in Convex Optimization

Lecture 3: Lagrangian duality and algorithms for the Lagrangian dual problem

Primal-Dual Nonlinear Rescaling Method for Convex Optimization 1

A null-space primal-dual interior-point algorithm for nonlinear optimization with nice convergence properties

A SHIFTED PRIMAL-DUAL PENALTY-BARRIER METHOD FOR NONLINEAR OPTIMIZATION

Algorithms for Constrained Optimization

MS&E 318 (CME 338) Large-Scale Numerical Optimization

Infeasibility Detection and an Inexact Active-Set Method for Large-Scale Nonlinear Optimization

Convex Optimization. Dani Yogatama. School of Computer Science, Carnegie Mellon University, Pittsburgh, PA, USA. February 12, 2014

Nonsymmetric potential-reduction methods for general cones

Constrained Optimization and Lagrangian Duality

Some Properties of the Augmented Lagrangian in Cone Constrained Optimization

5. Duality. Lagrangian

4TE3/6TE3. Algorithms for. Continuous Optimization

A PRIMAL-DUAL TRUST REGION ALGORITHM FOR NONLINEAR OPTIMIZATION

Primal-dual Subgradient Method for Convex Problems with Functional Constraints

Lecture 6: Conic Optimization September 8

CONVERGENCE ANALYSIS OF AN INTERIOR-POINT METHOD FOR NONCONVEX NONLINEAR PROGRAMMING

5 Handling Constraints

Lecture 18: Optimization Programming

A Second-Order Path-Following Algorithm for Unconstrained Convex Optimization

On well definedness of the Central Path

Lecture 13: Constrained optimization

Inexact Newton Methods and Nonlinear Constrained Optimization

Algorithms for constrained local optimization

LECTURE 25: REVIEW/EPILOGUE LECTURE OUTLINE

Primal-dual relationship between Levenberg-Marquardt and central trajectories for linearly constrained convex optimization

An Inexact Newton Method for Optimization

Algorithms for nonlinear programming problems II

Introduction to Nonlinear Stochastic Programming

Interior Point Methods. We ll discuss linear programming first, followed by three nonlinear problems. Algorithms for Linear Programming Problems

Convex Optimization M2

CS711008Z Algorithm Design and Analysis

E5295/5B5749 Convex optimization with engineering applications. Lecture 8. Smooth convex unconstrained and equality-constrained minimization

Convergence analysis of a primal-dual interior-point method for nonlinear programming

An Inexact Newton Method for Nonlinear Constrained Optimization

A FULL-NEWTON STEP INFEASIBLE-INTERIOR-POINT ALGORITHM COMPLEMENTARITY PROBLEMS

Support Vector Machines: Maximum Margin Classifiers

Convex Optimization & Lagrange Duality

Convex Optimization Boyd & Vandenberghe. 5. Duality

An Infeasible Interior-Point Algorithm with full-newton Step for Linear Optimization

Interior-Point Methods

Algorithms for nonlinear programming problems II

An Inexact Sequential Quadratic Optimization Method for Nonlinear Optimization

Lectures 9 and 10: Constrained optimization problems and their optimality conditions

Interior Point Algorithms for Constrained Convex Optimization

Convex Optimization and SVM

Enhanced Fritz John Optimality Conditions and Sensitivity Analysis

Lecture: Duality.

ICS-E4030 Kernel Methods in Machine Learning

CS-E4830 Kernel Methods in Machine Learning

SF2822 Applied Nonlinear Optimization. Preparatory question. Lecture 9: Sequential quadratic programming. Anders Forsgren

A new primal-dual path-following method for convex quadratic programming

Interior-Point Methods for Linear Optimization

Applications of Linear Programming

A Unified Analysis of Nonconvex Optimization Duality and Penalty Methods with General Augmenting Functions

arxiv: v1 [math.na] 8 Jun 2018

Lecture 3. Optimization Problems and Iterative Algorithms

Cubic regularization of Newton s method for convex problems with constraints

Convex Optimization. Newton s method. ENSAE: Optimisation 1/44

4TE3/6TE3. Algorithms for. Continuous Optimization

Generalization to inequality constrained problem. Maximize

LP. Kap. 17: Interior-point methods

LINEAR AND NONLINEAR PROGRAMMING

12. Interior-point methods

minimize x subject to (x 2)(x 4) u,

Derivative-Based Numerical Method for Penalty-Barrier Nonlinear Programming

A Regularized Interior-Point Method for Constrained Nonlinear Least Squares

A SHIFTED PRIMAL-DUAL INTERIOR METHOD FOR NONLINEAR OPTIMIZATION

Barrier Method. Javier Peña Convex Optimization /36-725

Lagrange Relaxation and Duality

Lecture: Duality of LP, SOCP and SDP

Lagrange Duality. Daniel P. Palomar. Hong Kong University of Science and Technology (HKUST)

Written Examination

Lecture 9 Sequential unconstrained minimization

Optimization for Machine Learning

A Trust-region-based Sequential Quadratic Programming Algorithm

Lecture 7: Convex Optimizations

14. Duality. ˆ Upper and lower bounds. ˆ General duality. ˆ Constraint qualifications. ˆ Counterexample. ˆ Complementary slackness.

Part 5: Penalty and augmented Lagrangian methods for equality constrained optimization. Nick Gould (RAL)

AM 205: lecture 19. Last time: Conditions for optimality, Newton s method for optimization Today: survey of optimization methods

I.3. LMI DUALITY. Didier HENRION EECI Graduate School on Control Supélec - Spring 2010

Primal-Dual Interior-Point Methods for Linear Programming based on Newton s Method

Duality. Lagrange dual problem weak and strong duality optimality conditions perturbation and sensitivity analysis generalized inequalities

NONLINEAR. (Hillier & Lieberman Introduction to Operations Research, 8 th edition)

Stochastic Optimization with Inequality Constraints Using Simultaneous Perturbations and Penalty Functions

Full Newton step polynomial time methods for LO based on locally self concordant barrier functions

Lecture 16: October 22

1. Gradient method. gradient method, first-order methods. quadratic bounds on convex functions. analysis of gradient method

Transcription:

PRIMAL DUAL METHODS FOR NONLINEAR CONSTRAINED OPTIMIZATION IGOR GRIVA Department of Mathematical Sciences, George Mason University, Fairfax, Virginia ROMAN A. POLYAK Department of SEOR and Mathematical Sciences, George Mason University, Fairfax, Virginia HISTORICAL NOTE In 1797 in his book Theorie des functions analytiques Joseph Luis Lagrange (1736 1813) introduced the Lagrangian and Lagrange multipliers rule for solving optimization problems with equality constraints.... If a function of several variables should be a maximum or minimum and there are between these variables one or several equations, then it will be suffice to add to the proposed function the functions that should be zero, each multiplied by an undetermined quantity, and then to look for the maximum and the minimum as if the variables were independent; the equation that one will find combined with the given equations, will serve to determine all the unknowns. J.-L. Lagrange The Lagrange multipliers rule is only a necessary but not sufficient condition for an equality constrained optimum. The primal dual (PD) vector is neither a maximum nor a minimum of the Lagrangian: it is a saddle point. Nevertheless, for more than 200 years the Lagrangian has remained an invaluable tool for optimization. Moreover, with time, the great value of the Lagrangian has become increasingly evident. The Lagrangian is the main instrument for establishing optimality conditions, the basic ingredient in the Lagrange duality, and one of the most important tools in numerical constrained optimization. Any PD method for solving constrained optimization problems in one way or another uses the Lagrangian and the Lagrange multipliers. OPTIMIZATION WITH EQUALITY CONSTRAINTS: LAGRANGE SYSTEM OF EQUATIONS Consider q + 1 smooth enough functions f, g j : IR n IR, j = 1,..., q and the feasible set = { x : g j (x) = 0, j = 1,..., q }. The equality constrained optimization (ECO) problem consists of finding (ECO) f (x ) = min{f (x) x }. The Lagrangian L :IR n IR q IR 1 for (ECO) is given by the formula L(x, v) = f (x) v j g j (x). ( Let us consider the vector function g T (x) = g1 (x),..., g q (x) ), the Jacobian J(g(x)) = g(x) = ( g 1 (x),..., g q (x) ) T and assume rank g(x ) = q < n. (1) Then for x to be a (ECO) solution, it is necessary the existence of v IR q that the pair (x, v ) is a solution to the following Lagrange system of equations: x L(x, v) = f (x) g T (x)v = 0, (2) g i (x) = 0, i = 1,..., q. (3) We consider the Hessian 2 xx L(x, v) = 2 f (x) v i 2 g i (x) Wiley Encyclopedia of Operations Research and Management Science, edited by James J. Cochran Copyright 2010 John Wiley & Sons, Inc. 1

2 PRIMAL DUAL METHODS FOR NONLINEAR CONSTRAINED OPTIMIZATION of the Lagrangian L(x, v). The regularity condition (1), together with sufficient condition for the minimizer x to be isolated that is xx 2 L(x, v )ξ, ξ > m ξ, ξ, ξ : g(x ) ξ = 0, m > 0, (4) comprise the standard second-order optimality conditions for the (ECO) problem. Application of Newton s method to the nonlinear PD system of Equations (2), (3) leads to one of the first PD methods for constrained optimization [1]. By linearizing the system (2), (3) and ignoring terms of the second and higher order, one obtains the following linear PD system for finding the Newton direction ( x, v): [ 2 xx L( ) g T ( ) g( ) 0 ][ x v ] [ x L( ) = g( ) ]. (5) Let S(y, δ) ={y = (x, v) : y y δ}, where is the Eucledian norm, δ small enough, and 0 <ɛ δ is the desired accuracy. The following merit function ν(y) = ν(x, v) { } = max x L(x, v), max g i(x), (6) 1 i q measures the distance between the PD approximation (x, v) and the solution (x, v ). If f and g i are smooth enough and the second-order optimality conditions for (ECO) are satisfied, then for y S(y, δ) we have ν(y) = 0 y = y, (7) and there are 0 < m 0 < M 0 that m 0 y y ν(y) M 0 y y. (8) The PD method for (ECO) consists of the following operations: PDECOM-Primal Dual Equality Constrained Optimization Method Step 0. Let y S(y, δ). Step 1. If ν(y) ɛ, Outputy as the solution. Step 2. Find x and v from (5). Step 3. Update the primal dual pair by the following formulas x := x + x, v := v + v. Step 4. Goto Step 1. If the standard second-order optimality for conditions for (ECO) are satisfied, the Lipschitz conditions for the Hessians 2 f (x), 2 g i, i = 1,..., q hold, y = (x, v) S(y, δ)and δ > 0 is small enough, then the PDECOM generates the PD sequence that converges to the PD solution y = (x, v )with quadratic rate; that is, the following bound holds: AUGMENTED LAGRANGIAN AND PRIMAL DUAL AUGMENTED LAGRANGIAN METHOD For a given k > 0, the augmented Lagrangian (AL) L :IR n IR q IR + IR 1 is defined by the following formula [2,3]: ŷ y C y y 2, (9) where C > 0 is independent from y S(y, δ) and depends only on the problem data [1]. L(x, v, k) = f (x) v j g j (x) + k 2 g 2 j (x). (10)

PRIMAL DUAL METHODS FOR NONLINEAR CONSTRAINED OPTIMIZATION 3 The AL method alternates the unconstrained minimization of the AL L(x, v, k) in the primal space with a Lagrange multipliers update. Let δ > 0 be sufficiently small, 0 <ɛ δ be the desired accuracy, k > 0 be sufficiently large, and y = (x, v) S(y, δ). One step of the classical AL method consists of finding the primal minimizer ˆx = arg min L(x, v, k), (11) n x IR followed by the Lagrange multipliers update. In other words, for a given scaling parameter k > 0 and a starting point y = (x, v) S(y, δ), one step of the AL method is equivalent to solving the following nonlinear PD system for ˆx and ˆv: x L(ˆx, ˆv) = f (ˆx) ˆv j g j (ˆx) = 0, (12) ˆv v + kg(ˆx) = 0. (13) Application of Newton s method for solving the PD system (12) (13) for (ˆx, ˆv) together with a proper update of the penalty parameter k > 0 leads to the primal dual augmented Lagrangian (PDAL) method for solving (ECO) problems [4]. By linearizing the system (12) (13) and ignoring terms of the second and higher order, we obtain the following linear PD system for finding the Newton direction ( x, v): [ 2 xx L( ) g T ][ ] ( ) x g( ) = 1 k I q v [ x L( ) g( ) ], (14) where I q is the identity matrices in IR q. Note that if k, then the system (14) gets close to the system (5). Therefore, by changing k > 0 properly it is possible to achieve a quadratic rate of convergence for the PDAL method. PDALM - Primal Dual Augmented Lagrangian Method Step 0. Let y S(y, δ) andk > 0 is large enough. Step 1. If ν(y) ɛ, Outputy as a solution. Step 2. Set k = ν(x, v) 1. Step 3. Find x and v from (14). Step 4. Update the primal dual pair by the formulas x := x + x, v := v + v. Step 5. Goto Step 1. If the standard second-order optimality conditions for (ECO) are satisfied, the Lipschitz conditions for the Hessians 2 f (x), 2 g i, i = 1,..., q hold, y = (x, v) S(y, δ), and δ > 0 is sufficiently small, then the PDALM generates a PD sequence that converges to the PD solution y = (x, v ) with a quadratic rate, that is Equation (9) holds [4]. OPTIMIZATION WITH INEQUALITY CONSTRAINTS Let f :IR n IR 1 be convex and all c i :IR n IR 1, i = 1,..., p be concave and smooth functions. Consider a convex set ={x IR n : c i (x) 0, i = 1,..., p} and the following convex inequality constrained optimization (ICO) problem: (ICO) Let us assume that f (x ) = min{f (x) x }. (a) The primal optimal set X is not empty and bounded. (b) The Slater s condition holds; that is, there exists ˆx IR n : c i (ˆx) > 0, i = 1,..., p.

4 PRIMAL DUAL METHODS FOR NONLINEAR CONSTRAINED OPTIMIZATION DuetotheassumptionB, thekarush Ku hn Tucker s (KKT s) conditions hold true; that is, there exists a vector u = (u 1,..., u p ) IRp + such that x L(x, u ) = f (x ) u i c i(x ) = 0, (15) c i (x ) 0, u i 0, i = 1,..., p, (16) and the complementary slackness conditions u i c i(x ) = 0, i = 1,..., p (17) hold true. Let I ={i : c i (x ) = 0} ={1,..., r} be the set of indices of the active at x constraints, c T (r) (x) = ( c 1 (x),..., c r (x) ) be the vector function of the active at x constraints and c (r) (x) = J ( c (r) (x) ) be the corresponding Jacobian. The standard second-order optimality conditions for (ICO) consists of existence u IR p + and m > 0 such that and rank ( c (r) (x ) ) = r, u i > 0, i = 1,..., r (18) xx 2 L(x, u )ξ, ξ m ξ, ξ, ξ : c (r) (x )ξ = 0. (19) The dual to (ICO) problem consists of finding (D) d(u ) = max { d(u) u IR p +, }, where d(u) = inf x IR n L(x, u) is the dual function. For convex (ICO) that satisfies the Slater condition we have f (x ) = d(u ). The PD methods generate PD sequences {x s, u s } s=1 that f (x ) = lim s f (x s ) = lim s d(u s ) = d(u ). LOG-BARRIER FUNCTION AND INTERIOR-POINT METHODS The (ECO) problem is not combinatorial by nature because all constraints are active at the solution (by definition, active constraints are those that are satisfied as equalities). At the same time, the methods for (ECO) require an initial approximation from the neighborhood S(y, δ). The (ICO) in general and convex (ICO) in particular, are combinatorial by nature because the set of active at the solution constraints is unknown apriori. On the other hand, for convex (ICO) there is no need to have an initial approximation y S(y, δ). Consider PD interior-point methods for convex optimization that are based on the classical log-barrier function β :int IR + IR defined by the formula β(x, μ) = f (x) μ ln c i (x). Let ln t = for t 0; then for any given μ > 0 the Frisch log-barrier function F :IR n IR + IR is defined as follows: { β(x, μ), x int ; F(x, μ) =, x / int. The classical sequential unconstrained minimization technique (SUMT) [5] finds the PD trajectory {y(μ) = (x(μ), u(μ))} by the following formulas: and x F(x(μ), μ) = f (x(μ)) μ ( c i (x(μ)) ) 1 ci (x(μ)) = 0 (20) ( u(μ) = u i (μ)) = μ ( c i (x(μ)) ) 1, ) i = 1,..., p. (21) Equation (20) is the optimality condition for the unconstrained problem x(μ) = arg min { F(x, μ) x IR n}.

PRIMAL DUAL METHODS FOR NONLINEAR CONSTRAINED OPTIMIZATION 5 Thus, for a given μ > 0, finding the PD approximation from Equations (20), (21) is equivalent to solving the following system for (ˆx,û): x L(ˆx,û) = f (ˆx) û i c i (ˆx) = 0, û i c i (ˆx) = μ, i = 1,..., p. (22) It follows from Equation (20) that x(μ) int. It follows from Equation (21) that u(μ) IR p ++, that is x(μ) and u(μ) are primal and dual interior points. It follows from Equations (20), (21) that or x F(x(μ), μ) = x L(x(μ), u(μ)) = 0, d(u(μ)) = min L(x, u(μ)) = L(x(μ), u(μ)) n x IR and the PD gap (i.e., the difference between the values of the primal and dual objective functions) is that is (μ) = f (x(μ)) d(u(μ)) = u i (μ)c i (x(μ)) = pμ, lim x(μ) = μ 0 x and lim u(μ) = u. μ 0 (23) Finding (ˆx, û) = (x(μ), u(μ)) or its approximation from Equation (22) and reducing μ > 0 is the main idea of the SUMT [5]. Solving the nonlinear PD system (22) is an infinite procedure in general. The main idea of the PD interior-point methods is to replace the nonlinear PD system (22) by one Newton step toward solution of the system (22) and follow that by the barrier parameter update. For a given approximation y = (x, u) andthe barrier parameter μ > 0, the application of Newton s method to the nonlinear PD system (22) leads to the following linear PD system: [ 2 xx L(x, u) c(x) T ][ x ] U c(x) u C(x) [ = L(x, u) Uc(x) + μe ], (24) where C(x) = diag ( c i (x) ) p, U = diag (u i) p, and e = (1,...,1) T IR p. The system (24) finds the Newton direction y = ( x, u), which is used to update the current approximation y = (x, u) : x = x + α x; u = u + α u. (25) One must determine the step length α > 0 in such a way that a new approximation y = (x, u) not only remains primal and dual interior, but also remains in the area for which Newton s method for PD system (22) is well defined for the updated barrier parameter [6]. For some classes of constrained optimization problems, including linear programming problems (LPs) and quadratic programming problems (QPs), it is possible to take α = 1 in Equation (25) and update the barrier parameter μ by the formula μ = μ(1 ρ/ n), where 0 <ρ<1 is independent on n. The new approximation belongs to the neighborhood of the solution of the system (22) with μ replaced by μ. Moreover, each step reduces the PD gap by the same factor (1 ρ/ n). This leads to polynomial complexity of the PD interior-point methods for LP and QP [6]. LP, QP, and QP with quadratic constraints problems are well structured, which means that constraints and epigraph of the objective function can be equipped with a self-concordant (SC) barrier [7,8]. If (ICO) problem is not well structured, establishing polynomial complexity of the path-following methods becomes problematic, if not impossible. Nevertheless, the PD interior-point approach remains productive and leads to globally convergent and numerically efficient PD methods [9 13]. TherearetwomainclassesofPDmethods for (ICO): interior-point PD method based on the path-following idea (which goes back to SUMT) and exterior-point PD methods (which are based on nonlinear rescaling (NR) theory [14 16]).

6 PRIMAL DUAL METHODS FOR NONLINEAR CONSTRAINED OPTIMIZATION PRIMAL DUAL INTERIOR-POINT METHODS Consider the following problem min f (x), s.t. c(x) w = 0, w 0, (26) which is equivalent to (ICO) problem. The log-barrier function F(x, w, μ) = f (x) μ p ln w i is used to handle the nonnegativity of the slack vector w IR p +, which guarantees the primal feasibility. One step of the path-following method consists of solving the following (ECO) problem: min F(x, w, μ), s.t. c(x) w = 0, (27) followed by the barrier parameter μ > 0 update. The Lagrangian for the problem (27) is defined by formula L(x, w, u) = f (x) μ log w i u i (c i (x) w i ). Let W = diag (w i ) p and U = diag (u i) p. The following Lagrange system of equations corresponds to (ECO) (27). f (x) c(x) T u = 0 μe + WUe = 0 (28) c(x) w = 0 Application of Newton s method to nonlinear PD system (28) leads to the following linear PD system for finding the Newton directions: xx 2 L(x, u) 0 c(x)t 0 U W c(x) I p 0 = x w u f (x) + c(x) T u μe WUe c(x) + w, (29) where xx 2 L(x, u) = 2 f (x) p u i 2 c i (x) is the Hessian in x of the Lagrangian for (ICO) problem. For convex (ICO) the merit function ν(y) ν(x, w, u) = max { x L(x, w, u), c(x) w, WUe, } is used. Under the standard second-order optimality conditions, the merit function ν(y) satisfies Equations (7) and (8). Take δ > 0 to be sufficiently small and 0 < ɛ δ to be the defined accuracy. Consider the following PD interior-point method: PDIPM - Primal Dual Interior-Point Method Step 0. Let y S(y, δ) ={y : y y δ} be the initial approximation. Step 1. If ν(x, w, u) ɛ, Outputy = (x, u) asasolution. Step 2. Calculate the barrier parameter μ = min { θμ, ν 2 (x, w, u) },0<θ<1. Step 3. Find x, w, and u from (29). Step 4. Calculate the parameter κ and the step lengths α P and α D by formulas κ = max { κ,1 ν(x, w, u) }, 0 < κ<1, { } α P = min 1; κ (ws ) i 1 i m ( w s :( w s ) i < 0, ) i α D = min 1 i m { 1; κ (us ) i ( u s ) i :( u s ) i < 0 }.

PRIMAL DUAL METHODS FOR NONLINEAR CONSTRAINED OPTIMIZATION 7 Step 5. Update the primal dual pair by the formulas Step 6. Goto Step 1. ˆx := x + α P x, ŵ := w + α P w, û := u + α D u. If the standard second-order optimality conditions are satisfied for (ICO), the Hessians 2 f (x) and 2 c i (x), i = 1,..., p are Lipschitz continuous, δ > 0 is sufficiently small, and a starting point y S(y, δ), then the PD sequence generated by the PDIPM converges to the KKT s point with a quadratic rate [10,17]. PRIMAL DUAL NONLINEAR RESCALING METHOD Consider a class of strictly concave and twice continuously differential functions ψ :(t 0, t 1 ) IR, < t 0 < 0 < t 1 < that satisfy the following properties: 1 0.ψ(0) = 0. 2 0.ψ (t) > 0. 3 0.ψ (0) = 1. 4 0.ψ (t) < 0. 5 0.there is a > 0thatψ(t) at 2, t 0. 6 0.a)ψ (t) b 1 t 1, b) ψ (t) b 2 t 2, t > 0, b 1 > 0, b 2 > 0. The following transformations ψ satisfy the above properties: 1. Exponential transformation [18] ψ 1 (t) = 1 e t. 2. Logarithmic MBF [14] ψ 2 (t) = ln(t + 1). 3. Hyperbolic MBF [14] ψ 3 (t) = t 1 + t. Each of the above transformations can be modified in the following way. For a given 1 <τ <0 we define quadratic extrapolation of the transformations 1 3 by the formulas 4. { ψi (t), t τ, ψ qi (t) = q i (t) = a i t 2 + b i t + c i, t τ, where a i, b i, c i one finds from the following equations: ψ i (τ) = q i (τ), ψ i (τ) = q i (τ), ψ i (τ) = q i (τ). Modification 4 leads to transformations that are defined on (, ) and, along with penalty function properties, have some additional important features [15,16,19]. Dueto1 0 3 0 for any ψ and any k > 0, the following problem { f (x ) = min f (x) k 1 ψ(kc i (x)) 0, } i = 1,..., p (30) is equivalent to the original (ICO) problem. The classical Lagrangian L :IR n IR p + IR + IR 1 for the equivalent problem (30) is defined by the formula L(x, u, k) = f (x) k 1 u i ψ(kc i (x)). The NR principle consists of transforming the original (ICO) problem into an equivalent one and then using the classical Lagrangian for the equivalent problem for theoretical analysis and numerical methods. In contrast to SUMT, the NR principle leads to exteriorpoint methods. Convergence of the NR methods is due to both, the Lagrange multipliers and the scaling parameter updates.

8 PRIMAL DUAL METHODS FOR NONLINEAR CONSTRAINED OPTIMIZATION We use the following merit function: { ν ν(x, u) = max x L(x, u), min c i(x), 1 i p } u i c i (x), min u i,. (31) 1 i p For convex (ICO) under the standard second-order optimality condition, the merit function (31) satisfies Equations (7) and (8). For a given u IR p ++ and k > 0, one step of the NR method with scaling parameter update consists in finding or ˆx = arg min { L(x, u, k) x IR n}, (32) ˆx : x L(ˆx, u, k) = f (ˆx) u i c i (ˆx) = 0, ψ ( kc i (ˆx) ) followed by the Lagrange multipliers update by formula û i = ψ ( kc i (ˆx) ) u i, i = 1,..., p, or û = ( kc(ˆx) ) u, (33) where ( k s c(x) ) = diag ( ψ ( kc i (x) )) p,and scaling parameter update by formula ˆk := ν(ˆx,û) 0.5. (34) In other words, for a given Lagrange multipliers vector u IR p ++ and the scaling parameter k > 0, one step of the NR method is equivalent to solving the following nonlinear PD system: x L(ˆx, u, k) = f (ˆx) ψ ( kc i (ˆx) ) u i c i (ˆx) = x L(ˆx,û) = 0, (35) û = ( kc(ˆx) ) u, (36) for ˆx and û, followed by the scaling parameter k > 0 update by Equation (34). Application of Newton s method for solving the nonlinear PD system (35) (36) for ˆx and û leads to the primal dual nonlinear rescaling (PDNR) method for solving (ICO) problem [20]. By linearizing Equations (35) (36) and assuming that u = ( kc(x) ) u, we obtain the following linear PD system for finding the Newton direction ( x, u): [ 2 xx L( ) c T ][ ] ( ) x U ( ) c( ) 1 k I p u [ x L( ) = 1 k (u u) ], (37) where c( ) = c(x), ( ) = ( kc(x) ) = diag ( ψ ( kc i (x) )) p, U = diag (u i) p and I p is an identity matrix in IR p. Let δ > 0 be small enough and 0 <ɛ δ be the desired accuracy. Then for a given x IR n, Lagrange multipliers vector u IR p ++, and scaling parameter k > 0, one step of the PDNR method consists of the following operations: PDNRM - Primal Dual Nonlinear Rescaling Method Step 0. Let y = (x, u) S(y, δ). Step 1. If ν(y) ɛ, then output y as a solution. Step 2. Set k = ν(x, u) 0.5. Step 3. Find x and v from (37). Step 4. Update the primal dual pair by the formulas ˆx := x + x, û := u + u. Step 5. Goto Step 1.

PRIMAL DUAL METHODS FOR NONLINEAR CONSTRAINED OPTIMIZATION 9 If the standard second-order optimality conditions are satisfied, the Lipschitz conditions for the Hessians 2 f (x), 2 c i, i = 1,..., p hold, and y = (x, u) S(y, δ), then the PDNR algorithm generates a PD sequence that converges to the PD solution with 1.5-Q-superlinear rate. A small modification of the PDNR algorithm generates a PD sequence that converges from any starting point y = (x, u) IR n IR p ++ with asymptotic 1.5-Q-superlinear rate [20]. PRIMAL DUAL EXTERIOR-POINT METHOD FOR INEQUALITY CONSTRAINED OPTIMIZATION Due to properties 1 0 and 2 0 for any given vector k = (k 1,..., k p ) IR p ++, wehave c i (x) 0 k i ψ(k i c i (x)) 0, i = 1,..., p. Consider the Lagrangian L :IR n IR p + IR p ++ IR L(x, u, k) = f (x) k 1 i u i ψ(kc i (x)) for the equivalent problem. Let k > 0, x IR n, u IR p and k = (k i = k(u i ) 1, i = 1,..., p). One step of the NR method with dynamic scaling parameters update maps the given triple (x, u, k) IR n IR p + IR p ++ into the triple (ˆx,û, ˆk) IR n IR p + IR p ++ defined by formulas ˆx : x L(ˆx, u, k) = f (ˆx) ψ (k i c i (ˆx))u i c i (ˆx) = f (ˆx) û i c i (ˆx), (38) û :û i = u i ψ (k i c i (ˆx)), i = 1,..., p, (39) ˆk : ˆk i = kû 1 i, i = 1,..., p. (40) Such a method was considered in Ref. 18 for exponential transformation ψ 1. By removing the formula for the scaling vector update (40) from the system (38), (39), (40) we obtain the following nonlinear PD system: x L(ˆx,û) = f (ˆx) û i c i (ˆx), (41) û = (kc(ˆx))u, (42) where (kc(ˆx)) = diag(ψ (k i c i (ˆx))) p. To measure the distance between the current approximation y = (x, u) and the solution y, the merit function (31) is used. Due to Equation (40), the Lagrange multipliers that correspond to the passive constraints (LMPC) converge to zero with at least a quadratic rate. Therefore, for 0 < ɛ 1, it requires at most s = O(ln ln ɛ 1 ) Lagrange multipliers updates for the LMPC to become of the order of O(ɛ 2 ). So the PD system (41), (42) is reduced to the following nonlinear PD system: x L(ˆx,û) = f (ˆx) r û i c i (ˆx), (43) û i = u i ψ (k i c i (ˆx)), i = 1,..., r, (44) where I ={i : c i (x ) = 0} ={1,..., r}. Application of Newton s method for solving the PD system (43), (44) leads to a local PDEP method for (ICO) problems. Start with linearization of the system (44); due to the property 2 0 of the transformation ψ, the inverse ψ 1 exists. Therefore using the identity ψ 1 = ψ, where ψ (s) = inf t {st ψ(t) t IR} is the conjugate of ψ, and denoting ϕ = ψ, we can rewrite Equation (44) as follows: c i (ˆx) = k 1 u i ψ 1 ( ûi u i ( ) = k 1 u i ϕ ûi. u i ) ( ) = k 1 u i ψ ûi It follows from property 3 0 of transformation ψ that ϕ (1) = 0. Assuming ˆx = x + x and û = u + u, and ignoring terms of the second and higher order, we obtain c i (ˆx) = c i (x) + c i (x) x u i

10 PRIMAL DUAL METHODS FOR NONLINEAR CONSTRAINED OPTIMIZATION ( ) = k 1 u i ϕ ui + u i u i ( = k 1 u i ϕ 1 + u ) i = k 1 ϕ (1) u i, i = 1,..., r, u i finding the PD Newton directions: [ xx L( ) c T ( ) ][ x c( ) k 1 ϕ (1)I r u [ ] x L( ) = c( ) ] (45) or c i (x) + c i (x) x + k 1 ϕ (1) u i = 0, i = 1,..., r. By linearizing the system (43) at y = (x, u), we obtain the following linear PD system for where I r is the identity matrix in IR r and c(x) = J(c(x)) is the Jacobian of the vector function c(x). Let δ > 0 be small enough and 0 <ɛ δ be the desired accuracy. Then the PDEP method for (ICO) problems consists of the following operations. PDEPICOM - Primal Dual Exterior-Point Method for ICO Step 0. Let y = (x, u) S(y, δ). Step 1. If ν(y) ɛ, then output y as a solution. Step 2. Set k = ν(x, u) 1. Step 3. Find x and v from (45). Step 4. Update the primal dual pair by the formulas Step 5. Goto Step 1. ˆx := x + x, û := u + u. Under the standard second-order optimality conditions and Lipschitz conditions for the Hessians 2 f (x), 2 c i, i = 1,..., r, the PDEP algorithm generates a PD sequence that converges to the PD solution y = (x, u ) with quadratic rate. The globally convergent PDEP method with an asymptotic quadratic rate is given in [21]. PRIMAL DUAL EXTERIOR-POINT METHOD FOR OPTIMIZATION PROBLEMS WITH BOTH EQUALITY AND INEQUALITY CONSTRAINTS Consider p + q + 1 twice continuously differential functions f, c i, g j : IR n IR, i = 1,..., p, j = 1,..., q and the feasible set = { x : c i (x) 0, i = 1,..., p; g j (x) = 0, j = 1,..., q }. The problem with both inequality and equality constraints consists of finding (IECO) f (x ) = min{f (x) x }. We use ψ to transform the inequality constraints c i (x) 0, i = 1,..., p into an equivalent set of constraints. For any fixed k > 0, the following problem is equivalent to the original (IECO) problem due to the properties of ψ ; that is, f (x ) = min{f (x) k 1 ψ(kc i (x)) 0, i = 1,..., p; g j (x) = 0, j = 1,..., q}. For a given k > 0, the AL L k : IR n IR p + IR q IR 1 for the equivalent problem is defined by the formula L k (x, u, v) = f (x) k 1 v j g j (x) + k 2 u i ψ(kc i (x)) g 2 j (x). (46) The first two terms define the classical Lagrangian for the equivalent problem in the absence of equality constraints. The last two

PRIMAL DUAL METHODS FOR NONLINEAR CONSTRAINED OPTIMIZATION 11 terms coincide with the AL terms associated with equality constraints. For a given k > 0andy = (x, u, v) S(y, δ), a step of nonlinear rescaling-al method maps the triple y = (x, u, v) into the triple ŷ = (ˆx,û, ˆv), where or ˆx = arg min x IR n L k(x, u, v), (47) ˆx : f (ˆx) u i ψ ( kc i (ˆx) ) c i (ˆx) ( vj kg j (ˆx) ) g j (ˆx) = 0, (48) û i = u i ψ (kc i (ˆx)), i = 1,..., p; (49) ˆv j = v j kg j (ˆx), j = 1,..., q. (50) The system (48) (50) can be replaced by the following nonlinear PD system: x L(ˆx,û, ˆv) = f (ˆx) û i c i (ˆx) ˆv j g j (ˆx) = 0, (51) û ( kc(ˆx) ) u = 0, (52) ˆv v + kg(ˆx) = 0, (53) where (kc(ˆx)) = diag(ψ (kc i (ˆx))) p. We use the following merit function: ν(y) = ν(x, u, v) = max{ x L(x, u, v), min c i(x), 1 i p max 1 i q g i(x), u i c i (x), min 1 i p u i}. (54) Under the standard second-order optimality conditions, ν(y) possesses properties (7), (8). Application of Newton s method to the system (51), (52), (53) for ˆx, û, andˆv from the starting point y = (x, u, v) S(y, δ) leads to the PD exterior-point method (PDEPM) [22]. By linearizing the system (51), (52), (53) and ignoring terms of the second and higher order, we obtain the following linear PD system for finding the Newton direction y = ( x, u, v): xx 2 L( ) ct ( ) g T ( ) U 1 ( ) c( ) k I p 0 g( ) 0 x u v x L( ) = 1 k (u u) g( ) 1 k I q, (55) where c( ) = c(x), g( ) = g(x), ( ) = ( kc(x) ) =diag ( ψ ( kc i (x) )) p, u = ( kc(x) ) u, U = diag (u i ) p,andi p, I q are the identity matrices in IR p and IR q respectively. The PD exterior-point method for constrained optimization problems with both inequality and equality constraints consists of the following operations. PDEPM - Primal Dual Exterior-Point Method Step 0. Let y S(y, δ).. Step 1. If μ(y) ɛ, then output y as a solution. Step 2. Set k = ν(x, u, v) 0.5. Step 3. Find x, u, and v from Equation (55). Step 4. Update the primal dual pair by the formulas ˆx := x + x, û := u + u. ˆv := v + v. Step 5. Goto Step 1.

12 PRIMAL DUAL METHODS FOR NONLINEAR CONSTRAINED OPTIMIZATION Under the standard second-order optimality conditions and Lipschitz conditions for the Hessians 2 f (x), 2 c i, i = 1,..., m, 2 g i, i = 1,..., q, foranyy S(y, δ), the PDEPM algorithm generates the PD sequence that converges to the PD solution with 1.5-Qsuperlinear rate [22]. CONCLUDING REMARKS As we have seen, any PD method for constrained optimization is associated with particular PD nonlinear system of equations. Application of Newton s method to the system is equivalent to one step of either an interior- or an exterior-point method. Therefore, the computational process is defined by a particular PD nonlinear system and excellent convergence properties of Newton s method. As a result, under the standard second-order optimality conditions, the PD methods generate sequences that converge to the PD solution with an asymptotic quadratic rate. As a practical matter it is worth mentioning that in the neighborhood of the solution the PD interior-point method generates a sequence similar to that generated by Newton s method for solving KKT s system, while the PD exterior-point method generates a sequence similar to that generated by Newton s method for solving the Lagrange system of equations that correspond to the active constraints. Unlike the former nonlinear system, the latter system does not contain complementarity equations. Therefore, the practical PD exterior-point method often is capable of finding the PD solution with a very high of accuracy. REFERENCES 1. Polyak BT. Introduction to optimization. New York: Software Inc.; 1987. 2. Hestenes MR. Multipliers and gradient methods. J Optim Theory Appl 1969;4(5):303 320. 3. Powell MJD. A method for nonlinear constraints in minimization problems. In: Fletcher R, editor. Optimization. London: Acadamic Press; 1969. pp. 283 298. 4. Polyak RA. On the local quadratic convergence of the primal dual augmented Lagrangian method. Optim Method Soft 2009;24(3):369 379. 5. Fiacco AV, McCormick GP. Nonlinear programming: sequential unconstrained minimization techniques. New York: John Wiley & Sons, Inc.; 1968. 6. Wright SJ. Primal dual interior-point methods. Philadelphia (PA): SIAM; 1997. 7. Nesterov Y, Nemirovsky A. Interior-point polynomial algorithm in convex programming. Philadelphia (PA): SIAM; 1994. 8. Nesterov Y. Introductory lectures on convex optimization. Norwell (MA): Kluwer Academic Publishers; 2004. 9. Armand P, Benoist J, Orban D. Dynamic updates of the barrier parameter in primal dual methods for nonlinear programming. Comput Optim Appl 2008;41(1): 1 25. 10. Griva I, Shanno DF, Vanderbei RJ, et al. Global convergence of a primal dual interiorpoint method for nonlinear programming. Algorithm Oper Res 2008;3(1):12 29. 11. Forsgren A, Gill P. Primal dual interior methods for nonconvex nonlinear programming. SIAM J Optim 1998;8(4):1132 1152. 12. Wachter A, Biegler LT. On the implementation of an interior-point filter line-search algorithm for large-scale nonlinear programming. Math Program 2006;106(1):25 57. 13. Waltz RA, Morales JL, Nocedal J, et al. An interior algorithm for nonlinear optimization that combines line search and trust region steps. Math Program 2006;107:391 408. 14. Polyak RA. Modified barrier functions theory and methods. Math Program 1992;54:177 222. 15. Polyak R, Teboulle M. Nonlinear rescaling and proximal-like methods in convex optimization. Math Program 1997;76:265 284. 16. Polyak RA. Nonlinear rescaling vs. smoothing technique in convex optimization. Math Program 2002;92:197 235. 17. Yamashita H, Yabe H. Superlinear and quadratic convergence of some primal dual interior point methods for constrained optimization. Math Program 1996;75:377 397. 18. Tseng P, Bertsekas D. On the convergence of exponential multipliers method for convex programming. Math Program 1993;60:1 19. 19. Ben-Tal A, Zibulevsky M. Penalty-barrier methods for convex optimization. SIAM J Control Optim 1997;7:347 366.

PRIMAL DUAL METHODS FOR NONLINEAR CONSTRAINED OPTIMIZATION 13 20. Griva I, Polyak RA. Primal dual nonlinear rescaling method with dynamic scaling parameter update. Math Program 2006;106(2):237 259. 21. Polyak RA. Primal dual exterior-point method for convex optimization. Optim Method Softw 2008;23(1):141 160. 22. Griva I, Polyak RA. 1.5-Q-superlinear convergence of an exterior-point method for constrained optimization. J Glob Optim 2008;40(4):679 695.