An Inexact Sequential Quadratic Optimization Method for Nonlinear Optimization

Size: px
Start display at page:

Download "An Inexact Sequential Quadratic Optimization Method for Nonlinear Optimization"

Transcription

1 An Inexact Sequential Quadratic Optimization Method for Nonlinear Optimization Frank E. Curtis, Lehigh University involving joint work with Travis Johnson, Northwestern University Daniel P. Robinson, Johns Hopkins University Andreas Wächter, Northwestern University SIAM Conference on Optimization San Diego, CA 22 May 2014 An Inexact Sequential Quadratic Optimization Method for Nonlinear Optimization 1 of 30

2 Outline Motivation Algorithm Description Numerical Experiments Summary An Inexact Sequential Quadratic Optimization Method for Nonlinear Optimization 2 of 30

3 Problem formulation Our goal is to solve a large-scale constrained nonlinear optimization problem: min f(x) x R n s.t. c(x) = 0, c(x) 0. (NLP) If (NLP) is infeasible, then at least we want to minimize constraint violation: min v(x), where v(x) := c(x) 1 + [ c(x)] + 1. (FP) x R n Any feasible point for (NLP) is an optimal solution of (FP). An Inexact Sequential Quadratic Optimization Method for Nonlinear Optimization 3 of 30

4 Algorithm design objectives 1 Demands on algorithms for large-scale constrained nonlinear optimization: Scalable step computation Effectively handles negative curvature Superlinear convergence in primal-dual space Asymptotic monotonicity (consistent progress toward solution) Active-set detection and warm-starting Traditional methods (SQO, IP, AL) only satisfy subsets of these demands. I can solve my problem quickly enough with an IP method using matrix factorizations... then you re not interested in what I have to say. If you can only afford a few NLP iterations, are solving sequences of similar problems, and/or need a method that can converge fast locally,... 1 Paraphrased from Zavala and Anitescu (2014) An Inexact Sequential Quadratic Optimization Method for Nonlinear Optimization 4 of 30

5 Sequential quadratic optimization (SQO) SQO advantages: Parameter free search direction computation (ideally) Strong global convergence properties and behavior Active-set identification = Newton-like local convergence SQO disadvantages: No best way to handle negative curvature/inconsistent subproblems Quadratic subproblems (QPs) are expensive to solve exactly not scalable! Questions Employing scalable QP solvers, can we exploit inexact solves? From inexactness, do we lose global/local convergence? active set detection? Is there any benefit in applying a scalable method for the QP subproblems, as opposed to applying a similar scalable method to (NLP) directly? Sit and solve QP? Or move on nonlinear problem as in filtersd; Fletcher (2012) An Inexact Sequential Quadratic Optimization Method for Nonlinear Optimization 5 of 30

6 Inexact SQO vs. inexact SQO Inexact SQO may mean inexact gradients with (in)exact QP solves: Dennis, El-Alem, Maciel (1997) Diehl, Walther, Bock, Kostina (2010) Jäger, Sachs (1997) Heinkenschloss, Vicente (2000) Walther, Vetukuri, Biegler (2012) Inexact SQO may also mean exact gradients with inexact QP solves: Byrd, Curtis, Nocedal (2008) Izmailov, Solodov (2010) Leibfritz, Sachs (1999) Morales, Nocedal, Wu (2010) In this talk, we are interested in the latter type of inexact SQO. An Inexact Sequential Quadratic Optimization Method for Nonlinear Optimization 6 of 30

7 Algorithmic framework: Classic NLP solver approximation model solution QP solver approximation model solution linear solver An Inexact Sequential Quadratic Optimization Method for Nonlinear Optimization 7 of 30

8 Algorithmic framework: Inexact NLP solver approximation model termination conditions approximate solution step type QP solver approximation model termination conditions approximate solution step type linear solver An Inexact Sequential Quadratic Optimization Method for Nonlinear Optimization 8 of 30

9 SQO w/ inexact subproblem solves Contributions: Broad, but consistent, framework; Gill, Robinson (2013), pdal/ssqo Implementable termination conditions for inexact QP solves No specific QP solver required Global convergence guarantees (feasible and infeasible problems) Future work: Fast local convergence (feasible and infeasible problems) 2 Algorithmic features: Allows generic inexactness in QP solutions Convex combination of optimality and feasibility steps Negative curvature handled with dynamic Hessian modifications Separate multipliers for (NLP) and (FP) Dynamic updates for penalty parameter and Lagrange multipliers 2 Avoid using Cauchy points that only yield minimal progress for global convergence. An Inexact Sequential Quadratic Optimization Method for Nonlinear Optimization 9 of 30

10 Fritz John and penalty functions (NLP): min f(x) x s.t. c(x) = 0, c(x) 0 (FP):» + min v(x) := c(x) x [ c(x)] 1 Define the Fritz John (FJ) function F(x, y, ȳ, µ) := µf(x) + c(x) T y + c(x) T ȳ and the l 1 -norm exact penalty function φ(x, µ) := µf(x) + v(x). µ 0 acts as objective multiplier/penalty parameter. An Inexact Sequential Quadratic Optimization Method for Nonlinear Optimization 10 of 30

11 Optimality conditions (NLP): min f(x) x s.t. c(x) = 0, c(x) 0 (FP):» + min v(x) := c(x) x [ c(x)] 1 (PP): min x (FJ): φ(x, µ) := µf(x)+v(x) F(x, y, ȳ, µ) := µf(x) + c(x) T y + c(x) T ȳ KKT conditions for (FP) and (PP) expressed with residual 2 µg(x) + J(x)y + J(x)ȳ 3 min{[c(x)] +, e y} ρ(x, y, ȳ, µ) := 6 min{[c(x)], e + y} 7 4 min{[ c(x)] +, e ȳ} 5 min{[ c(x)], ȳ} FJ point: KKT point: ρ(x, y, ȳ, µ) = 0, v(x) = 0, (y, ȳ, µ) 0 ρ(x, y, ȳ, µ) = 0, v(x) = 0, µ > 0 Infeasible stationary point: ρ(x, y, ȳ, 0) = 0, v(x) > 0 An Inexact Sequential Quadratic Optimization Method for Nonlinear Optimization 11 of 30

12 Penalty function model and QP subproblem (NLP): min f(x) x s.t. c(x) = 0, c(x) 0 (FP):» + min v(x) := c(x) x [ c(x)] 1 (PP): min x (FJ): φ(x, µ) := µf(x)+v(x) F(x, y, ȳ, µ) := Define a local model of φ(, µ) at x k : l k (d, µ) := µ(f k + g T k d) + c k + J T k d 1 + [ c k + J T k d]+ 1 Reduction in this model yielded by a given d: Subproblem of interest: l k (d, µ) := l(0, µ) l(d, µ) min d l k (d, µ) dt Hd (QP) µf(x) + c(x) T y + c(x) T ȳ KKT residual: ρ(x, y, ȳ, µ) l k (d, µ) > 0 implies d is a direction of strict descent for φ(, µ) from x k An Inexact Sequential Quadratic Optimization Method for Nonlinear Optimization 12 of 30

13 Optimality conditions (for QP) (NLP): min f(x) x s.t. c(x) = 0, c(x) 0 (FP):» + min v(x) := c(x) x [ c(x)] 1 (PP): min x (FJ): φ(x, µ) := µf(x)+v(x) F(x, y, ȳ, µ) := µf(x) + c(x) T y + c(x) T ȳ KKT residual: ρ(x, y, ȳ, µ) Local model of φ at x k : l k (d, µ) KKT conditions for (FP) and (PP) expressed with residual 2 µg(x) + J(x)y + J(x)ȳ 3 min{[c(x)] +, e y} ρ(x, y, ȳ, µ) := 6 min{[c(x)], e + y} 7 4 min{[ c(x)] +, e ȳ} 5 min{[ c(x)], ȳ} KKT conditions for (QP) expressed with 2 µg k + Hd + J k y + J 3 k ȳ min{[c k + Jk T d]+, e y} ρ k (d, y, ȳ, µ, H) := min{[c 6 k + Jk T d], e + y} 4min{[ c k + J T 7 k d]+, e ȳ} 5 min{[ c k + J T k d], ȳ} Exact solution of (QP): ρ k (d, y, ȳ, µ, H) = 0 An Inexact Sequential Quadratic Optimization Method for Nonlinear Optimization 13 of 30

14 Assumptions and well-posedness Assumption (1) The functions f, c, and c and their first derivatives are bounded and Lipschitz continuous in an open convex set containing {x k } and {x k + d k }. (2) The QP solver can solve (QP) arbitrarily accurately for any µ 0. Theorem (Well-posedness) One of the following holds for our method, isqo: 1. isqo terminates finitely with a KKT point or infeasible stationary point. 2. isqo generates an infinite sequence of iterates»» y x k, k y, k y ȳ k ȳ k, µ k where k y k [ e, e],»ȳ k [0, e], and µ k > 0. ȳ k An Inexact Sequential Quadratic Optimization Method for Nonlinear Optimization 14 of 30

15 Global convergence Theorem (Global convergence) One of the following holds: (a) µ k = µ for some µ > 0 for all large k and either every limit point of {x k } corresponds to a KKT point or is an infeasible stationary point; (b) µ k 0 and every limit point of {x k } is an infeasible stationary point; (c) µ k 0, all limit points of {x k } are feasible, and, with K µ := {k : µ k+1 < µ k }, every limit point of {x k } k Kµ corresponds to an FJ point where the MFCQ fails. Corollary If {x k } is bounded and every limit point of this sequence is a feasible point at which the MFCQ holds, then µ k = µ for some µ > 0 for all large k and every limit point of {x k } corresponds to a KKT point. An Inexact Sequential Quadratic Optimization Method for Nonlinear Optimization 15 of 30

16 Direct scenario (NLP): min f(x) x s.t. c(x) = 0, c(x) 0 (FP):» + min v(x) := c(x) x [ c(x)] 1 (PP): min x (FJ): φ(x, µ) := µf(x)+v(x) F(x, y, ȳ, µ) := µf(x) + c(x) T y + c(x) T ȳ KKT residuals: ρ(x, y, ȳ, µ) ρ k (d, y, ȳ, µ, H) Local model of φ at x k : l k (d, µ) Terminate the QP solver when the solution (d k, y k+1, ȳ k+1 ) of (QP) with µ = µ k satisfies If y k+1 [ e, e], ȳ k+1 [0, e] l k (d k, µ k ) θ d k 2 > 0 for θ (0, 1) ρ k (d k, y k+1, ȳ k+1, µ k, H k ) κ ρ(x k, y k, ȳ k, µ k ) l k (d k, µ k ) ɛv k for ɛ (0, 1) then d k d k is the search direction µ k+1 µ k An Inexact Sequential Quadratic Optimization Method for Nonlinear Optimization 16 of 30

17 Reference scenario (NLP): min f(x) x s.t. c(x) = 0, c(x) 0 (FP):» + min v(x) := c(x) x [ c(x)] 1 (PP): min x (FJ): φ(x, µ) := µf(x)+v(x) F(x, y, ȳ, µ) := µf(x) + c(x) T y + c(x) T ȳ KKT residuals: ρ(x, y, ȳ, µ) ρ k (d, y, ȳ, µ, H) Local model of φ at x k : l k (d, µ) Terminate the QP solver when the solution (d k, y k+1, ȳ k+1 ) of (QP) with µ = 0 satisfies If y k+1 [ e, e], ȳ k+1 [0, e] l k (d k, 0) θ d k 2 for θ (0, 1) ρ k (d k, y k+1, ȳ k+1, 0, H k ) κ ρ(x k, y k, ȳ k, 0) l k (d k, µ k ) ɛ l k (d k, 0) for ɛ (0, 1) then d k d k is the search direction µ k+1 µ k An Inexact Sequential Quadratic Optimization Method for Nonlinear Optimization 17 of 30

18 Combination scenario (NLP): min f(x) x s.t. c(x) = 0, c(x) 0 (FP):» + min v(x) := c(x) x [ c(x)] 1 (PP): min x (FJ): φ(x, µ) := µf(x)+v(x) F(x, y, ȳ, µ) := Choose the largest τ [0, 1] such that d k τd k + (1 τ)d k yields l k (d k, 0) ɛ l k (d k, 0) then choose µ k+1 < µ k such that l k (d k, µ k+1 ) β l k (d k, 0) for β (0, 1) µf(x) + c(x) T y + c(x) T ȳ KKT residuals: ρ(x, y, ȳ, µ) ρ k (d, y, ȳ, µ, H) Local model of φ at x k : l k (d, µ) An Inexact Sequential Quadratic Optimization Method for Nonlinear Optimization 18 of 30

19 isqo framework repeat (1) Check whether KKT point or infeasible stationary point has been obtained. (2) Compute an inexact solution of (QP) with µ = µ k. (a) If Direct scenario occurs, then go to step 4. (3) Compute an inexact solution of (QP) with µ = 0. (a) If Reference scenario occurs, then go to step 4. (b) If Combination scenario occurs, then go to step 4. (4) Perform a backtracking line search to reduce φ(, µ k+1 ). endrepeat An Inexact Sequential Quadratic Optimization Method for Nonlinear Optimization 19 of 30

20 Complicating factors A few special cases make our actual algorithm slightly ;-) more complicated Landing on stationary points for φ(, µ k ) We allow only a multiplier and/or penalty parameter update A tightened accuracy tolerance is needed in combination scenarios We may require certain multipliers to be close to their bounds (Think of identifying violated constraints) H k and/or H k may not be positive definite We ask the QP solver to check the curvature along trial directions (Dynamic inertia correction if trial curvature is too small/negative) Actual algorithm involves six scenarios, but we have presented the core ideas An Inexact Sequential Quadratic Optimization Method for Nonlinear Optimization 20 of 30

21 Implementation details: Experiments in paper Matlab implementation Test set involves 309 CUTEr problems with at least one free variable at least one general (non-bound) constraint #variables + #constraints 20,000 (because it s Matlab!) no failures due to BQPD BQPD for QP solves with indefinite Hessians; see (Fletcher, 2000) Simulated inexactness by perturbing QP solutions Termination conditions (ɛ tol = 10 6 and ɛ µ = 10 8 ): ρ(x k, y k, ȳ k, µ k ) ɛ tol and v k ɛ tol ; (Optimal) ρ(x k, y k, ȳ k, 0) = 0 and v k > 0; (Infeasible) ρ(x k, y k, ȳ k, 0) ɛ tol and v k > ɛ tol and µ k ɛ µ (Infeasible) Investigate performance of inexact algorithm with κ = 0.01, 0.1, and 0.5 An Inexact Sequential Quadratic Optimization Method for Nonlinear Optimization 21 of 30

22 Success statistics Counts of termination messages for exact and three variants of inexact algorithm: Termination message Exact Inexact κ = 0.01 κ = 0.1 κ = 0.5 Optimal solution found Infeasible stationary point found Iteration limit reached Termination statistics and reliability do not degrade with inexactness! An Inexact Sequential Quadratic Optimization Method for Nonlinear Optimization 22 of 30

23 Inexactness levels Observe induced relative residuals for QP solves: κ I := ρ k ρ For problem j, we compute minimum (κ I (j)) and mean ( κ I (j)) values over run: min κ κ I,min κi(j) [ ) [ ) [ ) [ ) [ ) [ ) [ ) e e e mean κ κ I,mean e e e κi(j) [0.5 1) [1 ) Relative residuals generally need only be moderately smaller than parameter κ! An Inexact Sequential Quadratic Optimization Method for Nonlinear Optimization 23 of 30

24 Iteration comparison Considering the logarithmic outperforming factor r j := log 2 (iter j inexact /iterj exact ), we compare iteration counts of our inexact (κ = 0.01) and exact algorithms: Iteration counts do not degrade significantly with inexactness! An Inexact Sequential Quadratic Optimization Method for Nonlinear Optimization 24 of 30

25 Implementation details: Experiments from last night Matlab implementation Test set involves 60 CUTEr problems that were all successfully solved with augmented Lagrangian (AL) method sequential quadratic optimization (SQO) method isqo method with AL method for QP subproblems AL method or CPLEX for QP solves: switch when ρ(x k, y k, ȳ k, µ k ) 10 2 and v k 10 2 Terminate solve for (NLP) when ρ(x k, y k, ȳ k, µ k ) 10 6 and v k 10 6 An Inexact Sequential Quadratic Optimization Method for Nonlinear Optimization 25 of 30

26 Iteration comparison: AL vs. SQO AL (left) with cheap iterations vs. SQO (right) with expensive iterations An Inexact Sequential Quadratic Optimization Method for Nonlinear Optimization 26 of 30

27 Iteration comparison: SQO vs. isqo SQO (left) with expensive iterations vs. isqo (right) with cheaper iterations An Inexact Sequential Quadratic Optimization Method for Nonlinear Optimization 27 of 30

28 Iteration comparison: AL vs. isqo(subproblems) AL (left) with cheap iterations vs. isqo (right) with few expensive iterations An Inexact Sequential Quadratic Optimization Method for Nonlinear Optimization 28 of 30

29 Summary Contributions: Developed, analyzed, and experimented with an inexact SQO method Allows generic inexactness in QP subproblem solves No specific QP solver required Global convergence guarantees established Numerical experiments suggest inexact algorithm is reliable Inexact solutions allowed without degradation of performance An Inexact Sequential Quadratic Optimization Method for Nonlinear Optimization 29 of 30

30 Thanks! Exact Algorithms: J. V. Burke, F. E. Curtis, and H. Wang, A Sequential Quadratic Optimization Algorithm with Rapid Infeasibility Detection, to appear in SIAM Journal on Optimization, R. H. Byrd, F. E. Curtis, and J. Nocedal, Infeasibility Detection and SQP Methods for Nonlinear Optimization, SIAM Journal on Optimization, Volume 20, Issue 5, pg , Inexact Algorithms: F. E. Curtis, T. C. Johnson, D. P. Robinson, and A. Wächter, An Inexact Sequential Quadratic Optimization Algorithm for Large-Scale Nonlinear Optimization, to appear in SIAM Journal on Optimization, F. E. Curtis, J. Huber, O. Schenk, and A. Wächter, A Note on the Implementation of an Interior-Point Algorithm for Nonlinear Optimization with Inexact Step Computations, Mathematical Programming, Series B, Volume 136, Issue 1, pg , F. E. Curtis, O. Schenk, and A. Wächter, An Interior-Point Algorithm for Large-Scale Nonlinear Optimization with Inexact Step Computations, SIAM Journal on Scientific Computing, Volume 32, Issue 6, pg , R. H. Byrd, F. E. Curtis, and J. Nocedal, An Inexact Newton Method for Nonconvex Equality Constrained Optimization, Mathematical Programming, Volume 122, Issue 2, pg , F. E. Curtis, J. Nocedal, and A. Wächter, A Matrix-free Algorithm for Equality Constrained Optimization Problems with Rank-Deficient Jacobians, SIAM Journal on Optimization, Volume 20, Issue 3, pg , R. H. Byrd, F. E. Curtis, and J. Nocedal, An Inexact SQP Method for Equality Constrained Optimization, SIAM Journal on Optimization, Volume 19, Issue 1, pg , An Inexact Sequential Quadratic Optimization Method for Nonlinear Optimization 30 of 30

Infeasibility Detection and an Inexact Active-Set Method for Large-Scale Nonlinear Optimization

Infeasibility Detection and an Inexact Active-Set Method for Large-Scale Nonlinear Optimization Infeasibility Detection and an Inexact Active-Set Method for Large-Scale Nonlinear Optimization Frank E. Curtis, Lehigh University involving joint work with James V. Burke, University of Washington Daniel

More information

Recent Adaptive Methods for Nonlinear Optimization

Recent Adaptive Methods for Nonlinear Optimization Recent Adaptive Methods for Nonlinear Optimization Frank E. Curtis, Lehigh University involving joint work with James V. Burke (U. of Washington), Richard H. Byrd (U. of Colorado), Nicholas I. M. Gould

More information

An Inexact Newton Method for Optimization

An Inexact Newton Method for Optimization New York University Brown Applied Mathematics Seminar, February 10, 2009 Brief biography New York State College of William and Mary (B.S.) Northwestern University (M.S. & Ph.D.) Courant Institute (Postdoc)

More information

A New Penalty-SQP Method

A New Penalty-SQP Method Background and Motivation Illustration of Numerical Results Final Remarks Frank E. Curtis Informs Annual Meeting, October 2008 Background and Motivation Illustration of Numerical Results Final Remarks

More information

An Inexact Newton Method for Nonlinear Constrained Optimization

An Inexact Newton Method for Nonlinear Constrained Optimization An Inexact Newton Method for Nonlinear Constrained Optimization Frank E. Curtis Numerical Analysis Seminar, January 23, 2009 Outline Motivation and background Algorithm development and theoretical results

More information

Inexact Newton Methods and Nonlinear Constrained Optimization

Inexact Newton Methods and Nonlinear Constrained Optimization Inexact Newton Methods and Nonlinear Constrained Optimization Frank E. Curtis EPSRC Symposium Capstone Conference Warwick Mathematics Institute July 2, 2009 Outline PDE-Constrained Optimization Newton

More information

A Trust Funnel Algorithm for Nonconvex Equality Constrained Optimization with O(ɛ 3/2 ) Complexity

A Trust Funnel Algorithm for Nonconvex Equality Constrained Optimization with O(ɛ 3/2 ) Complexity A Trust Funnel Algorithm for Nonconvex Equality Constrained Optimization with O(ɛ 3/2 ) Complexity Mohammadreza Samadi, Lehigh University joint work with Frank E. Curtis (stand-in presenter), Lehigh University

More information

Infeasibility Detection in Nonlinear Optimization

Infeasibility Detection in Nonlinear Optimization Infeasibility Detection in Nonlinear Optimization Frank E. Curtis, Lehigh University Hao Wang, Lehigh University SIAM Conference on Optimization 2011 May 16, 2011 (See also Infeasibility Detection in Nonlinear

More information

PDE-Constrained and Nonsmooth Optimization

PDE-Constrained and Nonsmooth Optimization Frank E. Curtis October 1, 2009 Outline PDE-Constrained Optimization Introduction Newton s method Inexactness Results Summary and future work Nonsmooth Optimization Sequential quadratic programming (SQP)

More information

Algorithms for Constrained Optimization

Algorithms for Constrained Optimization 1 / 42 Algorithms for Constrained Optimization ME598/494 Lecture Max Yi Ren Department of Mechanical Engineering, Arizona State University April 19, 2015 2 / 42 Outline 1. Convergence 2. Sequential quadratic

More information

A Primal-Dual Augmented Lagrangian Penalty-Interior-Point Filter Line Search Algorithm

A Primal-Dual Augmented Lagrangian Penalty-Interior-Point Filter Line Search Algorithm Journal name manuscript No. (will be inserted by the editor) A Primal-Dual Augmented Lagrangian Penalty-Interior-Point Filter Line Search Algorithm Rene Kuhlmann Christof Büsens Received: date / Accepted:

More information

Hot-Starting NLP Solvers

Hot-Starting NLP Solvers Hot-Starting NLP Solvers Andreas Wächter Department of Industrial Engineering and Management Sciences Northwestern University waechter@iems.northwestern.edu 204 Mixed Integer Programming Workshop Ohio

More information

Trust-Region SQP Methods with Inexact Linear System Solves for Large-Scale Optimization

Trust-Region SQP Methods with Inexact Linear System Solves for Large-Scale Optimization Trust-Region SQP Methods with Inexact Linear System Solves for Large-Scale Optimization Denis Ridzal Department of Computational and Applied Mathematics Rice University, Houston, Texas dridzal@caam.rice.edu

More information

A STABILIZED SQP METHOD: SUPERLINEAR CONVERGENCE

A STABILIZED SQP METHOD: SUPERLINEAR CONVERGENCE A STABILIZED SQP METHOD: SUPERLINEAR CONVERGENCE Philip E. Gill Vyacheslav Kungurtsev Daniel P. Robinson UCSD Center for Computational Mathematics Technical Report CCoM-14-1 June 30, 2014 Abstract Regularized

More information

arxiv: v1 [math.oc] 20 Aug 2014

arxiv: v1 [math.oc] 20 Aug 2014 Adaptive Augmented Lagrangian Methods: Algorithms and Practical Numerical Experience Detailed Version Frank E. Curtis, Nicholas I. M. Gould, Hao Jiang, and Daniel P. Robinson arxiv:1408.4500v1 [math.oc]

More information

A GLOBALLY CONVERGENT STABILIZED SQP METHOD

A GLOBALLY CONVERGENT STABILIZED SQP METHOD A GLOBALLY CONVERGENT STABILIZED SQP METHOD Philip E. Gill Daniel P. Robinson July 6, 2013 Abstract Sequential quadratic programming SQP methods are a popular class of methods for nonlinearly constrained

More information

A GLOBALLY CONVERGENT STABILIZED SQP METHOD: SUPERLINEAR CONVERGENCE

A GLOBALLY CONVERGENT STABILIZED SQP METHOD: SUPERLINEAR CONVERGENCE A GLOBALLY CONVERGENT STABILIZED SQP METHOD: SUPERLINEAR CONVERGENCE Philip E. Gill Vyacheslav Kungurtsev Daniel P. Robinson UCSD Center for Computational Mathematics Technical Report CCoM-14-1 June 30,

More information

Adaptive augmented Lagrangian methods: algorithms and practical numerical experience

Adaptive augmented Lagrangian methods: algorithms and practical numerical experience Optimization Methods and Software ISSN: 1055-6788 (Print) 1029-4937 (Online) Journal homepage: http://www.tandfonline.com/loi/goms20 Adaptive augmented Lagrangian methods: algorithms and practical numerical

More information

1. Introduction. In this paper we discuss an algorithm for equality constrained optimization problems of the form. f(x) s.t.

1. Introduction. In this paper we discuss an algorithm for equality constrained optimization problems of the form. f(x) s.t. AN INEXACT SQP METHOD FOR EQUALITY CONSTRAINED OPTIMIZATION RICHARD H. BYRD, FRANK E. CURTIS, AND JORGE NOCEDAL Abstract. We present an algorithm for large-scale equality constrained optimization. The

More information

A STABILIZED SQP METHOD: GLOBAL CONVERGENCE

A STABILIZED SQP METHOD: GLOBAL CONVERGENCE A STABILIZED SQP METHOD: GLOBAL CONVERGENCE Philip E. Gill Vyacheslav Kungurtsev Daniel P. Robinson UCSD Center for Computational Mathematics Technical Report CCoM-13-4 Revised July 18, 2014, June 23,

More information

Large-Scale Nonlinear Optimization with Inexact Step Computations

Large-Scale Nonlinear Optimization with Inexact Step Computations Large-Scale Nonlinear Optimization with Inexact Step Computations Andreas Wächter IBM T.J. Watson Research Center Yorktown Heights, New York andreasw@us.ibm.com IPAM Workshop on Numerical Methods for Continuous

More information

Numerical Methods for PDE-Constrained Optimization

Numerical Methods for PDE-Constrained Optimization Numerical Methods for PDE-Constrained Optimization Richard H. Byrd 1 Frank E. Curtis 2 Jorge Nocedal 2 1 University of Colorado at Boulder 2 Northwestern University Courant Institute of Mathematical Sciences,

More information

5 Handling Constraints

5 Handling Constraints 5 Handling Constraints Engineering design optimization problems are very rarely unconstrained. Moreover, the constraints that appear in these problems are typically nonlinear. This motivates our interest

More information

An Inexact Newton Method for Nonconvex Equality Constrained Optimization

An Inexact Newton Method for Nonconvex Equality Constrained Optimization Noname manuscript No. (will be inserted by the editor) An Inexact Newton Method for Nonconvex Equality Constrained Optimization Richard H. Byrd Frank E. Curtis Jorge Nocedal Received: / Accepted: Abstract

More information

HYBRID FILTER METHODS FOR NONLINEAR OPTIMIZATION. Yueling Loh

HYBRID FILTER METHODS FOR NONLINEAR OPTIMIZATION. Yueling Loh HYBRID FILTER METHODS FOR NONLINEAR OPTIMIZATION by Yueling Loh A dissertation submitted to The Johns Hopkins University in conformity with the requirements for the degree of Doctor of Philosophy. Baltimore,

More information

Algorithms for constrained local optimization

Algorithms for constrained local optimization Algorithms for constrained local optimization Fabio Schoen 2008 http://gol.dsi.unifi.it/users/schoen Algorithms for constrained local optimization p. Feasible direction methods Algorithms for constrained

More information

MS&E 318 (CME 338) Large-Scale Numerical Optimization

MS&E 318 (CME 338) Large-Scale Numerical Optimization Stanford University, Management Science & Engineering (and ICME) MS&E 318 (CME 338) Large-Scale Numerical Optimization 1 Origins Instructor: Michael Saunders Spring 2015 Notes 9: Augmented Lagrangian Methods

More information

arxiv: v1 [math.na] 8 Jun 2018

arxiv: v1 [math.na] 8 Jun 2018 arxiv:1806.03347v1 [math.na] 8 Jun 2018 Interior Point Method with Modified Augmented Lagrangian for Penalty-Barrier Nonlinear Programming Martin Neuenhofen June 12, 2018 Abstract We present a numerical

More information

A Trust-region-based Sequential Quadratic Programming Algorithm

A Trust-region-based Sequential Quadratic Programming Algorithm Downloaded from orbit.dtu.dk on: Oct 19, 2018 A Trust-region-based Sequential Quadratic Programming Algorithm Henriksen, Lars Christian; Poulsen, Niels Kjølstad Publication date: 2010 Document Version

More information

A Trust-Funnel Algorithm for Nonlinear Programming

A Trust-Funnel Algorithm for Nonlinear Programming for Nonlinear Programming Daniel P. Robinson Johns Hopins University Department of Applied Mathematics and Statistics Collaborators: Fran E. Curtis (Lehigh University) Nic I. M. Gould (Rutherford Appleton

More information

A Penalty-Interior-Point Algorithm for Nonlinear Constrained Optimization

A Penalty-Interior-Point Algorithm for Nonlinear Constrained Optimization Noname manuscript No. (will be inserted by the editor) A Penalty-Interior-Point Algorithm for Nonlinear Constrained Optimization Frank E. Curtis April 26, 2011 Abstract Penalty and interior-point methods

More information

On the use of piecewise linear models in nonlinear programming

On the use of piecewise linear models in nonlinear programming Math. Program., Ser. A (2013) 137:289 324 DOI 10.1007/s10107-011-0492-9 FULL LENGTH PAPER On the use of piecewise linear models in nonlinear programming Richard H. Byrd Jorge Nocedal Richard A. Waltz Yuchen

More information

A globally and quadratically convergent primal dual augmented Lagrangian algorithm for equality constrained optimization

A globally and quadratically convergent primal dual augmented Lagrangian algorithm for equality constrained optimization Optimization Methods and Software ISSN: 1055-6788 (Print) 1029-4937 (Online) Journal homepage: http://www.tandfonline.com/loi/goms20 A globally and quadratically convergent primal dual augmented Lagrangian

More information

CONVERGENCE ANALYSIS OF AN INTERIOR-POINT METHOD FOR NONCONVEX NONLINEAR PROGRAMMING

CONVERGENCE ANALYSIS OF AN INTERIOR-POINT METHOD FOR NONCONVEX NONLINEAR PROGRAMMING CONVERGENCE ANALYSIS OF AN INTERIOR-POINT METHOD FOR NONCONVEX NONLINEAR PROGRAMMING HANDE Y. BENSON, ARUN SEN, AND DAVID F. SHANNO Abstract. In this paper, we present global and local convergence results

More information

SEQUENTIAL QUADRATIC PROGAMMING METHODS FOR PARAMETRIC NONLINEAR OPTIMIZATION

SEQUENTIAL QUADRATIC PROGAMMING METHODS FOR PARAMETRIC NONLINEAR OPTIMIZATION SEQUENTIAL QUADRATIC PROGAMMING METHODS FOR PARAMETRIC NONLINEAR OPTIMIZATION Vyacheslav Kungurtsev Moritz Diehl July 2013 Abstract Sequential quadratic programming (SQP) methods are known to be efficient

More information

Penalty and Barrier Methods General classical constrained minimization problem minimize f(x) subject to g(x) 0 h(x) =0 Penalty methods are motivated by the desire to use unconstrained optimization techniques

More information

On the Implementation of an Interior-Point Algorithm for Nonlinear Optimization with Inexact Step Computations

On the Implementation of an Interior-Point Algorithm for Nonlinear Optimization with Inexact Step Computations On the Implementation of an Interior-Point Algorithm for Nonlinear Optimization with Inexact Step Computations Frank E. Curtis Lehigh University Johannes Huber University of Basel Olaf Schenk University

More information

A null-space primal-dual interior-point algorithm for nonlinear optimization with nice convergence properties

A null-space primal-dual interior-point algorithm for nonlinear optimization with nice convergence properties A null-space primal-dual interior-point algorithm for nonlinear optimization with nice convergence properties Xinwei Liu and Yaxiang Yuan Abstract. We present a null-space primal-dual interior-point algorithm

More information

REGULARIZED SEQUENTIAL QUADRATIC PROGRAMMING METHODS

REGULARIZED SEQUENTIAL QUADRATIC PROGRAMMING METHODS REGULARIZED SEQUENTIAL QUADRATIC PROGRAMMING METHODS Philip E. Gill Daniel P. Robinson UCSD Department of Mathematics Technical Report NA-11-02 October 2011 Abstract We present the formulation and analysis

More information

Constraint Identification and Algorithm Stabilization for Degenerate Nonlinear Programs

Constraint Identification and Algorithm Stabilization for Degenerate Nonlinear Programs Preprint ANL/MCS-P865-1200, Dec. 2000 (Revised Nov. 2001) Mathematics and Computer Science Division Argonne National Laboratory Stephen J. Wright Constraint Identification and Algorithm Stabilization for

More information

Complexity analysis of second-order algorithms based on line search for smooth nonconvex optimization

Complexity analysis of second-order algorithms based on line search for smooth nonconvex optimization Complexity analysis of second-order algorithms based on line search for smooth nonconvex optimization Clément Royer - University of Wisconsin-Madison Joint work with Stephen J. Wright MOPTA, Bethlehem,

More information

Some new facts about sequential quadratic programming methods employing second derivatives

Some new facts about sequential quadratic programming methods employing second derivatives To appear in Optimization Methods and Software Vol. 00, No. 00, Month 20XX, 1 24 Some new facts about sequential quadratic programming methods employing second derivatives A.F. Izmailov a and M.V. Solodov

More information

Constrained Nonlinear Optimization Algorithms

Constrained Nonlinear Optimization Algorithms Department of Industrial Engineering and Management Sciences Northwestern University waechter@iems.northwestern.edu Institute for Mathematics and its Applications University of Minnesota August 4, 2016

More information

AN AUGMENTED LAGRANGIAN AFFINE SCALING METHOD FOR NONLINEAR PROGRAMMING

AN AUGMENTED LAGRANGIAN AFFINE SCALING METHOD FOR NONLINEAR PROGRAMMING AN AUGMENTED LAGRANGIAN AFFINE SCALING METHOD FOR NONLINEAR PROGRAMMING XIAO WANG AND HONGCHAO ZHANG Abstract. In this paper, we propose an Augmented Lagrangian Affine Scaling (ALAS) algorithm for general

More information

Part 5: Penalty and augmented Lagrangian methods for equality constrained optimization. Nick Gould (RAL)

Part 5: Penalty and augmented Lagrangian methods for equality constrained optimization. Nick Gould (RAL) Part 5: Penalty and augmented Lagrangian methods for equality constrained optimization Nick Gould (RAL) x IR n f(x) subject to c(x) = Part C course on continuoue optimization CONSTRAINED MINIMIZATION x

More information

Interior Methods for Mathematical Programs with Complementarity Constraints

Interior Methods for Mathematical Programs with Complementarity Constraints Interior Methods for Mathematical Programs with Complementarity Constraints Sven Leyffer, Gabriel López-Calva and Jorge Nocedal July 14, 25 Abstract This paper studies theoretical and practical properties

More information

Part 4: Active-set methods for linearly constrained optimization. Nick Gould (RAL)

Part 4: Active-set methods for linearly constrained optimization. Nick Gould (RAL) Part 4: Active-set methods for linearly constrained optimization Nick Gould RAL fx subject to Ax b Part C course on continuoue optimization LINEARLY CONSTRAINED MINIMIZATION fx subject to Ax { } b where

More information

A Penalty-Interior-Point Algorithm for Nonlinear Constrained Optimization

A Penalty-Interior-Point Algorithm for Nonlinear Constrained Optimization Noname manuscript No. (will be inserted by the editor) A Penalty-Interior-Point Algorithm for Nonlinear Constrained Optimization Fran E. Curtis February, 22 Abstract Penalty and interior-point methods

More information

A stabilized SQP method: superlinear convergence

A stabilized SQP method: superlinear convergence Math. Program., Ser. A (2017) 163:369 410 DOI 10.1007/s10107-016-1066-7 FULL LENGTH PAPER A stabilized SQP method: superlinear convergence Philip E. Gill 1 Vyacheslav Kungurtsev 2 Daniel P. Robinson 3

More information

CONVEXIFICATION SCHEMES FOR SQP METHODS

CONVEXIFICATION SCHEMES FOR SQP METHODS CONVEXIFICAION SCHEMES FOR SQP MEHODS Philip E. Gill Elizabeth Wong UCSD Center for Computational Mathematics echnical Report CCoM-14-06 July 18, 2014 Abstract Sequential quadratic programming (SQP) methods

More information

arxiv: v1 [math.oc] 10 Apr 2017

arxiv: v1 [math.oc] 10 Apr 2017 A Method to Guarantee Local Convergence for Sequential Quadratic Programming with Poor Hessian Approximation Tuan T. Nguyen, Mircea Lazar and Hans Butler arxiv:1704.03064v1 math.oc] 10 Apr 2017 Abstract

More information

1. Introduction. We consider nonlinear optimization problems of the form. f(x) ce (x) = 0 c I (x) 0,

1. Introduction. We consider nonlinear optimization problems of the form. f(x) ce (x) = 0 c I (x) 0, AN INTERIOR-POINT ALGORITHM FOR LARGE-SCALE NONLINEAR OPTIMIZATION WITH INEXACT STEP COMPUTATIONS FRANK E. CURTIS, OLAF SCHENK, AND ANDREAS WÄCHTER Abstract. We present a line-search algorithm for large-scale

More information

1 Computing with constraints

1 Computing with constraints Notes for 2017-04-26 1 Computing with constraints Recall that our basic problem is minimize φ(x) s.t. x Ω where the feasible set Ω is defined by equality and inequality conditions Ω = {x R n : c i (x)

More information

10 Numerical methods for constrained problems

10 Numerical methods for constrained problems 10 Numerical methods for constrained problems min s.t. f(x) h(x) = 0 (l), g(x) 0 (m), x X The algorithms can be roughly divided the following way: ˆ primal methods: find descent direction keeping inside

More information

MS&E 318 (CME 338) Large-Scale Numerical Optimization

MS&E 318 (CME 338) Large-Scale Numerical Optimization Stanford University, Management Science & Engineering (and ICME) MS&E 318 (CME 338) Large-Scale Numerical Optimization Instructor: Michael Saunders Spring 2015 Notes 11: NPSOL and SNOPT SQP Methods 1 Overview

More information

POWER SYSTEMS in general are currently operating

POWER SYSTEMS in general are currently operating TO APPEAR IN IEEE TRANSACTIONS ON POWER SYSTEMS 1 Robust Optimal Power Flow Solution Using Trust Region and Interior-Point Methods Andréa A. Sousa, Geraldo L. Torres, Member IEEE, Claudio A. Cañizares,

More information

ALADIN An Algorithm for Distributed Non-Convex Optimization and Control

ALADIN An Algorithm for Distributed Non-Convex Optimization and Control ALADIN An Algorithm for Distributed Non-Convex Optimization and Control Boris Houska, Yuning Jiang, Janick Frasch, Rien Quirynen, Dimitris Kouzoupis, Moritz Diehl ShanghaiTech University, University of

More information

An augmented Lagrangian method for equality constrained optimization with rapid infeasibility detection capabilities

An augmented Lagrangian method for equality constrained optimization with rapid infeasibility detection capabilities An augmented Lagrangian method for equality constrained optimization with rapid infeasibility detection capabilities Paul Armand, Ngoc Nguyen Tran To cite this version: Paul Armand, Ngoc Nguyen Tran. An

More information

Survey of NLP Algorithms. L. T. Biegler Chemical Engineering Department Carnegie Mellon University Pittsburgh, PA

Survey of NLP Algorithms. L. T. Biegler Chemical Engineering Department Carnegie Mellon University Pittsburgh, PA Survey of NLP Algorithms L. T. Biegler Chemical Engineering Department Carnegie Mellon University Pittsburgh, PA NLP Algorithms - Outline Problem and Goals KKT Conditions and Variable Classification Handling

More information

An Active Set Strategy for Solving Optimization Problems with up to 200,000,000 Nonlinear Constraints

An Active Set Strategy for Solving Optimization Problems with up to 200,000,000 Nonlinear Constraints An Active Set Strategy for Solving Optimization Problems with up to 200,000,000 Nonlinear Constraints Klaus Schittkowski Department of Computer Science, University of Bayreuth 95440 Bayreuth, Germany e-mail:

More information

What s New in Active-Set Methods for Nonlinear Optimization?

What s New in Active-Set Methods for Nonlinear Optimization? What s New in Active-Set Methods for Nonlinear Optimization? Philip E. Gill Advances in Numerical Computation, Manchester University, July 5, 2011 A Workshop in Honor of Sven Hammarling UCSD Center for

More information

Direct Methods. Moritz Diehl. Optimization in Engineering Center (OPTEC) and Electrical Engineering Department (ESAT) K.U.

Direct Methods. Moritz Diehl. Optimization in Engineering Center (OPTEC) and Electrical Engineering Department (ESAT) K.U. Direct Methods Moritz Diehl Optimization in Engineering Center (OPTEC) and Electrical Engineering Department (ESAT) K.U. Leuven Belgium Overview Direct Single Shooting Direct Collocation Direct Multiple

More information

INTERIOR-POINT METHODS FOR NONCONVEX NONLINEAR PROGRAMMING: CONVERGENCE ANALYSIS AND COMPUTATIONAL PERFORMANCE

INTERIOR-POINT METHODS FOR NONCONVEX NONLINEAR PROGRAMMING: CONVERGENCE ANALYSIS AND COMPUTATIONAL PERFORMANCE INTERIOR-POINT METHODS FOR NONCONVEX NONLINEAR PROGRAMMING: CONVERGENCE ANALYSIS AND COMPUTATIONAL PERFORMANCE HANDE Y. BENSON, ARUN SEN, AND DAVID F. SHANNO Abstract. In this paper, we present global

More information

Lecture 15: SQP methods for equality constrained optimization

Lecture 15: SQP methods for equality constrained optimization Lecture 15: SQP methods for equality constrained optimization Coralia Cartis, Mathematical Institute, University of Oxford C6.2/B2: Continuous Optimization Lecture 15: SQP methods for equality constrained

More information

Methods for Unconstrained Optimization Numerical Optimization Lectures 1-2

Methods for Unconstrained Optimization Numerical Optimization Lectures 1-2 Methods for Unconstrained Optimization Numerical Optimization Lectures 1-2 Coralia Cartis, University of Oxford INFOMM CDT: Modelling, Analysis and Computation of Continuous Real-World Problems Methods

More information

Derivative-Based Numerical Method for Penalty-Barrier Nonlinear Programming

Derivative-Based Numerical Method for Penalty-Barrier Nonlinear Programming Derivative-Based Numerical Method for Penalty-Barrier Nonlinear Programming Martin Peter Neuenhofen October 26, 2018 Abstract We present an NLP solver for nonlinear optimization with quadratic penalty

More information

Constrained optimization: direct methods (cont.)

Constrained optimization: direct methods (cont.) Constrained optimization: direct methods (cont.) Jussi Hakanen Post-doctoral researcher jussi.hakanen@jyu.fi Direct methods Also known as methods of feasible directions Idea in a point x h, generate a

More information

Nonmonotone Trust Region Methods for Nonlinear Equality Constrained Optimization without a Penalty Function

Nonmonotone Trust Region Methods for Nonlinear Equality Constrained Optimization without a Penalty Function Nonmonotone Trust Region Methods for Nonlinear Equality Constrained Optimization without a Penalty Function Michael Ulbrich and Stefan Ulbrich Zentrum Mathematik Technische Universität München München,

More information

A globally convergent Levenberg Marquardt method for equality-constrained optimization

A globally convergent Levenberg Marquardt method for equality-constrained optimization Computational Optimization and Applications manuscript No. (will be inserted by the editor) A globally convergent Levenberg Marquardt method for equality-constrained optimization A. F. Izmailov M. V. Solodov

More information

Lecture 13: Constrained optimization

Lecture 13: Constrained optimization 2010-12-03 Basic ideas A nonlinearly constrained problem must somehow be converted relaxed into a problem which we can solve (a linear/quadratic or unconstrained problem) We solve a sequence of such problems

More information

MODIFYING SQP FOR DEGENERATE PROBLEMS

MODIFYING SQP FOR DEGENERATE PROBLEMS PREPRINT ANL/MCS-P699-1097, OCTOBER, 1997, (REVISED JUNE, 2000; MARCH, 2002), MATHEMATICS AND COMPUTER SCIENCE DIVISION, ARGONNE NATIONAL LABORATORY MODIFYING SQP FOR DEGENERATE PROBLEMS STEPHEN J. WRIGHT

More information

The use of second-order information in structural topology optimization. Susana Rojas Labanda, PhD student Mathias Stolpe, Senior researcher

The use of second-order information in structural topology optimization. Susana Rojas Labanda, PhD student Mathias Stolpe, Senior researcher The use of second-order information in structural topology optimization Susana Rojas Labanda, PhD student Mathias Stolpe, Senior researcher What is Topology Optimization? Optimize the design of a structure

More information

A PRIMAL-DUAL TRUST REGION ALGORITHM FOR NONLINEAR OPTIMIZATION

A PRIMAL-DUAL TRUST REGION ALGORITHM FOR NONLINEAR OPTIMIZATION Optimization Technical Report 02-09, October 2002, UW-Madison Computer Sciences Department. E. Michael Gertz 1 Philip E. Gill 2 A PRIMAL-DUAL TRUST REGION ALGORITHM FOR NONLINEAR OPTIMIZATION 7 October

More information

Preprint ANL/MCS-P , Dec 2002 (Revised Nov 2003, Mar 2004) Mathematics and Computer Science Division Argonne National Laboratory

Preprint ANL/MCS-P , Dec 2002 (Revised Nov 2003, Mar 2004) Mathematics and Computer Science Division Argonne National Laboratory Preprint ANL/MCS-P1015-1202, Dec 2002 (Revised Nov 2003, Mar 2004) Mathematics and Computer Science Division Argonne National Laboratory A GLOBALLY CONVERGENT LINEARLY CONSTRAINED LAGRANGIAN METHOD FOR

More information

arxiv: v2 [math.oc] 11 Jan 2018

arxiv: v2 [math.oc] 11 Jan 2018 A one-phase interior point method for nonconvex optimization Oliver Hinder, Yinyu Ye January 12, 2018 arxiv:1801.03072v2 [math.oc] 11 Jan 2018 Abstract The work of Wächter and Biegler [40] suggests that

More information

Determination of Feasible Directions by Successive Quadratic Programming and Zoutendijk Algorithms: A Comparative Study

Determination of Feasible Directions by Successive Quadratic Programming and Zoutendijk Algorithms: A Comparative Study International Journal of Mathematics And Its Applications Vol.2 No.4 (2014), pp.47-56. ISSN: 2347-1557(online) Determination of Feasible Directions by Successive Quadratic Programming and Zoutendijk Algorithms:

More information

Some Inexact Hybrid Proximal Augmented Lagrangian Algorithms

Some Inexact Hybrid Proximal Augmented Lagrangian Algorithms Some Inexact Hybrid Proximal Augmented Lagrangian Algorithms Carlos Humes Jr. a, Benar F. Svaiter b, Paulo J. S. Silva a, a Dept. of Computer Science, University of São Paulo, Brazil Email: {humes,rsilva}@ime.usp.br

More information

AM 205: lecture 19. Last time: Conditions for optimality Today: Newton s method for optimization, survey of optimization methods

AM 205: lecture 19. Last time: Conditions for optimality Today: Newton s method for optimization, survey of optimization methods AM 205: lecture 19 Last time: Conditions for optimality Today: Newton s method for optimization, survey of optimization methods Optimality Conditions: Equality Constrained Case As another example of equality

More information

A SHIFTED PRIMAL-DUAL PENALTY-BARRIER METHOD FOR NONLINEAR OPTIMIZATION

A SHIFTED PRIMAL-DUAL PENALTY-BARRIER METHOD FOR NONLINEAR OPTIMIZATION A SHIFTED PRIMAL-DUAL PENALTY-BARRIER METHOD FOR NONLINEAR OPTIMIZATION Philip E. Gill Vyacheslav Kungurtsev Daniel P. Robinson UCSD Center for Computational Mathematics Technical Report CCoM-19-3 March

More information

Newton s Method. Ryan Tibshirani Convex Optimization /36-725

Newton s Method. Ryan Tibshirani Convex Optimization /36-725 Newton s Method Ryan Tibshirani Convex Optimization 10-725/36-725 1 Last time: dual correspondences Given a function f : R n R, we define its conjugate f : R n R, Properties and examples: f (y) = max x

More information

1. Introduction. We consider the general smooth constrained optimization problem:

1. Introduction. We consider the general smooth constrained optimization problem: OPTIMIZATION TECHNICAL REPORT 02-05, AUGUST 2002, COMPUTER SCIENCES DEPT, UNIV. OF WISCONSIN TEXAS-WISCONSIN MODELING AND CONTROL CONSORTIUM REPORT TWMCC-2002-01 REVISED SEPTEMBER 2003. A FEASIBLE TRUST-REGION

More information

A DECOMPOSITION PROCEDURE BASED ON APPROXIMATE NEWTON DIRECTIONS

A DECOMPOSITION PROCEDURE BASED ON APPROXIMATE NEWTON DIRECTIONS Working Paper 01 09 Departamento de Estadística y Econometría Statistics and Econometrics Series 06 Universidad Carlos III de Madrid January 2001 Calle Madrid, 126 28903 Getafe (Spain) Fax (34) 91 624

More information

NUMERICAL OPTIMIZATION. J. Ch. Gilbert

NUMERICAL OPTIMIZATION. J. Ch. Gilbert NUMERICAL OPTIMIZATION J. Ch. Gilbert Numerical optimization (past) The discipline deals with the classical smooth (nonconvex) problem min {f(x) : c E (x) = 0, c I (x) 0}. Applications: variable added

More information

A Quasi-Newton Algorithm for Nonconvex, Nonsmooth Optimization with Global Convergence Guarantees

A Quasi-Newton Algorithm for Nonconvex, Nonsmooth Optimization with Global Convergence Guarantees Noname manuscript No. (will be inserted by the editor) A Quasi-Newton Algorithm for Nonconvex, Nonsmooth Optimization with Global Convergence Guarantees Frank E. Curtis Xiaocun Que May 26, 2014 Abstract

More information

Convex Optimization with ALADIN

Convex Optimization with ALADIN Mathematical Programming manuscript No. (will be inserted by the editor) Convex Optimization with ALADIN Boris Houska Dimitris Kouzoupis Yuning Jiang Moritz Diehl the date of receipt and acceptance should

More information

Scientific Computing: An Introductory Survey

Scientific Computing: An Introductory Survey Scientific Computing: An Introductory Survey Chapter 6 Optimization Prof. Michael T. Heath Department of Computer Science University of Illinois at Urbana-Champaign Copyright c 2002. Reproduction permitted

More information

Scientific Computing: An Introductory Survey

Scientific Computing: An Introductory Survey Scientific Computing: An Introductory Survey Chapter 6 Optimization Prof. Michael T. Heath Department of Computer Science University of Illinois at Urbana-Champaign Copyright c 2002. Reproduction permitted

More information

Rejoinder on: Critical Lagrange multipliers: what we currently know about them, how they spoil our lives, and what we can do about it

Rejoinder on: Critical Lagrange multipliers: what we currently know about them, how they spoil our lives, and what we can do about it TOP manuscript No. (will be inserted by the editor) Rejoinder on: Critical Lagrange multipliers: what we currently know about them, how they spoil our lives, and what we can do about it A. F. Izmailov

More information

12. Interior-point methods

12. Interior-point methods 12. Interior-point methods Convex Optimization Boyd & Vandenberghe inequality constrained minimization logarithmic barrier function and central path barrier method feasibility and phase I methods complexity

More information

minimize x subject to (x 2)(x 4) u,

minimize x subject to (x 2)(x 4) u, Math 6366/6367: Optimization and Variational Methods Sample Preliminary Exam Questions 1. Suppose that f : [, L] R is a C 2 -function with f () on (, L) and that you have explicit formulae for

More information

Newton s Method. Javier Peña Convex Optimization /36-725

Newton s Method. Javier Peña Convex Optimization /36-725 Newton s Method Javier Peña Convex Optimization 10-725/36-725 1 Last time: dual correspondences Given a function f : R n R, we define its conjugate f : R n R, f ( (y) = max y T x f(x) ) x Properties and

More information

IBM Research Report. Line Search Filter Methods for Nonlinear Programming: Motivation and Global Convergence

IBM Research Report. Line Search Filter Methods for Nonlinear Programming: Motivation and Global Convergence RC23036 (W0304-181) April 21, 2003 Computer Science IBM Research Report Line Search Filter Methods for Nonlinear Programming: Motivation and Global Convergence Andreas Wächter, Lorenz T. Biegler IBM Research

More information

Lecture 11 and 12: Penalty methods and augmented Lagrangian methods for nonlinear programming

Lecture 11 and 12: Penalty methods and augmented Lagrangian methods for nonlinear programming Lecture 11 and 12: Penalty methods and augmented Lagrangian methods for nonlinear programming Coralia Cartis, Mathematical Institute, University of Oxford C6.2/B2: Continuous Optimization Lecture 11 and

More information

4TE3/6TE3. Algorithms for. Continuous Optimization

4TE3/6TE3. Algorithms for. Continuous Optimization 4TE3/6TE3 Algorithms for Continuous Optimization (Algorithms for Constrained Nonlinear Optimization Problems) Tamás TERLAKY Computing and Software McMaster University Hamilton, November 2005 terlaky@mcmaster.ca

More information

AM 205: lecture 19. Last time: Conditions for optimality, Newton s method for optimization Today: survey of optimization methods

AM 205: lecture 19. Last time: Conditions for optimality, Newton s method for optimization Today: survey of optimization methods AM 205: lecture 19 Last time: Conditions for optimality, Newton s method for optimization Today: survey of optimization methods Quasi-Newton Methods General form of quasi-newton methods: x k+1 = x k α

More information

Written Examination

Written Examination Division of Scientific Computing Department of Information Technology Uppsala University Optimization Written Examination 202-2-20 Time: 4:00-9:00 Allowed Tools: Pocket Calculator, one A4 paper with notes

More information

On sequential optimality conditions for constrained optimization. José Mario Martínez martinez

On sequential optimality conditions for constrained optimization. José Mario Martínez  martinez On sequential optimality conditions for constrained optimization José Mario Martínez www.ime.unicamp.br/ martinez UNICAMP, Brazil 2011 Collaborators This talk is based in joint papers with Roberto Andreani

More information

Evaluation complexity for nonlinear constrained optimization using unscaled KKT conditions and high-order models by E. G. Birgin, J. L. Gardenghi, J. M. Martínez, S. A. Santos and Ph. L. Toint Report NAXYS-08-2015

More information

Inexact-Restoration Method with Lagrangian Tangent Decrease and New Merit Function for Nonlinear Programming 1 2

Inexact-Restoration Method with Lagrangian Tangent Decrease and New Merit Function for Nonlinear Programming 1 2 Inexact-Restoration Method with Lagrangian Tangent Decrease and New Merit Function for Nonlinear Programming 1 2 J. M. Martínez 3 December 7, 1999. Revised June 13, 2007. 1 This research was supported

More information

1. Introduction. We consider the mathematical programming problem

1. Introduction. We consider the mathematical programming problem SIAM J. OPTIM. Vol. 15, No. 1, pp. 210 228 c 2004 Society for Industrial and Applied Mathematics NEWTON-TYPE METHODS FOR OPTIMIZATION PROBLEMS WITHOUT CONSTRAINT QUALIFICATIONS A. F. IZMAILOV AND M. V.

More information