Infeasibility Detection in Nonlinear Optimization
|
|
- Hester Perry
- 5 years ago
- Views:
Transcription
1 Infeasibility Detection in Nonlinear Optimization Frank E. Curtis, Lehigh University Hao Wang, Lehigh University SIAM Conference on Optimization 2011 May 16, 2011 (See also Infeasibility Detection in Nonlinear Programming, Figen Oztoprak, MS71) Infeasibility Detection in Nonlinear Optimization 1 of 35
2 Outline Motivation Penalty-SQO Method Penalty-Interior-Point Method Unified Framework(?) Infeasibility Detection in Nonlinear Optimization 2 of 35
3 Outline Motivation Penalty-SQO Method Penalty-Interior-Point Method Unified Framework(?) Infeasibility Detection in Nonlinear Optimization 3 of 35
4 Constrained minimization Is this how we should formulate optimization problems? min x s.t. f (x) ( c E (x) = 0 c I (x) 0 (OP) Infeasibility Detection in Nonlinear Optimization 4 of 35
5 Feasibility violation minimization Suppose that if (OP) is infeasible, then we want to solve» min v(x) := c E (x) x max{c I (FP) (x), 0} Many algorithms/codes do this already, by either switching from solving one problem to the other; transitioning from solving one problem to the other. But are they doing it efficiently? Infeasibility Detection in Nonlinear Optimization 5 of 35
6 Numerical experiments: Infeasible optimization problems Iterations and f evaluations for 8 infeasible optimization problems (2-3 variables): Prob. Ipopt Knitro-Direct Knitro-Active Filter Iter. Eval. Iter. Eval. Iter. Eval. Iter. Eval *10000 * *1060 * *76 * *10000 *43652 *10000 * Problems also run with Knitro-CG and LOQO, but they failed every time. (We want your infeasible test problems!) Infeasibility Detection in Nonlinear Optimization 6 of 35
7 Numerical experiments: Feasibility problems (solved directly) Iterations and f evaluations for 8 feasibility problems (2-3 variables): Problem Ipopt Knitro-Direct Knitro-Active Filter Iter. Eval. Iter. Eval. Iter. Eval. Iter. Eval = If we can switch/transition efficiently, then our current tools work well. Infeasibility Detection in Nonlinear Optimization 7 of 35
8 Rapid detection of infeasibility Problem type Global convergence Fast local convergence Feasible Infeasible? Infeasibility Detection in Nonlinear Optimization 8 of 35
9 Two-phase strategy Infeasibility Detection in Nonlinear Optimization 9 of 35
10 Two-phase strategy Exploratory step determines possible progress toward feasibility. Rapid infeasibility detection requires rapid transition/switch to feasibility steps. (Important to have consistency between ways the two steps measure feasibility.) Infeasibility Detection in Nonlinear Optimization 10 of 35
11 Outline Motivation Penalty-SQO Method Penalty-Interior-Point Method Unified Framework(?) Infeasibility Detection in Nonlinear Optimization 11 of 35
12 Penalty methods Penalization provides one mechanism for transitioning to infeasibility detection: min ρf (x) + x,r,s,t et r + e T s + e T t 8 >< c E (x) = r s s.t. c >: I (x) t (r, s, t) 0 (PP) How should ρ be updated? Infeasibility Detection in Nonlinear Optimization 12 of 35
13 Literature (Rich history of SQP methods) Fletcher, Leyffer (2002) Byrd, Gould, Nocedal (2005) Byrd, Nocedal, Waltz (2008) Byrd, Curtis, Nocedal (2010) Byrd, López-Calva, Nocedal (2010) Gould, Robinson (2010) Morales, Nocedal, Wu (2010) Infeasibility Detection in Nonlinear Optimization 13 of 35
14 Steering and FILTER strategies Steering methods solve a sequence of constrained subproblems: FILTER:... we make use of a property of the phase I algorithm in our QP solver. If an infeasible QP is detected, the solver exists with a solution of [an LP minimizing the l 1 norm of violated constraints]. Infeasibility Detection in Nonlinear Optimization 14 of 35
15 SQuID strategy Compute v k as the optimal value for min d,r,s,t et r + e T s + e T t 8 c E (x k ) + c E (x k ) T d = r s >< c I (x s.t. k ) + c I (x k ) T d t d >: (r, s, t) 0 Update ρ for rapid infeasibility detection (Byrd, Curtis, Nocedal, 2010) Compute d k as the search direction from min ρ(f (x k) + f (x k ) T d) + e T r + e T s + e T t + 1 d,r,s,t 2 dt H(x k, λ k ; ρ)d 8 c E (x k ) + c E (x k ) T d = r s >< c I (x k ) + c I (x k ) T d t s.t. e >: T r + e T s + e T t v k (r, s, t) 0 Update ρ for descent Infeasibility Detection in Nonlinear Optimization 15 of 35
16 Numerical experiments: Infeasible optimization problems Iterations and f evaluations for 8 infeasible optimization problems (2-3 variables): Prob. Filter SQuID Iter. Eval. Iter. Eval Infeasibility Detection in Nonlinear Optimization 16 of 35
17 Outline Motivation Penalty-SQO Method Penalty-Interior-Point Method Unified Framework(?) Infeasibility Detection in Nonlinear Optimization 17 of 35
18 Penalty and interior-point methods Constrained subproblems in penalty methods can be expensive: min ρf (x) + x,r,s,t et r + e T s + e T t 8 >< c E (x) = r s s.t. c >: I (x) t (r, s, t) 0 (PP) Interior-point methods are more efficient for large-scale problems: min f (x) µ X ln u i x,u 8 >< c E (x) = 0 s.t. c >: I (x) = u u 0 (IP) Infeasibility Detection in Nonlinear Optimization 18 of 35
19 Penalty-interior-point method Applying an interior-point reformulation to (PP): X min ρf (x) µ (ln r i + ln s i ) + X (ln t i + ln u i ) + e T r + e T s + e T t x,r,s,t,u ( c E (x) = r s s.t. c I (x) = t u (PIP) The optimization problem (OP) and feasibility problem (FP) can be solved via (PIP): µ 0 and ρ ρ > 0 to solve (OP). µ 0 and ρ 0 to solve (FP). Infeasibility Detection in Nonlinear Optimization 19 of 35
20 Literature Previous work with similar motivations: Jittorntrum and Osborne (1980) Polyak (1982, 1992, 2008) Breitfeld and Shanno (1994, 1996) Goldfarb, Polyak, Scheinberg, and Yuzefovich (1999) Gould, Orban, and Toint (2003) Chen and Goldfarb (2006, 2006) Benson, Sen, and Shanno (2008) Infeasibility Detection in Nonlinear Optimization 20 of 35
21 Literature Previous work with similar motivations: Jittorntrum and Osborne (1980) Polyak (1982, 1992, 2008) Breitfeld and Shanno (1994, 1996) Goldfarb, Polyak, Scheinberg, and Yuzefovich (1999) Gould, Orban, and Toint (2003) Chen and Goldfarb (2006, 2006) Benson, Sen, and Shanno (2008) Parameter updates are essential to have a practical algorithm. Conservative strategy: 1. Fix ρ > 0 and (approximately) solve (PIP) for µ If infeasible, then reduce ρ and go to step 1. (ρ reduced too slowly to ensure fast convergence for infeasible problems.) Infeasibility Detection in Nonlinear Optimization 21 of 35
22 PIPAL strategy Newton system has the form A ρ,µd = ρq 1 + µq 2 + q 3. Steering strategy now efficient! (Only one factorization required.) Solution for (ρ, µ) = (0, µ) indicates progress possible toward primal feasibility. Adjust ρ to aim for primal feasibility. Adjust µ to aim for dual feasibility and complementarity. Infeasibility Detection in Nonlinear Optimization 22 of 35
23 Conservative strategy ρ μ Each dot may require at least a few iterations. Each row may require the computational effort of an entire interior-point solve. For an infeasible problem, we need both ρ and µ to reduce to zero. Infeasibility Detection in Nonlinear Optimization 23 of 35
24 PIPAL strategy ρ μ Blue line separates acceptable and unacceptable ρ values (for primal feasibility). µ then set to ensure progress toward dual feasibility and complementarity. For an infeasible problem, ρ and µ can reduce rapidly. Infeasibility Detection in Nonlinear Optimization 24 of 35
25 Numerical results: Feasible problems (sample size = 422) Infeasibility Detection in Nonlinear Optimization 25 of 35
26 Numerical results: Feasible problems w/ ρ decrease (sample size = 136) Infeasibility Detection in Nonlinear Optimization 26 of 35
27 Numerical results: Degenerate problems (sample size = 114) Added constraints: c i (x) 2 0 Infeasibility Detection in Nonlinear Optimization 27 of 35
28 Numerical results: Infeasible problems (sample size = 90) Added constraints: c i (x) 2 1 Infeasibility Detection in Nonlinear Optimization 28 of 35
29 Outline Motivation Penalty-SQO Method Penalty-Interior-Point Method Unified Framework(?) Infeasibility Detection in Nonlinear Optimization 29 of 35
30 Constrained minimization Is this how we should formulate optimization problems? min x s.t. f (x) ( c E (x) = 0 c I (x) 0 (OP) That is, should the strategy be the following: Attempt to solve (OP), but... if it is infeasible, then solve» min v(x) := c E (x) x max{c I (FP) (x), 0} Infeasibility Detection in Nonlinear Optimization 30 of 35
31 Minimization over set of feasibility violation minimizers Let X be the solution set of feasibility violation minimization:» min v(x) := c E (x) x max{c I (FP) (x), 0} We should solve min x f (x) s.t. x X. (Even better would be a single problem or set of conditions for solving (UP).) (UP) Infeasibility Detection in Nonlinear Optimization 31 of 35
32 1 0 Consider the following model (AMPL presolver turned off): var x; minimize f: (x-1)^2; s.t. c: 1 <= 0; Solver Inf. Declared? Iterations Final (x, y) CONOPT No FILTER Yes IPOPT Yes KNITRO Yes LANCELOT Yes LOQO No MINOS Yes MOSEK Yes PENNON No SNOPT Yes Infeasibility Detection in Nonlinear Optimization 32 of 35
33 y 2 1 Consider the following model (AMPL presolver turned off): var x; var y; minimize f: (x-1)^2; s.t. c: y^2 + 1 <= 0; Solver Inf. Declared? Iterations Final (x, y) CONOPT Yes 3 (0.0000,0.0000) FILTER Yes 2 (0.0000,0.0000) IPOPT Yes 6 (0.0099,0.0000) KNITRO Yes 4 (0.3765,0.0000) LANCELOT Yes 1 (1.0000,0.0000) LOQO No 500 (0.4569,0.0000) MINOS Yes 2 (1.0000,0.0000) MOSEK Yes 1 (0.0000,0.0000) PENNON No 0 (1.0000,0.0000) SNOPT Yes 3 (1.0000,0.0000) Infeasibility Detection in Nonlinear Optimization 33 of 35
34 Penalty and switching methods Penalization provides one mechanism for transitioning to infeasibility detection: min ρf (x) + x,r,s,t et r + e T s + e T t 8 >< c E (x) = r s s.t. c >: I (x) t (r, s, t) 0 (PP) However, once x X is found, f (x) has been thrown out (since ρ 0). Switching methods throw out f (x) even earlier. Infeasibility Detection in Nonlinear Optimization 34 of 35
35 What is needed? General-purpose algorithms that: require no constraint qualifications to do something useful minimize the objective over the set of feasibility violation minimizers are as efficient and robust as contemporary methods when problems are nice Infeasibility Detection in Nonlinear Optimization 35 of 35
Infeasibility Detection and an Inexact Active-Set Method for Large-Scale Nonlinear Optimization
Infeasibility Detection and an Inexact Active-Set Method for Large-Scale Nonlinear Optimization Frank E. Curtis, Lehigh University involving joint work with James V. Burke, University of Washington Daniel
More informationA New Penalty-SQP Method
Background and Motivation Illustration of Numerical Results Final Remarks Frank E. Curtis Informs Annual Meeting, October 2008 Background and Motivation Illustration of Numerical Results Final Remarks
More informationRecent Adaptive Methods for Nonlinear Optimization
Recent Adaptive Methods for Nonlinear Optimization Frank E. Curtis, Lehigh University involving joint work with James V. Burke (U. of Washington), Richard H. Byrd (U. of Colorado), Nicholas I. M. Gould
More informationAn Inexact Sequential Quadratic Optimization Method for Nonlinear Optimization
An Inexact Sequential Quadratic Optimization Method for Nonlinear Optimization Frank E. Curtis, Lehigh University involving joint work with Travis Johnson, Northwestern University Daniel P. Robinson, Johns
More informationNonlinear Optimization Solvers
ICE05 Argonne, July 19, 2005 Nonlinear Optimization Solvers Sven Leyffer leyffer@mcs.anl.gov Mathematics & Computer Science Division, Argonne National Laboratory Nonlinear Optimization Methods Optimization
More informationA Trust Funnel Algorithm for Nonconvex Equality Constrained Optimization with O(ɛ 3/2 ) Complexity
A Trust Funnel Algorithm for Nonconvex Equality Constrained Optimization with O(ɛ 3/2 ) Complexity Mohammadreza Samadi, Lehigh University joint work with Frank E. Curtis (stand-in presenter), Lehigh University
More informationSteering Exact Penalty Methods for Nonlinear Programming
Steering Exact Penalty Methods for Nonlinear Programming Richard H. Byrd Jorge Nocedal Richard A. Waltz April 10, 2007 Technical Report Optimization Technology Center Northwestern University Evanston,
More informationSurvey of NLP Algorithms. L. T. Biegler Chemical Engineering Department Carnegie Mellon University Pittsburgh, PA
Survey of NLP Algorithms L. T. Biegler Chemical Engineering Department Carnegie Mellon University Pittsburgh, PA NLP Algorithms - Outline Problem and Goals KKT Conditions and Variable Classification Handling
More informationINTERIOR-POINT METHODS FOR NONCONVEX NONLINEAR PROGRAMMING: CONVERGENCE ANALYSIS AND COMPUTATIONAL PERFORMANCE
INTERIOR-POINT METHODS FOR NONCONVEX NONLINEAR PROGRAMMING: CONVERGENCE ANALYSIS AND COMPUTATIONAL PERFORMANCE HANDE Y. BENSON, ARUN SEN, AND DAVID F. SHANNO Abstract. In this paper, we present global
More informationAlgorithms for Constrained Optimization
1 / 42 Algorithms for Constrained Optimization ME598/494 Lecture Max Yi Ren Department of Mechanical Engineering, Arizona State University April 19, 2015 2 / 42 Outline 1. Convergence 2. Sequential quadratic
More informationA Primal-Dual Augmented Lagrangian Penalty-Interior-Point Filter Line Search Algorithm
Journal name manuscript No. (will be inserted by the editor) A Primal-Dual Augmented Lagrangian Penalty-Interior-Point Filter Line Search Algorithm Rene Kuhlmann Christof Büsens Received: date / Accepted:
More informationA Trust-Funnel Algorithm for Nonlinear Programming
for Nonlinear Programming Daniel P. Robinson Johns Hopins University Department of Applied Mathematics and Statistics Collaborators: Fran E. Curtis (Lehigh University) Nic I. M. Gould (Rutherford Appleton
More informationMS&E 318 (CME 338) Large-Scale Numerical Optimization
Stanford University, Management Science & Engineering (and ICME) MS&E 318 (CME 338) Large-Scale Numerical Optimization 1 Origins Instructor: Michael Saunders Spring 2015 Notes 9: Augmented Lagrangian Methods
More informationCONVERGENCE ANALYSIS OF AN INTERIOR-POINT METHOD FOR NONCONVEX NONLINEAR PROGRAMMING
CONVERGENCE ANALYSIS OF AN INTERIOR-POINT METHOD FOR NONCONVEX NONLINEAR PROGRAMMING HANDE Y. BENSON, ARUN SEN, AND DAVID F. SHANNO Abstract. In this paper, we present global and local convergence results
More informationA Penalty-Interior-Point Algorithm for Nonlinear Constrained Optimization
Noname manuscript No. (will be inserted by the editor) A Penalty-Interior-Point Algorithm for Nonlinear Constrained Optimization Frank E. Curtis April 26, 2011 Abstract Penalty and interior-point methods
More informationWhat s New in Active-Set Methods for Nonlinear Optimization?
What s New in Active-Set Methods for Nonlinear Optimization? Philip E. Gill Advances in Numerical Computation, Manchester University, July 5, 2011 A Workshop in Honor of Sven Hammarling UCSD Center for
More informationInexact Newton Methods and Nonlinear Constrained Optimization
Inexact Newton Methods and Nonlinear Constrained Optimization Frank E. Curtis EPSRC Symposium Capstone Conference Warwick Mathematics Institute July 2, 2009 Outline PDE-Constrained Optimization Newton
More informationImplementation of an Interior Point Multidimensional Filter Line Search Method for Constrained Optimization
Proceedings of the 5th WSEAS Int. Conf. on System Science and Simulation in Engineering, Tenerife, Canary Islands, Spain, December 16-18, 2006 391 Implementation of an Interior Point Multidimensional Filter
More informationAlgorithms for constrained local optimization
Algorithms for constrained local optimization Fabio Schoen 2008 http://gol.dsi.unifi.it/users/schoen Algorithms for constrained local optimization p. Feasible direction methods Algorithms for constrained
More informationEffective reformulations of the truss topology design problem
Effective reformulations of the truss topology design problem Michal Kočvara and Jiří V. Outrata Abstract We present a new formulation of the truss topology problem that results in unique design and unique
More informationA SHIFTED PRIMAL-DUAL PENALTY-BARRIER METHOD FOR NONLINEAR OPTIMIZATION
A SHIFTED PRIMAL-DUAL PENALTY-BARRIER METHOD FOR NONLINEAR OPTIMIZATION Philip E. Gill Vyacheslav Kungurtsev Daniel P. Robinson UCSD Center for Computational Mathematics Technical Report CCoM-19-3 March
More informationA null-space primal-dual interior-point algorithm for nonlinear optimization with nice convergence properties
A null-space primal-dual interior-point algorithm for nonlinear optimization with nice convergence properties Xinwei Liu and Yaxiang Yuan Abstract. We present a null-space primal-dual interior-point algorithm
More informationAn Inexact Newton Method for Optimization
New York University Brown Applied Mathematics Seminar, February 10, 2009 Brief biography New York State College of William and Mary (B.S.) Northwestern University (M.S. & Ph.D.) Courant Institute (Postdoc)
More informationPDE-Constrained and Nonsmooth Optimization
Frank E. Curtis October 1, 2009 Outline PDE-Constrained Optimization Introduction Newton s method Inexactness Results Summary and future work Nonsmooth Optimization Sequential quadratic programming (SQP)
More informationThe use of second-order information in structural topology optimization. Susana Rojas Labanda, PhD student Mathias Stolpe, Senior researcher
The use of second-order information in structural topology optimization Susana Rojas Labanda, PhD student Mathias Stolpe, Senior researcher What is Topology Optimization? Optimize the design of a structure
More informationSolving MPECs Implicit Programming and NLP Methods
Solving MPECs Implicit Programming and NLP Methods Michal Kočvara Academy of Sciences of the Czech Republic September 2005 1 Mathematical Programs with Equilibrium Constraints Mechanical motivation Mechanical
More informationAn Inexact Newton Method for Nonlinear Constrained Optimization
An Inexact Newton Method for Nonlinear Constrained Optimization Frank E. Curtis Numerical Analysis Seminar, January 23, 2009 Outline Motivation and background Algorithm development and theoretical results
More informationA Penalty-Interior-Point Algorithm for Nonlinear Constrained Optimization
Noname manuscript No. (will be inserted by the editor) A Penalty-Interior-Point Algorithm for Nonlinear Constrained Optimization Fran E. Curtis February, 22 Abstract Penalty and interior-point methods
More informationNUMERICAL OPTIMIZATION. J. Ch. Gilbert
NUMERICAL OPTIMIZATION J. Ch. Gilbert Numerical optimization (past) The discipline deals with the classical smooth (nonconvex) problem min {f(x) : c E (x) = 0, c I (x) 0}. Applications: variable added
More informationA Trust-region-based Sequential Quadratic Programming Algorithm
Downloaded from orbit.dtu.dk on: Oct 19, 2018 A Trust-region-based Sequential Quadratic Programming Algorithm Henriksen, Lars Christian; Poulsen, Niels Kjølstad Publication date: 2010 Document Version
More informationInterior Methods for Mathematical Programs with Complementarity Constraints
Interior Methods for Mathematical Programs with Complementarity Constraints Sven Leyffer, Gabriel López-Calva and Jorge Nocedal July 14, 25 Abstract This paper studies theoretical and practical properties
More informationarxiv: v1 [math.na] 8 Jun 2018
arxiv:1806.03347v1 [math.na] 8 Jun 2018 Interior Point Method with Modified Augmented Lagrangian for Penalty-Barrier Nonlinear Programming Martin Neuenhofen June 12, 2018 Abstract We present a numerical
More informationEffective reformulations of the truss topology design problem
Effective reformulations of the truss topology design problem Michal Kočvara and Jiří V. Outrata Abstract We present a new formulation of the truss topology problem that results in unique design and unique
More informationLecture 15: SQP methods for equality constrained optimization
Lecture 15: SQP methods for equality constrained optimization Coralia Cartis, Mathematical Institute, University of Oxford C6.2/B2: Continuous Optimization Lecture 15: SQP methods for equality constrained
More informationFlexible Penalty Functions for Nonlinear Constrained Optimization
IMA Journal of Numerical Analysis (2005) Page 1 of 19 doi: 10.1093/imanum/dri017 Flexible Penalty Functions for Nonlinear Constrained Optimization FRANK E. CURTIS Department of Industrial Engineering and
More informationApolynomialtimeinteriorpointmethodforproblemswith nonconvex constraints
Apolynomialtimeinteriorpointmethodforproblemswith nonconvex constraints Oliver Hinder, Yinyu Ye Department of Management Science and Engineering Stanford University June 28, 2018 The problem I Consider
More informationAn Active Set Strategy for Solving Optimization Problems with up to 200,000,000 Nonlinear Constraints
An Active Set Strategy for Solving Optimization Problems with up to 200,000,000 Nonlinear Constraints Klaus Schittkowski Department of Computer Science, University of Bayreuth 95440 Bayreuth, Germany e-mail:
More informationMS&E 318 (CME 338) Large-Scale Numerical Optimization
Stanford University, Management Science & Engineering (and ICME) MS&E 318 (CME 338) Large-Scale Numerical Optimization Instructor: Michael Saunders Spring 2015 Notes 11: NPSOL and SNOPT SQP Methods 1 Overview
More informationHYBRID FILTER METHODS FOR NONLINEAR OPTIMIZATION. Yueling Loh
HYBRID FILTER METHODS FOR NONLINEAR OPTIMIZATION by Yueling Loh A dissertation submitted to The Johns Hopkins University in conformity with the requirements for the degree of Doctor of Philosophy. Baltimore,
More informationDerivative-Based Numerical Method for Penalty-Barrier Nonlinear Programming
Derivative-Based Numerical Method for Penalty-Barrier Nonlinear Programming Martin Peter Neuenhofen October 26, 2018 Abstract We present an NLP solver for nonlinear optimization with quadratic penalty
More informationINTERIOR-POINT ALGORITHMS, PENALTY METHODS AND EQUILIBRIUM PROBLEMS
INTERIOR-POINT ALGORITHMS, PENALTY METHODS AND EQUILIBRIUM PROBLEMS HANDE Y. BENSON, ARUN SEN, DAVID F. SHANNO, AND ROBERT J. VANDERBEI Abstract. In this paper we consider the question of solving equilibrium
More informationSF2822 Applied Nonlinear Optimization. Preparatory question. Lecture 9: Sequential quadratic programming. Anders Forsgren
SF2822 Applied Nonlinear Optimization Lecture 9: Sequential quadratic programming Anders Forsgren SF2822 Applied Nonlinear Optimization, KTH / 24 Lecture 9, 207/208 Preparatory question. Try to solve theory
More informationNumerical Nonlinear Optimization with WORHP
Numerical Nonlinear Optimization with WORHP Christof Büskens Optimierung & Optimale Steuerung London, 8.9.2011 Optimization & Optimal Control Nonlinear Optimization WORHP Concept Ideas Features Results
More informationTrust-Region SQP Methods with Inexact Linear System Solves for Large-Scale Optimization
Trust-Region SQP Methods with Inexact Linear System Solves for Large-Scale Optimization Denis Ridzal Department of Computational and Applied Mathematics Rice University, Houston, Texas dridzal@caam.rice.edu
More informationINTERIOR-POINT ALGORITHMS, PENALTY METHODS AND EQUILIBRIUM PROBLEMS
INTERIOR-POINT ALGORITHMS, PENALTY METHODS AND EQUILIBRIUM PROBLEMS HANDE Y. BENSON, ARUN SEN, DAVID F. SHANNO, AND ROBERT J. VANDERBEI Abstract. In this paper we consider the question of solving equilibrium
More informationINTERIOR-POINT METHODS FOR NONLINEAR, SECOND-ORDER CONE, AND SEMIDEFINITE PROGRAMMING. Hande Yurttan Benson
INTERIOR-POINT METHODS FOR NONLINEAR, SECOND-ORDER CONE, AND SEMIDEFINITE PROGRAMMING Hande Yurttan Benson A DISSERTATION PRESENTED TO THE FACULTY OF PRINCETON UNIVERSITY IN CANDIDACY FOR THE DEGREE OF
More informationAdaptive augmented Lagrangian methods: algorithms and practical numerical experience
Optimization Methods and Software ISSN: 1055-6788 (Print) 1029-4937 (Online) Journal homepage: http://www.tandfonline.com/loi/goms20 Adaptive augmented Lagrangian methods: algorithms and practical numerical
More informationAM 205: lecture 19. Last time: Conditions for optimality Today: Newton s method for optimization, survey of optimization methods
AM 205: lecture 19 Last time: Conditions for optimality Today: Newton s method for optimization, survey of optimization methods Optimality Conditions: Equality Constrained Case As another example of equality
More informationA PRIMAL-DUAL TRUST REGION ALGORITHM FOR NONLINEAR OPTIMIZATION
Optimization Technical Report 02-09, October 2002, UW-Madison Computer Sciences Department. E. Michael Gertz 1 Philip E. Gill 2 A PRIMAL-DUAL TRUST REGION ALGORITHM FOR NONLINEAR OPTIMIZATION 7 October
More informationSecond-Order Cone Program (SOCP) Detection and Transformation Algorithms for Optimization Software
and Second-Order Cone Program () and Algorithms for Optimization Software Jared Erickson JaredErickson2012@u.northwestern.edu Robert 4er@northwestern.edu Northwestern University INFORMS Annual Meeting,
More informationPenalty and Barrier Methods General classical constrained minimization problem minimize f(x) subject to g(x) 0 h(x) =0 Penalty methods are motivated by the desire to use unconstrained optimization techniques
More informationA STABILIZED SQP METHOD: SUPERLINEAR CONVERGENCE
A STABILIZED SQP METHOD: SUPERLINEAR CONVERGENCE Philip E. Gill Vyacheslav Kungurtsev Daniel P. Robinson UCSD Center for Computational Mathematics Technical Report CCoM-14-1 June 30, 2014 Abstract Regularized
More informationOn the use of piecewise linear models in nonlinear programming
Math. Program., Ser. A (2013) 137:289 324 DOI 10.1007/s10107-011-0492-9 FULL LENGTH PAPER On the use of piecewise linear models in nonlinear programming Richard H. Byrd Jorge Nocedal Richard A. Waltz Yuchen
More informationSeminal papers in nonlinear optimization
Seminal papers in nonlinear optimization Nick Gould, CSED, RAL, Chilton, OX11 0QX, England (n.gould@rl.ac.uk) December 7, 2006 The following papers are classics in the field. Although many of them cover
More information10 Numerical methods for constrained problems
10 Numerical methods for constrained problems min s.t. f(x) h(x) = 0 (l), g(x) 0 (m), x X The algorithms can be roughly divided the following way: ˆ primal methods: find descent direction keeping inside
More informationA GLOBALLY CONVERGENT STABILIZED SQP METHOD: SUPERLINEAR CONVERGENCE
A GLOBALLY CONVERGENT STABILIZED SQP METHOD: SUPERLINEAR CONVERGENCE Philip E. Gill Vyacheslav Kungurtsev Daniel P. Robinson UCSD Center for Computational Mathematics Technical Report CCoM-14-1 June 30,
More informationHot-Starting NLP Solvers
Hot-Starting NLP Solvers Andreas Wächter Department of Industrial Engineering and Management Sciences Northwestern University waechter@iems.northwestern.edu 204 Mixed Integer Programming Workshop Ohio
More informationAn augmented Lagrangian method for equality constrained optimization with rapid infeasibility detection capabilities
An augmented Lagrangian method for equality constrained optimization with rapid infeasibility detection capabilities Paul Armand, Ngoc Nguyen Tran To cite this version: Paul Armand, Ngoc Nguyen Tran. An
More informationUsing Interior-Point Methods within Mixed-Integer Nonlinear Programming
Using Interior-Point Methods within Mixed-Integer Nonlinear Programming. Hande Y. Benson Drexel University IMA - MINLP - p. 1/34 Motivation: Discrete Variables Motivation: Discrete Variables Interior-Point
More informationAn Interior Algorithm for Nonlinear Optimization That Combines Line Search and Trust Region Steps
An Interior Algorithm for Nonlinear Optimization That Combines Line Search and Trust Region Steps R.A. Waltz J.L. Morales J. Nocedal D. Orban September 8, 2004 Abstract An interior-point method for nonlinear
More informationINTERIOR-POINT METHODS FOR NONCONVEX NONLINEAR PROGRAMMING: COMPLEMENTARITY CONSTRAINTS
INTERIOR-POINT METHODS FOR NONCONVEX NONLINEAR PROGRAMMING: COMPLEMENTARITY CONSTRAINTS HANDE Y. BENSON, DAVID F. SHANNO, AND ROBERT J. VANDERBEI Operations Research and Financial Engineering Princeton
More informationA SHIFTED PRIMAL-DUAL INTERIOR METHOD FOR NONLINEAR OPTIMIZATION
A SHIFTED RIMAL-DUAL INTERIOR METHOD FOR NONLINEAR OTIMIZATION hilip E. Gill Vyacheslav Kungurtsev Daniel. Robinson UCSD Center for Computational Mathematics Technical Report CCoM-18-1 February 1, 2018
More informationUniversity of Erlangen-Nürnberg and Academy of Sciences of the Czech Republic. Solving MPECs by Implicit Programming and NLP Methods p.
Solving MPECs by Implicit Programming and NLP Methods Michal Kočvara University of Erlangen-Nürnberg and Academy of Sciences of the Czech Republic Solving MPECs by Implicit Programming and NLP Methods
More information5 Handling Constraints
5 Handling Constraints Engineering design optimization problems are very rarely unconstrained. Moreover, the constraints that appear in these problems are typically nonlinear. This motivates our interest
More informationLecture 13: Constrained optimization
2010-12-03 Basic ideas A nonlinearly constrained problem must somehow be converted relaxed into a problem which we can solve (a linear/quadratic or unconstrained problem) We solve a sequence of such problems
More informationAM 205: lecture 19. Last time: Conditions for optimality, Newton s method for optimization Today: survey of optimization methods
AM 205: lecture 19 Last time: Conditions for optimality, Newton s method for optimization Today: survey of optimization methods Quasi-Newton Methods General form of quasi-newton methods: x k+1 = x k α
More informationMultidisciplinary System Design Optimization (MSDO)
Multidisciplinary System Design Optimization (MSDO) Numerical Optimization II Lecture 8 Karen Willcox 1 Massachusetts Institute of Technology - Prof. de Weck and Prof. Willcox Today s Topics Sequential
More informationREGULARIZED SEQUENTIAL QUADRATIC PROGRAMMING METHODS
REGULARIZED SEQUENTIAL QUADRATIC PROGRAMMING METHODS Philip E. Gill Daniel P. Robinson UCSD Department of Mathematics Technical Report NA-11-02 October 2011 Abstract We present the formulation and analysis
More informationarxiv: v1 [math.oc] 20 Aug 2014
Adaptive Augmented Lagrangian Methods: Algorithms and Practical Numerical Experience Detailed Version Frank E. Curtis, Nicholas I. M. Gould, Hao Jiang, and Daniel P. Robinson arxiv:1408.4500v1 [math.oc]
More informationLarge-Scale Nonlinear Optimization with Inexact Step Computations
Large-Scale Nonlinear Optimization with Inexact Step Computations Andreas Wächter IBM T.J. Watson Research Center Yorktown Heights, New York andreasw@us.ibm.com IPAM Workshop on Numerical Methods for Continuous
More informationNumerical Methods for PDE-Constrained Optimization
Numerical Methods for PDE-Constrained Optimization Richard H. Byrd 1 Frank E. Curtis 2 Jorge Nocedal 2 1 University of Colorado at Boulder 2 Northwestern University Courant Institute of Mathematical Sciences,
More informationThe Squared Slacks Transformation in Nonlinear Programming
Technical Report No. n + P. Armand D. Orban The Squared Slacks Transformation in Nonlinear Programming August 29, 2007 Abstract. We recall the use of squared slacks used to transform inequality constraints
More informationAN EXACT PRIMAL-DUAL PENALTY METHOD APPROACH TO WARMSTARTING INTERIOR-POINT METHODS FOR LINEAR PROGRAMMING
AN EXACT PRIMAL-DUAL PENALTY METHOD APPROACH TO WARMSTARTING INTERIOR-POINT METHODS FOR LINEAR PROGRAMMING HANDE Y. BENSON AND DAVID F. SHANNO Abstract. One perceived deficiency of interior-point methods
More informationQuasi-Newton Methods. Zico Kolter (notes by Ryan Tibshirani, Javier Peña, Zico Kolter) Convex Optimization
Quasi-Newton Methods Zico Kolter (notes by Ryan Tibshirani, Javier Peña, Zico Kolter) Convex Optimization 10-725 Last time: primal-dual interior-point methods Given the problem min x f(x) subject to h(x)
More informationQuasi-Newton Methods. Javier Peña Convex Optimization /36-725
Quasi-Newton Methods Javier Peña Convex Optimization 10-725/36-725 Last time: primal-dual interior-point methods Consider the problem min x subject to f(x) Ax = b h(x) 0 Assume f, h 1,..., h m are convex
More informationPart 4: Active-set methods for linearly constrained optimization. Nick Gould (RAL)
Part 4: Active-set methods for linearly constrained optimization Nick Gould RAL fx subject to Ax b Part C course on continuoue optimization LINEARLY CONSTRAINED MINIMIZATION fx subject to Ax { } b where
More informationNewton s Method. Ryan Tibshirani Convex Optimization /36-725
Newton s Method Ryan Tibshirani Convex Optimization 10-725/36-725 1 Last time: dual correspondences Given a function f : R n R, we define its conjugate f : R n R, Properties and examples: f (y) = max x
More informationNonlinear Programming, Elastic Mode, SQP, MPEC, MPCC, complementarity
Preprint ANL/MCS-P864-1200 ON USING THE ELASTIC MODE IN NONLINEAR PROGRAMMING APPROACHES TO MATHEMATICAL PROGRAMS WITH COMPLEMENTARITY CONSTRAINTS MIHAI ANITESCU Abstract. We investigate the possibility
More informationPOWER SYSTEMS in general are currently operating
TO APPEAR IN IEEE TRANSACTIONS ON POWER SYSTEMS 1 Robust Optimal Power Flow Solution Using Trust Region and Interior-Point Methods Andréa A. Sousa, Geraldo L. Torres, Member IEEE, Claudio A. Cañizares,
More informationHow the augmented Lagrangian algorithm can deal with an infeasible convex quadratic optimization problem. Motivation, analysis, implementation
Journées Franciliennes de Recherche Opérationnelle Conservatoire National des Arts et Métiers, Paris, France How the augmented Lagrangian algorithm can deal with an infeasible convex quadratic optimization
More informationLecture 3: Lagrangian duality and algorithms for the Lagrangian dual problem
Lecture 3: Lagrangian duality and algorithms for the Lagrangian dual problem Michael Patriksson 0-0 The Relaxation Theorem 1 Problem: find f := infimum f(x), x subject to x S, (1a) (1b) where f : R n R
More information1. Introduction. We consider nonlinear optimization problems of the form. f(x) ce (x) = 0 c I (x) 0,
AN INTERIOR-POINT ALGORITHM FOR LARGE-SCALE NONLINEAR OPTIMIZATION WITH INEXACT STEP COMPUTATIONS FRANK E. CURTIS, OLAF SCHENK, AND ANDREAS WÄCHTER Abstract. We present a line-search algorithm for large-scale
More informationSolving a Signalized Traffic Intersection Problem with NLP Solvers
Solving a Signalized Traffic Intersection Problem with NLP Solvers Teófilo Miguel M. Melo, João Luís H. Matias, M. Teresa T. Monteiro CIICESI, School of Technology and Management of Felgueiras, Polytechnic
More informationOn the behavior of Lagrange multipliers in convex and non-convex infeasible interior point methods
On the behavior of Lagrange multipliers in convex and non-convex infeasible interior point methods Gabriel Haeser Oliver Hinder Yinyu Ye July 23, 2017 Abstract This paper analyzes sequences generated by
More informationPreprint ANL/MCS-P , Dec 2002 (Revised Nov 2003, Mar 2004) Mathematics and Computer Science Division Argonne National Laboratory
Preprint ANL/MCS-P1015-1202, Dec 2002 (Revised Nov 2003, Mar 2004) Mathematics and Computer Science Division Argonne National Laboratory A GLOBALLY CONVERGENT LINEARLY CONSTRAINED LAGRANGIAN METHOD FOR
More informationOverview of course. Introduction to Optimization, DIKU Monday 12 November David Pisinger
Introduction to Optimization, DIKU 007-08 Monday November David Pisinger Lecture What is OR, linear models, standard form, slack form, simplex repetition, graphical interpretation, extreme points, basic
More information2.098/6.255/ Optimization Methods Practice True/False Questions
2.098/6.255/15.093 Optimization Methods Practice True/False Questions December 11, 2009 Part I For each one of the statements below, state whether it is true or false. Include a 1-3 line supporting sentence
More informationConvergence analysis of a primal-dual interior-point method for nonlinear programming
Convergence analysis of a primal-dual interior-point method for nonlinear programming Igor Griva David F. Shanno Robert J. Vanderbei August 19, 2005 Abstract We analyze a primal-dual interior-point method
More informationA New Low Rank Quasi-Newton Update Scheme for Nonlinear Programming
A New Low Rank Quasi-Newton Update Scheme for Nonlinear Programming Roger Fletcher Numerical Analysis Report NA/223, August 2005 Abstract A new quasi-newton scheme for updating a low rank positive semi-definite
More informationEttore Majorana Centre for Scientific Culture International School of Mathematics G. Stampacchia. Erice, Italy. 33rd Workshop
Ettore Majorana Centre for Scientific Culture International School of Mathematics G. Stampacchia Erice, Italy 33rd Workshop HIGH PERFORMANCE ALGORITHMS AND SOFTWARE FOR NONLINEAR OPTIMIZATION 30 June -
More informationOn Lagrange multipliers of trust region subproblems
On Lagrange multipliers of trust region subproblems Ladislav Lukšan, Ctirad Matonoha, Jan Vlček Institute of Computer Science AS CR, Prague Applied Linear Algebra April 28-30, 2008 Novi Sad, Serbia Outline
More information1 1 Introduction Some of the most successful algorithms for large-scale, generally constrained, nonlinear optimization fall into one of two categories
An Active-Set Algorithm for Nonlinear Programming Using Linear Programming and Equality Constrained Subproblems Richard H. Byrd Nicholas I.M. Gould y Jorge Nocedal z Richard A. Waltz z September 30, 2002
More informationA penalty-interior-point algorithm for nonlinear constrained optimization
Math. Prog. Comp. (2012) 4:181 209 DOI 10.1007/s12532-012-0041-4 FULL LENGTH PAPER A penalty-interior-point algorithm for nonlinear constrained optimization Fran E. Curtis Received: 26 April 2011 / Accepted:
More informationOn the complexity of an Inexact Restoration method for constrained optimization
On the complexity of an Inexact Restoration method for constrained optimization L. F. Bueno J. M. Martínez September 18, 2018 Abstract Recent papers indicate that some algorithms for constrained optimization
More informationarxiv: v2 [math.oc] 11 Jan 2018
A one-phase interior point method for nonconvex optimization Oliver Hinder, Yinyu Ye January 12, 2018 arxiv:1801.03072v2 [math.oc] 11 Jan 2018 Abstract The work of Wächter and Biegler [40] suggests that
More informationFollowing The Central Trajectory Using The Monomial Method Rather Than Newton's Method
Following The Central Trajectory Using The Monomial Method Rather Than Newton's Method Yi-Chih Hsieh and Dennis L. Bricer Department of Industrial Engineering The University of Iowa Iowa City, IA 52242
More informationA Branch-and-Refine Method for Nonconvex Mixed-Integer Optimization
A Branch-and-Refine Method for Nonconvex Mixed-Integer Optimization Sven Leyffer 2 Annick Sartenaer 1 Emilie Wanufelle 1 1 University of Namur, Belgium 2 Argonne National Laboratory, USA IMA Workshop,
More information2 P. E. GILL, W. MURRAY AND M. A. SAUNDERS We assume that the nonlinear functions are smooth and that their rst derivatives are available (and possibl
SNOPT: AN SQP ALGORITHM FOR LARGE-SCALE CONSTRAINED OPTIMIZATION PHILIP E. GILL y, WALTER MURRAY z, AND MICHAEL A. SAUNDERS z Abstract. Sequential quadratic programming (SQP) methods have proved highly
More informationA STABILIZED SQP METHOD: GLOBAL CONVERGENCE
A STABILIZED SQP METHOD: GLOBAL CONVERGENCE Philip E. Gill Vyacheslav Kungurtsev Daniel P. Robinson UCSD Center for Computational Mathematics Technical Report CCoM-13-4 Revised July 18, 2014, June 23,
More informationThe Surprisingly Complicated Case of [Convex] Quadratic Optimization
The Surprisingly Complicated Case of [Convex] Quadratic Optimization Robert Fourer 4er@ampl.com AMPL Optimization Inc. www.ampl.com +1 773-336-AMPL U.S.-Mexico Workshop on Optimization and Its Applications
More information