SOLVING LARGE-SCALE OPTIMIZATION PROBLEMS WITH GAMS AND MINOS
|
|
- Baldric Sparks
- 6 years ago
- Views:
Transcription
1 SOLVING LARGE-SCALE OPTIMIZATION PROBLEMS WITH GAMS AND MINOS Michael A. SAUNDERS Department of Operations Research Stanford University Stanford, California OPERATIONAL RESEARCH SOCIETY OF NEW ZEALAND 24TH ANNUAL CONFERENCE University of Auckland August 18 19, 1988 Abstract Operations researchers have long used modeling languages to construct large linear programming models. The modeling system GAMS and the optimization system MINOS now allow the development of large nonlinear models with relative ease. We discuss some applications of GAMS and MINOS. At the same time we promote the philosophy of putting all variables and equations into the optimization model, rather than eliminating variables.
2 2 TEMPORARY VARIABLES World War 2 Fortran DO 20 J=1,N CS=A(J,J)/SQRT(A(J,J)*A(J,J)+Y(J)*Y(J)) SN=Y(J)/SQRT(A(J,J)*A(J,J)+Y(J)*Y(J)) JM1=J-1 IF (JM1.LT.0) GOTO 20 DO 10 I=1,JM1 A(I,J)=CS*A(I,J)+SN*Y(J) 10 Y(J)=SN*A(I,J)-CS*Y(J) 20 CONTINUE Spot the deliberate error! Fortran 77 DO 20 J = 1, N X = SQRT( A(J,J)**2 + Y(J)**2 ) CS = A(J,J) / X SN = Y(J) / X DO 10 I = 1, J - 1 AIJ = A(I,J) A(I,J) = CS*AIJ + SN*Y(J) Y(J) = SN*AIJ - CS*Y(J) 10 CONTINUE 20 CONTINUE
3 3 TWO TYPES OF VARIABLE Linear least squares Given matrix A and vector b, find y such that Ay b: Alternatively: minimize y (b Ay) T (b Ay). minimize x, y φ(x, y) = x T x subject to x + Ay = b, where x is the residual vector. Generalized least squares minimize y (b Ay) T W 1 (b Ay). Assume W = LL T is positive definite (or semi-definite): This is a (special) QP. minimize x, y φ(x, y) = x T x subject to Lx + Ay = b. The data A and b appear only once. The partial derivatives φ x, φ y Does not matter if W is singular. are trivial. Often L is known anyway (rather than W ). x, y sometimes called state variables and control variables.
4 4 BLACK BOXES For given control variables y, there is often black box software available to compute x satisfying the nonlinear equations f(x, y) = b. Electrical power transmission ( Load Flow ) Oil refinery models Steady-state flowsheeting Dynamic process simulation Design of pipe with leaks (Water Studies, Stanford) Aquifer reclamation (maybe) Obvious wish: to optimize with respect to y.
5 5 PROCESS OPTIMIZATION minimize φ(x, y) subject to f(x, y) = b. The natural inclination is to keep using the immensely complicated and expensive black boxes, and simply build an optimization system around them: 0. Set k = 0. Choose y Solve f(x, y k ) = b to obtain x = x k. 2. Evaluate objective function φ k = φ(x k, y k ). 3. Estimate derivatives of objective function wrto y; i.e., for each y j, ȳ y k + δe j. Solve f( x, ȳ) = b. g j (φ( x, ȳ) φ k )/δ. 4. Use gradient vector g to obtain a search direction p. 5. y k+1 y k + αp for some steplength α. 6. k k + 1. Repeat from Step 1. This approach is far from ideal.
6 6 DISADVANTAGES OF BLACK BOX OPTIMIZATION The nonlinear equations f(x, y) = b have to be solved many times. Gradients are only estimates (not analytic). Computation of p and α is therefore less reliable. Are World War 2 black boxes worth saving? Some of them already have to juggle things to be able to solve f(x, y) = b. Might as well let the optimizer do the juggling. Example in electrical power industry: New York Power Authority. Load flows are terminated when the iterates start diverge(!). Optimal power flows can have errors of 300,000 watts (error in satisfying Kirchoff s nonlinear equations). Typical error for the General Electic OPF code: 1 watt.
7 7 PUT EVERYTHING INTO THE OPTIMIZATION MODEL A large-scale optimization system (e.g., MINOS) should be able to treat the black box constraints directly: minimize φ(x, y) subject to f(x, y) = b, and l x y u. As in LP, bounds on the variables can be handled easily. (These are what some of the black boxes are busy juggling in an ad hoc fashion.) The functions in f(x, y) are typically simple. 1 A modelling language (e.g., GAMS) can obtain the partial derivatives of f(x, y) analytically. f(x, y)/ (x, y) = J(x, y), the Jacobian matrix. Certain oil companies now use GAMS/MINOS (and SLP). Refinery models are typically 70 80% nonlinear. A three-refinery model has about 900 equations, 1000 variables, 4000 entries in J(x, y). Cold-start solve on an IBM 3090 takes minutes. 1 Well... In a preliminary GAMS refinery model, one equation (i.e., one GAMS statement) filled 20 pages!
8 8 GAMS, GAMS/MINOS, GAMS/ZOOM General Algebraic Modeling System. An equation-oriented modeling language for LP, NLP, MIP. GAMS: A User s Guide, Anthony Brooke, David Kendrick and Alex Meeraus, with tutorial by Richard Rosenthal, The Scientific Press, Redwood City, California, IBM PC, PS/2 and compatibles (DOS). IBM mainframes (MVS and CMS), DEC VAX systems (VMS and Ultrix), CDC (if you must!). GAMS is in Pascal, MINOS and ZOOM are in Fortran 77.
9 9 STRUCTURED CONTROLLER DESIGN A GAMS Example Lyapunov s Theorem (for linear systems) Solutions of dz/dt = Az are bounded for t 0 There exists X > 0 such that A T X + XA 0. Illustration A large chemical process. A controller might form each control signal (e.g., a pump voltage) based on local information (e.g., temperatures and flow rates near the pump). Structured controller means there are constraints on X or A. Example 1: X should be block-diagonal. Example 2: A should be a linear combination of some given matrices: A = A 0 + y 1 A y k A k. Typical objective: minimize Trace 0 e A(y)T t e A(y)t dt.
10 10 Optimization Problem Find y and a positive-definite matrix X that is Lyapunov stable for a matrix of the form A = A 0 + y 1 A y k A k (where A 0,..., A k are given). Minimize Trace(X). minimize X, y, A, U i X ii subject to A = A 0 + y 1 A y k A k X = U T U A T X + X T A = I U is upper triangular (Cholesky). To make U nonsingular, include bounds on its diagonals: U ii δ. A and X are n n. Problem is significant even for n = 5, k = 3.
11 22 ALGORITHMS FOR NONLINEAR CONSTRAINTS minimize φ(x) subject to f(x) = b, l x u. Solve a sequence of linearly constrained subproblems Assume x k and λ k are the kth estimates of x and λ. Approximate Lagrangian: F k (x) φ(x) λ T k f(x). Linearize constraints: f k (x) f(x k ) + J k (x x k ). Solve subproblem: min F k (x) subject to f k (x) = b, l x u. Form search direction: p x x k, q λ λ k. Step: x k+1 λ k+1 x k λ k + α p q. SLC and SQP algorithms SLC: SQP: Subproblem objective is the Lagrangian. Objective is a quadratic approx n to the Lagrangian.
12 23 OPTIMAL POWER FLOW EVOLUTION OF OPF ALGORITHMS Rob Burchett at General Electric, Schenectady, NY SOL Algorithms Group at Stanford 1980 MINOS 4.0 used directly (SLC method). Subproblems solved with the reduced-gradient method. Combined LP simplex technology with quasi-newton method. (Needs first derivatives; approximates reduced Hessian.) Jul 81 OPF problem solved at GE with 1200 nonlinear constraints. 500 iterations, 1000 evaluations of functions and gradients, 11 hours on VAX 11/780, 0.1 billion page faults (??!). Never before feasible, let alone optimal. Aug 81 Replaced P 4 basis factorization with Reid s LA05 package. (P 4 assumed bases were nearly triangular, but most of the Jacobian is symmetric.) 1200-row problem: 1 hour on VAX 11/780. Oct 81 EPRI awarded contract to ESCA (not GE). Feb row problem: 90 mins on VAX. Mar 82 Philadelphia Electric Company test OPF. Quote: After using the only other two existing OPF programs available to the industry (BPA and PG&E) we have found the GE program to be far superior. Nov 82 PG&E test OPF in San Francisco. Nonlinear constraints satisfied to machine precision. Many different starting point gave same solution! Mar 83 OPF problem solved with NPSOL (dense SQP method). 340-row problem: 6 QP subproblems, 300 QP iterations, 8 evaluations of functions + 1st and 2nd derivatives.
13 MINOS 5.0 adapted to solve QP subproblems with exact 1st and 2nd derivatives (OPF version 2.5). Still used quasi-newton to approximate the reduced Hessian row problem: 7 QP subproblems, 900 iterations, 4 hours on VAX OPF 3.0 solves QPs using full Kuhn-Tucker system: 0 J k Jk T H k λ p = 0 g Used Harwell code MA27 on symmetric indefinite system. Factors repacked to suit MINOS factorization routines to allow sparse LU updates each iteration. Up to 3 times faster than OPF 2.5. Sporadic trouble with indefiniteness, singularity OPF 3.1: Artificial column method for Phase 1. (Improves objective during Phase 1.) 1984 OPF 3.2: Schur-complement updates to KT factorization. Up to 5 times faster than OPF Good ordering for MA27 obtained using 1/4 of KT matrix SECURITY CONSTRAINED DISPATCH (SCD). Allow for K contingencies (transmission breakdowns). Problem is K times as large (K = 1 10). Benders decomposition: QP master problem as in OPF, LP subproblems generate cuts to be added to QP. Feb 85 Wrote to NZE Wellington asking if we could obtain data for NZ power network. Requested IEEE standard format if possible. Apr 85 NZE wrote to GE asking for definition of IEEE format. May 85 GE sent format to NZE (now Electricity Division, Ministry of Energy), asked for data on tape. Offered to do the conversion to IEEE if necessary..
14 25 Jan 86 Met with Power Transmission people at NZE. Happy there with Load Flows. (No optimization. Can t measure load voltages better than 1%.) Asked for NZ data MicroVAX II acquired for SCD project. Can handle problems with 5000 rows. (New Zealand system would be about 500 rows.) Feb 87 Met with Power Transmission people at NZE (now Electricorp). Still happy with Load Flows. They estimate 3% saving in transmission losses = $NZ 2.4 million annually. Asked for NZ data. Apr 87 Moved office at Stanford, threw away lots of listings. Acquired 4 VAXstation IIs at SOL (courtesy of US Army). Experience with Rob s machine and VMS a large influence. Apr 87 Wrote to NZE apologizing for short notice last visit. Asked for NZ data. Apr 87 NZE sent GE printouts(!) of typical NI and SI power flows, in IBM PSP format. Jul 87 Shock: Rob resigned from GE to set up his own company. Began work on totally new code. Recriminations rife. Dec 87 Trip to GE (Schenectady NY) to try to resurrect SCD. Hopeless without Rob. (One line of code changed.) May 88 Prospective first sale for Rob s new system to NY company who market a large Energy Management System containing a Load Flow code but no OPF. Current customers 150, including NZE! Jun 88 More successful effort to resurrect SCD. (Recovery when MA27 says KT system is singular.) Jun 88 Temporary truce reached between Rob and GE Optimal Power Flow in NZ?
15 26 MAIN THEMES Remember your temporary variables (x). More equations but simpler functions, sparser Jacobians: Bx + Cy = b Dx + Ey = c f(x, y) = d (B nonsingular) General sparse-matrix methods can handle several thousand variables and constraints, and the nonlinear constraints do not have to be satisfied throughout only in the limit (at an optimal solution). One day, GAMS might be able to generate both f/ (x, y) and 2 f/ x 2, 2 f/ x y, etc. Think large-scale (don t eliminate variables). Think nonlinear (let GAMS do the derivatives).
What s New in Active-Set Methods for Nonlinear Optimization?
What s New in Active-Set Methods for Nonlinear Optimization? Philip E. Gill Advances in Numerical Computation, Manchester University, July 5, 2011 A Workshop in Honor of Sven Hammarling UCSD Center for
More informationMS&E 318 (CME 338) Large-Scale Numerical Optimization
Stanford University, Management Science & Engineering (and ICME) MS&E 318 (CME 338) Large-Scale Numerical Optimization Instructor: Michael Saunders Spring 2015 Notes 11: NPSOL and SNOPT SQP Methods 1 Overview
More informationMS&E 318 (CME 338) Large-Scale Numerical Optimization
Stanford University, Management Science & Engineering (and ICME) MS&E 318 (CME 338) Large-Scale Numerical Optimization 1 Origins Instructor: Michael Saunders Spring 2015 Notes 9: Augmented Lagrangian Methods
More informationminimize x subject to (x 2)(x 4) u,
Math 6366/6367: Optimization and Variational Methods Sample Preliminary Exam Questions 1. Suppose that f : [, L] R is a C 2 -function with f () on (, L) and that you have explicit formulae for
More information1 Computing with constraints
Notes for 2017-04-26 1 Computing with constraints Recall that our basic problem is minimize φ(x) s.t. x Ω where the feasible set Ω is defined by equality and inequality conditions Ω = {x R n : c i (x)
More informationMS&E 318 (CME 338) Large-Scale Numerical Optimization
Stanford University, Management Science & Engineering (and ICME) MS&E 318 (CME 338) Large-Scale Numerical Optimization Instructor: Michael Saunders Spring 2017 Notes 10: MINOS Part 1: the Reduced-Gradient
More informationPart 4: Active-set methods for linearly constrained optimization. Nick Gould (RAL)
Part 4: Active-set methods for linearly constrained optimization Nick Gould RAL fx subject to Ax b Part C course on continuoue optimization LINEARLY CONSTRAINED MINIMIZATION fx subject to Ax { } b where
More informationAM 205: lecture 19. Last time: Conditions for optimality Today: Newton s method for optimization, survey of optimization methods
AM 205: lecture 19 Last time: Conditions for optimality Today: Newton s method for optimization, survey of optimization methods Optimality Conditions: Equality Constrained Case As another example of equality
More informationMS&E 318 (CME 338) Large-Scale Numerical Optimization
Stanford University, Management Science & Engineering (and ICME MS&E 38 (CME 338 Large-Scale Numerical Optimization Course description Instructor: Michael Saunders Spring 28 Notes : Review The course teaches
More informationE5295/5B5749 Convex optimization with engineering applications. Lecture 8. Smooth convex unconstrained and equality-constrained minimization
E5295/5B5749 Convex optimization with engineering applications Lecture 8 Smooth convex unconstrained and equality-constrained minimization A. Forsgren, KTH 1 Lecture 8 Convex optimization 2006/2007 Unconstrained
More informationNonlinear Optimization: What s important?
Nonlinear Optimization: What s important? Julian Hall 10th May 2012 Convexity: convex problems A local minimizer is a global minimizer A solution of f (x) = 0 (stationary point) is a minimizer A global
More informationGAMINGRE 8/1/ of 7
FYE 09/30/92 JULY 92 0.00 254,550.00 0.00 0 0 0 0 0 0 0 0 0 254,550.00 0.00 0.00 0.00 0.00 254,550.00 AUG 10,616,710.31 5,299.95 845,656.83 84,565.68 61,084.86 23,480.82 339,734.73 135,893.89 67,946.95
More informationMatrix Assembly in FEA
Matrix Assembly in FEA 1 In Chapter 2, we spoke about how the global matrix equations are assembled in the finite element method. We now want to revisit that discussion and add some details. For example,
More informationCONVEXIFICATION SCHEMES FOR SQP METHODS
CONVEXIFICAION SCHEMES FOR SQP MEHODS Philip E. Gill Elizabeth Wong UCSD Center for Computational Mathematics echnical Report CCoM-14-06 July 18, 2014 Abstract Sequential quadratic programming (SQP) methods
More informationAn Introduction to Algebraic Multigrid (AMG) Algorithms Derrick Cerwinsky and Craig C. Douglas 1/84
An Introduction to Algebraic Multigrid (AMG) Algorithms Derrick Cerwinsky and Craig C. Douglas 1/84 Introduction Almost all numerical methods for solving PDEs will at some point be reduced to solving A
More informationConvex Quadratic Approximation
Convex Quadratic Approximation J. Ben Rosen 1 and Roummel F. Marcia 2 Abstract. For some applications it is desired to approximate a set of m data points in IR n with a convex quadratic function. Furthermore,
More informationMS&E 318 (CME 338) Large-Scale Numerical Optimization
Stanford University, Management Science & Engineering (and ICME) MS&E 318 (CME 338) Large-Scale Numerical Optimization Instructor: Michael Saunders Spring 2015 Notes 4: The Primal Simplex Method 1 Linear
More informationGAMS/MINOS MINOS 1 GAMS/MINOS. Table of Contents:
GAMS/MINOS MINOS 1 GAMS/MINOS Table of Contents:... 1. INTRODUCTION... 2 2. HOW TO RUN A MODEL WITH GAMS/MINOS... 2 3. OVERVIEW OF GAMS/MINOS... 2 3.1. Linear programming... 3 3.2. Problems with a Nonlinear
More informationA convex QP solver based on block-lu updates
Block-LU updates p. 1/24 A convex QP solver based on block-lu updates PP06 SIAM Conference on Parallel Processing for Scientific Computing San Francisco, CA, Feb 22 24, 2006 Hanh Huynh and Michael Saunders
More informationPascal s Triangle on a Budget. Accuracy, Precision and Efficiency in Sparse Grids
Covering : Accuracy, Precision and Efficiency in Sparse Grids https://people.sc.fsu.edu/ jburkardt/presentations/auburn 2009.pdf... John Interdisciplinary Center for Applied Mathematics & Information Technology
More informationA Trust-region-based Sequential Quadratic Programming Algorithm
Downloaded from orbit.dtu.dk on: Oct 19, 2018 A Trust-region-based Sequential Quadratic Programming Algorithm Henriksen, Lars Christian; Poulsen, Niels Kjølstad Publication date: 2010 Document Version
More informationConstrained Nonlinear Optimization Algorithms
Department of Industrial Engineering and Management Sciences Northwestern University waechter@iems.northwestern.edu Institute for Mathematics and its Applications University of Minnesota August 4, 2016
More informationAM 205: lecture 19. Last time: Conditions for optimality, Newton s method for optimization Today: survey of optimization methods
AM 205: lecture 19 Last time: Conditions for optimality, Newton s method for optimization Today: survey of optimization methods Quasi-Newton Methods General form of quasi-newton methods: x k+1 = x k α
More informationOutline Introduction: Problem Description Diculties Algebraic Structure: Algebraic Varieties Rank Decient Toeplitz Matrices Constructing Lower Rank St
Structured Lower Rank Approximation by Moody T. Chu (NCSU) joint with Robert E. Funderlic (NCSU) and Robert J. Plemmons (Wake Forest) March 5, 1998 Outline Introduction: Problem Description Diculties Algebraic
More informationOptimization Methods
Optimization Methods Decision making Examples: determining which ingredients and in what quantities to add to a mixture being made so that it will meet specifications on its composition allocating available
More informationSF2822 Applied Nonlinear Optimization. Preparatory question. Lecture 9: Sequential quadratic programming. Anders Forsgren
SF2822 Applied Nonlinear Optimization Lecture 9: Sequential quadratic programming Anders Forsgren SF2822 Applied Nonlinear Optimization, KTH / 24 Lecture 9, 207/208 Preparatory question. Try to solve theory
More informationA review of sparsity vs stability in LU updates
A review of sparsity vs stability in LU updates Michael Saunders Dept of Management Science and Engineering (MS&E) Systems Optimization Laboratory (SOL) Institute for Computational Mathematics and Engineering
More informationNumerical Optimization
Numerical Optimization Unit 2: Multivariable optimization problems Che-Rung Lee Scribe: February 28, 2011 (UNIT 2) Numerical Optimization February 28, 2011 1 / 17 Partial derivative of a two variable function
More informationBindel, Fall 2016 Matrix Computations (CS 6210) Notes for
1 Logistics Notes for 2016-09-14 1. There was a goof in HW 2, problem 1 (now fixed) please re-download if you have already started looking at it. 2. CS colloquium (4:15 in Gates G01) this Thurs is Margaret
More informationLUSOL: A basis package for constrained optimization
LUSOL: A basis package for constrained optimization IFORS Triennial Conference on OR/MS Honolulu, HI, July 11 15, 2005 Michael O Sullivan and Michael Saunders Dept of Engineering Science Dept of Management
More informationCOMP 558 lecture 18 Nov. 15, 2010
Least squares We have seen several least squares problems thus far, and we will see more in the upcoming lectures. For this reason it is good to have a more general picture of these problems and how to
More informationECEN 615 Methods of Electric Power Systems Analysis Lecture 18: Least Squares, State Estimation
ECEN 615 Methods of Electric Power Systems Analysis Lecture 18: Least Squares, State Estimation Prof. om Overbye Dept. of Electrical and Computer Engineering exas A&M University overbye@tamu.edu Announcements
More informationAn Inexact Sequential Quadratic Optimization Method for Nonlinear Optimization
An Inexact Sequential Quadratic Optimization Method for Nonlinear Optimization Frank E. Curtis, Lehigh University involving joint work with Travis Johnson, Northwestern University Daniel P. Robinson, Johns
More informationAlgorithms for Constrained Optimization
1 / 42 Algorithms for Constrained Optimization ME598/494 Lecture Max Yi Ren Department of Mechanical Engineering, Arizona State University April 19, 2015 2 / 42 Outline 1. Convergence 2. Sequential quadratic
More informationLecture 11: CMSC 878R/AMSC698R. Iterative Methods An introduction. Outline. Inverse, LU decomposition, Cholesky, SVD, etc.
Lecture 11: CMSC 878R/AMSC698R Iterative Methods An introduction Outline Direct Solution of Linear Systems Inverse, LU decomposition, Cholesky, SVD, etc. Iterative methods for linear systems Why? Matrix
More informationCS 542G: Conditioning, BLAS, LU Factorization
CS 542G: Conditioning, BLAS, LU Factorization Robert Bridson September 22, 2008 1 Why some RBF Kernel Functions Fail We derived some sensible RBF kernel functions, like φ(r) = r 2 log r, from basic principles
More informationLINEAR AND NONLINEAR PROGRAMMING
LINEAR AND NONLINEAR PROGRAMMING Stephen G. Nash and Ariela Sofer George Mason University The McGraw-Hill Companies, Inc. New York St. Louis San Francisco Auckland Bogota Caracas Lisbon London Madrid Mexico
More informationComputational Finance
Department of Mathematics at University of California, San Diego Computational Finance Optimization Techniques [Lecture 2] Michael Holst January 9, 2017 Contents 1 Optimization Techniques 3 1.1 Examples
More informationarxiv: v1 [math.oc] 10 Apr 2017
A Method to Guarantee Local Convergence for Sequential Quadratic Programming with Poor Hessian Approximation Tuan T. Nguyen, Mircea Lazar and Hans Butler arxiv:1704.03064v1 math.oc] 10 Apr 2017 Abstract
More informationB553 Lecture 5: Matrix Algebra Review
B553 Lecture 5: Matrix Algebra Review Kris Hauser January 19, 2012 We have seen in prior lectures how vectors represent points in R n and gradients of functions. Matrices represent linear transformations
More informationSoftware for Integer and Nonlinear Optimization
Software for Integer and Nonlinear Optimization Sven Leyffer, leyffer@mcs.anl.gov Mathematics & Computer Science Division Argonne National Laboratory Roger Fletcher & Jeff Linderoth Advanced Methods and
More information1.2 Derivation. d p f = d p f(x(p)) = x fd p x (= f x x p ). (1) Second, g x x p + g p = 0. d p f = f x g 1. The expression f x gx
PDE-constrained optimization and the adjoint method Andrew M. Bradley November 16, 21 PDE-constrained optimization and the adjoint method for solving these and related problems appear in a wide range of
More informationLinear Solvers. Andrew Hazel
Linear Solvers Andrew Hazel Introduction Thus far we have talked about the formulation and discretisation of physical problems...... and stopped when we got to a discrete linear system of equations. Introduction
More informationHot-Starting NLP Solvers
Hot-Starting NLP Solvers Andreas Wächter Department of Industrial Engineering and Management Sciences Northwestern University waechter@iems.northwestern.edu 204 Mixed Integer Programming Workshop Ohio
More informationImplementation of a KKT-based active-set QP solver
Block-LU updates p. 1/27 Implementation of a KKT-based active-set QP solver ISMP 2006 19th International Symposium on Mathematical Programming Rio de Janeiro, Brazil, July 30 August 4, 2006 Hanh Huynh
More informationPenalty and Barrier Methods. So we again build on our unconstrained algorithms, but in a different way.
AMSC 607 / CMSC 878o Advanced Numerical Optimization Fall 2008 UNIT 3: Constrained Optimization PART 3: Penalty and Barrier Methods Dianne P. O Leary c 2008 Reference: N&S Chapter 16 Penalty and Barrier
More informationAn Active Set Strategy for Solving Optimization Problems with up to 200,000,000 Nonlinear Constraints
An Active Set Strategy for Solving Optimization Problems with up to 200,000,000 Nonlinear Constraints Klaus Schittkowski Department of Computer Science, University of Bayreuth 95440 Bayreuth, Germany e-mail:
More informationLecture 15: SQP methods for equality constrained optimization
Lecture 15: SQP methods for equality constrained optimization Coralia Cartis, Mathematical Institute, University of Oxford C6.2/B2: Continuous Optimization Lecture 15: SQP methods for equality constrained
More informationSparse least squares and Q-less QR
Notes for 2016-02-29 Sparse least squares and Q-less QR Suppose we want to solve a full-rank least squares problem in which A is large and sparse. In principle, we could solve the problem via the normal
More informationEAD 115. Numerical Solution of Engineering and Scientific Problems. David M. Rocke Department of Applied Science
EAD 115 Numerical Solution of Engineering and Scientific Problems David M. Rocke Department of Applied Science Taylor s Theorem Can often approximate a function by a polynomial The error in the approximation
More informationLinear Algebra in IPMs
School of Mathematics H E U N I V E R S I Y O H F E D I N B U R G Linear Algebra in IPMs Jacek Gondzio Email: J.Gondzio@ed.ac.uk URL: http://www.maths.ed.ac.uk/~gondzio Paris, January 2018 1 Outline Linear
More informationComputational Methods. Least Squares Approximation/Optimization
Computational Methods Least Squares Approximation/Optimization Manfred Huber 2011 1 Least Squares Least squares methods are aimed at finding approximate solutions when no precise solution exists Find the
More informationAccuracy, Precision and Efficiency in Sparse Grids
John, Information Technology Department, Virginia Tech.... http://people.sc.fsu.edu/ jburkardt/presentations/ sandia 2009.pdf... Computer Science Research Institute, Sandia National Laboratory, 23 July
More informationCS 542G: Robustifying Newton, Constraints, Nonlinear Least Squares
CS 542G: Robustifying Newton, Constraints, Nonlinear Least Squares Robert Bridson October 29, 2008 1 Hessian Problems in Newton Last time we fixed one of plain Newton s problems by introducing line search
More informationNumerical Methods I Solving Nonlinear Equations
Numerical Methods I Solving Nonlinear Equations Aleksandar Donev Courant Institute, NYU 1 donev@courant.nyu.edu 1 MATH-GA 2011.003 / CSCI-GA 2945.003, Fall 2014 October 16th, 2014 A. Donev (Courant Institute)
More informationA DECOMPOSITION PROCEDURE BASED ON APPROXIMATE NEWTON DIRECTIONS
Working Paper 01 09 Departamento de Estadística y Econometría Statistics and Econometrics Series 06 Universidad Carlos III de Madrid January 2001 Calle Madrid, 126 28903 Getafe (Spain) Fax (34) 91 624
More informationAn introduction to PDE-constrained optimization
An introduction to PDE-constrained optimization Wolfgang Bangerth Department of Mathematics Texas A&M University 1 Overview Why partial differential equations? Why optimization? Examples of PDE optimization
More informationWritten Examination
Division of Scientific Computing Department of Information Technology Uppsala University Optimization Written Examination 202-2-20 Time: 4:00-9:00 Allowed Tools: Pocket Calculator, one A4 paper with notes
More informationIntroduction, basic but important concepts
Introduction, basic but important concepts Felix Kubler 1 1 DBF, University of Zurich and Swiss Finance Institute October 7, 2017 Felix Kubler Comp.Econ. Gerzensee, Ch1 October 7, 2017 1 / 31 Economics
More informationAn Introduction to NeRDS (Nearly Rank Deficient Systems)
(Nearly Rank Deficient Systems) BY: PAUL W. HANSON Abstract I show that any full rank n n matrix may be decomposento the sum of a diagonal matrix and a matrix of rank m where m < n. This decomposition
More informationIntroduction to Forecasting
Introduction to Forecasting Introduction to Forecasting Predicting the future Not an exact science but instead consists of a set of statistical tools and techniques that are supported by human judgment
More informationNumerical Nonlinear Optimization with WORHP
Numerical Nonlinear Optimization with WORHP Christof Büskens Optimierung & Optimale Steuerung London, 8.9.2011 Optimization & Optimal Control Nonlinear Optimization WORHP Concept Ideas Features Results
More informationQuasi-Newton Methods
Newton s Method Pros and Cons Quasi-Newton Methods MA 348 Kurt Bryan Newton s method has some very nice properties: It s extremely fast, at least once it gets near the minimum, and with the simple modifications
More informationDirect Methods. Moritz Diehl. Optimization in Engineering Center (OPTEC) and Electrical Engineering Department (ESAT) K.U.
Direct Methods Moritz Diehl Optimization in Engineering Center (OPTEC) and Electrical Engineering Department (ESAT) K.U. Leuven Belgium Overview Direct Single Shooting Direct Collocation Direct Multiple
More informationL2-7 Some very stylish matrix decompositions for solving Ax = b 10 Oct 2015
L-7 Some very stylish matrix decompositions for solving Ax = b 10 Oct 015 Marty McFly: Wait a minute, Doc. Ah... Are you telling me you built a time machine... out of a DeLorean? Doc Brown: The way I see
More informationLinear Models Review
Linear Models Review Vectors in IR n will be written as ordered n-tuples which are understood to be column vectors, or n 1 matrices. A vector variable will be indicted with bold face, and the prime sign
More informationSNOPT: An SQP Algorithm for Large-Scale Constrained Optimization
SNOPT: An SQP Algorithm for Large-Scale Constrained Optimization Philip E. Gill; Department of Mathematics, University of California, San Diego, La Jolla, CA Walter Murray, Michael A. Saunders; Department
More informationOptimization Tutorial 1. Basic Gradient Descent
E0 270 Machine Learning Jan 16, 2015 Optimization Tutorial 1 Basic Gradient Descent Lecture by Harikrishna Narasimhan Note: This tutorial shall assume background in elementary calculus and linear algebra.
More informationMath 273a: Optimization Netwon s methods
Math 273a: Optimization Netwon s methods Instructor: Wotao Yin Department of Mathematics, UCLA Fall 2015 some material taken from Chong-Zak, 4th Ed. Main features of Newton s method Uses both first derivatives
More informationConstrained optimization. Unconstrained optimization. One-dimensional. Multi-dimensional. Newton with equality constraints. Active-set method.
Optimization Unconstrained optimization One-dimensional Multi-dimensional Newton s method Basic Newton Gauss- Newton Quasi- Newton Descent methods Gradient descent Conjugate gradient Constrained optimization
More informationCME 302: NUMERICAL LINEAR ALGEBRA FALL 2005/06 LECTURE 6
CME 302: NUMERICAL LINEAR ALGEBRA FALL 2005/06 LECTURE 6 GENE H GOLUB Issues with Floating-point Arithmetic We conclude our discussion of floating-point arithmetic by highlighting two issues that frequently
More informationOPTIMISATION 2007/8 EXAM PREPARATION GUIDELINES
General: OPTIMISATION 2007/8 EXAM PREPARATION GUIDELINES This points out some important directions for your revision. The exam is fully based on what was taught in class: lecture notes, handouts and homework.
More informationOptimization. Charles J. Geyer School of Statistics University of Minnesota. Stat 8054 Lecture Notes
Optimization Charles J. Geyer School of Statistics University of Minnesota Stat 8054 Lecture Notes 1 One-Dimensional Optimization Look at a graph. Grid search. 2 One-Dimensional Zero Finding Zero finding
More informationInfeasibility Detection and an Inexact Active-Set Method for Large-Scale Nonlinear Optimization
Infeasibility Detection and an Inexact Active-Set Method for Large-Scale Nonlinear Optimization Frank E. Curtis, Lehigh University involving joint work with James V. Burke, University of Washington Daniel
More informationOutline. Relaxation. Outline DMP204 SCHEDULING, TIMETABLING AND ROUTING. 1. Lagrangian Relaxation. Lecture 12 Single Machine Models, Column Generation
Outline DMP204 SCHEDULING, TIMETABLING AND ROUTING 1. Lagrangian Relaxation Lecture 12 Single Machine Models, Column Generation 2. Dantzig-Wolfe Decomposition Dantzig-Wolfe Decomposition Delayed Column
More informationProcess Behavior Analysis Understanding Variation
Process Behavior Analysis Understanding Variation Steven J Mazzuca ASQ 2015-11-11 Why Process Behavior Analysis? Every day we waste valuable resources because we misunderstand or misinterpret what our
More information7.2 Linear equation systems. 7.3 Linear least square fit
72 Linear equation systems In the following sections, we will spend some time to solve linear systems of equations This is a tool that will come in handy in many di erent places during this course For
More informationSurvey of NLP Algorithms. L. T. Biegler Chemical Engineering Department Carnegie Mellon University Pittsburgh, PA
Survey of NLP Algorithms L. T. Biegler Chemical Engineering Department Carnegie Mellon University Pittsburgh, PA NLP Algorithms - Outline Problem and Goals KKT Conditions and Variable Classification Handling
More informationLinear programming II
Linear programming II Review: LP problem 1/33 The standard form of LP problem is (primal problem): max z = cx s.t. Ax b, x 0 The corresponding dual problem is: min b T y s.t. A T y c T, y 0 Strong Duality
More informationHomework 4. Convex Optimization /36-725
Homework 4 Convex Optimization 10-725/36-725 Due Friday November 4 at 5:30pm submitted to Christoph Dann in Gates 8013 (Remember to a submit separate writeup for each problem, with your name at the top)
More informationSMO vs PDCO for SVM: Sequential Minimal Optimization vs Primal-Dual interior method for Convex Objectives for Support Vector Machines
vs for SVM: Sequential Minimal Optimization vs Primal-Dual interior method for Convex Objectives for Support Vector Machines Ding Ma Michael Saunders Working paper, January 5 Introduction In machine learning,
More informationThe Solution of Linear Systems AX = B
Chapter 2 The Solution of Linear Systems AX = B 21 Upper-triangular Linear Systems We will now develop the back-substitution algorithm, which is useful for solving a linear system of equations that has
More informationECE133A Applied Numerical Computing Additional Lecture Notes
Winter Quarter 2018 ECE133A Applied Numerical Computing Additional Lecture Notes L. Vandenberghe ii Contents 1 LU factorization 1 1.1 Definition................................. 1 1.2 Nonsingular sets
More informationAMS526: Numerical Analysis I (Numerical Linear Algebra)
AMS526: Numerical Analysis I (Numerical Linear Algebra) Lecture 3: Positive-Definite Systems; Cholesky Factorization Xiangmin Jiao Stony Brook University Xiangmin Jiao Numerical Analysis I 1 / 11 Symmetric
More informationISM206 Lecture Optimization of Nonlinear Objective with Linear Constraints
ISM206 Lecture Optimization of Nonlinear Objective with Linear Constraints Instructor: Prof. Kevin Ross Scribe: Nitish John October 18, 2011 1 The Basic Goal The main idea is to transform a given constrained
More informationGaussian Elimination without/with Pivoting and Cholesky Decomposition
Gaussian Elimination without/with Pivoting and Cholesky Decomposition Gaussian Elimination WITHOUT pivoting Notation: For a matrix A R n n we define for k {,,n} the leading principal submatrix a a k A
More informationOptimization 2. CS5240 Theoretical Foundations in Multimedia. Leow Wee Kheng
Optimization 2 CS5240 Theoretical Foundations in Multimedia Leow Wee Kheng Department of Computer Science School of Computing National University of Singapore Leow Wee Kheng (NUS) Optimization 2 1 / 38
More informationInterior Point Methods. We ll discuss linear programming first, followed by three nonlinear problems. Algorithms for Linear Programming Problems
AMSC 607 / CMSC 764 Advanced Numerical Optimization Fall 2008 UNIT 3: Constrained Optimization PART 4: Introduction to Interior Point Methods Dianne P. O Leary c 2008 Interior Point Methods We ll discuss
More informationOPTIMISATION /09 EXAM PREPARATION GUIDELINES
General: OPTIMISATION 2 2008/09 EXAM PREPARATION GUIDELINES This points out some important directions for your revision. The exam is fully based on what was taught in class: lecture notes, handouts and
More informationPreprint ANL/MCS-P , Dec 2002 (Revised Nov 2003, Mar 2004) Mathematics and Computer Science Division Argonne National Laboratory
Preprint ANL/MCS-P1015-1202, Dec 2002 (Revised Nov 2003, Mar 2004) Mathematics and Computer Science Division Argonne National Laboratory A GLOBALLY CONVERGENT LINEARLY CONSTRAINED LAGRANGIAN METHOD FOR
More informationBindel, Spring 2016 Numerical Analysis (CS 4220) Notes for
Life beyond Newton Notes for 2016-04-08 Newton s method has many attractive properties, particularly when we combine it with a globalization strategy. Unfortunately, Newton steps are not cheap. At each
More informationN. L. P. NONLINEAR PROGRAMMING (NLP) deals with optimization models with at least one nonlinear function. NLP. Optimization. Models of following form:
0.1 N. L. P. Katta G. Murty, IOE 611 Lecture slides Introductory Lecture NONLINEAR PROGRAMMING (NLP) deals with optimization models with at least one nonlinear function. NLP does not include everything
More information1 GSW Sets of Systems
1 Often, we have to solve a whole series of sets of simultaneous equations of the form y Ax, all of which have the same matrix A, but each of which has a different known vector y, and a different unknown
More information5 Handling Constraints
5 Handling Constraints Engineering design optimization problems are very rarely unconstrained. Moreover, the constraints that appear in these problems are typically nonlinear. This motivates our interest
More informationMultidisciplinary System Design Optimization (MSDO)
Multidisciplinary System Design Optimization (MSDO) Numerical Optimization II Lecture 8 Karen Willcox 1 Massachusetts Institute of Technology - Prof. de Weck and Prof. Willcox Today s Topics Sequential
More informationLinear algebra I Homework #1 due Thursday, Oct Show that the diagonals of a square are orthogonal to one another.
Homework # due Thursday, Oct. 0. Show that the diagonals of a square are orthogonal to one another. Hint: Place the vertices of the square along the axes and then introduce coordinates. 2. Find the equation
More informationSELECT TWO PROBLEMS (OF A POSSIBLE FOUR) FROM PART ONE, AND FOUR PROBLEMS (OF A POSSIBLE FIVE) FROM PART TWO. PART ONE: TOTAL GRAND
1 56:270 LINEAR PROGRAMMING FINAL EXAMINATION - MAY 17, 1985 SELECT TWO PROBLEMS (OF A POSSIBLE FOUR) FROM PART ONE, AND FOUR PROBLEMS (OF A POSSIBLE FIVE) FROM PART TWO. PART ONE: 1 2 3 4 TOTAL GRAND
More informationCAISO Participating Intermittent Resource Program for Wind Generation
CAISO Participating Intermittent Resource Program for Wind Generation Jim Blatchford CAISO Account Manager Agenda CAISO Market Concepts Wind Availability in California How State Supports Intermittent Resources
More informationGAUSSIAN ELIMINATION AND LU DECOMPOSITION (SUPPLEMENT FOR MA511)
GAUSSIAN ELIMINATION AND LU DECOMPOSITION (SUPPLEMENT FOR MA511) D. ARAPURA Gaussian elimination is the go to method for all basic linear classes including this one. We go summarize the main ideas. 1.
More informationA FACILITY MANAGER S INTRODUCTION TO WEATHER CORRECTION FOR UTILITY BILL TRACKING. John Avina, Director Abraxas Energy Consulting
A FACILITY MANAGER S INTRODUCTION TO WEATHER CORRECTION FOR UTILITY BILL TRACKING John Avina, Director Abraxas Energy Consulting ABSTRACT Utility bill tracking is at the heart of an effective energy management
More information