Some modelling aspects for the Matlab implementation of MMA

Similar documents
MMA and GCMMA two methods for nonlinear optimization

Solutions to exam in SF1811 Optimization, Jan 14, 2015

U.C. Berkeley CS294: Beyond Worst-Case Analysis Luca Trevisan September 5, 2017

College of Computer & Information Science Fall 2009 Northeastern University 20 October 2009

CHAPTER 5 NUMERICAL EVALUATION OF DYNAMIC RESPONSE

The Study of Teaching-learning-based Optimization Algorithm

princeton univ. F 17 cos 521: Advanced Algorithm Design Lecture 7: LP Duality Lecturer: Matt Weinberg

COS 521: Advanced Algorithms Game Theory and Linear Programming

Some Comments on Accelerating Convergence of Iterative Sequences Using Direct Inversion of the Iterative Subspace (DIIS)

The Minimum Universal Cost Flow in an Infeasible Flow Network

Additional Codes using Finite Difference Method. 1 HJB Equation for Consumption-Saving Problem Without Uncertainty

FIRST AND SECOND ORDER NECESSARY OPTIMALITY CONDITIONS FOR DISCRETE OPTIMAL CONTROL PROBLEMS

Lecture 21: Numerical methods for pricing American type derivatives

Chapter 5. Solution of System of Linear Equations. Module No. 6. Solution of Inconsistent and Ill Conditioned Systems

VARIATION OF CONSTANT SUM CONSTRAINT FOR INTEGER MODEL WITH NON UNIFORM VARIABLES

Perfect Competition and the Nash Bargaining Solution

PHYS 705: Classical Mechanics. Calculus of Variations II

Mean Field / Variational Approximations

CHAPTER 7 CONSTRAINED OPTIMIZATION 2: SQP AND GRG

Convex Optimization. Optimality conditions. (EE227BT: UC Berkeley) Lecture 9 (Optimality; Conic duality) 9/25/14. Laurent El Ghaoui.

Assortment Optimization under MNL

Lecture 20: November 7

General viscosity iterative method for a sequence of quasi-nonexpansive mappings

Module 9. Lecture 6. Duality in Assignment Problems

NON-CENTRAL 7-POINT FORMULA IN THE METHOD OF LINES FOR PARABOLIC AND BURGERS' EQUATIONS

ON A DETERMINATION OF THE INITIAL FUNCTIONS FROM THE OBSERVED VALUES OF THE BOUNDARY FUNCTIONS FOR THE SECOND-ORDER HYPERBOLIC EQUATION

CSci 6974 and ECSE 6966 Math. Tech. for Vision, Graphics and Robotics Lecture 21, April 17, 2006 Estimating A Plane Homography

A New Algorithm for Finding a Fuzzy Optimal. Solution for Fuzzy Transportation Problems

PROBLEM SET 7 GENERAL EQUILIBRIUM

Numerical Heat and Mass Transfer

Lecture 12: Discrete Laplacian

Difference Equations

Feature Selection: Part 1

Inexact Newton Methods for Inverse Eigenvalue Problems

EEL 6266 Power System Operation and Control. Chapter 3 Economic Dispatch Using Dynamic Programming

Solutions HW #2. minimize. Ax = b. Give the dual problem, and make the implicit equality constraints explicit. Solution.

Appendix for Causal Interaction in Factorial Experiments: Application to Conjoint Analysis

Chapter 2 - The Simple Linear Regression Model S =0. e i is a random error. S β2 β. This is a minimization problem. Solution is a calculus exercise.

Generalized Linear Methods

EEE 241: Linear Systems

Maximal Margin Classifier

6.854J / J Advanced Algorithms Fall 2008

Dynamic Programming. Preview. Dynamic Programming. Dynamic Programming. Dynamic Programming (Example: Fibonacci Sequence)

Interactive Bi-Level Multi-Objective Integer. Non-linear Programming Problem

A 2D Bounded Linear Program (H,c) 2D Linear Programming

Yong Joon Ryang. 1. Introduction Consider the multicommodity transportation problem with convex quadratic cost function. 1 2 (x x0 ) T Q(x x 0 )

A PROBABILITY-DRIVEN SEARCH ALGORITHM FOR SOLVING MULTI-OBJECTIVE OPTIMIZATION PROBLEMS

IJRSS Volume 2, Issue 2 ISSN:

Combining Constraint Programming and Integer Programming

e - c o m p a n i o n

Introduction to Vapor/Liquid Equilibrium, part 2. Raoult s Law:

4DVAR, according to the name, is a four-dimensional variational method.

DUE: WEDS FEB 21ST 2018

Maximizing Overlap of Large Primary Sampling Units in Repeated Sampling: A comparison of Ernst s Method with Ohlsson s Method

Polynomial Regression Models

On the Global Linear Convergence of the ADMM with Multi-Block Variables

Solving the Quadratic Eigenvalue Complementarity Problem by DC Programming

Lecture 17: Lee-Sidford Barrier

Linear Approximation with Regularization and Moving Least Squares

An Interactive Optimisation Tool for Allocation Problems

NUMERICAL DIFFERENTIATION

Lecture Notes on Linear Regression

ELASTIC WAVE PROPAGATION IN A CONTINUOUS MEDIUM

The Second Anti-Mathima on Game Theory

The Geometry of Logit and Probit

Markov Chain Monte Carlo (MCMC), Gibbs Sampling, Metropolis Algorithms, and Simulated Annealing Bioinformatics Course Supplement

On a direct solver for linear least squares problems

10.34 Numerical Methods Applied to Chemical Engineering Fall Homework #3: Systems of Nonlinear Equations and Optimization

Beyond Zudilin s Conjectured q-analog of Schmidt s problem

Hidden Markov Models

CS : Algorithms and Uncertainty Lecture 17 Date: October 26, 2016

2E Pattern Recognition Solutions to Introduction to Pattern Recognition, Chapter 2: Bayesian pattern classification

Simultaneous Optimization of Berth Allocation, Quay Crane Assignment and Quay Crane Scheduling Problems in Container Terminals

A Bayes Algorithm for the Multitask Pattern Recognition Problem Direct Approach

Estimation: Part 2. Chapter GREG estimation

Module 3 LOSSY IMAGE COMPRESSION SYSTEMS. Version 2 ECE IIT, Kharagpur

DETERMINATION OF TEMPERATURE DISTRIBUTION FOR ANNULAR FINS WITH TEMPERATURE DEPENDENT THERMAL CONDUCTIVITY BY HPM

Problem Set 9 Solutions

Lecture 10 Support Vector Machines. Oct

Lagrange Multipliers Kernel Trick

Linear Feature Engineering 11

Supplement: Proofs and Technical Details for The Solution Path of the Generalized Lasso

ECE559VV Project Report

Credit Card Pricing and Impact of Adverse Selection

SOLVING CAPACITATED VEHICLE ROUTING PROBLEMS WITH TIME WINDOWS BY GOAL PROGRAMMING APPROACH

Fuzzy Approaches for Multiobjective Fuzzy Random Linear Programming Problems Through a Probability Maximization Model

coordinates. Then, the position vectors are described by

e i is a random error

A New Refinement of Jacobi Method for Solution of Linear System Equations AX=b

n α j x j = 0 j=1 has a nontrivial solution. Here A is the n k matrix whose jth column is the vector for all t j=0

Lecture 10 Support Vector Machines II

Chapter 2 A Class of Robust Solution for Linear Bilevel Programming

What is LP? LP is an optimization technique that allocates limited resources among competing activities in the best possible manner.

CIS526: Machine Learning Lecture 3 (Sept 16, 2003) Linear Regression. Preparation help: Xiaoying Huang. x 1 θ 1 output... θ M x M

An Integrated OR/CP Method for Planning and Scheduling

Singular Value Decomposition: Theory and Applications

On the Multicriteria Integer Network Flow Problem

Affine transformations and convexity

Errors for Linear Systems

Week 5: Neural Networks

Transcription:

Some modellng aspects for the Matlab mplementaton of MMA Krster Svanberg krlle@math.kth.se Optmzaton and Systems Theory Department of Mathematcs KTH, SE 10044 Stockholm September 2004 1. Consdered optmzaton problem. The Matlab verson of the author s MMA code s based on the assumpton that the users optmzaton problem s wrtten on the followng form, where the optmzaton varables are x = (x 1,..., x n ) T, y = (y 1,..., y m ) T and z. m f 0 (x) a 0 z (c y 1 2 d y 2 ) subect to f (x) a z y 0, = 1,..., m y 0, z 0. x x max, = 1,..., n = 1,..., m (1.1) Here, x 1,..., x n are the true optmzaton varables, whle y 1,..., y m and z are artfcal optmzaton varables whch wll be motvated below. f 0, f 1,..., f m are gven, contnuously dfferentable, real-valued functons. and x max are gven real numbers whch satsfy < x max. a 0 and a are gven real numbers whch satsfy a 0 > 0 and a 0. c and d are gven real numbers whch satsfy c 0, d 0 and c d > 0. 2. Ordnary NLP problems. Assume that the user wants to solve a problem on the followng standard form for nonlnear programmng. f 0 (x) subect to f (x) 0, = 1,..., m x x max, = 1,..., n (2.1) To put (1.1) nto ths form (2.1), frst let a 0 = 1 and a = 0 for all. Then z = 0 n any optmal soluton of (1.1). Further, for each, let d = 0 and c = a large number, so that the varables y become expensve. Then typcally y =0 n any optmal soluton of (1.1), and the correspondng x s an optmal soluton of (2.1). 1

It should be noted that the problem (1.1) always has feasble solutons, and n fact also at least one optmal soluton. Ths holds even f the user s problem (2.1) does not have any feasble solutons, n whch case some y > 0 n the optmal soluton of (1.1). Ths s ust one advantage of the formulaton (1.1) compared to the formulaton (2.1). Now some practcal consderatons and recommendatons. In many applcatons, the constrants are on the form σ (x) σ max, where σ (x) stands for e.g. a certan stress, whle σ max s the largest permtted value on ths stress. Ths means that f (x) = σ (x) σ max (n (1.1) as well as n (2.1)). The user should then preferably scale the constrants n such a way that 1 σ max 100 for each (and not σ max = 10 10 ). The obectve functon f 0 (x) should preferably be scaled such that 1 f 0 (x) 100 for reasonable values on the varables. The varables x should preferably be scaled such that 0.1 x max 100, for all. Concernng the large numbers on the coeffcents c (mentoned above), the user should for numercal reasons try to avod extremely large values on these coeffcents (lke 10 10 ). It s better to start wth reasonably large values and then, f t turns out that not all y = 0 n the optmal soluton of (1.1), ncrease the correspondng values of c by e.g. a factor 100 and solve the problem agan, etc. If the functons and the varables have been scaled accordng to above, then resonably large values on the parameters c could be, say, c = 1000 or 10000. Fnally, concernng the smple bound constrants x x max, t may sometmes be the case that some varables x do not have any prescrbed upper and/or lower bounds. In that case, t s n practce always possble to choose artfcal bounds and x max such that every realstc soluton x satsfes the correspondng bound constrants. The user should then preferably avod choosng x max unnecessarly large. It s better to try some reasonable bounds and then, f t turns out that some varable x becomes equal to such an artfcal bound n the optmal soluton of (1.1), change ths bound and solve the problem agan (startng from the recently obtaned soluton), etc. 3. Least squares problems. (Mnmum 2 norm problems.) Assume that the user wants to solve a constraned least squares problem on the form p (h (x)) 2 subect to g (x) 0, = 1,..., q x x max, = 1,..., n (3.1) where h and g are gven dfferentable functons. 2

The functons f and the parameters a, c and d should then be chosen as follows n problem (1.1). m = 2p q, f p (x) = h (x), = 1,..., p f 2p (x) = g (x), = 1,..., q a = 0, = 1,..., m d = 2, = 1,..., 2p d 2p = 0, = 1,..., q c = 0, = 1,..., 2p c 2p = large number, = 1,..., q 4. Mnmum 1 norm problems. Assume that the user wants to solve a mnmum 1 norm problem on the form p h (x) subect to g (x) 0, = 1,..., q x x max, = 1,..., n (4.1) where h and g are gven dfferentable functons. The functons f and the parameters a, c and d should then be chosen as follows n problem (1.1). m = 2p q, f p (x) = h (x), = 1,..., p f 2p (x) = g (x), = 1,..., q a = 0, = 1,..., m d = 0, = 1,..., m c = 1, = 1,..., 2p c 2p = large number, = 1,..., q 5. Mnmax problem. (Mnmum norm problems.) Assume that the user wants to solve a mnmax problem on the form max (x) },..,p subect to g (x) 0, = 1,..., q where h and g are gven dfferentable functons. x x max, = 1,..., n (5.1) 3

The functons f and the parameters a, c and d should then be chosen as follows n problem (1.1). m = 2p q, f p (x) = h (x), = 1,..., p f 2p (x) = g (x), = 1,..., q a = 1, = 1,..., 2p a 2p = 0, = 1,..., q d = 0, = 1,..., m c = large number, = 1,..., m 6. The MMA subproblem MMA s a method for solvng problems on the form (1.1), usng the followng approach: In each teraton, the current teraton pont (x, y, z ) s gven. Then an approxmatng explct subproblem s generated. In ths subproblem, the functons f (x) are replaced by approxmatng convex functons f (x). These approxmatons are based manly on gradent nformaton at the current teraton pont, but also (mplctly) on nformaton from prevous teraton ponts. The subproblem s solved, and the unque optmal soluton becomes the next teraton pont (x (k1), y (k1), z (k1) ). Then a new subproblem s generated, etc. The subproblem mentoned above looks as follows. subect to m f 0 (x) a 0 z (c y 1 2 d y 2 ) f (x) a z y 0, = 1,..., m α y 0, z 0. x β, = 1,..., n = 1,..., m (6.1) The approxmatng functons f (x) = n =1 f p u x (x) are chosen as q x l r, = 0, 1,..., m, where p q = (u = (x ) x ) ( 2 f (x ) κ, x ) l ) ( 2 f (x ) κ, x 4

Here, r = f (x ) α β n =1 u p x x q = max{, 0.9l 0.1x }, = mn{x max, 0.9u 0.1x }. l, ( ) f (x ) = max{0, f (x )} and x x ( ) f (x ) = max{0, f (x )}. x x The default rules for updatng the lower asymptotes l and the upper asymptotes u are as follows. The frst two teratons, when k = 1 and k = 2, l u In later teratons, when k 3, l u = x = x = x = x 0.5(x max 0.5(x max ), ). γ (x (k 1) l (k 1) ), γ (u (k 1) ), where γ = 0.7 f (x 1.2 f (x 1 f (x )(x (k 1) x (k 2) ) < 0, )(x (k 1) x (k 2) ) > 0, )(x (k 1) x (k 2) ) = 0. The default values of the parameters κ are κ = 10 3 f (x ) x 10 6 u l, for = 0, 1,.., m and = 1,.., n. (6.2) Ths mples that all the approxmatng functons are strctly convex, whch n turn mples that there s always a unque optmal soluton of the MMA subproblem. Regardless of the values of the parameters κ, the functons f are always frst order approxmatons of the orgnal functons f at the current teraton pont,.e. f f (x ) = f (x ) and f x (x ) = f x (x ). (6.3) 5