Nonlinear programming
|
|
- Victoria Weaver
- 5 years ago
- Views:
Transcription
1 htt://staff.chemeng.lth.se/~berntn/courses/otimps.htm Otimization of Process Systems Nonlinear rogramming PhD course 08 Bernt Nilsson, Det of Chemical Engineering, Lund University Content Unconstraint Nonlinear Programming Problem formulation Conve roblems Methods Direct methods Search methods, Nelder-Mead and Pattern search Indirect methods and Newton based methods Line search acobian calculations and Hessian udates
2 Unconstraint minimization min X Minimize objective function, by changing the decision variables Maimize => Minimize Does a minima eist? -* = htt://en.wikiedia.org/wiki/nonlinear_rogramming Unconstraint minimization min X Elicit model based otimization min, y X objective dy dt f y, rocess model Parameterized otimization min, y, X objective dy dt f y,, rocess model
3 Unconstraint minimization - conve Does a minima eist? => Is the roblem conve? Yes, if f >0 from one-dim math Multi dimension second order derivative = Hessian matri, H i Hij here eist a minima if H is ositive definite => H has ositive eigenvalues j min X Other roblems Multile otima Local and global solution Noisy objective Robust solver 3
4 Methods overview Direct methods single-oint, derivative-free Fundamental aroach: Probing of the objective function Eamle: Nelder-Mead method Indirect methods single-oint, derivative-based Fundamental aroach: Derivative-based iteration Eamle: quasi-newton metods Evolutionary methods multi-oint, derivative-free Fundamental aroach: Survival of the fittest Eamle: Genetic algorithm Direct methods Fundamental aroach: Probing of the objective function Uses only direct function calls Characteristics: Easy to use Can handle comle object functions discontinuous, noisy, Slow convergence Eamle: Nelder-Mead Pattern search 4
5 Direct method: Nelder-Mead method fminsearch Nelder-Mead simle method Based on a regular figure a simle riangle in -D etrahedron in 3-D Function calls in all corners Steing Reflect = ste away Eand = far away Contract = otima inside 3 4 Reflection = steing away Contraction = steing inside htt://en.wikiedia.org/wiki/nelder%e%80%93mead_method Direct method: Nelder-Mead method. Order all values at the vertices Calculate 0, the center of gravity. Reflection Comute reflected oint, r. If better than second worst, then accet new oint 3. Eansion If reflected oint is best then comute eanded oint. If eanded oint is better than reflected, then accet eanded oint. 4. Contraction If reflected oint is worst then comute contraction oint. If contraction oint is better than worst oint, then accet contraction oint. 5. Reduction For all but the best oint, relace with a reduced oint. f f... f n f r 0 0 n f r f n n r f r f e 0 0 n f f e r n e f n f r c 0 0 n f c f n n c i i 5
6 Direct method: Nelder-Mead method. Order all values at the vertices Calculate 0, the center of gravity r 0 0 n f f. Reflection f r n f r f e 0 0 n f f n e 3. Eansion e r 4. Contraction c 0 0 n f c f n f n r f f f... f n n r n c 5. Reduction i i Direct method: Pattern search method atternsearch Multile function calls in one ste Better descrition of objective function Parallell comuting Polling, new oints in mesh Better oint => eand No better oints => refine htt://en.wikiedia.org/wiki/pattern_search_otimization 6
7 Indirect methods Derivative-based search Fundamental aroach: Newton s method min ' 0 '' ' Newton method I one variable Fundamental aroach: Newton s method Newton in one variable c 0 c c c' c' c Minimization using Newton 0 min ' 0 ' ' '' '' ' 0 Note! Newton s method search for etreem oint => can t see the difference between minima or maima 7
8 Newton method II multile variables acobian gradient vector and Hessian matri second order derivative Newton minimization H H H H 0 n n n n n H Ste: called a Newton ste quasi-newton method Newton ste Problems handled in quasi-newton:. acobian Finite difference Sensitivity functions. Hessian is comutational eensive and have low accuracy Hessian aroimations is the search direction 3. uasi-newton ste length Line search methods garantee decent H B B H, htt://en.wikiedia.org/wiki/uasi-newton_method
9 quasi-newton acobian B acobian, = Partial derivative of the objective function,, with resect to the decision variables, n Numerical aroimation finite difference using a small disturbance - disturbance size have to be bigger then the error in -calculation = +ε ε htt://en.wikiedia.org/wiki/bfgs quasi-newton Hessian B Secant here in D Broyden udate BFGS udate Broyden Fletcher Goldfarb - Shannon k B k k B B k B B B B B becomes not ositive definite after 5 udates => restart B DFP udate of the B-invers Davidson Fletcher Powell B B B B B start B 0 = I => steeest decent = htt://en.wikiedia.org/wiki/bfgs 9
10 dogleg quasi-newton 3 ste length B Near the solution the Newton method is very good! Far from solution B is wrong, making the Newton ste bad large Line search methods Use, and + to make a quadratic estimate of the minimum, find that fullfill, Goldstein-Armijo 0 ' ' where 0 4 & 0.9 rust-region where quadratic aroimation is good Ratio actual/redicted H 0 reduction of, ar/r H I 0 => small, decrease tr => medium, unchange tr H I => large, increase tr Summary Nonlinear rogramming of unconstraint roblem Conve roblem => a solution eists! Method classes: Direct methods E: Nelder-Mead and attern search Indirect methods E: uasi-newton with i Numerical acobian, ii BFGS udate and iii line search ste size method Evolutionary methods GA/DE 0
Image Alignment Computer Vision (Kris Kitani) Carnegie Mellon University
Lucas Kanade Image Alignment 16-385 Comuter Vision (Kris Kitani) Carnegie Mellon University htt://www.humansensing.cs.cmu.edu/intraface/ How can I find in the image? Idea #1: Temlate Matching Slow, combinatory,
More informationOptimization: Nonlinear Optimization without Constraints. Nonlinear Optimization without Constraints 1 / 23
Optimization: Nonlinear Optimization without Constraints Nonlinear Optimization without Constraints 1 / 23 Nonlinear optimization without constraints Unconstrained minimization min x f(x) where f(x) is
More information5 Quasi-Newton Methods
Unconstrained Convex Optimization 26 5 Quasi-Newton Methods If the Hessian is unavailable... Notation: H = Hessian matrix. B is the approximation of H. C is the approximation of H 1. Problem: Solve min
More informationMATH 4211/6211 Optimization Quasi-Newton Method
MATH 4211/6211 Optimization Quasi-Newton Method Xiaojing Ye Department of Mathematics & Statistics Georgia State University Xiaojing Ye, Math & Stat, Georgia State University 0 Quasi-Newton Method Motivation:
More informationNONLINEAR OPTIMIZATION WITH CONVEX CONSTRAINTS. The Goldstein-Levitin-Polyak algorithm
- (23) NLP - NONLINEAR OPTIMIZATION WITH CONVEX CONSTRAINTS The Goldstein-Levitin-Polya algorithm We consider an algorithm for solving the otimization roblem under convex constraints. Although the convexity
More informationLecture V. Numerical Optimization
Lecture V Numerical Optimization Gianluca Violante New York University Quantitative Macroeconomics G. Violante, Numerical Optimization p. 1 /19 Isomorphism I We describe minimization problems: to maximize
More informationEAD 115. Numerical Solution of Engineering and Scientific Problems. David M. Rocke Department of Applied Science
EAD 115 Numerical Solution of Engineering and Scientific Problems David M. Rocke Department of Applied Science Multidimensional Unconstrained Optimization Suppose we have a function f() of more than one
More informationConvex Optimization CMU-10725
Convex Optimization CMU-10725 Quasi Newton Methods Barnabás Póczos & Ryan Tibshirani Quasi Newton Methods 2 Outline Modified Newton Method Rank one correction of the inverse Rank two correction of the
More informationUnconstrained Multivariate Optimization
Unconstrained Multivariate Optimization Multivariate optimization means optimization of a scalar function of a several variables: and has the general form: y = () min ( ) where () is a nonlinear scalar-valued
More information(One Dimension) Problem: for a function f(x), find x 0 such that f(x 0 ) = 0. f(x)
Solving Nonlinear Equations & Optimization One Dimension Problem: or a unction, ind 0 such that 0 = 0. 0 One Root: The Bisection Method This one s guaranteed to converge at least to a singularity, i not
More informationQuasi-Newton Methods
Newton s Method Pros and Cons Quasi-Newton Methods MA 348 Kurt Bryan Newton s method has some very nice properties: It s extremely fast, at least once it gets near the minimum, and with the simple modifications
More information1 Numerical optimization
Contents 1 Numerical optimization 5 1.1 Optimization of single-variable functions............ 5 1.1.1 Golden Section Search................... 6 1.1. Fibonacci Search...................... 8 1. Algorithms
More informationOptimization II: Unconstrained Multivariable
Optimization II: Unconstrained Multivariable CS 205A: Mathematical Methods for Robotics, Vision, and Graphics Justin Solomon CS 205A: Mathematical Methods Optimization II: Unconstrained Multivariable 1
More informationNumerical Methods: Structured vs. unstructured grids. General Introduction: Why numerical methods? Numerical methods and their fields of application
Numerical Methods: Structured vs. unstructured grids The goals o this course General : Why numerical methods? Numerical methods and their ields o alication Review o inite dierences Goals: Understanding
More information2. Quasi-Newton methods
L. Vandenberghe EE236C (Spring 2016) 2. Quasi-Newton methods variable metric methods quasi-newton methods BFGS update limited-memory quasi-newton methods 2-1 Newton method for unconstrained minimization
More information+++ Modeling of Structural-dynamic Systems by UML Statecharts in AnyLogic +++ Modeling of Structural-dynamic Systems by UML Statecharts in AnyLogic
Modeling of Structural-dynamic Systems by UML Statecharts in AnyLogic Daniel Leitner, Johannes Krof, Günther Zauner, TU Vienna, Austria, dleitner@osiris.tuwien.ac.at Yuri Karov, Yuri Senichenkov, Yuri
More informationOptimal array pattern synthesis with desired magnitude response
Otimal array attern synthesis with desired magnitude resonse A.M. Pasqual a, J.R. Arruda a and P. erzog b a Universidade Estadual de Caminas, Rua Mendeleiev, 00, Cidade Universitária Zeferino Vaz, 13083-970
More informationNumerical Optimization: Basic Concepts and Algorithms
May 27th 2015 Numerical Optimization: Basic Concepts and Algorithms R. Duvigneau R. Duvigneau - Numerical Optimization: Basic Concepts and Algorithms 1 Outline Some basic concepts in optimization Some
More informationProgramming, numerics and optimization
Programming, numerics and optimization Lecture C-3: Unconstrained optimization II Łukasz Jankowski ljank@ippt.pan.pl Institute of Fundamental Technological Research Room 4.32, Phone +22.8261281 ext. 428
More informationResonances in high-contrast gratings with complex unit cell topology
Resonances in high-contrast gratings with comle unit cell toology Milan Maksimovic Focal-Vision & Otics, Oldenzaal, The Netherlands The XXI International Worksho on Otical Wave & Waveguide Theory and Numerical
More information1 Numerical optimization
Contents Numerical optimization 5. Optimization of single-variable functions.............................. 5.. Golden Section Search..................................... 6.. Fibonacci Search........................................
More informationImproving L-BFGS Initialization for Trust-Region Methods in Deep Learning
Improving L-BFGS Initialization for Trust-Region Methods in Deep Learning Jacob Rafati http://rafati.net jrafatiheravi@ucmerced.edu Ph.D. Candidate, Electrical Engineering and Computer Science University
More informationStatic unconstrained optimization
Static unconstrained optimization 2 In unconstrained optimization an objective function is minimized without any additional restriction on the decision variables, i.e. min f(x) x X ad (2.) with X ad R
More informationLecture 7 Unconstrained nonlinear programming
Lecture 7 Unconstrained nonlinear programming Weinan E 1,2 and Tiejun Li 2 1 Department of Mathematics, Princeton University, weinan@princeton.edu 2 School of Mathematical Sciences, Peking University,
More informationQuasi-Newton methods: Symmetric rank 1 (SR1) Broyden Fletcher Goldfarb Shanno February 6, / 25 (BFG. Limited memory BFGS (L-BFGS)
Quasi-Newton methods: Symmetric rank 1 (SR1) Broyden Fletcher Goldfarb Shanno (BFGS) Limited memory BFGS (L-BFGS) February 6, 2014 Quasi-Newton methods: Symmetric rank 1 (SR1) Broyden Fletcher Goldfarb
More informationChapter 4. Unconstrained optimization
Chapter 4. Unconstrained optimization Version: 28-10-2012 Material: (for details see) Chapter 11 in [FKS] (pp.251-276) A reference e.g. L.11.2 refers to the corresponding Lemma in the book [FKS] PDF-file
More informationMethods that avoid calculating the Hessian. Nonlinear Optimization; Steepest Descent, Quasi-Newton. Steepest Descent
Nonlinear Optimization Steepest Descent and Niclas Börlin Department of Computing Science Umeå University niclas.borlin@cs.umu.se A disadvantage with the Newton method is that the Hessian has to be derived
More informationMultidisciplinary System Design Optimization (MSDO)
Multiiscilinary System Design Otimization (MSDO) Graient Calculation an Sensitivity Analysis Lecture 9 Olivier e Weck Karen Willco Massachusetts Institute of Technology - Prof. e Weck an Prof. Willco Toay
More informationOptimization II: Unconstrained Multivariable
Optimization II: Unconstrained Multivariable CS 205A: Mathematical Methods for Robotics, Vision, and Graphics Doug James (and Justin Solomon) CS 205A: Mathematical Methods Optimization II: Unconstrained
More informationOptimization Methods
Optimization Methods Decision making Examples: determining which ingredients and in what quantities to add to a mixture being made so that it will meet specifications on its composition allocating available
More informationShiqian Ma, MAT-258A: Numerical Optimization 1. Chapter 3. Gradient Method
Shiqian Ma, MAT-258A: Numerical Optimization 1 Chapter 3 Gradient Method Shiqian Ma, MAT-258A: Numerical Optimization 2 3.1. Gradient method Classical gradient method: to minimize a differentiable convex
More informationNumerical Methods I Solving Nonlinear Equations
Numerical Methods I Solving Nonlinear Equations Aleksandar Donev Courant Institute, NYU 1 donev@courant.nyu.edu 1 MATH-GA 2011.003 / CSCI-GA 2945.003, Fall 2014 October 16th, 2014 A. Donev (Courant Institute)
More informationMAP Estimation Algorithms in Computer Vision - Part II
MAP Estimation Algorithms in Comuter Vision - Part II M. Pawan Kumar, University of Oford Pushmeet Kohli, Microsoft Research Eamle: Image Segmentation E() = c i i + c ij i (1- j ) i i,j E: {0,1} n R 0
More informationOptimization 2. CS5240 Theoretical Foundations in Multimedia. Leow Wee Kheng
Optimization 2 CS5240 Theoretical Foundations in Multimedia Leow Wee Kheng Department of Computer Science School of Computing National University of Singapore Leow Wee Kheng (NUS) Optimization 2 1 / 38
More informationIntroduction to unconstrained optimization - direct search methods
Introduction to unconstrained optimization - direct search methods Jussi Hakanen Post-doctoral researcher jussi.hakanen@jyu.fi Structure of optimization methods Typically Constraint handling converts the
More informationOptimization and Root Finding. Kurt Hornik
Optimization and Root Finding Kurt Hornik Basics Root finding and unconstrained smooth optimization are closely related: Solving ƒ () = 0 can be accomplished via minimizing ƒ () 2 Slide 2 Basics Root finding
More informationNonlinearOptimization
1/35 NonlinearOptimization Pavel Kordík Department of Computer Systems Faculty of Information Technology Czech Technical University in Prague Jiří Kašpar, Pavel Tvrdík, 2011 Unconstrained nonlinear optimization,
More information4. Score normalization technical details We now discuss the technical details of the score normalization method.
SMT SCORING SYSTEM This document describes the scoring system for the Stanford Math Tournament We begin by giving an overview of the changes to scoring and a non-technical descrition of the scoring rules
More informationToday. Introduction to optimization Definition and motivation 1-dimensional methods. Multi-dimensional methods. General strategies, value-only methods
Optimization Last time Root inding: deinition, motivation Algorithms: Bisection, alse position, secant, Newton-Raphson Convergence & tradeos Eample applications o Newton s method Root inding in > 1 dimension
More informationStatistics 580 Optimization Methods
Statistics 580 Optimization Methods Introduction Let fx be a given real-valued function on R p. The general optimization problem is to find an x ɛ R p at which fx attain a maximum or a minimum. It is of
More informationNonlinear Optimization: What s important?
Nonlinear Optimization: What s important? Julian Hall 10th May 2012 Convexity: convex problems A local minimizer is a global minimizer A solution of f (x) = 0 (stationary point) is a minimizer A global
More informationIntroduction to Black-Box Optimization in Continuous Search Spaces. Definitions, Examples, Difficulties
1 Introduction to Black-Box Optimization in Continuous Search Spaces Definitions, Examples, Difficulties Tutorial: Evolution Strategies and CMA-ES (Covariance Matrix Adaptation) Anne Auger & Nikolaus Hansen
More informationGaussian processes A hands-on tutorial
Gaussian rocesses A hands-on tutorial Slides and code: htts://github.com/araklas/gptutorial Paris Perdikaris Massachusetts Institute of Technology Deartment of Mechanical Engineering Web: htt://web.mit.edu/aris/www/
More informationC:\Users\Leonardo\Desktop\README_ELMO_NOTES.txt Montag, 7. Juli :48
*********** * Input * *********** Besides the normal GAMESS-UK input directives, to perform an ELMO calculation one needs to specify the elmo keyword and to provide additional instructions that are contained
More informationComparative study of Optimization methods for Unconstrained Multivariable Nonlinear Programming Problems
International Journal of Scientific and Research Publications, Volume 3, Issue 10, October 013 1 ISSN 50-3153 Comparative study of Optimization methods for Unconstrained Multivariable Nonlinear Programming
More informationNumerical Optimization Professor Horst Cerjak, Horst Bischof, Thomas Pock Mat Vis-Gra SS09
Numerical Optimization 1 Working Horse in Computer Vision Variational Methods Shape Analysis Machine Learning Markov Random Fields Geometry Common denominator: optimization problems 2 Overview of Methods
More informationL p Norms and the Sinc Function
L Norms and the Sinc Function D. Borwein, J. M. Borwein and I. E. Leonard March 6, 9 Introduction The sinc function is a real valued function defined on the real line R by the following eression: sin if
More informationCONTROL SYSTEMS, ROBOTICS, AND AUTOMATION Vol. III Stability Theory - Peter C. Müller
STABILITY THEORY Peter C. Müller University of Wuertal, Germany Keywords: Asymtotic stability, Eonential stability, Linearization, Linear systems, Lyaunov equation, Lyaunov function, Lyaunov stability,
More informationECS550NFB Introduction to Numerical Methods using Matlab Day 2
ECS550NFB Introduction to Numerical Methods using Matlab Day 2 Lukas Laffers lukas.laffers@umb.sk Department of Mathematics, University of Matej Bel June 9, 2015 Today Root-finding: find x that solves
More informationFeedback-error control
Chater 4 Feedback-error control 4.1 Introduction This chater exlains the feedback-error (FBE) control scheme originally described by Kawato [, 87, 8]. FBE is a widely used neural network based controller
More informationNumerical Optimization Algorithms
Numerical Optimization Algorithms 1. Overview. Calculus of Variations 3. Linearized Supersonic Flow 4. Steepest Descent 5. Smoothed Steepest Descent Overview 1 Two Main Categories of Optimization Algorithms
More informationMagnetospheric Physics - Homework, 4/04/2014
Magnetosheric hysics - Homework, // 7. Fast wave, fast shock, erendicular shock, entroy, jum relation. Consider the fast erendicular shock. a) Determine the ositive root of the solution for the comression
More informationQUIZ ON CHAPTER 4 - SOLUTIONS APPLICATIONS OF DERIVATIVES; MATH 150 FALL 2016 KUNIYUKI 105 POINTS TOTAL, BUT 100 POINTS = 100%
QUIZ ON CHAPTER - SOLUTIONS APPLICATIONS OF DERIVATIVES; MATH 150 FALL 016 KUNIYUKI 105 POINTS TOTAL, BUT 100 POINTS = 100% = x + 5 1) Consider f x and the grah of y = f x in the usual xy-lane in 16 x
More informationE-companion to A risk- and ambiguity-averse extension of the max-min newsvendor order formula
e-comanion to Han Du and Zuluaga: Etension of Scarf s ma-min order formula ec E-comanion to A risk- and ambiguity-averse etension of the ma-min newsvendor order formula Qiaoming Han School of Mathematics
More informationHigher-Order Methods
Higher-Order Methods Stephen J. Wright 1 2 Computer Sciences Department, University of Wisconsin-Madison. PCMI, July 2016 Stephen Wright (UW-Madison) Higher-Order Methods PCMI, July 2016 1 / 25 Smooth
More informationA Primer on Multidimensional Optimization
A Primer on Multidimensional Optimization Prof. Dr. Florian Rupp German University of Technology in Oman (GUtech) Introduction to Numerical Methods for ENG & CS (Mathematics IV) Spring Term 2016 Eercise
More informationMultivariate Newton Minimanization
Multivariate Newton Minimanization Optymalizacja syntezy biosurfaktantu Rhamnolipid Rhamnolipids are naturally occuring glycolipid produced commercially by the Pseudomonas aeruginosa species of bacteria.
More informationOptimization Methods for Circuit Design
Technische Universität München Department of Electrical Engineering and Information Technology Institute for Electronic Design Automation Optimization Methods for Circuit Design Compendium H. Graeb Version
More informationShort course A vademecum of statistical pattern recognition techniques with applications to image and video analysis. Agenda
Short course A vademecum of statistical attern recognition techniques with alications to image and video analysis Lecture 6 The Kalman filter. Particle filters Massimo Piccardi University of Technology,
More informationConvex Optimization methods for Computing Channel Capacity
Convex Otimization methods for Comuting Channel Caacity Abhishek Sinha Laboratory for Information and Decision Systems (LIDS), MIT sinhaa@mit.edu May 15, 2014 We consider a classical comutational roblem
More informationUnderstanding DPMFoam/MPPICFoam
Understanding DPMFoam/MPPICFoam Jeroen Hofman March 18, 2015 In this document I intend to clarify the flow solver and at a later stage, the article-fluid and article-article interaction forces as imlemented
More informationAlgorithms for Constrained Optimization
1 / 42 Algorithms for Constrained Optimization ME598/494 Lecture Max Yi Ren Department of Mechanical Engineering, Arizona State University April 19, 2015 2 / 42 Outline 1. Convergence 2. Sequential quadratic
More informationLecture 7: Linear Classification Methods
Homeork Homeork Lecture 7: Linear lassification Methods Final rojects? Grous oics Proosal eek 5 Lecture is oster session, Jacobs Hall Lobby, snacks Final reort 5 June. What is linear classification? lassification
More informationMethods for Unconstrained Optimization Numerical Optimization Lectures 1-2
Methods for Unconstrained Optimization Numerical Optimization Lectures 1-2 Coralia Cartis, University of Oxford INFOMM CDT: Modelling, Analysis and Computation of Continuous Real-World Problems Methods
More informationSecond Order Symmetric and Maxmin Symmetric Duality with Cone Constraints
International Journal of Oerations Research International Journal of Oerations Research Vol. 4, No. 4, 99 5 7) Second Order Smmetric Mamin Smmetric Dualit with Cone Constraints I. Husain,, Abha Goel, M.
More informationUnconstrained optimization
Chapter 4 Unconstrained optimization An unconstrained optimization problem takes the form min x Rnf(x) (4.1) for a target functional (also called objective function) f : R n R. In this chapter and throughout
More informationSemi-Infinite Optimization with Implicit Functions
ubs.acs.org/iecr Semi-Infinite Otimization with Imlicit Functions Matthew D. Stuber and Paul I. Barton* Process Systems Engineering Laboratory, Det. of Chemical Engineering, Massachusetts Institute of
More informationCMA-ES a Stochastic Second-Order Method for Function-Value Free Numerical Optimization
CMA-ES a Stochastic Second-Order Method for Function-Value Free Numerical Optimization Nikolaus Hansen INRIA, Research Centre Saclay Machine Learning and Optimization Team, TAO Univ. Paris-Sud, LRI MSRC
More informationPart III. for energy minimization
ICCV 2007 tutorial Part III Message-assing algorithms for energy minimization Vladimir Kolmogorov University College London Message assing ( E ( (,, Iteratively ass messages between nodes... Message udate
More informationJorge Marques, Image motion
Image motion finding a temlate Suose we wish to find a known temlate ) in a given image I). his roblem is known as temlate matching. temlate image FC Barcelona the temlate can be small or large alignment
More informationUniform Law on the Unit Sphere of a Banach Space
Uniform Law on the Unit Shere of a Banach Sace by Bernard Beauzamy Société de Calcul Mathématique SA Faubourg Saint Honoré 75008 Paris France Setember 008 Abstract We investigate the construction of a
More informationData Mining (Mineria de Dades)
Data Mining (Mineria de Dades) Lluís A. Belanche belanche@lsi.upc.edu Soft Computing Research Group Dept. de Llenguatges i Sistemes Informàtics (Software department) Universitat Politècnica de Catalunya
More informationBayesian Networks for Modeling and Managing Risks of Natural Hazards
[National Telford Institute and Scottish Informatics and Comuter Science Alliance, Glasgow University, Set 8, 200 ] Bayesian Networks for Modeling and Managing Risks of Natural Hazards Daniel Straub Engineering
More informationRadial Basis Function Networks: Algorithms
Radial Basis Function Networks: Algorithms Introduction to Neural Networks : Lecture 13 John A. Bullinaria, 2004 1. The RBF Maing 2. The RBF Network Architecture 3. Comutational Power of RBF Networks 4.
More informationEfficient & Robust LK for Mobile Vision
Efficient & Robust LK for Mobile Vision Instructor - Simon Lucey 16-623 - Designing Comuter Vision As Direct Method (ours) Indirect Method (ORB+RANSAC) H. Alismail, B. Browning, S. Lucey Bit-Planes: Dense
More informationLBFGS. John Langford, Large Scale Machine Learning Class, February 5. (post presentation version)
LBFGS John Langford, Large Scale Machine Learning Class, February 5 (post presentation version) We are still doing Linear Learning Features: a vector x R n Label: y R Goal: Learn w R n such that ŷ w (x)
More informationShort Solutions to Practice Material for Test #2 MATH 2421
Short Solutions to Practice Material for Test # MATH 4 Kawai (#) Describe recisely the D surfaces listed here (a) + y + z z = Shere ( ) + (y ) + (z ) = 4 = The center is located at C (; ; ) and the radius
More informationMachine Learning: Homework 4
10-601 Machine Learning: Homework 4 Due 5.m. Monday, February 16, 2015 Instructions Late homework olicy: Homework is worth full credit if submitted before the due date, half credit during the next 48 hours,
More informationPart 2: NLP Constrained Optimization
Part 2: NLP Constrained Optimization James G. Shanahan 2 Independent Consultant and Lecturer UC Santa Cruz EMAIL: James_DOT_Shanahan_AT_gmail_DOT_com WIFI: SSID Student USERname ucsc-guest Password EnrollNow!
More informationA = Chapter 6. Linear Programming: The Simplex Method. + 21x 3 x x 2. C = 16x 1. + x x x 1. + x 3. 16,x 2.
Chapter 6 Linear rogramming: The Simple Method Section The Dual roblem: Minimization with roblem Constraints of the Form Learning Objectives for Section 6. Dual roblem: Minimization with roblem Constraints
More informationApplied Fitting Theory VI. Formulas for Kinematic Fitting
Alied Fitting heory VI Paul Avery CBX 98 37 June 9, 1998 Ar. 17, 1999 (rev.) I Introduction Formulas for Kinematic Fitting I intend for this note and the one following it to serve as mathematical references,
More informationOPER 627: Nonlinear Optimization Lecture 14: Mid-term Review
OPER 627: Nonlinear Optimization Lecture 14: Mid-term Review Department of Statistical Sciences and Operations Research Virginia Commonwealth University Oct 16, 2013 (Lecture 14) Nonlinear Optimization
More informationQuasi-Newton methods for minimization
Quasi-Newton methods for minimization Lectures for PHD course on Numerical optimization Enrico Bertolazzi DIMS Universitá di Trento November 21 December 14, 2011 Quasi-Newton methods for minimization 1
More informationNotes on Numerical Optimization
Notes on Numerical Optimization University of Chicago, 2014 Viva Patel October 18, 2014 1 Contents Contents 2 List of Algorithms 4 I Fundamentals of Optimization 5 1 Overview of Numerical Optimization
More informationLecture 14: October 17
1-725/36-725: Convex Optimization Fall 218 Lecture 14: October 17 Lecturer: Lecturer: Ryan Tibshirani Scribes: Pengsheng Guo, Xian Zhou Note: LaTeX template courtesy of UC Berkeley EECS dept. Disclaimer:
More informationQuasi-Newton Methods
Quasi-Newton Methods Werner C. Rheinboldt These are excerpts of material relating to the boos [OR00 and [Rhe98 and of write-ups prepared for courses held at the University of Pittsburgh. Some further references
More informationPROFIT MAXIMIZATION. π = p y Σ n i=1 w i x i (2)
PROFIT MAXIMIZATION DEFINITION OF A NEOCLASSICAL FIRM A neoclassical firm is an organization that controls the transformation of inuts (resources it owns or urchases into oututs or roducts (valued roducts
More informationConvexity and SOS-Convexity
Conveit and SOS-Conveit Amir Ali Ahmadi Pablo A. Parrilo Laborator for Information and Decision Sstems Massachusetts Institute of Technolog SIAM Conference on Alied Algebraic Geometr NCSU, October 0 Deciding
More informationSTAT Advanced Bayesian Inference
1 / 8 STAT 625 - Advanced Bayesian Inference Meng Li Department of Statistics March 5, 2018 Distributional approximations 2 / 8 Distributional approximations are useful for quick inferences, as starting
More informationSquare-Root Information Filtering and Fixed-Interval Smoothing with Singularities
Square-Root Information Filtering and Fied-Interval Smoothing with Singularities Abstract he square-root information filter and smoother algorithms have been generalized to handle singular state transition
More informationGOOD MODELS FOR CUBIC SURFACES. 1. Introduction
GOOD MODELS FOR CUBIC SURFACES ANDREAS-STEPHAN ELSENHANS Abstract. This article describes an algorithm for finding a model of a hyersurface with small coefficients. It is shown that the aroach works in
More informationSolved Problems. (a) (b) (c) Figure P4.1 Simple Classification Problems First we draw a line between each set of dark and light data points.
Solved Problems Solved Problems P Solve the three simle classification roblems shown in Figure P by drawing a decision boundary Find weight and bias values that result in single-neuron ercetrons with the
More informationSTUDYING THE BASIN OF CONVERGENCE OF METHODS FOR COMPUTING PERIODIC ORBITS
Tutorials and Reviews International Journal of Bifurcation and Chaos, Vol. 21, No. 8 (2011) 2079 2106 c World Scientific Publishing Company DOI: 10.1142/S0218127411029653 STUDYING THE BASIN OF CONVERGENCE
More informationtransformation, and nonlinear deformations containing local The deformation can be decomposed into a global affine progress).
Image Waring for Forecast Verification Johan Lindström, Eric Gilleland 2 & Finn Lindgren Centre for Mathematical Sciences, Lund University 2 Research Alications Laboratory, National Center for Atmosheric
More informationALTERNATIVE SOLUTION TO THE QUARTIC EQUATION by Farid A. Chouery 1, P.E. 2006, All rights reserved
ALTERNATIVE SOLUTION TO THE QUARTIC EQUATION b Farid A. Chouer, P.E. 006, All rights reserved Abstract A new method to obtain a closed form solution of the fourth order olnomial equation is roosed in this
More informationMinimization of Static! Cost Functions!
Minimization of Static Cost Functions Robert Stengel Optimal Control and Estimation, MAE 546, Princeton University, 2017 J = Static cost function with constant control parameter vector, u Conditions for
More informationOnline Learning for Sparse PCA in High Dimensions: Exact Dynamics and Phase Transitions
Online Learning for Sarse PCA in High Dimensions: Eact Dynamics and Phase Transitions Chuang Wang and Yue M. Lu John A. Paulson School of Engineering and Alied Sciences Harvard University Abstract We study
More informationNumerical optimization
THE UNIVERSITY OF WESTERN ONTARIO LONDON ONTARIO Paul Klein Office: SSC 408 Phone: 661-111 ext. 857 Email: paul.klein@uwo.ca URL: www.ssc.uwo.ca/economics/faculty/klein/ Numerical optimization In these
More informationLecture 7: Optimization methods for non linear estimation or function estimation
Lecture 7: Optimization methods for non linear estimation or function estimation Y. Favennec 1, P. Le Masson 2 and Y. Jarny 1 1 LTN UMR CNRS 6607 Polytetch Nantes 44306 Nantes France 2 LIMATB Université
More informationFUZZY CONTROL. Khurshid Ahmad, Professor of Computer Science, Department of Computer Science Trinity College, Dublin-2, IRELAND Nov 15-16th, 2011
: Mamdani & akagi-sugeno Controllers Khurshid Ahmad, Professor of Comuter Science, Deartment of Comuter Science rinit College, Dublin-, IRELAND Nov 5-6th, 0 htts://www.cs.tcd.ie/khurshid.ahmad/eaching/eaching.html
More information