Introduction to Unconstrained Optimization: Part 2

Similar documents
Lecture Notes: Geometric Considerations in Unconstrained Optimization

Nonlinear Optimization: What s important?

, b = 0. (2) 1 2 The eigenvectors of A corresponding to the eigenvalues λ 1 = 1, λ 2 = 3 are

MATH 5720: Unconstrained Optimization Hung Phan, UMass Lowell September 13, 2018

Positive Definite Matrix

Unconstrained Optimization

1 Overview. 2 A Characterization of Convex Functions. 2.1 First-order Taylor approximation. AM 221: Advanced Optimization Spring 2016

Introduction to unconstrained optimization - direct search methods

Optimization. Escuela de Ingeniería Informática de Oviedo. (Dpto. de Matemáticas-UniOvi) Numerical Computation Optimization 1 / 30

Non-convex optimization. Issam Laradji

An Iterative Descent Method

ECE580 Exam 1 October 4, Please do not write on the back of the exam pages. Extra paper is available from the instructor.

AM 205: lecture 18. Last time: optimization methods Today: conditions for optimality

Gradient Descent. Dr. Xiaowei Huang

Numerical Optimization

HW3 - Due 02/06. Each answer must be mathematically justified. Don t forget your name. 1 2, A = 2 2

MATH529 Fundamentals of Optimization Unconstrained Optimization II

NonlinearOptimization

Econ Slides from Lecture 8

Numerical Optimization

Appendix A Taylor Approximations and Definite Matrices

Math (P)refresher Lecture 8: Unconstrained Optimization

IE 5531: Engineering Optimization I

EAD 115. Numerical Solution of Engineering and Scientific Problems. David M. Rocke Department of Applied Science

Functions of Several Variables

446 CHAP. 8 NUMERICAL OPTIMIZATION. Newton's Search for a Minimum of f(x,y) Newton s Method

Optimization: Nonlinear Optimization without Constraints. Nonlinear Optimization without Constraints 1 / 23

Lecture Unconstrained optimization. In this lecture we will study the unconstrained problem. minimize f(x), (2.1)

Linear Algebra And Its Applications Chapter 6. Positive Definite Matrix

Constrained Optimization

Convex Optimization. Problem set 2. Due Monday April 26th

Chapter 2. Optimization. Gradients, convexity, and ALS

CHAPTER 2: QUADRATIC PROGRAMMING

ECE 595, Section 10 Numerical Simulations Lecture 7: Optimization and Eigenvalues. Prof. Peter Bermel January 23, 2013

Multivariate Newton Minimanization

Math 273a: Optimization Basic concepts

Performance Surfaces and Optimum Points

Matrix Derivatives and Descent Optimization Methods

Symmetric Matrices and Eigendecomposition

10. Unconstrained minimization

Constrained optimization. Unconstrained optimization. One-dimensional. Multi-dimensional. Newton with equality constraints. Active-set method.

A Quick Tour of Linear Algebra and Optimization for Machine Learning

Image restoration: numerical optimisation

LINEAR AND NONLINEAR PROGRAMMING

Math 5BI: Problem Set 6 Gradient dynamical systems

Introduction to gradient descent

8 Numerical methods for unconstrained problems

Optimization Methods

x +3y 2t = 1 2x +y +z +t = 2 3x y +z t = 7 2x +6y +z +t = a

Trajectory-based optimization

Nonlinear Optimization for Optimal Control

Computational Methods. Least Squares Approximation/Optimization

Convex Optimization. Dani Yogatama. School of Computer Science, Carnegie Mellon University, Pittsburgh, PA, USA. February 12, 2014

Lagrange multipliers. Portfolio optimization. The Lagrange multipliers method for finding constrained extrema of multivariable functions.

Coordinate Update Algorithm Short Course Proximal Operators and Algorithms

Conjugate Gradient (CG) Method

IPAM Summer School Optimization methods for machine learning. Jorge Nocedal

Optimization. Next: Curve Fitting Up: Numerical Analysis for Chemical Previous: Linear Algebraic and Equations. Subsections

Here each term has degree 2 (the sum of exponents is 2 for all summands). A quadratic form of three variables looks as

Structural and Multidisciplinary Optimization. P. Duysinx and P. Tossings

Practical Optimization: Basic Multidimensional Gradient Methods

Computational Optimization. Convexity and Unconstrained Optimization 1/29/08 and 2/1(revised)

Neural Network Training

Written Examination

4 Newton Method. Unconstrained Convex Optimization 21. H(x)p = f(x). Newton direction. Why? Recall second-order staylor series expansion:

Optimization and Calculus

A geometric proof of the spectral theorem for real symmetric matrices

OPER 627: Nonlinear Optimization Lecture 14: Mid-term Review

Math 3013 Solutions to Problem Set 5

14. Nonlinear equations

Economics 101A (Lecture 3) Stefano DellaVigna

A A x i x j i j (i, j) (j, i) Let. Compute the value of for and

4TE3/6TE3. Algorithms for. Continuous Optimization

5 Quasi-Newton Methods

IE 5531: Engineering Optimization I

Higher-Order Methods

Introduction to Optimization

Optimization Methods. Lecture 18: Optimality Conditions and. Gradient Methods. for Unconstrained Optimization

IFT Lecture 2 Basics of convex analysis and gradient descent

Math 118, Fall 2014 Final Exam

Lec3p1, ORF363/COS323

NUMERICAL MATHEMATICS AND COMPUTING

Complexity analysis of second-order algorithms based on line search for smooth nonconvex optimization

Math 273a: Optimization Netwon s methods

ECE580 Fall 2015 Solution to Midterm Exam 1 October 23, Please leave fractions as fractions, but simplify them, etc.

Chapter 10 Conjugate Direction Methods

Optimization II: Unconstrained Multivariable

10.34 Numerical Methods Applied to Chemical Engineering Fall Quiz #1 Review

Two hours. To be provided by Examinations Office: Mathematical Formula Tables. THE UNIVERSITY OF MANCHESTER. xx xxxx 2017 xx:xx xx.

Nonlinear equations and optimization

2.098/6.255/ Optimization Methods Practice True/False Questions

age for ^^Undergraduates Gregory V. Bard Providence, Rhode Island AMERICAN MATHEMATICAL SOCIETY

Numerical optimization

Chapter 3 Numerical Methods

Math 240, 4.3 Linear Independence; Bases A. DeCelles. 1. definitions of linear independence, linear dependence, dependence relation, basis

Scientific Computing: An Introductory Survey

Non-Convex Optimization in Machine Learning. Jan Mrkos AIC

Unconstrained minimization of smooth functions

Second Order Optimization Algorithms I

Mathematical optimization

Transcription:

Introduction to Unconstrained Optimization: Part 2 James Allison ME 555 January 29, 2007

Overview Recap Recap selected concepts from last time (with examples) Use of quadratic functions Tests for positive definiteness Demonstration: tire clearance problem using gradient descent Convex sets and functions Demonstration: Newton s method for non-quadratic problem

Use of Function approximations Numerical Examples

Use of Function approximations Derivation of optimality conditions Development of optimization algorithms Numerical Examples

Gradient of a Quadratic Function Example: f (x) = f 0 + b x + 1 2 x Ax where: Gradient: [ a1 a A = 2 a 3 a 4 ] [ b1, b = b 2 ]

Gradient of a Quadratic Function Example: where: Gradient: f (x) = f (x) = f 0 + b x + 1 2 x Ax [ a1 a A = 2 a 3 a 4 [ b1 b 2 ] [ b1, b = b 2 ] [ a1 x + 1 a 2 x 2 a 3 x 1 a 4 x 2 ] ] = b + Ax

Hessian of a Quadratic Function Hessian: f (x) = [ b1 b 2 ] [ a1 x + 1 a 2 x 2 a 3 x 1 a 4 x 2 ]

Hessian of a Quadratic Function Hessian: f (x) = [ b1 b 2 ] [ a1 x + 1 a 2 x 2 a 3 x 1 a 4 x 2 [ ] a1 a H = 2 = A a 3 a 4 ]

Jacobian Recap Used in discussion on Newton s method Used again in Ch. 5 First derivative of a vector-valued, multivariate function If f = [f 1, f 2,..., f m ] T and x = [x 1, x 2,..., x n ] T, then f 1 / x 1... f 1 / x n J..... f m / x 1... f m / x n

Jacobian Recap f 1 / x 1... f 1 / x n J..... f m / x 1... f m / x n written in terms of the gradients of f:

Jacobian Recap f 1 / x 1... f 1 / x n J..... f m / x 1... f m / x n written in terms of the gradients of f: f 1 (x) T f 2 (x) T J =. f n (x) T

Tests for 1 λ i > 0 i = 1, 2,..., n 2 Determinant of all leading principal minors is positive 3 All pivots of A in rref are positive

Tests for 1 λ i > 0 i = 1, 2,..., n Practical, fast algorithms available, eigenvalues provide insight (How?) 2 Determinant of all leading principal minors is positive 3 All pivots of A in rref are positive

Tests for 1 λ i > 0 i = 1, 2,..., n Practical, fast algorithms available, eigenvalues provide insight (How?) 2 Determinant of all leading principal minors is positive What is a leading principal minor? Determinant? Useful for quick checks on small matrices Equivalent to all eigenvalues positive 3 All pivots of A in rref are positive

Tests for 1 λ i > 0 i = 1, 2,..., n Practical, fast algorithms available, eigenvalues provide insight (How?) 2 Determinant of all leading principal minors is positive What is a leading principal minor? Determinant? Useful for quick checks on small matrices Equivalent to all eigenvalues positive 3 All pivots of A in rref are positive Also equivalent to all eigenvalues positive (May as well compute eigenvalues)

Example Gradient: min x f (x) = 4x 1 + 2x 2 + 4x 2 1 4x 1 x 2 + x 2 2 Hessian:

Example Gradient: Hessian: min x f (x) = 4x 1 + 2x 2 + 4x 2 1 4x 1 x 2 + x 2 2 [ ] 4 + 8x1 4x f (x) = 2 2 4x 1 + 2x 2

Example Gradient: Hessian: min x f (x) = 4x 1 + 2x 2 + 4x 2 1 4x 1 x 2 + x 2 2 [ ] 4 + 8x1 4x f (x) = 2 2 4x 1 + 2x 2 H = [ 8 4 4 2 ]

Saddle Point Descent Example How can you determine what directions will result in descent after moving from a saddle point?

Saddle Point Descent Example How can you determine what directions will result in descent after moving from a saddle point? Approach: find pertubations ( f = f (x) f (x 0 )) that result in a function decrease

Saddle Point Descent Example How can you determine what directions will result in descent after moving from a saddle point? Approach: find pertubations ( f = f (x) f (x 0 )) that result in a function decrease Example (from Monday): [ ] f (x) = x 5 2.6 Ax, wherea = 2.6 2 v 1 = [.949.314 ] [.314, λ 1 = 5.86, v 2 =.949 ], λ 2 = 2.86

* ), * ), * ), Recap! #"$ % 0 - %1"2./ 0 34$5 76 8 %9 /% #"$ % '() '() '() & & & -! #"$./ 0 - %1"2./ 0 34$5 76 8 %9 /% #"$ % +

& % & % & % Recap Saddle Point Descent Example! " '!(" ' " )+*", - ". ' " $& $& $& $% $% $% #& #& #& #$% #$% #$% #$& % #$% $% #$& % #$% $% #$& % #$% $% " λ 1 =.769 λ 2 = 7.23 λ 1 =.723 λ 2 =.769 λ 1 = 5.86 λ 2 = 2.86

Tire Clearance Design Problem Physical design optimization problem Simplified to two variables for visualization Solved using gradient descent What is the minimum clearance between a vehicle tire and its surrounding wheel well throughout its range of motion?

Orthogonal Gradient Descent Directions

Set and Function Convex Set: a line connecting any two points in the set is completely in the set x(λ) = λx 2 + (1 λ)x 1, 0 λ 1 Convex Function: a line connecting any two points on a function lies on or below the function f (x(λ)) λf (x 2 ) + (1 λ)f (x 1 )

Newton s Method for a Cubic Objective gradient: min x f (x) = x 3 1 9x 2 1 + 23x 1 + x 3 2 9x 2 2 + 23x 2 30 Hessian:

Newton s Method for a Cubic Objective gradient: Hessian: min x f (x) = x 3 1 9x 2 1 + 23x 1 + x 3 2 9x 2 2 + 23x 2 30 f (x) = [ 3x 2 1 18x 1 + 23 3x 2 2 18x 2 + 23 ]

Newton s Method for a Cubic Objective gradient: Hessian: min x f (x) = x 3 1 9x 2 1 + 23x 1 + x 3 2 9x 2 2 + 23x 2 30 f (x) = H = [ 3x 2 1 18x 1 + 23 3x 2 2 18x 2 + 23 [ 6x1 18 0 0 6x 2 18 ] ]