MATH 4211/6211 Optimization Basics of Optimization Problems

Similar documents
Lecture 3: Basics of set-constrained and unconstrained optimization

Math 273a: Optimization Basic concepts

MATH 4211/6211 Optimization Constrained Optimization

ECE580 Fall 2015 Solution to Midterm Exam 1 October 23, Please leave fractions as fractions, but simplify them, etc.

Numerical Optimization

CE 191: Civil and Environmental Engineering Systems Analysis. LEC 05 : Optimality Conditions

Optimization. Escuela de Ingeniería Informática de Oviedo. (Dpto. de Matemáticas-UniOvi) Numerical Computation Optimization 1 / 30

MATH 4211/6211 Optimization Quasi-Newton Method

ECE 680 Modern Automatic Control. Gradient and Newton s Methods A Review

Optimization Part 1 P. Agius L2.1, Spring 2008

ECE580 Solution to Problem Set 3: Applications of the FONC, SONC, and SOSC

ECE580 Exam 1 October 4, Please do not write on the back of the exam pages. Extra paper is available from the instructor.

CE 191: Civil & Environmental Engineering Systems Analysis. LEC 17 : Final Review

Math 273a: Optimization Netwon s methods

1 Newton s Method. Suppose we want to solve: x R. At x = x, f (x) can be approximated by:

OPER 627: Nonlinear Optimization Lecture 14: Mid-term Review

Constrained optimization: direct methods (cont.)

Optimization Tutorial 1. Basic Gradient Descent

Optimization. Yuh-Jye Lee. March 21, Data Science and Machine Intelligence Lab National Chiao Tung University 1 / 29

Optimality conditions for unconstrained optimization. Outline

Scientific Computing: Optimization

Optimization Problems with Constraints - introduction to theory, numerical Methods and applications

OPER 627: Nonlinear Optimization Lecture 2: Math Background and Optimality Conditions


Line Search Methods. Shefali Kulkarni-Thaker

FALL 2018 MATH 4211/6211 Optimization Homework 4

E5295/5B5749 Convex optimization with engineering applications. Lecture 8. Smooth convex unconstrained and equality-constrained minimization

Constrained optimization. Unconstrained optimization. One-dimensional. Multi-dimensional. Newton with equality constraints. Active-set method.

Introduction to unconstrained optimization - direct search methods

ECE580 Solution to Problem Set 6

Numerical Optimization

Chapter 2: Unconstrained Extrema

MATH529 Fundamentals of Optimization Unconstrained Optimization II

Computational Optimization. Constrained Optimization Part 2

Lecture 15 Newton Method and Self-Concordance. October 23, 2008

Computational Optimization. Convexity and Unconstrained Optimization 1/29/08 and 2/1(revised)

Introduction to Nonlinear Optimization Paul J. Atzberger

Algorithms for nonlinear programming problems II

CHAPTER 2: QUADRATIC PROGRAMMING

6.252 NONLINEAR PROGRAMMING LECTURE 10 ALTERNATIVES TO GRADIENT PROJECTION LECTURE OUTLINE. Three Alternatives/Remedies for Gradient Projection

Numerisches Rechnen. (für Informatiker) M. Grepl P. Esser & G. Welper & L. Zhang. Institut für Geometrie und Praktische Mathematik RWTH Aachen

Math (P)refresher Lecture 8: Unconstrained Optimization

Computational Finance

Motivation: We have already seen an example of a system of nonlinear equations when we studied Gaussian integration (p.8 of integration notes)

Test 2 Review Math 1111 College Algebra

Optimality Conditions for Constrained Optimization

CoE 3SK3 Computer Aided Engineering Tutorial: Unconstrained Optimization

Examination paper for TMA4180 Optimization I

Algorithms for nonlinear programming problems II

Math 164 (Lec 1): Optimization Instructor: Alpár R. Mészáros

Part 5: Penalty and augmented Lagrangian methods for equality constrained optimization. Nick Gould (RAL)

MIT Manufacturing Systems Analysis Lecture 14-16

Numerical Optimization. Review: Unconstrained Optimization

Numerical Optimization

IE 5531: Engineering Optimization I

Constrained Optimization

Week 4: Calculus and Optimization (Jehle and Reny, Chapter A2)

Solution Methods. Richard Lusby. Department of Management Engineering Technical University of Denmark

Unconstrained minimization of smooth functions

Nonlinear Optimization: What s important?

Unconstrained Optimization

SYSTEMS OF NONLINEAR EQUATIONS

CONSTRAINED NONLINEAR PROGRAMMING

Math 409/509 (Spring 2011)

Numerical solutions of nonlinear systems of equations

Lecture 3. Optimization Problems and Iterative Algorithms

ISM206 Lecture Optimization of Nonlinear Objective with Linear Constraints

Multidisciplinary System Design Optimization (MSDO)

Written Examination

On the convergence properties of the modified Polak Ribiére Polyak method with the standard Armijo line search

LECTURE 25: REVIEW/EPILOGUE LECTURE OUTLINE

Arc Search Algorithms

ECE580 Partial Solution to Problem Set 3

minimize x subject to (x 2)(x 4) u,

Mathematical optimization

2.098/6.255/ Optimization Methods Practice True/False Questions

The Steepest Descent Algorithm for Unconstrained Optimization

Constrained Optimization and Lagrangian Duality

5 Quasi-Newton Methods

Numerical Optimization: Basic Concepts and Algorithms

Part 2: Linesearch methods for unconstrained optimization. Nick Gould (RAL)

14. Nonlinear equations

1. Method 1: bisection. The bisection methods starts from two points a 0 and b 0 such that

MATH2070 Optimisation

GENG2140, S2, 2012 Week 7: Curve fitting

Numerical Optimization

LECTURE 22: SWARM INTELLIGENCE 3 / CLASSICAL OPTIMIZATION

10 Numerical methods for constrained problems

An Introduction to Model-based Predictive Control (MPC) by

Algorithms for constrained local optimization

Optimization Methods. Lecture 18: Optimality Conditions and. Gradient Methods. for Unconstrained Optimization

1. Introduction. We analyze a trust region version of Newton s method for the optimization problem

Optimization with Scipy (2)

2.3 Linear Programming

Methods for Unconstrained Optimization Numerical Optimization Lectures 1-2

Line Search Methods for Unconstrained Optimisation

Interior Point Methods for Mathematical Programming

You should be able to...

Nonlinear Programming

Optimization II: Unconstrained Multivariable

Transcription:

MATH 4211/6211 Optimization Basics of Optimization Problems Xiaojing Ye Department of Mathematics & Statistics Georgia State University Xiaojing Ye, Math & Stat, Georgia State University 0

A standard minimization problem is written as where Ω is a subset of R n. minimize x Ω f(x) f(x) is called the objective function or cost function; Ω = R n : unconstrained minimization; Ω R n explicitly given: set constrained minimization; Ω = {x R n : g(x) = 0, h(x) 0}: functional constrained minimization. Xiaojing Ye, Math & Stat, Georgia State University 1

Local minimizer: x is called a local minimizer of f if ɛ > 0 such that f(x) f(x ) for all x B(x, ɛ). Global minimizer: x is called a global minimizer of f if f(x) f(x ) for all x Ω. Strict local minimizer: x is called a strict local minimizer of f if ɛ > 0 such that f(x) > f(x ) for all x B(x, ɛ) \ {x }. Strict global minimizer: x is called a strict global minimizer of f if f(x) > f(x ) for all x Ω \ {x }. (Strict) local/global maximizers are defined similarly. Xiaojing Ye, Math & Stat, Georgia State University 2

d R n is called a feasible direction at x Ω if ɛ > 0 such that x + αd Ω for all α [0, ɛ]. Directional derivative of f in the direction d is defined by lim α 0 f(x + αd) f(x) α = ( f(x)) T d Example. Let f : R 3 R defined by f(x) = x 1 x 2 x 3, and d = Compute the directional derivative of f in the direction of d: ( 1 2, 1 2, 1 2 ) T. f (x) T d = (x 2 x 3, x 1 x 3, x 1 x 2 ) ( 1 2, 1 ) T 2, 1 2 = x 2x 3 + x 1 x 3 + 2x 1 x 2 2 Xiaojing Ye, Math & Stat, Georgia State University 3

Suppose f is differentiable. First order necessary condition (FONC): x is a local minimizer, then for any feasible direction d at x, there is d T f(x ) 0. If x is an interior point of Ω, then FONC reduces to f(x ) = 0. Xiaojing Ye, Math & Stat, Georgia State University 4

Example. Consider the problem minimize f(x) := x 2 1 + 0.5x2 2 + 3x 2 + 4.5, subject to x 1, x 2 0. 1. Is the FONC for a local minimizer satisfied at x = (1, 3) T? 2. Is the FONC for a local minimizer satisfied at x = (0, 3) T? 3. Is the FONC for a local minimizer satisfied at x = (1, 0) T? 4. Is the FONC for a local minimizer satisfied at x = (0, 0) T? Idea: First compute f(x) = (2x 1, x 2 + 3) T. If x is interior point, check if f(x) = 0; otherwise, check if d T f(x) 0 for all feasible direction d. Xiaojing Ye, Math & Stat, Georgia State University 5

Suppose f is twice differentiable. Second order necessary condition (SONC): x is a local minimizer, then for any feasible direction d at x such that d T f(x ) = 0, there is d T 2 f(x )d 0. If x is an interior point of Ω, then SONC reduces to f(x ) = 0 and 2 f(x ) 0. Xiaojing Ye, Math & Stat, Georgia State University 6

Note that these conditions are necessary but not sufficient. This means x may not be a local minimizer even if FONC or SONC is satisfied. Example. Consider f(x) = x 3 for x R, then f (0) = 0, f (0) = 0. So x = 0 satisfies FONC (and even SONC), but it is not a local minimizer. Example. Let f(x) = x 2 1 x2 2, then the gradient is f(x) = (2x 1, 2x 2 ) T and the Hessian is [ ] 2 2 0 f(x) = 0 2 which is not positive semi-definite. So x = (0, 0) T does not satisfy SONC and hence is not a local minimizer of f. Xiaojing Ye, Math & Stat, Georgia State University 7

Let s see a sufficient condition of interior local minimizer. Suppose f is twice differentiable. Second order sufficient condition (SOSC): if x is an interior point and satisfies both f(x ) = 0 and 2 f(x ) 0, then x is a strict local minimizer of f. Xiaojing Ye, Math & Stat, Georgia State University 8

Example. Let f(x) = x 2 1 + x2 2, then the gradient is f(x) = (2x 1, 2x 2 ) T and the Hessian is [ ] 2 2 0 f(x) = 0 2 So x = (0, 0) T satisfies SOSC and hence is a strict local minimizer of f. Xiaojing Ye, Math & Stat, Georgia State University 9

Let s take a look of the Newton s method for univariate functions: minimize x R f(x) Suppose we obtained x (k). Now approximate f(x) nearby x (k) by a quadratic function q(x): q(x) = f(x (k) ) + f (x (k) )(x x (k) ) + 1 2 f (x (k) )(x x (k) ) 2 Note that q(x (k) ) = f(x (k) ), q (x (k) ) = f (x (k) ) and q (x (k) ) = f (x (k) ). Xiaojing Ye, Math & Stat, Georgia State University 10

If q approximates f well (we know it s true nearby x (k) ), then we can minimize q(x) instead of f(x). To this end, we compute q (x) = f (x (k) ) + f (x (k) )(x x (k) ) = 0 and solve for x to use as our next iterate x (k+1), i.e., x (k+1) = x (k) f (x (k) ) f (x (k) ) In practice, we give an initial guess x (0), and iterate the formula above to generate {x (k) : k = 0, 1,... }. Then x (k) x quickly (under some conditions). Xiaojing Ye, Math & Stat, Georgia State University 11

Example. Use Newton s method to find the minimizer of f(x) = 1 2 x2 sin x with initial value x (0) = 0.5. Stop when x (k+1) x (k) < 10 5. Solution. We first compute f (x) = x cos x, f (x) = 1 + sin x So the Newton s method gives x (k+1) = x (k) x(k) cos x (k) 1 + sin x (k) Apply this iteratively, we obtain x (0) = 0.5, x (1) = 0.7552, x (2) = 0.7391, x (3) = 0.7390, x (4) = 0.7390 and terminate. Note that f (x (4) ) = 1.673 > 0, so we expect x (4) to be a strict local minimizer. Xiaojing Ye, Math & Stat, Georgia State University 12

Newton s method can also be used for root-finding: if g(x) = f (x) for all x, then finding the roots of g is equivalent to finding the critical points of f. So for root-finding, Newton s method iterates x (k+1) = x (k) g(x(k) ) g (x (k) ) and x (k) x (under conditions) such that g(x ) = 0. Xiaojing Ye, Math & Stat, Georgia State University 13

A variation of the Newton s method is called the secant method, where f (x (k) ) is replaced by f (x (k) ) f (x (k 1) ) x (k) x (k 1) : x (k+1) = x (k) (x(k) x (k 1) )f (x (k) ) f (x (k) ) f (x (k 1) ) Need initials x (0) and x (1) to start. = x(k 1) f (x (k) ) x (k) f (x (k 1) ) f (x (k) ) f (x (k 1) ) The secant method does not need computations of f (x) at the expense of more iterations. Xiaojing Ye, Math & Stat, Georgia State University 14

Consider a general optimization problem minimize x Ω f(x) If we have x (k), then we want to find d (k) (descent direction) and α k (step size) to obtain x (k+1) = x (k) + α k d (k) such that f(x (k+1) ) reduces from f(x (k) ) as much as possible. Xiaojing Ye, Math & Stat, Georgia State University 15

How to find d (k)? Often times d (k) = f(x (k) ) but better choices may exist. Given d (k), how to find α k? This is an optimization with α k R + : minimize α>0 which is called line search. φ(α) := f(x (k) + αd (k) ) We need φ (α) = f(x (k) + αd (k) ) T d (k) and probably also φ (α) = (d (k) ) T 2 f(x (k) + αd (k) )d (k). This can be computationally expensive. We may not need to find the best α every step. Sometimes we should allocate more cost to find a better d (k). These issues will be considered in depth later in the class. Xiaojing Ye, Math & Stat, Georgia State University 16