Optimality conditions for unconstrained optimization. Outline

Similar documents
Numerical Optimization

Math (P)refresher Lecture 8: Unconstrained Optimization

Computational Optimization. Convexity and Unconstrained Optimization 1/29/08 and 2/1(revised)

OPTIMIZATION THEORY IN A NUTSHELL Daniel McFadden, 1990, 2003

MATH 4211/6211 Optimization Basics of Optimization Problems

MATHEMATICAL ECONOMICS: OPTIMIZATION. Contents

Unconstrained Optimization

Optimization Theory. Lectures 4-6

1 Newton s Method. Suppose we want to solve: x R. At x = x, f (x) can be approximated by:

Nonlinear Programming (NLP)

Math 273a: Optimization Basic concepts

Lecture 3: Basics of set-constrained and unconstrained optimization

Lectures 9 and 10: Constrained optimization problems and their optimality conditions

Optimization Methods. Lecture 19: Line Searches and Newton s Method

Examination paper for TMA4180 Optimization I

Optimality Conditions for Constrained Optimization

CE 191: Civil and Environmental Engineering Systems Analysis. LEC 05 : Optimality Conditions

Many of you got these steps reversed or otherwise out of order.

Optimality Conditions

Optimization. Yuh-Jye Lee. March 21, Data Science and Machine Intelligence Lab National Chiao Tung University 1 / 29

Optimization Tutorial 1. Basic Gradient Descent

Unconstrained optimization

Optimization. Escuela de Ingeniería Informática de Oviedo. (Dpto. de Matemáticas-UniOvi) Numerical Computation Optimization 1 / 30

CE 191: Civil & Environmental Engineering Systems Analysis. LEC 17 : Final Review

Chapter 2: Unconstrained Extrema

Course Notes for EE227C (Spring 2018): Convex Optimization and Approximation

Part 2: Linesearch methods for unconstrained optimization. Nick Gould (RAL)

1 Overview. 2 A Characterization of Convex Functions. 2.1 First-order Taylor approximation. AM 221: Advanced Optimization Spring 2016

Constrained Optimization Theory

EC9A0: Pre-sessional Advanced Mathematics Course. Lecture Notes: Unconstrained Optimisation By Pablo F. Beker 1

Fundamentals of Unconstrained Optimization

8. Constrained Optimization

The Steepest Descent Algorithm for Unconstrained Optimization

MS 2001: Test 1 B Solutions

MAT 419 Lecture Notes Transcribed by Eowyn Cenek 6/1/2012

AM 205: lecture 18. Last time: optimization methods Today: conditions for optimality

Chapter 2. Optimization. Gradients, convexity, and ALS

Quadratic Programming

Constrained Optimization and Lagrangian Duality

Math Review for Exam Answer each of the following questions as either True or False. Circle the correct answer.

Zangwill s Global Convergence Theorem

Static Problem Set 2 Solutions

Lecture 3 Introduction to optimality conditions

September Math Course: First Order Derivative

Second Order Optimality Conditions for Constrained Nonlinear Programming

Gradient Descent. Dr. Xiaowei Huang

Nonlinear Optimization: What s important?

MATH 5720: Unconstrained Optimization Hung Phan, UMass Lowell September 13, 2018

Lecture 2 - Unconstrained Optimization Definition[Global Minimum and Maximum]Let f : S R be defined on a set S R n. Then

OPER 627: Nonlinear Optimization Lecture 2: Math Background and Optimality Conditions

Structural and Multidisciplinary Optimization. P. Duysinx and P. Tossings

Lecture 3. Optimization Problems and Iterative Algorithms

Chapter 1. Optimality Conditions: Unconstrained Optimization. 1.1 Differentiable Problems

Optimization. A first course on mathematics for economists

11.10a Taylor and Maclaurin Series

Calculus 2502A - Advanced Calculus I Fall : Local minima and maxima

Computational Optimization. Mathematical Programming Fundamentals 1/25 (revised)

Optimization Methods. Lecture 18: Optimality Conditions and. Gradient Methods. for Unconstrained Optimization

Continuous Optimisation, Chpt 6: Solution methods for Constrained Optimisation

HOMEWORK ASSIGNMENT 5

Math 421, Homework #9 Solutions

Week 4: Calculus and Optimization (Jehle and Reny, Chapter A2)

Lakehead University ECON 4117/5111 Mathematical Economics Fall 2002

Lecture 1: Entropy, convexity, and matrix scaling CSE 599S: Entropy optimality, Winter 2016 Instructor: James R. Lee Last updated: January 24, 2016

8. Conjugate functions

Functions of Several Variables

Chapter 8: Taylor s theorem and L Hospital s rule

Unconstrained optimization I Gradient-type methods

Module 04 Optimization Problems KKT Conditions & Solvers

Optimization and Optimal Control in Banach Spaces

Seminars on Mathematics for Economics and Finance Topic 5: Optimization Kuhn-Tucker conditions for problems with inequality constraints 1

Convex Optimization. Problem set 2. Due Monday April 26th

Generalization to inequality constrained problem. Maximize

Introduction to unconstrained optimization - direct search methods

SECTION C: CONTINUOUS OPTIMISATION LECTURE 9: FIRST ORDER OPTIMALITY CONDITIONS FOR CONSTRAINED NONLINEAR PROGRAMMING

March 8, 2010 MATH 408 FINAL EXAM SAMPLE

This manuscript is for review purposes only.

Numerical Analysis: Interpolation Part 1

Algorithms for constrained local optimization

Lecture 12 Unconstrained Optimization (contd.) Constrained Optimization. October 15, 2008

Infeasibility Detection and an Inexact Active-Set Method for Large-Scale Nonlinear Optimization

Conditional Gradient (Frank-Wolfe) Method

Chapter 2 BASIC PRINCIPLES. 2.1 Introduction. 2.2 Gradient Information

Introduction to gradient descent

Lecture 8. Strong Duality Results. September 22, 2008

Numerisches Rechnen. (für Informatiker) M. Grepl P. Esser & G. Welper & L. Zhang. Institut für Geometrie und Praktische Mathematik RWTH Aachen

Optimization. Sherif Khalifa. Sherif Khalifa () Optimization 1 / 50

Lecture Unconstrained optimization. In this lecture we will study the unconstrained problem. minimize f(x), (2.1)

The Randomized Newton Method for Convex Optimization

NOTES ON DIOPHANTINE APPROXIMATION

Lecture 5: September 15

To get horizontal and slant asymptotes algebraically we need to know about end behaviour for rational functions.

Economics 204 Fall 2013 Problem Set 5 Suggested Solutions

FIXED POINT ITERATIONS

Paul Schrimpf. October 18, UBC Economics 526. Unconstrained optimization. Paul Schrimpf. Notation and definitions. First order conditions

CHAPTER 2: QUADRATIC PROGRAMMING

Constrained Optimization

Part 5: Penalty and augmented Lagrangian methods for equality constrained optimization. Nick Gould (RAL)

Problem 3. Give an example of a sequence of continuous functions on a compact domain converging pointwise but not uniformly to a continuous function

1. Which one of the following points is a singular point of. f(x) = (x 1) 2/3? f(x) = 3x 3 4x 2 5x + 6? (C)

Transcription:

Optimality conditions for unconstrained optimization Daniel P. Robinson Department of Applied Mathematics and Statistics Johns Hopkins University September 13, 2018 Outline 1 The problem and definitions 2 First-order optimality conditions 3 Second-order optimality conditions

The basic problem minimize x R n f(x) Definition (global minimizer) The vector x is a global minimizer if f(x ) f(x) for all x R n Definition (local minimizer) The vector x is a local minimizer if there exists ε > 0 such that f(x ) f(x) for all x B(x, ε) := {x R n : x x ε} Definition (strict local minimizer) The vector x is a strict local minimizer if there exists ε > 0 such that f(x ) < f(x) for all x x such that x B(x, ε) Definition (isolated local minimizer) The vector x is an isolated local minimizer if there exists ε > 0 such that x is the only local minimizer in B(x, ε) If x is an isolated local minimizer then x is a strict local minimizer One-dimensional example f(x) strict isolated local minimizer local minimizers x strict isolated global minimizer If we assume that f is continuously differentiable, then we can derive verifiable local optimality conditions for determining whether a point is a local minimizer We rarely can verify that a point is a global minimizer Theorem If x is a local minimizer of a convex function f defined on R n, then x is a global minimizer of f

We are interested in optimality conditions because they provide a means of guaranteeing when a candidate solution x is indeed optimal (sufficient conditions) indicate when a point is not optimal (necessary conditions) guide in the design of algorithms since lack of optimality indication of improvement Recall the following notation g(x) = x f(x) H(x) = 2 xx f(x) Theorem (First-order necessary condition) Suppose that f : R n R is once continuously differentiable. If x is a local minimizer of f, then g(x ) = 0 Proof: Suppose that g(x ) 0. A Taylor expansion in the direction g(x ) gives f ( x αg(x ) ) = f(x ) α g(x ) 2 2 + o ( α ). We also have that there exists an ᾱ > 0 such that so that o ( α ) α 2 g(x ) 2 2 for all 0 < α < ᾱ, f ( x αg(x ) ) f(x ) 1 2 α g(x ) 2 2 < f(x ) for all 0 < α < ᾱ, which contradicts the hypothesis that x is a local minimizer.

if g(x ) 0, then x is not a local minimizer we can limit our search to points x such that g(x ) = 0 a stationary point is any point x that satisfies g(x) = 0 IMPORTANT: if g(x ) = 0, it does not imply that we have found a local minimizer f(x) stationary point stationary point stationary point curvature is important! x In the proof of the previous theorem, the direction g(x ) 0 was a descent direction. Definition (descent direction) We say that the direction s is a descent direction for the continuously differentiable function f at the point x if g(x) T s < 0 Note: when the directional derivative of f at x in the direction d exists, then it equals g(x) T s, i.e., f (x; s) = def f(x + ts) f(x) lim = g(x) T s t 0 t Question: Why do we call them descent directions? Answer: If f is twice continuously differentiable, then by the Mean Value Theorem we have f(x + αs) = f(x) + αg(x) T s + α2 2 st H(ζ α)s for some ζ α [x, x + αs]. Thus, if g(x) T s < 0, then f(x + αs) < f(x) for all α > 0 sufficiently small where we have used that H(ζ α) H(x) as α 0.

Theorem (Second-order necessary conditions) Suppose that f : R n R is twice continuously differentiable. If x is a local minimizer of f, then g(x ) = 0 and H(x ) is positive semi-definite, i.e., s T H(x )s 0 for all s R n Proof: We know from the previous theorem that g(x ) = 0. Suppose that s T H(x )s < 0. A Taylor expansion in the direction s gives f(x + αs) = f(x ) + 1 2 α2 s T H(x )s + o ( α 2), where we used the fact that g(x ) = 0. We are also guaranteed of an ᾱ > 0 such that so that o ( α 2) 1 4 α2 s T H(x )s for all 0 < α < ᾱ, f(x + αs) f(x ) + 1 4 α2 s T H(x )s < f(x ) for all 0 < α < ᾱ. This contradicts the hypothesis that x is a local minimizer. Note: these conditions are not sufficient for being a local minimizer f(x) = x 3 = x = 0 is a saddle point f(x) = x 4 = x = 0 is a maximizer Theorem (Second-order sufficient conditions) If f : R n R is twice continuously differentiable, the vector x satisfies g(x ) = 0, and the matrix H(x ) is positive definite, i.e., then x is a strict local minimizer of f. s T H(x )s > 0 for all s 0 Proof: Continuity implies that H(x) is positive definite for all x in an open ball B(x, ε) for some ε. For any s 0 satisfying x + s B(x, ε) we may use the fact that g(x ) = 0 and the Generalized Mean Value Theorem to conclude that there exists z between x and x + s such that f(x + s) = f(x ) + g(x ) T s + 1 2 st H(z)s = f(x ) + 1 2 st H(z)s > f(x ), which implies that x is a strict local minimizer.

In the previous theorem, the quantity s T H(x)s was important. Definition (direction of positive curvature) We say that the direction d for a twice-continuously differentiable function f is a direction of positive curvature at the point x if d T H(x)d > 0 Definition (direction of negative curvature) We say that the direction d for a twice-continuously differentiable function f is a direction of negative curvature at the point x if d T H(x)d < 0 Definition (direction of zero curvature) We say that the direction d for a twice-continuously differentiable function f is a direction of zero curvature at the point x if d T H(x)d = 0 Note: the quantity d T H(x)d provides second-order curvature information at the point x along the direction d