Examination paper for TMA4180 Optimization I

Similar documents
SECTION C: CONTINUOUS OPTIMISATION LECTURE 11: THE METHOD OF LAGRANGE MULTIPLIERS

Constrained Optimization Theory

Exam in TMA4180 Optimization Theory

Lectures 9 and 10: Constrained optimization problems and their optimality conditions

MATH2070 Optimisation

Algorithms for constrained local optimization

Continuous Optimisation, Chpt 6: Solution methods for Constrained Optimisation

Second Order Optimality Conditions for Constrained Nonlinear Programming

March 5, 2012 MATH 408 FINAL EXAM SAMPLE

Optimality Conditions for Constrained Optimization

Optimization. Escuela de Ingeniería Informática de Oviedo. (Dpto. de Matemáticas-UniOvi) Numerical Computation Optimization 1 / 30

Introduction to Nonlinear Stochastic Programming

Structural and Multidisciplinary Optimization. P. Duysinx and P. Tossings

Introduction to Machine Learning Lecture 7. Mehryar Mohri Courant Institute and Google Research

UNDERGROUND LECTURE NOTES 1: Optimality Conditions for Constrained Optimization Problems

Constrained Optimization

Computational Optimization. Convexity and Unconstrained Optimization 1/29/08 and 2/1(revised)

Numerical Optimization

Lagrange Multipliers

Lecture 3: Lagrangian duality and algorithms for the Lagrangian dual problem

SF2822 Applied Nonlinear Optimization. Preparatory question. Lecture 9: Sequential quadratic programming. Anders Forsgren

Constrained Optimization and Lagrangian Duality

Optimization Problems with Constraints - introduction to theory, numerical Methods and applications

Optimization: Problem Set Solutions

TMA 4180 Optimeringsteori KARUSH-KUHN-TUCKER THEOREM

1 Computing with constraints

SECTION C: CONTINUOUS OPTIMISATION LECTURE 9: FIRST ORDER OPTIMALITY CONDITIONS FOR CONSTRAINED NONLINEAR PROGRAMMING

x +3y 2t = 1 2x +y +z +t = 2 3x y +z t = 7 2x +6y +z +t = a

Unconstrained Optimization

EC /11. Math for Microeconomics September Course, Part II Problem Set 1 with Solutions. a11 a 12. x 2

AM 205: lecture 18. Last time: optimization methods Today: conditions for optimality

E5295/5B5749 Convex optimization with engineering applications. Lecture 8. Smooth convex unconstrained and equality-constrained minimization

Convex Functions and Optimization

5 Handling Constraints

DEPARTMENT OF MATHEMATICS AND STATISTICS UNIVERSITY OF MASSACHUSETTS. MATH 233 SOME SOLUTIONS TO EXAM 2 Fall 2018

Quiz Discussion. IE417: Nonlinear Programming: Lecture 12. Motivation. Why do we care? Jeff Linderoth. 16th March 2006

MATHEMATICAL ECONOMICS: OPTIMIZATION. Contents

Computational Optimization. Mathematical Programming Fundamentals 1/25 (revised)

Math 164-1: Optimization Instructor: Alpár R. Mészáros

EC /11. Math for Microeconomics September Course, Part II Lecture Notes. Course Outline

Algorithms for nonlinear programming problems II

Examination paper for TMA4215 Numerical Mathematics

5.6 Penalty method and augmented Lagrangian method

MATH 4211/6211 Optimization Basics of Optimization Problems

Numerisches Rechnen. (für Informatiker) M. Grepl P. Esser & G. Welper & L. Zhang. Institut für Geometrie und Praktische Mathematik RWTH Aachen

Algorithms for nonlinear programming problems II

Optimality Conditions

minimize x subject to (x 2)(x 4) u,

Generalization to inequality constrained problem. Maximize

Convex Optimization. Newton s method. ENSAE: Optimisation 1/44

1 Overview. 2 A Characterization of Convex Functions. 2.1 First-order Taylor approximation. AM 221: Advanced Optimization Spring 2016

Chapter 7. Extremal Problems. 7.1 Extrema and Local Extrema

Nonlinear Optimization: What s important?

Optimality conditions for unconstrained optimization. Outline

Seminars on Mathematics for Economics and Finance Topic 5: Optimization Kuhn-Tucker conditions for problems with inequality constraints 1

March 8, 2010 MATH 408 FINAL EXAM SAMPLE

Quadratic Programming

Written Examination

Lecture 3. Optimization Problems and Iterative Algorithms

Optimisation in Higher Dimensions

Examination paper for TMA4130 Matematikk 4N: SOLUTION

Math 273a: Optimization Basic concepts

1. Find and classify the extrema of h(x, y) = sin(x) sin(y) sin(x + y) on the square[0, π] [0, π]. (Keep in mind there is a boundary to check out).

2.3 Linear Programming

Computational Optimization. Augmented Lagrangian NW 17.3

ON LICQ AND THE UNIQUENESS OF LAGRANGE MULTIPLIERS

Convex Optimization Theory. Chapter 5 Exercises and Solutions: Extended Version

Part 5: Penalty and augmented Lagrangian methods for equality constrained optimization. Nick Gould (RAL)

Lecture 15 Newton Method and Self-Concordance. October 23, 2008

Name: Instructor: Lecture time: TA: Section time:

EC9A0: Pre-sessional Advanced Mathematics Course. Lecture Notes: Unconstrained Optimisation By Pablo F. Beker 1

Math 212-Lecture Interior critical points of functions of two variables

Math 5311 Constrained Optimization Notes

Differentiable exact penalty functions for nonlinear optimization with easy constraints. Takuma NISHIMURA

4 Newton Method. Unconstrained Convex Optimization 21. H(x)p = f(x). Newton direction. Why? Recall second-order staylor series expansion:

FIRST- AND SECOND-ORDER OPTIMALITY CONDITIONS FOR MATHEMATICAL PROGRAMS WITH VANISHING CONSTRAINTS 1. Tim Hoheisel and Christian Kanzow

BASICS OF CONVEX ANALYSIS

McGill University April Calculus 3. Tuesday April 29, 2014 Solutions

Two hours. To be provided by Examinations Office: Mathematical Formula Tables. THE UNIVERSITY OF MANCHESTER. xx xxxx 2017 xx:xx xx.

Lecture 11 and 12: Penalty methods and augmented Lagrangian methods for nonlinear programming

ISM206 Lecture Optimization of Nonlinear Objective with Linear Constraints

Summary Notes on Maximization

MATH 5720: Unconstrained Optimization Hung Phan, UMass Lowell September 13, 2018

SWFR ENG 4TE3 (6TE3) COMP SCI 4TE3 (6TE3) Continuous Optimization Algorithm. Convex Optimization. Computing and Software McMaster University

8 Barrier Methods for Constrained Optimization

0, otherwise. Find each of the following limits, or explain that the limit does not exist.

Additional Exercises for Introduction to Nonlinear Optimization Amir Beck March 16, 2017

Lecture 1: Introduction. Outline. B9824 Foundations of Optimization. Fall Administrative matters. 2. Introduction. 3. Existence of optima

Nonlinear Programming and the Kuhn-Tucker Conditions

Computational Optimization. Constrained Optimization Part 2

Lecture 1: Introduction. Outline. B9824 Foundations of Optimization. Fall Administrative matters. 2. Introduction. 3. Existence of optima

Scientific Computing: An Introductory Survey

Scientific Computing: An Introductory Survey

AM 205: lecture 19. Last time: Conditions for optimality Today: Newton s method for optimization, survey of optimization methods

1. Gradient method. gradient method, first-order methods. quadratic bounds on convex functions. analysis of gradient method

Faculty of Engineering, Mathematics and Science School of Mathematics

Lagrange Relaxation and Duality

Tangent spaces, normals and extrema

Solution Methods. Richard Lusby. Department of Management Engineering Technical University of Denmark

Some new facts about sequential quadratic programming methods employing second derivatives

Transcription:

Department of Mathematical Sciences Examination paper for TMA4180 Optimization I Academic contact during examination: Phone: Examination date: 26th May 2016 Examination time (from to): 09:00 13:00 Permitted examination support material: The textbook: Nocedal & Wright, Numerical Optimization including errata Rottmann, Mathematical formulae Handouts on Minimisers of unconstrained optimisation problems, Basic properties of convex functions Approved basic calculator Other information: All answers should be justified and include enough details to make it clear which methods or results have been used You may answer to the questions of the exam either in English or in Norwegian Language: English Number of pages: 8 Number of pages enclosed: 0 Checked by: Date Signature

TMA4180 Optimization I, 26th May 2016 Page 1 of 8 Problem 1 Consider the function f : R 2 R given by f(x, y) = 2x 2 2xy 2 12x + y 4 + 2y 2 + 36 a) Compute all stationary points of f and find all local or global minimizers of f We start by computing the gradient and the Hessian of f We have 4x 2y f(x, y) = 2 12 4xy + 4y 3 + 4y and 2 f(x, y) = 4 4y 4y 4x + 12y 2 + 4 The stationary points satisfy f(x, y) = 0 Now the second row of the gradient implies that 4y( x + y 2 + 1) = 0 and thus either y = 0 or y 2 = x 1 In the first case, the equation for the first row of the gradient implies that x = 3 Thus we have a stationary point (x, y) = (3, 0) On the other hand, if y 2 = x 1, then the equation for the first row of the gradient becomes 4x 2(x 1) 12 = 0 or x = 5, which implies that in this case y = ± 5 1 = ±2 Thus we have the two stationary points (x, y) = (5, ±2) In order to determine whether they are local minima, we look at the Hessian of f We have 2 4 0 f(3, 0) =, 0 8 which is obviously indefinite Thus (3, 0) is a saddle point of f On the other hand 2 4 8 f(5, ±2) = 8 48 Since the diagonal entries are positive and the determinant of this matrix is also positive, the matrix is positive definite, which implies that the points (5, ±2) are local minima

Page 2 of 8 TMA4180 Optimization I, 26th May 2016 Now note that we can write f(x, y) = (x y 2 ) 2 + (x 6) 2 + 2y 2, which is obviously coercive Thus the function f attains at least one global minimum Since f is symmetric with respect to y, it follows that f(5, 2) = f(5, 2) Thus the function f has the same function value at the only two candidates for the global minimum, which implies that both of these points are actually global minima b) Starting at the point (x, y) = (1, 2) compute one step of Newton s method with backtracking (Armijo) linesearch (see Algorithm 31 in Nocedal and Wright) Start with an initial step length ᾱ = 1 and use the parameters c = 1/8 (sufficient decrease parameter) and ρ = 1/2 (contraction factor) We need the gradient and the Hessian of f at (1, 2) We have f(1, 2) = 16 32 and 2 f(1, 2) = 4 8 8 48 Solving the system 2 f(1, 2)p = f(1, 2) yields the Newton direction p = The Armijo condition now requires that 4 0 With α = 1 we have f((1, 2) + αp) f(1, 2) + 1 8 αpt f(1, 2) = 42 8α f((1, 2) + p) = f(5, 2) = 10 34 = 42 8α Thus the step α = 1 is accepted by the line search, and as a consequence the next iterate in the Newton method would be the point (5, 2) (Since this is actually the global minimum, the iteration will stop there)

TMA4180 Optimization I, 26th May 2016 Page 3 of 8 Problem 2 Consider the constrained optimization problem x 3 min subject to x 2 + y 2 = 1 (1) Find (by whatever means) the solution of this problem In addition, formulate the augmented Lagrangian for this problem and determine all parameters λ R and µ > 0 for which the solution of (1) is a local minimizer of the augmented Lagrangian We want to minimize the function x 3 for (x, y) on the unit circle Since x x 3 is strictly increasing, it is obvious that the minimizer will be the point on the unit circle with minimal x-coordinate, which is the point ( 1, 0) Now the augmented Lagrangian of the problem is defined as L A (x, y; λ, µ) := x 3 λ(x 2 + y 2 1) + µ 2 (x2 + y 2 1) 2 In order to determine whether ( 1, 0) is a local minimizer of L A, we compute the gradient of L A We have 3x L A (x, y; λ, µ) = 2 2xλ + 2xµ(x 2 + y 2 1) 2yλ + 2yµ(x 2 + y 2 1) and thus 3 + 2λ L A ( 1, 0; λ, µ) = 0 Thus ( 1, 0) is a stationary point of the augmented Lagrangian, if and only if λ = 3/2 Next we need the Hessian of L A, which is 2 6x 2λ + 2µ(x L A = 2 + y 2 1) + 4µx 2 4µxy 4µxy 2λ + 2µ(x 2 + y 2 1) + 4µy 2 In particular, we have 2 L A ( 1, 0; 3/2, µ) = 4µ 3 0, 0 3 which is positive definite for µ > 3/4 and positive semi-definite for µ = 3/4 As a consequence, ( 1, 0) is a local minimizer for the augmented Lagrangian if λ = 3/2 and µ > 3/4 and is no local minimizer if either λ 3/2 or λ = 3/2 and µ < 3/4 The only parameter values where it is not yet clear whether ( 1, 0) minimizes L A are λ = 3/2 and µ = 3/4 Here the augmented Lagrangian is L A (x, y; 3/2, 3/4) = x 3 + 3 2 (x2 + y 2 1) + 3 8 (x2 + y 2 1) 2

Page 4 of 8 TMA4180 Optimization I, 26th May 2016 Moreover, for y = 0 we have L A (x, 0; 3/2, 3/4) = 3 8 x4 + x 3 + 3 4 x2 9 8 =: p(x) We have p ( 1) = 0 and p ( 1) = 0 However, p ( 1) 0, which implies that 1 is no local minimizer of p As a consequence, ( 1, 0) is no local minimizer of L A for λ = 3/2 and µ = 3/4 Problem 3 Consider the constrained optimization problem f(x, y) := (x + 1) 2 + (y 2) 2 min subject to (x, y) Ω, where the set Ω R 2 is given by the inequality constraints xy 0 and x y(y + 2) 0 a) Sketch the set Ω and find all points (x, y) Ω for which the LICQ holds Denoting c 1 (x, y) = xy and c 2 (x, y) = x y(y + 2), we have y 1 c 1 (x, y) = and c x 2 (x, y) = 2y 2 We consider first the points (0, 0) and (0, 2), where both constraints are active At (0, 0) we have c 1 (0, 0) = 0, and thus the constraints are not linearly independent At (0, 2) we have c 1 (0, 2) = ( 2, 0) and c 2 (0, 2) = (1, 2), which are obviously linearly independent Thus the LICQ holds Next we note that c 1 (x, y) 0 for (x, y) (0, 0), which implies that c 1 (x, y) 0 whenever only the constraint c 1 is active Thus the LICQ holds on each of these points Similarly, c 2 (x, y) 0 for all x and y, which implies that the LICQ holds whenever only the constraint c 2 is active Finally, the LICQ trivially holds if no constraints are active Thus the LICQ holds everywhere in Ω \ {(0, 0)} b) Determine the set F(0, 0) of linearized feasible direction at (0, 0) and show that the tangent cone to Ω at (0, 0) is different from F(0, 0) At the point (x, y) = (0, 0) the set of linearized feasible directions F(0, 0) equals F(0, 0) = { (p, q) R 2 : (p, q) c 1 (0, 0) 0 and (p, q) c 2 (0, 0) 0 } = { (p, q) R 2 : (p, q) ( 0 0 = { (p, q) R 2 : p 2q } ) 0 and (p, q) 1 2 0 }

TMA4180 Optimization I, 26th May 2016 Page 5 of 8 In particular, the vector (1, 1) is an element of F(0, 0) We show now that (1, 1) cannot be an element of the tangent cone to Ω at (0, 0) Indeed, assume to the contrary that (1, 1) is an element of this tangent cone Then there exists a sequence (x k, y k ) 0 and positive scalars t k converging to 0 such that (x k, y k ) Ω for all k, x k /t k 1, and y k /t k 1 In particular, this implies that x k > 0 for sufficiently large k whereas y k < 0 for sufficiently large k This, however, implies that x k y k < 0, which is a contradiction to the requirement that (x k, y k ) Ω Thus (1, 1) cannot be an element of the tangent cone at 0 c) Use the second order optimality conditions in order show that the point (0, 0) is a local solution of the constrained optimization problem We first have to show (0, 0) is a KKT point That is, we have to find λ 0 and µ 0 such that L(0, 0; λ, µ) = 0 (note that the point (0, 0) is an element of Ω and that both constraints are active; thus we need not care about the other KKT-conditions) The Lagrangian in this case reads as L(x, y; λ, µ) = (x + 1) 2 + (y 2) 2 λxy µ(x y(y + 2)) Thus ( ) 2(x + 1) λy µ L(x, y; λ, µ) = 2(y 2) λx + µ(2y + 2) For (x, y) = (0, 0) we have L(0, 0; λ, µ) = 2 µ, 4 + 2µ and we see that (0, 0) is a KKT point with Lagrange parameters µ = 2 and arbitrary λ 0 Next we note that the Hessian of the Lagrangian with respect to x and y is 2 2 λ L(x, y; λ, µ) = λ 2µ With the Lagrange parameters λ = 0 and µ = 2 we see that the Hessian of the Lagrangian at (0, 0) is positive definite Thus the second order sufficient conditions for a strict local solution of a constraint optimization problem are satisfied at the point (0, 0)

Page 6 of 8 TMA4180 Optimization I, 26th May 2016 Problem 4 A function f : R n R is called strongly convex, if there exists c > 0 such that the function x f(x) c 2 x 2 is convex A function F : R l R m R is called strongly separately convex, if there exists c > 0 such that for every ˆx R l and ŷ R m the functions are convex x F (x, ŷ) c 2 x 2 and y F (ˆx, y) c 2 y 2 a) Show that a twice differentiable function f : R n R is strongly convex, if and only if there exists c > 0 such that for every p R n and x R n p T 2 f(x)p c p 2 A twice differentiable function g is convex, if and only if 2 g(x) is positive semi-definite for all x, which is equivalent to p T 2 g(x)p 0 for all x and p Applying this characterization to the function g(x) = f(x) c 2 x 2 with Hessian 2 g(x) = 2 f(x) ci with I R n n being the identity matrix, we obtain that f is strongly convex, if and only if 2 f(x) ci is positive semi-definite for some c > 0 and all x R n, which in turn is equivalent to the condition that p T ( 2 f(x) ci)p 0 for all x R n and p R n, or, equivalently, for all x R n and p R n b) Show that the function F : R R R, p T 2 f(x)p c p 2 F (x, y) = x 4 + y 4 4xy + x 2 + y 2 is non-convex, but strongly separately convex We note that xx F (x, y) = 12x 2 + 2 2 and yy F (x, y) = 12y 2 + 2 2, showing the F is strongly separately convex On the other hand, 2 12x F (x, y) = 2 + 2 4 4 12y 2, + 2 which is not positive semi-definite for (x, y) = (0, 0), as det 2 F (0, 0) = 12 < 0 Thus F is non-convex

TMA4180 Optimization I, 26th May 2016 Page 7 of 8 c) Assume that the function f : R n R is twice continuously differentiable and strongly convex and assume that x is a minimizer of f Show that there exists c > 0 such that for every x R n f(x) f(x ) + c 2 x x 2 An application of Taylor s theorem yields f(x) = f(x ) + f(x ) T (x x ) + 1 2 (x x ) 2 f(x + t(x x ))(x x ) for some 0 t 1 Since x minimizes f, we have f(x ) = 0 Moreover, the application of part a) of this problem allows us to estimate (x x ) 2 f(x + t(x x ))(x x ) c x x 2 Therefore f(x) f(x ) + c 2 x x 2 d) Assume that the function F : R l R m R is strongly separately convex, twice continuously differentiable, and coercive Given y 0 R l we define iterates x k+1 R l, y k+1 R m by x k+1 minimizes the function x F (x, y k ), y k+1 minimizes the function y F (x k+1, y) Show that x k+1 x k 2 + y k+1 y k 2 < k N Let c > 0 be such that the conditions defining strong separate convexity of F are satisfied By construction, the point x k+1 minimizes the function x F (x, y k ) Thus part c) implies that F (x k, y k ) F (x k+1, y k ) + c 2 x k+1 x k 2 or x k+1 x k 2 2 c ( F (xk, y k ) F (x k+1, y k ) ) Similarly, y k+1 minimizes the function y F (x k+1, y), and therefore F (x k+1, y k ) F (x k+1, y k+1 ) + c 2 y k+1 y k 2

Page 8 of 8 TMA4180 Optimization I, 26th May 2016 or y k+1 y k 2 2 c ( F (xk+1, y k ) F (x k+1, y k+1 ) ) Thus x k+1 x k 2 + y k+1 y k 2 k N 2 F (x k, y k ) F (x k+1, y k ) + F (x k+1, y k ) F (x k+1, y k+1 ) c k N = 2 F (x k, y k ) F (x k+1, y k+1 ) c k N = 2 c ( F (x1, y 1 ) lim k F (x k, y k ) 2 c ( F (x1, y 1 ) inf F (x, y) ) Since F is coercive and twice differentiable, it follows that F attains a minimum, and therefore inf F (x, y) > Thus x k+1 x k 2 + y k+1 y k 2 < k N e) Show that every strongly convex and differentiable function f : R n R has a unique minimizer x R n Let c > 0 be such that g(x) := f(x) c 2 x 2 is convex Then the differential characterization of convexity implies that and therefore f(x) c 2 x 2 = g(x) g(0) + g(0) T x = f(0) + f(0) T x f(x) f(0) + f(0) T x + c 2 x 2 f(0) f(0) x + c 2 x 2 In particular, if x, then also f(x) ; that is, the function f is coercive Since f by assumption is differentiable and therefore continuous, it follows that f attains its minimum Next assume that x and y are minimizers of f From c) of this problem we obtain that f(x) f(y) + c 2 x y 2 However, f(x) = f(y) because by assumption both x and y are minimizers of f Thus x y 2 0, which is only possible if x = y This shows that the minimizer is unique