Optimization: Problem Set Solutions

Similar documents
x +3y 2t = 1 2x +y +z +t = 2 3x y +z t = 7 2x +6y +z +t = a

1 Overview. 2 A Characterization of Convex Functions. 2.1 First-order Taylor approximation. AM 221: Advanced Optimization Spring 2016

Partial Derivatives. w = f(x, y, z).

3 Applications of partial differentiation

LESSON 25: LAGRANGE MULTIPLIERS OCTOBER 30, 2017

MATH Midterm 1 Sample 4

Solutions to Homework 7

Implicit Functions, Curves and Surfaces

Math 10C - Fall Final Exam

0, otherwise. Find each of the following limits, or explain that the limit does not exist.

Examination paper for TMA4180 Optimization I

Taylor Series and stationary points

Let f(x) = x, but the domain of f is the interval 0 x 1. Note

Tangent spaces, normals and extrema

Mathematical Analysis II, 2018/19 First semester

MATH2070 Optimisation

McGill University April Calculus 3. Tuesday April 29, 2014 Solutions

DEPARTMENT OF MATHEMATICS AND STATISTICS UNIVERSITY OF MASSACHUSETTS. MATH 233 SOME SOLUTIONS TO EXAM 2 Fall 2018

UNDERGROUND LECTURE NOTES 1: Optimality Conditions for Constrained Optimization Problems

University of Alberta. Math 214 Sample Exam Math 214 Solutions

MATHEMATICAL ECONOMICS: OPTIMIZATION. Contents

MATH Max-min Theory Fall 2016

Lecture 4 September 15

Preliminary draft only: please check for final version

Lecture 4: Optimization. Maximizing a function of a single variable

Higher order derivative

MAT 211 Final Exam. Spring Jennings. Show your work!

MATH H53 : Final exam

7. Let X be a (general, abstract) metric space which is sequentially compact. Prove X must be complete.

Mathematical Economics. Lecture Notes (in extracts)

CALCULUS: Math 21C, Fall 2010 Final Exam: Solutions. 1. [25 pts] Do the following series converge or diverge? State clearly which test you use.

Structural and Multidisciplinary Optimization. P. Duysinx and P. Tossings

2015 Math Camp Calculus Exam Solution

14.7: Maxima and Minima

Seminars on Mathematics for Economics and Finance Topic 5: Optimization Kuhn-Tucker conditions for problems with inequality constraints 1

SOLUTIONS TO THE FINAL EXAM. December 14, 2010, 9:00am-12:00 (3 hours)

7a3 2. (c) πa 3 (d) πa 3 (e) πa3

SOLUTIONS to Problems for Review Chapter 15 McCallum, Hughes, Gleason, et al. ISBN by Vladimir A. Dobrushkin

Solutions to Sample Questions for Final Exam

1. Find and classify the extrema of h(x, y) = sin(x) sin(y) sin(x + y) on the square[0, π] [0, π]. (Keep in mind there is a boundary to check out).

e x3 dx dy. 0 y x 2, 0 x 1.

Lecture 8: Maxima and Minima

Math 210, Final Exam, Fall 2010 Problem 1 Solution. v cosθ = u. v Since the magnitudes of the vectors are positive, the sign of the dot product will

Optimization. Escuela de Ingeniería Informática de Oviedo. (Dpto. de Matemáticas-UniOvi) Numerical Computation Optimization 1 / 30

The University of British Columbia Midterm 1 Solutions - February 3, 2012 Mathematics 105, 2011W T2 Sections 204, 205, 206, 211.

MATH 5720: Unconstrained Optimization Hung Phan, UMass Lowell September 13, 2018

1. Let g(x) and h(x) be polynomials with real coefficients such that

Calculus - II Multivariable Calculus. M.Thamban Nair. Department of Mathematics Indian Institute of Technology Madras

Math 210, Final Exam, Spring 2012 Problem 1 Solution. (a) Find an equation of the plane passing through the tips of u, v, and w.

EC /11. Math for Microeconomics September Course, Part II Problem Set 1 with Solutions. a11 a 12. x 2

Lagrange Multipliers

1 Directional Derivatives and Differentiability

EC9A0: Pre-sessional Advanced Mathematics Course. Lecture Notes: Unconstrained Optimisation By Pablo F. Beker 1

CIRCLES PART - II Theorem: The condition that the straight line lx + my + n = 0 may touch the circle x 2 + y 2 = a 2 is

Maths Class 11 Chapter 5 Part -1 Quadratic equations

Introduction to the Calculus of Variations

Solutions to old Exam 3 problems

Calculus Example Exam Solutions

McGill University December Intermediate Calculus. Tuesday December 17, 2014 Time: 14:00-17:00

Convex Optimization Theory. Chapter 5 Exercises and Solutions: Extended Version

V. Graph Sketching and Max-Min Problems

Optimality Conditions

Optimization Theory. Lectures 4-6

Chapter 7. Extremal Problems. 7.1 Extrema and Local Extrema

1. For each function, find all of its critical points and then classify each point as a local extremum or saddle point.

Math 164-1: Optimization Instructor: Alpár R. Mészáros

Optimization using Calculus. Optimization of Functions of Multiple Variables subject to Equality Constraints

September Math Course: First Order Derivative

CHAPTER 1 Systems of Linear Equations

3. Minimization with constraints Problem III. Minimize f(x) in R n given that x satisfies the equality constraints. g j (x) = c j, j = 1,...

Convex Functions and Optimization

AP Calculus Testbank (Chapter 9) (Mr. Surowski)

A Note on Two Different Types of Matrices and Their Applications

YET ANOTHER ELEMENTARY SOLUTION OF THE BRACHISTOCHRONE PROBLEM

Algorithms for nonlinear programming problems II

Transpose & Dot Product

EC /11. Math for Microeconomics September Course, Part II Lecture Notes. Course Outline

Advanced Calculus Notes. Michael P. Windham

REVIEW OF DIFFERENTIAL CALCULUS

Functions of Several Variables

Second Order Optimality Conditions for Constrained Nonlinear Programming

(a) The points (3, 1, 2) and ( 1, 3, 4) are the endpoints of a diameter of a sphere.

Transpose & Dot Product

Part 5: Penalty and augmented Lagrangian methods for equality constrained optimization. Nick Gould (RAL)

Computational Optimization. Convexity and Unconstrained Optimization 1/29/08 and 2/1(revised)

f(x, y) = 1 2 x y2 xy 3

Lecture 9: Implicit function theorem, constrained extrema and Lagrange multipliers

Division of the Humanities and Social Sciences. Supergradients. KC Border Fall 2001 v ::15.45

Final Exam Advanced Mathematics for Economics and Finance

Introduction to Machine Learning Lecture 7. Mehryar Mohri Courant Institute and Google Research

Practice Problems for Final Exam

Math 5BI: Problem Set 6 Gradient dynamical systems

Ca Foscari University of Venice - Department of Management - A.A Luciano Battaia. December 14, 2017

Section 5.1 Composite Functions

Maxima and Minima. Marius Ionescu. November 5, Marius Ionescu () Maxima and Minima November 5, / 7

Physics 6010, Fall 2016 Constraints and Lagrange Multipliers. Relevant Sections in Text:

Faculty of Engineering, Mathematics and Science School of Mathematics

Variational Methods & Optimal Control

MA102: Multivariable Calculus

Econ Slides from Lecture 14

Transcription:

Optimization: Problem Set Solutions Annalisa Molino University of Rome Tor Vergata annalisa.molino@uniroma2.it Fall 20

Compute the maxima minima or saddle points of the following functions a. f(x y) = +x 2 + y 2 b. g(x y) =x + y + xy. Solution a. Since the function square root is strictly increasing on its domain we can study the function f(x y) =x 2 + y 2 where we left out the constant. The gradient of f(x y) is given by f(x y) =(2x 2y). Hence setting f(x y) =(0 0) we obtain the stationary point (0 0). In order to find the nature of the stationary point we have to compute the Hessian matrix 2 0 Hf(x y) =. 0 2 The matrix is positive definite thus the point (0 0) is a local minimum for the function f(x y) so is for f(x y). Moreover (0 0) is an absolute minimum for f(x y) because f(0 0) = 0 and f(x y) 0 for every (x y) R 2 so is for f(x y). b. The gradient of g(x y) is given by g(x y) =(x 2 + y y 2 + x). Hence setting g(x y) =(0 0) we obtain two stationary points (0 0) and ( ). In order to figure out the nature of the stationary points we have to compute the Hessian matrix 6x Hg(x y) = 6y which yields and Hg(0 0) = 0 0 Hg( 2 )=. 2 The matrix is indefinite at (0 0) and negative definite at ( ). Therefore (0 0) is a saddle point and ( ) is a local maximum for g.

2 Find the stationary points of the following functions and discuss the behavior of the functions in those points a. f(x y) =x xy 2 + y 4 b. g(x y) =x 4 + x 2 y 2 2x 2 +2y 2 8 c. h(x y) =x 2 + y xy. Solution a. The gradient of f(x y) is given by f(x y) =(x 2 y 2 6xy +4y ). Hence setting f(x y) =(0 0) we obtain the stationary points (0 0) ( 2 2 ) and ( 2 2 ). In order to figure out the nature of the stationary points we have to compute the Hessian matrix 6x 6y Hf(x y) = 6y 6x + 2y 2 which yields 0 0 Hf(0 0) = 0 0 Hf( 2 9 9 2 )= 9 8 Hf( 2 9 9 2 )=. 9 8 The matrix is positive definite at ( 2 2 ) and ( 2 2 ) therefore they are local minima for f. Since Hf(0 0) = 0 by the Hessian criterion nothing can be said about (0 0). In this case we can study the sign of one of the partial derivative of f the one that is analytically simpler. Let us study the sign of f x =x 2 y 2. We have that f x > 0 for x< y and x>ythusx = y is a curve of maxima and x = y is a curve of minima for y fixed. The function of one variable φ (y) =f(y y) =y (y 2) has a maximum in 0 while the function of one variable φ 2 (y) =f( y y) =y (y + 2) has a minimum in 0. It follows that (0 0) is a saddle point for f. Indeedf(0 0) = 0 f(y y) < 0 and f( y y) > 0 for all y = 0sufficiently small. 2

b. Leaving out the constant 8 the gradient of g(x y) is given by g(x y) =(4x +2xy 2 4x 2x 2 y +4y). Hence setting g(x y) =(0 0) we obtain the stationary points (0 0) ( 0) and ( 0). In order to figure out the nature of the stationary points we have to compute the Hessian matrix 2x Hg(x y) = 2 +2y 2 4 4xy 4xy 2x 2 +4 which yields and 0 0 Hg(0 0) = 0 0 8 0 Hg( 0) = 0 6 Hg( 0) = 8 0. 0 6 The matrix is positive definite at ( 0) and ( 0) therefore they are local minima for g. Since Hg(0 0) = 0 by the Hessian criterion nothing can be said about (0 0). We can proceed as in point a. Let us study the sign of f y =2x 2 y +4y =2y(x 2 + 2). f y > 0 for y>0 thus y = 0 is a curve of minima for x fixed. The function of one variable φ(x) =f(x 0) = x 4 2x 2 8 has a local maximum in 0. It follows that (0 0) is a saddle point for f. Indeedf(0 0) = 8 f(x 0) < 8 f(0y) > 8 for all x = 0 and y = 0sufficiently small. c. The gradient of h(x y) is given by h(x y) =(2x y y 2 x). Hence setting h(x y) =(0 0) we obtain the stationary points (0 0) and ( 2 6 ). In order to figure out the nature of the stationary points we have to compute the Hessian matrix 2 Hh(x y) = 6y

which yields and 2 Hh(0 0) = 0 Hh( 2 2 6 )=. The matrix is indefinite at (0 0) and positive definite at ( (0 0) is a saddle point and ( 2 6 ) is a local minimum point. 2 6 ). Therefore Find the minima maxima or saddle points of the following functions a. f(x x 2 x )=x x 2 x s.t. x 2 + x2 2 + x2 = 0 b. f(x x 2 )=x 2 +x 2 s.t. x 2 4 + x2 2 9 = 0. Solution a. f(x x 2 x )=x x 2 x s.t. x 2 + x2 2 + x2 = 0. Let us note first that the constraint {(x x 2 x ) R x 2 + x 2 2 + x 2 =0} S((0 0 0); ) is a sphere of R with centre (0 0 0) and radius. Hence it is a compact subset of R. Therefore since the function f is continuos on R by the Weierstrass theorem there exist a maximum and a minimum point for f on S((0 0 0); ). The Lagrangian of f is given by Thus the FOCs are It follows that L(x λ) =x x 2 x + λ(x 2 + x 2 2 + x 2 ). x = x 2 x +2λx =0 x 2 = x x +2λx 2 =0 x = x x 2 +2λx =0 λ = x 2 + x2 2 + x2 =0. 4

if λ =0then x 2 x =0 x x =0 x x 2 =0 x 2 + x2 2 + x2 =0 whose possible solutions are (± 0 0) (0 ± 0)(0 0 ±). If λ = 0 then subtracting the 2 nd equation from the st Therefore x 2 x +2λx x x 2λx 2 =0. (x x 2 ) (2λ x ) =0. A B CASE A = 0 and B = 0. Subtracting the nd equation from the 2 st we obtain as before (x 2 x ) (2λ x ) =0. C D CASE.a A = 0 and C = 0. We have x = x 2 = x. Substituting into the fourth equation (the constraint) we obtain x = x 2 = x = or x = x 2 = x =. CASE.b A = 0 and D = 0. We have x = x 2 =2λ. Substituting into the third equation we obtain 4λ 2 +2λx = 0. Therefore 2λ(2λ + x ) = 0. We do not consider the solution λ = 0 for our initial assumption λ = 0. The only part that could be equal to zero is 2λ + x = 0 i.e. x = 2λ. Substituting into the constraint x = x 2 = x =2λ we obtain λ = ± 2. Substituting this value of λ in order to obtain the values of x we obtain the following solutions x = x 2 = x = or x = x 2 = x =. 5

CASE 2 A = 0 and B = 0. In this case we have x =2λ. Substitute this result into the second equation we obtain 2λ(x + x 2 ) = 0. The possible solutions are λ = 0 (which we do not consider because we have impose λ = 0) or x = x 2. Now substituting x = x 2 into the third equation we obtain x 2 + 4λ 2 = 0 i.e. x = ±2λ. We have the possible combinations x = 2λx 2 = 2λx =2λ or x = 2λx 2 =2λx =2λ. As before substituting this result into the constraint we obtain λ = ± 2. This gives us the four possible solutions: if λ =+ 2 x = x 2 = x = or x = x 2 = x = if λ = 2 x = x 2 = x = or x = x 2 = x =. All the possible solutions are ( ( (± 0 0); (0 ± 0); (0 0 ±); ); ( ); ( ); ( ); ( ); ( ); ); ( ). Computing the values of the function f in all the stationary points we have f( )=f( = f( )= f( )=f( )=f( )= global maxima )=f( )= 6

= f( )= global minima f(± 0 0) = f(0 ± 0) = f(0 0 ±) = 0. x 2 b. f(x x 2 )=x 2 +x 2 s.t. 4 + x2 2 9 = 0. Let us note first that the constraint (x x 2 x ) R x2 4 + x2 2 9 =0 E((0 0 0); 2 ) is an ellipsis of R with centre (0 0 0) and semi-axes 2 and. Hence E((0 0 0); 2 ) is a compact subset of R. Therefore since the function f is continuos on R by the Weierstrass theorem there exist a maximum and a minimum point for f on E((0 0 0); 2 ). The Lagrangian of f is given by x L(λx)=x 2 2 +x 2 + λ 4 + x2 2 9. The FOCs are It follows that x =2x + 2 λx =0 x 2 =+ 2 9 λx 2 =0 λ = x2 4 + x2 2 9. if x =0 then λ = 4. Substituting λ = 4 into the second condition we have x 2 = 2. Thus the last constraint gives that is x 2 2 2 + 6 2 6 2 =0 x 2 = 26 4 2 4 = 7 6 which yield a non real root. if x = 0 then from the constraint we obtain x 2 = ±. Since f(0 ) = 9 and f(0 ) = 9 the point (0 ) is a global minimum and the point (0 ) is a global maximum for f on E((0 0 0); 2 ). 7

4 Find the maxima points of the following functions a. f(x y) =xy s.t. x 2 + y 2 = b. f(x y) =x α y α s.t. w = p x x + p y y. Solution a. f(x y) =xy s.t. x 2 + y 2 =. Since (x y) satisfy the relationship x 2 + y 2 = 0 on this curve the function f can be rewritten as f(x y) =xy = xy + 2 (x2 + y 2 ) = 2 (x2 + y 2 +2xy) 2 = 2 (x + y)2 2. In this way it is easy to check that the minimum is attained on the line x = y that intersects the curve in the points ( 2/2 2/2) and ( 2/2 2/2) and therefore these are the minima for f. Similarly f(x y) =xy = xy 2 (x2 + y 2 ) = 2 (x2 + y 2 ) = 2 (x y)2 + 2. In this way it is easy to check that he maximum is attained on the line x = y that intersects the curve in the points ( 2/2 2/2) and ( 2/2 2/2) and therefore these are the maxima for f. b. f(x y) =x α y α s.t. w = p x x + p y y. L(x y λ) =x α y α + λ(p x x + p y y w) x = αxα y α + λp x =0 y =( α)xα y α + λp y =0 λ = p xx + p y y w =0 αx α y α = λp x It follows that that is ( α)x α y α = λp y w = p x x + p y y. αx α y α ( α)x α y α = p x p y α y α x = p x. p y 8

Substitute y = x px α p y α into the constraint to get Therefore is a maximum point for f. w = xp x α. x(p x w)= wα p x y(p y w)= w( α) p y 5 Compute the maxima minima or saddle points of the following function f(x y) =x 2 +y s.t x2 4 + y2 9 =. Solution The constraint is an ellipsis thus is a compact set. Since the function f is continuous by the Weierstrass theorem it has a maximum and a minimum. Parametrize the ellipsis as x = 2 cos θ with θ [0 2π). It follows that y =sinθ f(θ) = 4 cos 2 θ +9sinθ = 4( sin 2 θ)+9sinθ = 4 sin 2 θ +9sinθ +4 Impose f (θ) = 0. We get f (θ) = 8 sin θ cos θ + 9 cos θ = cos θ(9 8 sin θ). cos θ = 0 therefore θ = π 2 θ = 2 π 9 8 sin θ = 0 that is sin θ = 9 8 > therefore no solutions. 9

Furthermore since 9 8 sin θ > 0 forall θ wehave f (θ) > 0 iff cos θ > 0. Therefore θ = π 2 is a global maximum and θ = 2π is a global minimum point for f. The global maximum correspond to the point (0 ) and the global minimum to the point (0 ). 6 Find the (local) maxima and minima of the function f(x y) =xy y 2 + subject to the constraint g(x y) =x + y 2 =0 using:. a parametric representation of the constraint; 2. Lagrange multipliers. Are they global? Solution First of all notice that since the constraint set is not compact we cannot use the Weierstrass theorem to claim that the function f assumes global maximum and minimum.. A parametrization of the parabola is x = t 2 y = t for <t<+. It follows that f(t) =( t 2 )t t 2 += t t 2 + t + is a function of one variable on the domain ( + ). Therefore solving f (t) = t 2 2t +=0 0

we get the stationary points t = andt 2 =. Since f (t) > 0for <t< t = isarelativeminimumandt 2 = is a relative maximum. Notice that this maximum and minimum are not global indeed lim f(t) = t + and lim f(t) =+. t In correspondence we have that (0 ) is a relative minimum and 8 is a relative maximum for f. 2. The Lagrangian for the problem is L(x y λ) =f(x y) λg(x y) =xy y 2 ++λ(x + y 2 ). Solving the first order conditions y + λ =0 x 2y +2λy =0 x + y 2 =0 we get the two critical points (0 ) and 8 9. In order to find the nature of the stationary points we have to study the bordered Hessian matrix 0 2y Hf(x y λ) = 0 2y 2+2λ in such points. 0 2 Hf(0 ) = 0. 2 0 Since det(hf(0 )) < 0 (0 ) is a relative minimum for f. 8 Hf 9 2 0 = 0. 2 8 9

Since det(hf 8 9 ) > 0 ( 8 )isarelativemaximumforf. These 9 maximum and minimum are not global because for example we have lim f(x y) =+ x + and lim f(x y) =. x 2