Optimization using Calculus. Optimization of Functions of Multiple Variables subject to Equality Constraints

Size: px
Start display at page:

Download "Optimization using Calculus. Optimization of Functions of Multiple Variables subject to Equality Constraints"

Transcription

1 Optimization using Calculus Optimization of Functions of Multiple Variables subject to Equality Constraints 1

2 Objectives Optimization of functions of multiple variables subjected to equality constraints using the method of constrained variation, and the method of Lagrange multipliers.

3 Constrained optimization A function of multiple variables, f(x), is to be optimized subject to one or more equality constraints of many variables. These equality constraints, g j (x), may or may not be linear. The problem statement is as follows: where Maximize (or minimize) f(x), subject to g j (X) = 0, j = 1,,, m X x1 x M x n = 3

4 Constrained optimization (contd.) m n With the condition that ; or else if m > n then the problem becomes an over defined one and there will be no solution. Of the many available methods, the method of constrained variation and the method of using Lagrange multipliers are discussed. 4

5 Solution by method of Constrained Variation For the optimization problem defined above, let us consider a specific case with n = and m = 1 before we proceed to find the necessary and sufficient conditions for a general problem using Lagrange multipliers. The problem statement is as follows: Minimize f(x 1,x ), subject to g(x 1,x ) = 0 For f(x 1,x ) to have a minimum at a point X * = [x 1*,x * ], a necessary condition is that the total derivative of f(x 1,x ) must be zero at [x 1*,x * ]. 5 f f df = dx + dx = x 1 1 x 0 (1)

6 Method of Constrained Variation (contd.) Since g(x 1*,x * ) = 0 at the minimum point, variations dx 1 and dx about the point [x 1*, x * ] must be admissible variations, i.e. the point lies on the constraint: g(x 1* + dx 1, x * + dx ) = 0 () assuming dx 1 and dx are small the Taylor series expansion of this gives us * gx ( * + dx, x + dx) 1 1 g g = gx (, x ) + (x,x ) dx+ (x,x ) dx= 0 * * * * * * x1 x (3) 6

7 Method of Constrained Variation (contd.) g g or dg = dx at [x 1*,x * ] (4) 1+ dx = 0 x x 1 which is the condition that must be satisfied for all admissible variations. Assuming, (4) can be rewritten as g/ x 0 g/ x dx = ( x, x ) dx 1 * * 1 1 g/ x (5) 7

8 Method of Constrained Variation (contd.) (5) indicates that once variation along x 1 (dx 1 ) is chosen arbitrarily, the variation along x (dx ) is decided automatically to satisfy the condition for the admissible variation. Substituting equation (5) in (1) we have: f g/ x1 f (6) df = dx1 = 0 x1 g/ x x * * (x 1, x ) The equation on the left hand side is called the constrained variation of f. Equation (5) has to be satisfied for all dx 1, hence we have 8 f g f g = 0 x x x x 1 1 (x, x ) This gives us the necessary condition to have [x 1*, x * ] as an extreme point (maximum or minimum) * * 1 (7)

9 Solution by method of Lagrange multipliers Continuing with the same specific case of the optimization problem with n = and m = 1 we define a quantity λ, called the Lagrange multiplier as f / x λ = g/ x (8) (x, x ) * * 1 Using this in (5) f x g + λ = 0 x 1 1 (x, x ) * * 1 (9) And (8) written as f x g + λ = 0 x (x, x ) * * 1 (10) 9

10 Solution by method of Lagrange multipliers contd. Also, the constraint equation has to be satisfied at the extreme point gx (, x ) = 0 1 ( x, ) * * 1 x (11) Hence equations (9) to (11) represent the necessary conditions for the point [x 1 *, x *] to be an extreme point. Note that λ could be expressed in terms of g / x1 as well and g / x1 has to be non-zero. Thus, these necessary conditions require that at least one of the partial derivatives of g(x 1, x ) be non-zero at an extreme point. 10

11 Solution by method of Lagrange multipliers contd. The conditions given by equations (9) to (11) can also be generated by constructing a functions L, known as the Lagrangian function, as Lx ( 1, x, λ) = f( x1, x) + λgx ( 1, x) (1) 11 Alternatively, treating L as a function of x 1,x and λ, the necessary conditions for its extremum are given by L f g ( x1, x, λ) = ( x1, x) + λ ( x1, x) = 0 x x x L f g ( x, x, λ) = ( x, x ) + λ ( x, x ) = 0 x x x L ( x 1, x, λ) = g ( x 1, x ) = 0 λ (13)

12 Necessary conditions for a general problem For a general problem with n variables and m equality constraints the problem is defined as shown earlier Maximize (or minimize) f(x), subject to g j (X) = 0, j = 1,,, m where X x1 x M x n = In this case the Lagrange function, L, will have one Lagrange multiplier λ j for each constraint as Lx (, x,..., xλ, λ,..., λ ) = f( X) + λ g( X) + λ g( X) λ g ( X) 1 n, 1 m 1 1 m m (14) 1

13 Necessary conditions for a general problem contd. L is now a function of n + m unknowns, x, and the 1, x,..., xn, λ1, λ,..., λm necessary conditions for the problem defined above are given by L f g = ( ) + ( X) = 0, i = 1,,..., n j = 1,,..., m x x x L λ m j X λ j i i j= 1 i j = g ( X) = 0, j = 1,,..., m j (15) 13 which represent n + m equations in terms of the n + m unknowns, x i and λ j. The solution to this set of equations gives us * * x 1 λ 1 (16) * * x and * λ X = M * x n λ = M * λ m

14 Sufficient conditions for a general problem A sufficient condition for f(x) to have a relative minimum at X * is that each root of the polynomial in Є, defined by the following determinant equation be positive. 14 L L L L g g L g n 11 1 m1 L L L g g g 1 n 1 m M O M M O M L L L L g g L g n1 n nn 1n n mn g g L g 0 L L n g g g 1 n g g L g 0 L L 0 m1 m mn M O M M O M M M = 0 (17)

15 Sufficient conditions for a general problem contd. where L * * Lij = ( X, λ ), for i = 1,,..., n and j = 1,,..., m x x i j g gpq = X p= m q= x p * ( ), where 1,,..., and 1,,..., q n (18) Similarly, a sufficient condition for f(x) to have a relative maximum at X * is that each root of the polynomial in Є, defined by equation (17) be negative. If equation (17), on solving yields roots, some of which are positive and others negative, then the point X * is neither a maximum nor a minimum. 15

16 Example Minimize f ( X) = 3x 6x x 5x + 7x + 5x 1 1 1, Subject to x 1+ x = 5 or 16

17 L x = 6x 10x + 5+ λ = => 3x1+ 5 x = (5 + λ1) 1 => 3( x1+ x) + x = (5 + λ1) x = 1 x = X* =, ; * = [ 3] λ L11 L1 g11 L1 L g1 = g11 g1 0 0

18

19 Thank you 19

Optimization Methods: Optimization using Calculus - Equality constraints 1. Module 2 Lecture Notes 4

Optimization Methods: Optimization using Calculus - Equality constraints 1. Module 2 Lecture Notes 4 Optimization Methods: Optimization using Calculus - Equality constraints Module Lecture Notes 4 Optimization of Functions of Multiple Variables subect to Equality Constraints Introduction In the previous

More information

Lecture 9: Implicit function theorem, constrained extrema and Lagrange multipliers

Lecture 9: Implicit function theorem, constrained extrema and Lagrange multipliers Lecture 9: Implicit function theorem, constrained extrema and Lagrange multipliers Rafikul Alam Department of Mathematics IIT Guwahati What does the Implicit function theorem say? Let F : R 2 R be C 1.

More information

MATH529 Fundamentals of Optimization Constrained Optimization I

MATH529 Fundamentals of Optimization Constrained Optimization I MATH529 Fundamentals of Optimization Constrained Optimization I Marco A. Montes de Oca Mathematical Sciences, University of Delaware, USA 1 / 26 Motivating Example 2 / 26 Motivating Example min cost(b)

More information

AP Calculus Testbank (Chapter 9) (Mr. Surowski)

AP Calculus Testbank (Chapter 9) (Mr. Surowski) AP Calculus Testbank (Chapter 9) (Mr. Surowski) Part I. Multiple-Choice Questions n 1 1. The series will converge, provided that n 1+p + n + 1 (A) p > 1 (B) p > 2 (C) p >.5 (D) p 0 2. The series

More information

Lagrange multipliers. Portfolio optimization. The Lagrange multipliers method for finding constrained extrema of multivariable functions.

Lagrange multipliers. Portfolio optimization. The Lagrange multipliers method for finding constrained extrema of multivariable functions. Chapter 9 Lagrange multipliers Portfolio optimization The Lagrange multipliers method for finding constrained extrema of multivariable functions 91 Lagrange multipliers Optimization problems often require

More information

Introduction to the Calculus of Variations

Introduction to the Calculus of Variations 236861 Numerical Geometry of Images Tutorial 1 Introduction to the Calculus of Variations Alex Bronstein c 2005 1 Calculus Calculus of variations 1. Function Functional f : R n R Example: f(x, y) =x 2

More information

Calculus Overview. f(x) f (x) is slope. I. Single Variable. A. First Order Derivative : Concept : measures slope of curve at a point.

Calculus Overview. f(x) f (x) is slope. I. Single Variable. A. First Order Derivative : Concept : measures slope of curve at a point. Calculus Overview I. Single Variable A. First Order Derivative : Concept : measures slope of curve at a point. Notation : Let y = f (x). First derivative denoted f ʹ (x), df dx, dy dx, f, etc. Example

More information

2015 Math Camp Calculus Exam Solution

2015 Math Camp Calculus Exam Solution 015 Math Camp Calculus Exam Solution Problem 1: x = x x +5 4+5 = 9 = 3 1. lim We also accepted ±3, even though it is not according to the prevailing convention 1. x x 4 x+4 =. lim 4 4+4 = 4 0 = 4 0 = We

More information

Mechanical Systems II. Method of Lagrange Multipliers

Mechanical Systems II. Method of Lagrange Multipliers Mechanical Systems II. Method of Lagrange Multipliers Rafael Wisniewski Aalborg University Abstract. So far our approach to classical mechanics was limited to finding a critical point of a certain functional.

More information

AP Calculus (BC) Chapter 9 Test No Calculator Section Name: Date: Period:

AP Calculus (BC) Chapter 9 Test No Calculator Section Name: Date: Period: WORKSHEET: Series, Taylor Series AP Calculus (BC) Chapter 9 Test No Calculator Section Name: Date: Period: 1 Part I. Multiple-Choice Questions (5 points each; please circle the correct answer.) 1. The

More information

Sufficient Conditions for Finite-variable Constrained Minimization

Sufficient Conditions for Finite-variable Constrained Minimization Lecture 4 It is a small de tour but it is important to understand this before we move to calculus of variations. Sufficient Conditions for Finite-variable Constrained Minimization ME 256, Indian Institute

More information

Advanced Calculus Math 127B, Winter 2005 Solutions: Final. nx2 1 + n 2 x, g n(x) = n2 x

Advanced Calculus Math 127B, Winter 2005 Solutions: Final. nx2 1 + n 2 x, g n(x) = n2 x . Define f n, g n : [, ] R by f n (x) = Advanced Calculus Math 27B, Winter 25 Solutions: Final nx2 + n 2 x, g n(x) = n2 x 2 + n 2 x. 2 Show that the sequences (f n ), (g n ) converge pointwise on [, ],

More information

Taylor Series and stationary points

Taylor Series and stationary points Chapter 5 Taylor Series and stationary points 5.1 Taylor Series The surface z = f(x, y) and its derivatives can give a series approximation for f(x, y) about some point (x 0, y 0 ) as illustrated in Figure

More information

Making Models with Polynomials: Taylor Series

Making Models with Polynomials: Taylor Series Making Models with Polynomials: Taylor Series Why polynomials? i a 3 x 3 + a 2 x 2 + a 1 x + a 0 How do we write a general degree n polynomial? n a i x i i=0 Why polynomials and not something else? We

More information

Paul Schrimpf. October 17, UBC Economics 526. Constrained optimization. Paul Schrimpf. First order conditions. Second order conditions

Paul Schrimpf. October 17, UBC Economics 526. Constrained optimization. Paul Schrimpf. First order conditions. Second order conditions UBC Economics 526 October 17, 2012 .1.2.3.4 Section 1 . max f (x) s.t. h(x) = c f : R n R, h : R n R m Draw picture of n = 2 and m = 1 At optimum, constraint tangent to level curve of function Rewrite

More information

Chapter 11. Taylor Series. Josef Leydold Mathematical Methods WS 2018/19 11 Taylor Series 1 / 27

Chapter 11. Taylor Series. Josef Leydold Mathematical Methods WS 2018/19 11 Taylor Series 1 / 27 Chapter 11 Taylor Series Josef Leydold Mathematical Methods WS 2018/19 11 Taylor Series 1 / 27 First-Order Approximation We want to approximate function f by some simple function. Best possible approximation

More information

3 Applications of partial differentiation

3 Applications of partial differentiation Advanced Calculus Chapter 3 Applications of partial differentiation 37 3 Applications of partial differentiation 3.1 Stationary points Higher derivatives Let U R 2 and f : U R. The partial derivatives

More information

11.10a Taylor and Maclaurin Series

11.10a Taylor and Maclaurin Series 11.10a 1 11.10a Taylor and Maclaurin Series Let y = f(x) be a differentiable function at x = a. In first semester calculus we saw that (1) f(x) f(a)+f (a)(x a), for all x near a The right-hand side of

More information

Constrained optimization.

Constrained optimization. ams/econ 11b supplementary notes ucsc Constrained optimization. c 2016, Yonatan Katznelson 1. Constraints In many of the optimization problems that arise in economics, there are restrictions on the values

More information

Section Taylor and Maclaurin Series

Section Taylor and Maclaurin Series Section.0 Taylor and Maclaurin Series Ruipeng Shen Feb 5 Taylor and Maclaurin Series Main Goal: How to find a power series representation for a smooth function us assume that a smooth function has a power

More information

SECTION C: CONTINUOUS OPTIMISATION LECTURE 11: THE METHOD OF LAGRANGE MULTIPLIERS

SECTION C: CONTINUOUS OPTIMISATION LECTURE 11: THE METHOD OF LAGRANGE MULTIPLIERS SECTION C: CONTINUOUS OPTIMISATION LECTURE : THE METHOD OF LAGRANGE MULTIPLIERS HONOUR SCHOOL OF MATHEMATICS OXFORD UNIVERSITY HILARY TERM 005 DR RAPHAEL HAUSER. Examples. In this lecture we will take

More information

8. Constrained Optimization

8. Constrained Optimization 8. Constrained Optimization Daisuke Oyama Mathematics II May 11, 2018 Unconstrained Maximization Problem Let X R N be a nonempty set. Definition 8.1 For a function f : X R, x X is a (strict) local maximizer

More information

Error Bounds in Power Series

Error Bounds in Power Series Error Bounds in Power Series When using a power series to approximate a function (usually a Taylor Series), we often want to know how large any potential error is on a specified interval. There are two

More information

Optimization. Escuela de Ingeniería Informática de Oviedo. (Dpto. de Matemáticas-UniOvi) Numerical Computation Optimization 1 / 30

Optimization. Escuela de Ingeniería Informática de Oviedo. (Dpto. de Matemáticas-UniOvi) Numerical Computation Optimization 1 / 30 Optimization Escuela de Ingeniería Informática de Oviedo (Dpto. de Matemáticas-UniOvi) Numerical Computation Optimization 1 / 30 Unconstrained optimization Outline 1 Unconstrained optimization 2 Constrained

More information

Math 651 Introduction to Numerical Analysis I Fall SOLUTIONS: Homework Set 1

Math 651 Introduction to Numerical Analysis I Fall SOLUTIONS: Homework Set 1 ath 651 Introduction to Numerical Analysis I Fall 2010 SOLUTIONS: Homework Set 1 1. Consider the polynomial f(x) = x 2 x 2. (a) Find P 1 (x), P 2 (x) and P 3 (x) for f(x) about x 0 = 0. What is the relation

More information

e x = 1 + x + x2 2! + x3 If the function f(x) can be written as a power series on an interval I, then the power series is of the form

e x = 1 + x + x2 2! + x3 If the function f(x) can be written as a power series on an interval I, then the power series is of the form Taylor Series Given a function f(x), we would like to be able to find a power series that represents the function. For example, in the last section we noted that we can represent e x by the power series

More information

1 Lagrange Multiplier Method

1 Lagrange Multiplier Method 1 Lagrange Multiplier Method Near a maximum the decrements on both sides are in the beginning only imperceptible. J. Kepler When a quantity is greatest or least, at that moment its flow neither increases

More information

APPM 2350 FINAL EXAM FALL 2017

APPM 2350 FINAL EXAM FALL 2017 APPM 25 FINAL EXAM FALL 27. ( points) Determine the absolute maximum and minimum values of the function f(x, y) = 2 6x 4y + 4x 2 + y. Be sure to clearly give both the locations and values of the absolute

More information

Optimization. A first course on mathematics for economists

Optimization. A first course on mathematics for economists Optimization. A first course on mathematics for economists Xavier Martinez-Giralt Universitat Autònoma de Barcelona xavier.martinez.giralt@uab.eu II.3 Static optimization - Non-Linear programming OPT p.1/45

More information

As f and g are differentiable functions such that. f (x) = 20e 2x, g (x) = 4e 2x + 4xe 2x,

As f and g are differentiable functions such that. f (x) = 20e 2x, g (x) = 4e 2x + 4xe 2x, srinivasan (rs7) Sample Midterm srinivasan (690) This print-out should have 0 questions. Multiple-choice questions may continue on the next column or page find all choices before answering. Determine if

More information

EC /11. Math for Microeconomics September Course, Part II Problem Set 1 with Solutions. a11 a 12. x 2

EC /11. Math for Microeconomics September Course, Part II Problem Set 1 with Solutions. a11 a 12. x 2 LONDON SCHOOL OF ECONOMICS Professor Leonardo Felli Department of Economics S.478; x7525 EC400 2010/11 Math for Microeconomics September Course, Part II Problem Set 1 with Solutions 1. Show that the general

More information

Final Exam - Answer key

Final Exam - Answer key Fall4 Final Exam - Answer key ARE Problem (Analysis) [7 points]: A) [ points] Let A = (,) R be endowed with the Euclidean metric. Consider the set O = {( n,) : n N}. Does the open cover O of A have a finite

More information

JUST THE MATHS UNIT NUMBER DIFFERENTIATION APPLICATIONS 5 (Maclaurin s and Taylor s series) A.J.Hobson

JUST THE MATHS UNIT NUMBER DIFFERENTIATION APPLICATIONS 5 (Maclaurin s and Taylor s series) A.J.Hobson JUST THE MATHS UNIT NUMBER.5 DIFFERENTIATION APPLICATIONS 5 (Maclaurin s and Taylor s series) by A.J.Hobson.5. Maclaurin s series.5. Standard series.5.3 Taylor s series.5.4 Exercises.5.5 Answers to exercises

More information

Math RE - Calculus II Antiderivatives and the Indefinite Integral Page 1 of 5

Math RE - Calculus II Antiderivatives and the Indefinite Integral Page 1 of 5 Math 201-203-RE - Calculus II Antiderivatives and the Indefinite Integral Page 1 of 5 What is the Antiderivative? In a derivative problem, a function f(x) is given and you find the derivative f (x) using

More information

TAYLOR AND MACLAURIN SERIES

TAYLOR AND MACLAURIN SERIES TAYLOR AND MACLAURIN SERIES. Introduction Last time, we were able to represent a certain restricted class of functions as power series. This leads us to the question: can we represent more general functions

More information

Here each term has degree 2 (the sum of exponents is 2 for all summands). A quadratic form of three variables looks as

Here each term has degree 2 (the sum of exponents is 2 for all summands). A quadratic form of three variables looks as Reading [SB], Ch. 16.1-16.3, p. 375-393 1 Quadratic Forms A quadratic function f : R R has the form f(x) = a x. Generalization of this notion to two variables is the quadratic form Q(x 1, x ) = a 11 x

More information

Completion Date: Monday February 11, 2008

Completion Date: Monday February 11, 2008 MATH 4 (R) Winter 8 Intermediate Calculus I Solutions to Problem Set #4 Completion Date: Monday February, 8 Department of Mathematical and Statistical Sciences University of Alberta Question. [Sec..9,

More information

ENGI Partial Differentiation Page y f x

ENGI Partial Differentiation Page y f x ENGI 3424 4 Partial Differentiation Page 4-01 4. Partial Differentiation For functions of one variable, be found unambiguously by differentiation: y f x, the rate of change of the dependent variable can

More information

E 600 Chapter 4: Optimization

E 600 Chapter 4: Optimization E 600 Chapter 4: Optimization Simona Helmsmueller August 8, 2018 Goals of this lecture: Every theorem in these slides is important! You should understand, remember and be able to apply each and every one

More information

Chapter 8: Taylor s theorem and L Hospital s rule

Chapter 8: Taylor s theorem and L Hospital s rule Chapter 8: Taylor s theorem and L Hospital s rule Theorem: [Inverse Mapping Theorem] Suppose that a < b and f : [a, b] R. Given that f (x) > 0 for all x (a, b) then f 1 is differentiable on (f(a), f(b))

More information

Lagrange Multipliers

Lagrange Multipliers Lagrange Multipliers (Com S 477/577 Notes) Yan-Bin Jia Nov 9, 2017 1 Introduction We turn now to the study of minimization with constraints. More specifically, we will tackle the following problem: minimize

More information

EC /11. Math for Microeconomics September Course, Part II Lecture Notes. Course Outline

EC /11. Math for Microeconomics September Course, Part II Lecture Notes. Course Outline LONDON SCHOOL OF ECONOMICS Professor Leonardo Felli Department of Economics S.478; x7525 EC400 20010/11 Math for Microeconomics September Course, Part II Lecture Notes Course Outline Lecture 1: Tools for

More information

MTH4101 CALCULUS II REVISION NOTES. 1. COMPLEX NUMBERS (Thomas Appendix 7 + lecture notes) ax 2 + bx + c = 0. x = b ± b 2 4ac 2a. i = 1.

MTH4101 CALCULUS II REVISION NOTES. 1. COMPLEX NUMBERS (Thomas Appendix 7 + lecture notes) ax 2 + bx + c = 0. x = b ± b 2 4ac 2a. i = 1. MTH4101 CALCULUS II REVISION NOTES 1. COMPLEX NUMBERS (Thomas Appendix 7 + lecture notes) 1.1 Introduction Types of numbers (natural, integers, rationals, reals) The need to solve quadratic equations:

More information

8.7 Taylor s Inequality Math 2300 Section 005 Calculus II. f(x) = ln(1 + x) f(0) = 0

8.7 Taylor s Inequality Math 2300 Section 005 Calculus II. f(x) = ln(1 + x) f(0) = 0 8.7 Taylor s Inequality Math 00 Section 005 Calculus II Name: ANSWER KEY Taylor s Inequality: If f (n+) is continuous and f (n+) < M between the center a and some point x, then f(x) T n (x) M x a n+ (n

More information

Midterm Review. Igor Yanovsky (Math 151A TA)

Midterm Review. Igor Yanovsky (Math 151A TA) Midterm Review Igor Yanovsky (Math 5A TA) Root-Finding Methods Rootfinding methods are designed to find a zero of a function f, that is, to find a value of x such that f(x) =0 Bisection Method To apply

More information

Math 5311 Constrained Optimization Notes

Math 5311 Constrained Optimization Notes ath 5311 Constrained Optimization otes February 5, 2009 1 Equality-constrained optimization Real-world optimization problems frequently have constraints on their variables. Constraints may be equality

More information

Nonlinear Programming (NLP)

Nonlinear Programming (NLP) Natalia Lazzati Mathematics for Economics (Part I) Note 6: Nonlinear Programming - Unconstrained Optimization Note 6 is based on de la Fuente (2000, Ch. 7), Madden (1986, Ch. 3 and 5) and Simon and Blume

More information

Finite Dimensional Optimization Part III: Convex Optimization 1

Finite Dimensional Optimization Part III: Convex Optimization 1 John Nachbar Washington University March 21, 2017 Finite Dimensional Optimization Part III: Convex Optimization 1 1 Saddle points and KKT. These notes cover another important approach to optimization,

More information

Differentiation. f(x + h) f(x) Lh = L.

Differentiation. f(x + h) f(x) Lh = L. Analysis in R n Math 204, Section 30 Winter Quarter 2008 Paul Sally, e-mail: sally@math.uchicago.edu John Boller, e-mail: boller@math.uchicago.edu website: http://www.math.uchicago.edu/ boller/m203 Differentiation

More information

Ma 530 Power Series II

Ma 530 Power Series II Ma 530 Power Series II Please note that there is material on power series at Visual Calculus. Some of this material was used as part of the presentation of the topics that follow. Operations on Power Series

More information

Math Real Analysis II

Math Real Analysis II Math 432 - Real Analysis II Solutions to Homework due February 3 In class, we learned that the n-th remainder for a smooth function f(x) defined on some open interval containing is given by f (k) () R

More information

Exam 2. Average: 85.6 Median: 87.0 Maximum: Minimum: 55.0 Standard Deviation: Numerical Methods Fall 2011 Lecture 20

Exam 2. Average: 85.6 Median: 87.0 Maximum: Minimum: 55.0 Standard Deviation: Numerical Methods Fall 2011 Lecture 20 Exam 2 Average: 85.6 Median: 87.0 Maximum: 100.0 Minimum: 55.0 Standard Deviation: 10.42 Fall 2011 1 Today s class Multiple Variable Linear Regression Polynomial Interpolation Lagrange Interpolation Newton

More information

Final Exam Advanced Mathematics for Economics and Finance

Final Exam Advanced Mathematics for Economics and Finance Final Exam Advanced Mathematics for Economics and Finance Dr. Stefanie Flotho Winter Term /5 March 5 General Remarks: ˆ There are four questions in total. ˆ All problems are equally weighed. ˆ This is

More information

Lagrange Relaxation and Duality

Lagrange Relaxation and Duality Lagrange Relaxation and Duality As we have already known, constrained optimization problems are harder to solve than unconstrained problems. By relaxation we can solve a more difficult problem by a simpler

More information

Lecture Notes for Chapter 9

Lecture Notes for Chapter 9 Lecture Notes for Chapter 9 Kevin Wainwright April 26, 2014 1 Optimization of One Variable 1.1 Critical Points A critical point occurs whenever the firest derivative of a function is equal to zero. ie.

More information

3.7 Constrained Optimization and Lagrange Multipliers

3.7 Constrained Optimization and Lagrange Multipliers 3.7 Constrained Optimization and Lagrange Multipliers 71 3.7 Constrained Optimization and Lagrange Multipliers Overview: Constrained optimization problems can sometimes be solved using the methods of the

More information

Bob Brown Math 251 Calculus 1 Chapter 4, Section 1 Completed 1 CCBC Dundalk

Bob Brown Math 251 Calculus 1 Chapter 4, Section 1 Completed 1 CCBC Dundalk Bob Brown Math 251 Calculus 1 Chapter 4, Section 1 Completed 1 Absolute (or Global) Minima and Maxima Def.: Let x = c be a number in the domain of a function f. f has an absolute (or, global ) minimum

More information

Lecture Notes: Math Refresher 1

Lecture Notes: Math Refresher 1 Lecture Notes: Math Refresher 1 Math Facts The following results from calculus will be used over and over throughout the course. A more complete list of useful results from calculus is posted on the course

More information

Partial Derivatives. w = f(x, y, z).

Partial Derivatives. w = f(x, y, z). Partial Derivatives 1 Functions of Several Variables So far we have focused our attention of functions of one variable. These functions model situations in which a variable depends on another independent

More information

Lecture 4: Optimization. Maximizing a function of a single variable

Lecture 4: Optimization. Maximizing a function of a single variable Lecture 4: Optimization Maximizing or Minimizing a Function of a Single Variable Maximizing or Minimizing a Function of Many Variables Constrained Optimization Maximizing a function of a single variable

More information

MS 2001: Test 1 B Solutions

MS 2001: Test 1 B Solutions MS 2001: Test 1 B Solutions Name: Student Number: Answer all questions. Marks may be lost if necessary work is not clearly shown. Remarks by me in italics and would not be required in a test - J.P. Question

More information

Homework 8 Solutions to Selected Problems

Homework 8 Solutions to Selected Problems Homework 8 Solutions to Selected Problems June 7, 01 1 Chapter 17, Problem Let f(x D[x] and suppose f(x is reducible in D[x]. That is, there exist polynomials g(x and h(x in D[x] such that g(x and h(x

More information

LESSON 25: LAGRANGE MULTIPLIERS OCTOBER 30, 2017

LESSON 25: LAGRANGE MULTIPLIERS OCTOBER 30, 2017 LESSON 5: LAGRANGE MULTIPLIERS OCTOBER 30, 017 Lagrange multipliers is another method of finding minima and maxima of functions of more than one variable. In fact, many of the problems from the last homework

More information

MATH Midterm 1 Sample 4

MATH Midterm 1 Sample 4 1. (15 marks) (a) (4 marks) Given the function: f(x, y) = arcsin x 2 + y, find its first order partial derivatives at the point (0, 3). Simplify your answers. Solution: Compute the first order partial

More information

Convergence of sequences and series

Convergence of sequences and series Convergence of sequences and series A sequence f is a map from N the positive integers to a set. We often write the map outputs as f n rather than f(n). Often we just list the outputs in order and leave

More information

Math 121 Winter 2010 Review Sheet

Math 121 Winter 2010 Review Sheet Math 121 Winter 2010 Review Sheet March 14, 2010 This review sheet contains a number of problems covering the material that we went over after the third midterm exam. These problems (in conjunction with

More information

REVIEW OF DIFFERENTIAL CALCULUS

REVIEW OF DIFFERENTIAL CALCULUS REVIEW OF DIFFERENTIAL CALCULUS DONU ARAPURA 1. Limits and continuity To simplify the statements, we will often stick to two variables, but everything holds with any number of variables. Let f(x, y) be

More information

Solutions to Homework 7

Solutions to Homework 7 Solutions to Homework 7 Exercise #3 in section 5.2: A rectangular box is inscribed in a hemisphere of radius r. Find the dimensions of the box of maximum volume. Solution: The base of the rectangular box

More information

Math 113 (Calculus 2) Exam 4

Math 113 (Calculus 2) Exam 4 Math 3 (Calculus ) Exam 4 November 0 November, 009 Sections 0, 3 7 Name Student ID Section Instructor In some cases a series may be seen to converge or diverge for more than one reason. For such problems

More information

Calculus 2502A - Advanced Calculus I Fall : Local minima and maxima

Calculus 2502A - Advanced Calculus I Fall : Local minima and maxima Calculus 50A - Advanced Calculus I Fall 014 14.7: Local minima and maxima Martin Frankland November 17, 014 In these notes, we discuss the problem of finding the local minima and maxima of a function.

More information

Generalization to inequality constrained problem. Maximize

Generalization to inequality constrained problem. Maximize Lecture 11. 26 September 2006 Review of Lecture #10: Second order optimality conditions necessary condition, sufficient condition. If the necessary condition is violated the point cannot be a local minimum

More information

(Kernels +) Support Vector Machines

(Kernels +) Support Vector Machines (Kernels +) Support Vector Machines Machine Learning Torsten Möller Reading Chapter 5 of Machine Learning An Algorithmic Perspective by Marsland Chapter 6+7 of Pattern Recognition and Machine Learning

More information

Seminars on Mathematics for Economics and Finance Topic 5: Optimization Kuhn-Tucker conditions for problems with inequality constraints 1

Seminars on Mathematics for Economics and Finance Topic 5: Optimization Kuhn-Tucker conditions for problems with inequality constraints 1 Seminars on Mathematics for Economics and Finance Topic 5: Optimization Kuhn-Tucker conditions for problems with inequality constraints 1 Session: 15 Aug 2015 (Mon), 10:00am 1:00pm I. Optimization with

More information

Basics of Calculus and Algebra

Basics of Calculus and Algebra Monika Department of Economics ISCTE-IUL September 2012 Basics of linear algebra Real valued Functions Differential Calculus Integral Calculus Optimization Introduction I A matrix is a rectangular array

More information

Taylor and Maclaurin Series

Taylor and Maclaurin Series Taylor and Maclaurin Series MATH 211, Calculus II J. Robert Buchanan Department of Mathematics Spring 2018 Background We have seen that some power series converge. When they do, we can think of them as

More information

Optimization: Problem Set Solutions

Optimization: Problem Set Solutions Optimization: Problem Set Solutions Annalisa Molino University of Rome Tor Vergata annalisa.molino@uniroma2.it Fall 20 Compute the maxima minima or saddle points of the following functions a. f(x y) =

More information

Mathematical Economics: Lecture 16

Mathematical Economics: Lecture 16 Mathematical Economics: Lecture 16 Yu Ren WISE, Xiamen University November 26, 2012 Outline 1 Chapter 21: Concave and Quasiconcave Functions New Section Chapter 21: Concave and Quasiconcave Functions Concave

More information

A Short Essay on Variational Calculus

A Short Essay on Variational Calculus A Short Essay on Variational Calculus Keonwook Kang, Chris Weinberger and Wei Cai Department of Mechanical Engineering, Stanford University Stanford, CA 94305-4040 May 3, 2006 Contents 1 Definition of

More information

Review for the Final Exam

Review for the Final Exam Math 171 Review for the Final Exam 1 Find the limits (4 points each) (a) lim 4x 2 3; x x (b) lim ( x 2 x x 1 )x ; (c) lim( 1 1 ); x 1 ln x x 1 sin (x 2) (d) lim x 2 x 2 4 Solutions (a) The limit lim 4x

More information

d(x n, x) d(x n, x nk ) + d(x nk, x) where we chose any fixed k > N

d(x n, x) d(x n, x nk ) + d(x nk, x) where we chose any fixed k > N Problem 1. Let f : A R R have the property that for every x A, there exists ɛ > 0 such that f(t) > ɛ if t (x ɛ, x + ɛ) A. If the set A is compact, prove there exists c > 0 such that f(x) > c for all x

More information

The Derivative. Appendix B. B.1 The Derivative of f. Mappings from IR to IR

The Derivative. Appendix B. B.1 The Derivative of f. Mappings from IR to IR Appendix B The Derivative B.1 The Derivative of f In this chapter, we give a short summary of the derivative. Specifically, we want to compare/contrast how the derivative appears for functions whose domain

More information

CALCULUS JIA-MING (FRANK) LIOU

CALCULUS JIA-MING (FRANK) LIOU CALCULUS JIA-MING (FRANK) LIOU Abstract. Contents. Power Series.. Polynomials and Formal Power Series.2. Radius of Convergence 2.3. Derivative and Antiderivative of Power Series 4.4. Power Series Expansion

More information

Chapter 3 Numerical Methods

Chapter 3 Numerical Methods Chapter 3 Numerical Methods Part 2 3.2 Systems of Equations 3.3 Nonlinear and Constrained Optimization 1 Outline 3.2 Systems of Equations 3.3 Nonlinear and Constrained Optimization Summary 2 Outline 3.2

More information

HOMEWORK 7 SOLUTIONS

HOMEWORK 7 SOLUTIONS HOMEWORK 7 SOLUTIONS MA11: ADVANCED CALCULUS, HILARY 17 (1) Using the method of Lagrange multipliers, find the largest and smallest values of the function f(x, y) xy on the ellipse x + y 1. Solution: The

More information

Constrained Optimization

Constrained Optimization Constrained Optimization Joshua Wilde, revised by Isabel Tecu, Takeshi Suzuki and María José Boccardi August 13, 2013 1 General Problem Consider the following general constrained optimization problem:

More information

Lagrange Multipliers

Lagrange Multipliers Optimization with Constraints As long as algebra and geometry have been separated, their progress have been slow and their uses limited; but when these two sciences have been united, they have lent each

More information

Numerical Optimization

Numerical Optimization Constrained Optimization Computer Science and Automation Indian Institute of Science Bangalore 560 012, India. NPTEL Course on Constrained Optimization Constrained Optimization Problem: min h j (x) 0,

More information

Examination paper for TMA4215 Numerical Mathematics

Examination paper for TMA4215 Numerical Mathematics Department of Mathematical Sciences Examination paper for TMA425 Numerical Mathematics Academic contact during examination: Trond Kvamsdal Phone: 93058702 Examination date: 6th of December 207 Examination

More information

Let f(x) = x, but the domain of f is the interval 0 x 1. Note

Let f(x) = x, but the domain of f is the interval 0 x 1. Note I.g Maximum and Minimum. Lagrange Multipliers Recall: Suppose we are given y = f(x). We recall that the maximum/minimum points occur at the following points: (1) where f = 0; (2) where f does not exist;

More information

Calculus Example Exam Solutions

Calculus Example Exam Solutions Calculus Example Exam Solutions. Limits (8 points, 6 each) Evaluate the following limits: p x 2 (a) lim x 4 We compute as follows: lim p x 2 x 4 p p x 2 x +2 x 4 p x +2 x 4 (x 4)( p x + 2) p x +2 = p 4+2

More information

Physics 105, Spring 2015, Reinsch. Homework Assignment 1 (This is a five page document. More hints added Feb. 2, 9pm. Problem statements unchanged.

Physics 105, Spring 2015, Reinsch. Homework Assignment 1 (This is a five page document. More hints added Feb. 2, 9pm. Problem statements unchanged. Physics 105, Spring 2015, Reinsch Homework Assignment 1 (This is a five page document. More hints added Feb. 2, 9pm. Problem statements unchanged.) Due Wednesday, February 4, 5:00 pm Problem 1 A particle

More information

Max Margin-Classifier

Max Margin-Classifier Max Margin-Classifier Oliver Schulte - CMPT 726 Bishop PRML Ch. 7 Outline Maximum Margin Criterion Math Maximizing the Margin Non-Separable Data Kernels and Non-linear Mappings Where does the maximization

More information

Power Series Solutions We use power series to solve second order differential equations

Power Series Solutions We use power series to solve second order differential equations Objectives Power Series Solutions We use power series to solve second order differential equations We use power series expansions to find solutions to second order, linear, variable coefficient equations

More information

Linearization and Extreme Values of Functions

Linearization and Extreme Values of Functions Linearization and Extreme Values of Functions 3.10 Linearization and Differentials Linear or Tangent Line Approximations of function values Equation of tangent to y = f(x) at (a, f(a)): Tangent line approximation

More information

Mathematical Economics. Lecture Notes (in extracts)

Mathematical Economics. Lecture Notes (in extracts) Prof. Dr. Frank Werner Faculty of Mathematics Institute of Mathematical Optimization (IMO) http://math.uni-magdeburg.de/ werner/math-ec-new.html Mathematical Economics Lecture Notes (in extracts) Winter

More information

Microeconomic Theory. Microeconomic Theory. Everyday Economics. The Course:

Microeconomic Theory. Microeconomic Theory. Everyday Economics. The Course: The Course: Microeconomic Theory This is the first rigorous course in microeconomic theory This is a course on economic methodology. The main goal is to teach analytical tools that will be useful in other

More information

Lecture 4 September 15

Lecture 4 September 15 IFT 6269: Probabilistic Graphical Models Fall 2017 Lecture 4 September 15 Lecturer: Simon Lacoste-Julien Scribe: Philippe Brouillard & Tristan Deleu 4.1 Maximum Likelihood principle Given a parametric

More information

Review of Optimization Methods

Review of Optimization Methods Review of Optimization Methods Prof. Manuela Pedio 20550 Quantitative Methods for Finance August 2018 Outline of the Course Lectures 1 and 2 (3 hours, in class): Linear and non-linear functions on Limits,

More information

(0, 0), (1, ), (2, ), (3, ), (4, ), (5, ), (6, ).

(0, 0), (1, ), (2, ), (3, ), (4, ), (5, ), (6, ). 1 Interpolation: The method of constructing new data points within the range of a finite set of known data points That is if (x i, y i ), i = 1, N are known, with y i the dependent variable and x i [x

More information

g(t) = f(x 1 (t),..., x n (t)).

g(t) = f(x 1 (t),..., x n (t)). Reading: [Simon] p. 313-333, 833-836. 0.1 The Chain Rule Partial derivatives describe how a function changes in directions parallel to the coordinate axes. Now we shall demonstrate how the partial derivatives

More information