Economics 204 Summer/Fall 2011 Lecture 11 Monday August 8, f(x + h) (f(x) + ah) = a = 0 = 0. f(x + h) (f(x) + T x (h))

Similar documents
Lecture 10. Econ August 21

Economics 204 Summer/Fall 2010 Lecture 10 Friday August 6, 2010

MA102: Multivariable Calculus

f(x) f(z) c x z > 0 1

Differentiation. f(x + h) f(x) Lh = L.

THE INVERSE FUNCTION THEOREM

Econ 204 Supplement to Section 3.6 Diagonalization and Quadratic Forms. 1 Diagonalization and Change of Basis

Problem Set 4. f(a + h) = P k (h) + o( h k ). (3)

Chapter 3. Differentiable Mappings. 1. Differentiable Mappings

1. Bounded linear maps. A linear map T : E F of real Banach

INVERSE FUNCTION THEOREM and SURFACES IN R n

Review of Multi-Calculus (Study Guide for Spivak s CHAPTER ONE TO THREE)

Economics 204 Summer/Fall 2017 Lecture 12 Tuesday August 1, 2017

Nonlinear equations. Norms for R n. Convergence orders for iterative methods

Economics 205 Exercises

Economics 204 Summer/Fall 2017 Lecture 7 Tuesday July 25, 2017

Chapter 8: Taylor s theorem and L Hospital s rule

1 Directional Derivatives and Differentiability

Economics 205, Fall 2002: Final Examination, Possible Answers

MAT 419 Lecture Notes Transcribed by Eowyn Cenek 6/1/2012

Economics 204 Summer/Fall 2011 Lecture 5 Friday July 29, 2011

Caculus 221. Possible questions for Exam II. March 19, 2002

Economics 204 Fall 2013 Problem Set 5 Suggested Solutions

Real Analysis. Joe Patten August 12, 2018

Metric Spaces and Topology

REAL ANALYSIS II: PROBLEM SET 2

j=1 u 1jv 1j. 1/ 2 Lemma 1. An orthogonal set of vectors must be linearly independent.

REVIEW OF DIFFERENTIAL CALCULUS

MAT 473 Intermediate Real Analysis II

Chapter 2: Unconstrained Extrema

Advanced Calculus Math 127B, Winter 2005 Solutions: Final. nx2 1 + n 2 x, g n(x) = n2 x

Functions. Chapter Continuous Functions

We denote the derivative at x by DF (x) = L. With respect to the standard bases of R n and R m, DF (x) is simply the matrix of partial derivatives,

x y More precisely, this equation means that given any ε > 0, there exists some δ > 0 such that

Sample Problems for the Second Midterm Exam

5 Integration with Differential Forms The Poincare Lemma Proper Maps and Degree Topological Invariance of Degree...

Real Analysis Math 131AH Rudin, Chapter #1. Dominique Abdi

MATH Max-min Theory Fall 2016

Analysis Finite and Infinite Sets The Real Numbers The Cantor Set

Math 115 Spring 11 Written Homework 10 Solutions

Functional Analysis. Franck Sueur Metric spaces Definitions Completeness Compactness Separability...

Math 291-2: Final Exam Solutions Northwestern University, Winter 2016

MATH 409 Advanced Calculus I Lecture 10: Continuity. Properties of continuous functions.

21-268: Multi-dimensional Calculus Spring 2012

E 600 Chapter 3: Multivariate Calculus

e x = 1 + x + x2 2! + x3 If the function f(x) can be written as a power series on an interval I, then the power series is of the form

Rank (df x ) min{m, n} We say

Chapter 3 Transformations

SIMPLE MULTIVARIATE OPTIMIZATION

MATH 409 Advanced Calculus I Lecture 16: Mean value theorem. Taylor s formula.

MAT137 Calculus! Lecture 10

Test 3 Review. y f(a) = f (a)(x a) y = f (a)(x a) + f(a) L(x) = f (a)(x a) + f(a)

November 18, 2013 ANALYTIC FUNCTIONAL CALCULUS

UNCONSTRAINED OPTIMIZATION PAUL SCHRIMPF OCTOBER 24, 2013

Analysis II - few selective results

Economics 204 Fall 2011 Problem Set 2 Suggested Solutions

Convex Optimization. Chapter Introduction Mathematical Optimization

Symmetric Matrices and Eigendecomposition

Definition 1.2. Let p R n be a point and v R n be a non-zero vector. The line through p in direction v is the set

Appendix E : Note on regular curves in Euclidean spaces

AN ELEMENTARY PROOF OF THE SPECTRAL RADIUS FORMULA FOR MATRICES

Course Summary Math 211

Mathematics for Economists

1. Nonlinear Equations. This lecture note excerpted parts from Michael Heath and Max Gunzburger. f(x) = 0

B553 Lecture 3: Multivariate Calculus and Linear Algebra Review

2.10 Saddles, Nodes, Foci and Centers

1.1 Limits and Continuity. Precise definition of a limit and limit laws. Squeeze Theorem. Intermediate Value Theorem. Extreme Value Theorem.

ARE211, Fall 2004 CONTENTS. 4. Univariate and Multivariate Differentiation (cont) Four graphical examples Taylor s Theorem 9

MATH 163 HOMEWORK Week 13, due Monday April 26 TOPICS. c n (x a) n then c n = f(n) (a) n!

Calculus 2502A - Advanced Calculus I Fall : Local minima and maxima

Tangent spaces, normals and extrema

(c) For each α R \ {0}, the mapping x αx is a homeomorphism of X.

The Derivative. Appendix B. B.1 The Derivative of f. Mappings from IR to IR

November 13, 2018 MAT186 Week 8 Justin Ko

Nonlinear Optimization

Calculus and Maximization I

MATH 5720: Unconstrained Optimization Hung Phan, UMass Lowell September 13, 2018

Transpose & Dot Product

Throughout these notes we assume V, W are finite dimensional inner product spaces over C.

Lecture 1 INF-MAT3350/ : Some Tridiagonal Matrix Problems

DIFFERENTIAL GEOMETRY Multivariable Calculus, a refresher

12. Perturbed Matrices

V. Graph Sketching and Max-Min Problems

Set, functions and Euclidean space. Seungjin Han

CHAPTER 7. Connectedness

Lecture 9: Taylor Series

LECTURE 5: SMOOTH MAPS. 1. Smooth Maps

Higher order derivative

Transformations from R m to R n.

Test 2 Review Math 1111 College Algebra

Transpose & Dot Product

DS-GA 1002 Lecture notes 0 Fall Linear Algebra. These notes provide a review of basic concepts in linear algebra.

Several variables. x 1 x 2. x n

Implicit Functions, Curves and Surfaces

M2PM1 Analysis II (2008) Dr M Ruzhansky List of definitions, statements and examples Preliminary version

EC9A0: Pre-sessional Advanced Mathematics Course. Lecture Notes: Unconstrained Optimisation By Pablo F. Beker 1

Vector Calculus. Lecture Notes

Spectral Graph Theory Lecture 2. The Laplacian. Daniel A. Spielman September 4, x T M x. ψ i = arg min

Continuity. Chapter 4

Transcription:

Economics 204 Summer/Fall 2011 Lecture 11 Monday August 8, 2011 Sections 41-43 (Unified) Definition 1 Let f : I R, where I R is an open interval f is differentiable at x I if for some a R f(x + h) f lim h 0 h = a This is equivalent to f(x + h) (f + ah) lim = 0 h 0 h ε > 0 δ > 0 st 0 < h < δ ε > 0 δ > 0 st 0 < h < δ f(x + h) (f + ah) lim = 0 h 0 h f(x + h) (f + ah) h f(x + h) (f + ah) h < ε < ε Recall that the limit considers h near zero, but not h = 0 Definition 2 If X R n is open, f : X R m is differentiable at x X if 1 T x L(R n,r m ) st f(x + h) (f + T x (h)) lim h 0,h R n h = 0 (1) f is differentiable if it is differentiable at all x X Note that T x is uniquely determined by Equation (1) h is a small, nonzero element of R n ; h 0 from any direction, from above, below, along a spiral, etc The definition requires that one linear operator T x works no matter how h approaches zero In this case, f + T x (h) is the best linear approximation to f(x + h) for small h Notation: y = O( h n ) as h 0 read y is big-oh of h n means 1 Recall denotes the Euclidean distance K, δ > 0 st h < δ y K h n 1

y = o( h n ) as h 0 read y is little-oh of h n means lim h 0 y h n = 0 Note that the statement y = O( h n+1 ) as h 0 implies y = o( h n ) as h 0 Also note that if y is either O( h n ) or o( h n ), then y 0 as h 0; the difference in whether y is big-oh or little-oh tells us something about the rate at which y 0 Using this notation, note that f is differentiable at x T x L(R n,r m ) such that f(x + h) = f + T x (h) + o(h) as h 0 Notation: df x is the linear transformation T x Df is the matrix of df x with respect to the standard basis This is called the Jacobian or Jacobian matrix of f at x E f (h) = f(x + h) (f + df x (h)) is the error term Using this notation, f is differentiable at x E f (h) = o(h) as h 0 Now compute Df = (a ij ) Let {e 1,, e n } be the standard basis of R n Look in direction e j (note that γe j = γ ) o(γ) = f(x + γe j ) (f + T x (γe j )) a 11 a 1j a 1n = f(x + γe j ) f + a m1 a mj a mn = f(x + γe j ) f + γa 1j γa mj 0 0 γ 0 0 2

For i = 1,, m, let f i denote the i th component of the function f: f i (x + γe j ) ( ) f i + γa ij = o(γ) so a ij = fi x j Theorem 3 (Thm 33) Suppose X R n is open and f : X R m is differentiable at x X Then fi x j exists at x for 1 i m, 1 j n, and Df = f 1 x 1 f m x 1 f 1 x n f m x n ie the Jacobian at x is the matrix of partial derivatives at x Remark: If f is differentiable at x, then all first-order partial derivatives fi x j exist at x However, the converse is false: existence of all the first-order partial derivatives does not imply that f is differentiable The missing piece is continuity of the partial derivatives: Theorem 4 (Thm 34) If all the first-order partial derivatives fi x j n) exist and are continuous at x, then f is differentiable at x (1 i m, 1 j Directional Derivatives: Suppose X R n open, f : X R m is differentiable at x, and u = 1 f(x + γu) (f + T x (γu)) = o(γ) as γ 0 f(x + γu) (f + γt x (u)) = o(γ) as γ 0 f(x + γu) f lim = T x (u) = Dfu γ 0 γ ie the directional derivative in the direction u (with u = 1) is Dfu R m Theorem 5 (Thm 35, Chain Rule) Let X R n, Y R m be open, f : X Y, g : Y R p Let x 0 X and F = g f If f is differentiable at x 0 and g is differentiable at f(x 0 ), then F = g f is differentiable at x 0 and df x0 = dg f(x0 ) df x0 (composition of linear transformations) DF(x 0 ) = Dg(f(x 0 ))Df(x 0 ) (matrix multiplication) 3

Remark: The statement is exactly the same as in the univariate case, except we replace the univariate derivative by a linear transformation The proof is more or less the same, with a bit of linear algebra added Theorem 6 (Thm 17, Mean Value Theorem, Univariate Case) Let a, b R Suppose f : [a, b] R is continuous on [a, b] and differentiable on (a, b) Then there exists c (a, b) such that f(b) f(a) = f (c) b a that is, such that f(b) f(a) = f (c)(b a) Proof: Consider the function g = f f(a) f(b) f(a) (x a) b a Then g(a) = 0 = g(b) See Figure 1 Note that for x (a, b), g = f so it suffices to find c (a, b) such that g (c) = 0 f(b) f(a) b a Case I: If g = 0 for all x [a, b], choose an arbitrary c (a, b), and note that g (c) = 0, so we are done Case II: Suppose g > 0 for some x [a, b] Since g is continuous on [a, b], it attains its maximum at some point c (a, b) Since g is differentiable at c and c is an interior point of the domain of g, we have g (c) = 0, and we are done Case III: If g < 0 for some x [a, b], the argument is similar to that in Case II Remark: The Mean Value Theorem is useful for estimating bounds on functions and error terms in approximation of functions Notation: is the line segment from x to y l(x, y) = {αx + (1 α)y : α [0, 1]} Theorem 7 (Mean Value Theorem) Suppose f : R n R is differentiable on an open set X R n, x, y X and l(x, y) X Then there exists z l(x, y) such that f(y) f = Df(z)(y x) 4

Remark: This statement is different from Theorem 37 in de la Fuente Notice that the statement is exactly the same as in the univariate case For f : R n R m, we can apply the Mean Value Theorem to each component, to obtain z 1,, z m l(x, y) such that f i (y) f i = Df i (z i )(y x) However, we cannot find a single z which works for every component Note that each z i l(x, y) R n ; there are m of them, one for each component in the range The following result plays the same role in estimating function values and error terms for functions taking values in R m as the Mean Value Theorem plays for functions from R to R Theorem 8 Suppose X R n is open and f : X R m is differentiable If x, y X and l(x, y) X, then there exists z l(x, y) such that f(y) f df z (y x) df z y x Remark: To understand why we don t get equality, consider f : [0, 1] R 2 defined by f(t) = (cos 2πt, sin2πt) f maps [0, 1] to the unit circle in R 2 Note that f(0) = f(1) = (1, 0), so f(1) f(0) = 0 However, for any z [0, 1], df z (1 0) = 2π( sin2πz, cos2πz) = 2π sin 2 2πz + cos 2 2πz = 2π Section 44 Taylor s Theorem Theorem 9 (Thm 19, Taylor s Theorem in R 1 ) Let f : I R be n-times differentiable, where I R is an open interval If x, x + h I, then f(x + h) = f + n 1 k=1 f (k) h k k! + E n where f (k) is the k th derivative of f and E n = f(n) (x + λh)h n n! for some λ (0, 1) 5

Motivation: Let T n (h) = f + n f (k) h k k=1 k! = f + f h + f h 2 T n (0) = f 2 + + f(n) h n n! T n (h) = f + f h + + f(n) h n 1 T n(0) = f T n (h) = f + + f(n) h n 2 (n 2)! T n(0) = f T n (n) (0) = f(n) so T n (h) is the unique n th degree polynomial such that T n (0) = f T n(0) = f T n (n) (0) = f(n) (n 1)! The proof of the formula for the remainder E n is essentially the Mean Value Theorem; the problem in applying it is that the point x + λh is not known in advance Theorem 10 (Alternate Taylor s Theorem in R 1 ) Let f : I R be n times differentiable, where I R is an open interval and x I Then n f (k) h k f(x + h) = f + k! k=1 + o(h n ) as h 0 If f is (n + 1) times continuously differentiable (ie all derivatives up to order n + 1 exist and are continuous), then n f (k) h k f(x + h) = f + k! k=1 + O ( h n+1) as h 0 Remark: The first equation in the statement of the theorem is essentially a restatement of the definition of the n th derivative The second statement is proven from Theorem 19, and the continuity of the derivative, hence the boundedness of the derivative on a neighborhood of x 6

Definition 11 Let X R n be open A function f : X R m is continuously differentiable on X if f is differentiable on X and df x is a continuous function of x from X to L(R n,r m ), with operator norm df x f is C k if all partial derivatives of order less than or equal to k exist and are continuous in X Theorem 12 (Thm 43) Suppose X R n is open and f : X R m Then f is continuously differentiable on X if and only if f is C 1 Remark: The notation in Taylor s Theorem is difficult If f : R n R m, the quadratic terms are not hard for m = 1; for m > 1, we handle each component separately For cubic and higher order terms, the notation is a mess Linear Terms: Theorem 13 Suppose X R n is open and x X If f : X R m is differentiable, then f(x + h) = f + Dfh + o(h) as h 0 The previous theorem is essentially a restatement of the definition of differentiability Theorem 14 (Corollary of 44) Suppose X R n is open and x X If f : X R m is C 2, then f(x + h) = f + Dfh + O ( h 2) as h 0 Quadratic Terms: We treat each component of the function separately, so consider f : X R, X R n an open set Let 2 f x1 2 f 2 x 2 x 1 2 f x n x 1 D 2 2 f f = x 1 x 2 2 f x2 2 f 2 x n x 2 2 f x 1 x n 2 f xn 2 f C 2 2 f x i x j = 2 f x j x i D 2 f is symmetric D 2 f has an orthonormal basis of eigenvectors and thus can be diagonalized 7

Theorem 15 (Stronger Version of Thm 44) Let X R n be open, f : X R, f C 2 (X), and x X Then If f C 3, f(x + h) = f + Dfh + 1 2 h (D 2 f)h + o ( h 2) as h 0 f(x + h) = f + Dfh + 1 2 h (D 2 f)h + O ( h 3) as h 0 Remark: de la Fuente assumes X is convex X is said to be convex if, for every x, y X and α [0, 1], αx + (1 α)y X Notice we don t need this Since X is open, and B δ is convex x X δ > 0 st B δ X Definition 16 We say f has a saddle at x if Df = 0 but f has neither a local maximum nor a local minimum at x Corollary 17 Suppose X R n is open and x X If f : X R is C 2, then there is an orthonormal basis {v 1,, v n } and corresponding eigenvalues λ 1,, λ n R of D 2 f such that where γ i = h v i f(x + h) = f(x + γ 1 v 1 + + γ n v n ) n = f + (Dfv i )γ i + 1 2 i=1 1 If f C 3, we may strengthen o( γ 2 ) to O ( γ 3 ) 2 If f has a local maximum or local minimum at x, then 3 If Df = 0, then Df = 0 n λ i γi 2 + o ( γ 2) i=1 λ 1,, λ n > 0 λ 1,, λ n < 0 λ i < 0 for some i, λ j > 0 for some j λ 1,, λ n 0, λ i > 0 for some i λ 1,, λ n 0, λ i < 0 for some i λ 1 = = λ n = 0 f has a local minimum at x f has a local maximum at x f has a saddle at x f has a local minimum or a saddle at x f has a local maximum or a saddle at x gives no information 8

Proof: (Sketch) From our study of quadratic forms, we know the behavior of the quadratic terms is determined by the signs of the eigenvalues If λ i = 0 for some i, then we know that the quadratic form arising from the second partial derivatives is identically zero in the direction v i, and the higher derivatives will determine the behavior of the function f in the direction v i For example, if f = x 3, then f (0) = 0, f (0) = 0, but we know that f has a saddle at x = 0; however, if f = x 4, then again f (0) = 0 and f (0) = 0 but f has a local (and global) minimum at x = 0 9

Figure 1: The Mean Value Theorem 10