Tangent Planes, Linear Approximations and Differentiability

Similar documents
JUST THE MATHS UNIT NUMBER INTEGRATION APPLICATIONS 13 (Second moments of a volume (A)) A.J.Hobson

MTH Calculus with Analytic Geom I TEST 1

Introduction to Differentials

ENGI Partial Differentiation Page y f x

Jim Lambers MAT 169 Fall Semester Lecture 6 Notes. a n. n=1. S = lim s k = lim. n=1. n=1

Functions of Several Variables

Integration - Past Edexcel Exam Questions

Lecture 10. (2) Functions of two variables. Partial derivatives. Dan Nichols February 27, 2018

e x3 dx dy. 0 y x 2, 0 x 1.

Lecture 5 - Logarithms, Slope of a Function, Derivatives

JUST THE MATHS UNIT NUMBER PARTIAL DIFFERENTIATION 1 (Partial derivatives of the first order) A.J.Hobson

MIDTERM 2 REVIEW: ADDITIONAL PROBLEMS. 1 2 x + 1. y = + 1 = x 1/ = 1. y = 1 2 x 3/2 = 1. into this equation would have then given. y 1.

Sec. 14.3: Partial Derivatives. All of the following are ways of representing the derivative. y dx

Math 10C - Fall Final Exam

REVIEW OF DIFFERENTIAL CALCULUS

1 Lecture 24: Linearization

Jim Lambers MAT 169 Fall Semester Practice Final Exam

Differentiation - Quick Review From Calculus

Math Review Night: Work and the Dot Product

MAC 2311 Calculus I Spring 2004

4 Partial Differentiation

MAT137 Calculus! Lecture 6

Second Order ODEs. Second Order ODEs. In general second order ODEs contain terms involving y, dy But here only consider equations of the form

18.02 Multivariable Calculus Fall 2007

Functions of Several Variables

Note: Each problem is worth 14 points except numbers 5 and 6 which are 15 points. = 3 2

Solutions for the Practice Final - Math 23B, 2016

Math 233. Directional Derivatives and Gradients Basics

a k 0, then k + 1 = 2 lim 1 + 1

Calculus III. Math 233 Spring Final exam May 3rd. Suggested solutions

MULTIVARIABLE INTEGRATION

MAC2313 Final A. (5 pts) 1. How many of the following are necessarily true? i. The vector field F = 2x + 3y, 3x 5y is conservative.

Tangent Plane. Linear Approximation. The Gradient

Jim Lambers MAT 460 Fall Semester Lecture 2 Notes

Answer Key. Calculus I Math 141 Fall 2003 Professor Ben Richert. Exam 2

x+ y = 50 Dividing both sides by 2 : ( ) dx By (7.2), x = 25m gives maximum area. Substituting this value into (*):

Lecture 7 - Separable Equations

DIFFERENTIAL EQUATIONS

Vectors in Function Spaces

SOLUTIONS TO THE FINAL EXAM. December 14, 2010, 9:00am-12:00 (3 hours)

PARTIAL DIFFERENTIATION

Student Session Topic: Average and Instantaneous Rates of Change

Ma 1c Practical - Solutions to Homework Set 7

Electromagnetism HW 1 math review

MA 510 ASSIGNMENT SHEET Spring 2009 Text: Vector Calculus, J. Marsden and A. Tromba, fifth edition

Math 265H: Calculus III Practice Midterm II: Fall 2014

This exam will be over material covered in class from Monday 14 February through Tuesday 8 March, corresponding to sections in the text.

Change Of Variable Theorem: Multiple Dimensions

Multiple Integrals and Vector Calculus (Oxford Physics) Synopsis and Problem Sets; Hilary 2015

CALCULUS III THE CHAIN RULE, DIRECTIONAL DERIVATIVES, AND GRADIENT

MATH 2554 (Calculus I)

1. For each function, find all of its critical points and then classify each point as a local extremum or saddle point.

Spring 2015, MA 252, Calculus II, Final Exam Preview Solutions

MATH20411 PDEs and Vector Calculus B

2. Which of the following is an equation of the line tangent to the graph of f(x) = x 4 + 2x 2 at the point where

MATH 18.01, FALL PROBLEM SET # 6 SOLUTIONS

Practice Midterm 2 Math 2153

We have seen that for a function the partial derivatives whenever they exist, play an important role. This motivates the following definition.

Tangent Lines and Derivatives

= π + sin π = π + 0 = π, so the object is moving at a speed of π feet per second after π seconds. (c) How far does it go in π seconds?

Pullbacks, Isometries & Conformal Maps

g(t) = f(x 1 (t),..., x n (t)).

The Derivative of a Function Measuring Rates of Change of a function. Secant line. f(x) f(x 0 ) Average rate of change of with respect to over,

MA 123 September 8, 2016

The Derivative. Appendix B. B.1 The Derivative of f. Mappings from IR to IR

Math Exam III - Spring

2.12: Derivatives of Exp/Log (cont d) and 2.15: Antiderivatives and Initial Value Problems

Review for the Final Exam

AP Calculus BC Chapter 4 AP Exam Problems. Answers

Applications of the Maximum Principle

Directional Derivatives and the Gradient

Volume: The Disk Method. Using the integral to find volume.

Lecture XVII. Abstract We introduce the concept of directional derivative of a scalar function and discuss its relation with the gradient operator.

Properties of Linear Transformations from R n to R m

Math 142 (Summer 2018) Business Calculus 5.8 Notes

UNIT 3: DERIVATIVES STUDY GUIDE

Partial Derivatives October 2013

PRACTICE PROBLEMS FOR MIDTERM I

Calculus I: Practice Midterm II

Algebraic Properties of Solutions of Linear Systems

Calculus II - Fall 2013

February 4, 2019 LECTURE 4: THE DERIVATIVE.

2.2 The derivative as a Function

Math 180, Final Exam, Fall 2012 Problem 1 Solution

EXAM 3 MAT 167 Calculus I Spring is a composite function of two functions y = e u and u = 4 x + x 2. By the. dy dx = dy du = e u x + 2x.

DO NOT BEGIN THIS TEST UNTIL INSTRUCTED TO START

Spring 2015 Sample Final Exam

Mathematical Economics: Lecture 9

Transformations from R m to R n.

Calculus 2502A - Advanced Calculus I Fall : Local minima and maxima

Vector Calculus. A primer

MSc Mas6002, Introductory Material Mathematical Methods Exercises

Example: Limit definition. Geometric meaning. Geometric meaning, y. Notes. Notes. Notes. f (x, y) = x 2 y 3 :

Lecture 4 Backpropagation

Sometimes the domains X and Z will be the same, so this might be written:

Introduction and Vectors Lecture 1

MATH1013 Calculus I. Derivatives II (Chap. 3) 1

8-1: Backpropagation Prof. J.C. Kao, UCLA. Backpropagation. Chain rule for the derivatives Backpropagation graphs Examples

P1 Calculus II. Partial Differentiation & Multiple Integration. Prof David Murray. dwm/courses/1pd

Math Numerical Analysis

Transcription:

Jim Lambers MAT 80 Spring Semester 009-10 Lecture 5 Notes These notes correspond to Section 114 in Stewart and Section 3 in Marsden and Tromba Tangent Planes, Linear Approximations and Differentiability Now that we have learned how to compute partial derivatives of functions of several independent variables, in order to measure their instantaneous rates of change with respect to these variables, we will discuss another essential application of derivatives: the approximation of functions by linear functions Linear functions are the simplest to work with, and for this reason, there are many instances in which functions are replaced by a linear approximation in the context of solving a problem such as solving a differential equation Tangent Planes and Linear Approximations In single-variable calculus, we learned that the graph of a function f(x) can be approximated near a point x 0 by its tangent line, which has the equation y f(x 0 ) + f (x 0 )(x x 0 ) For this reason, the function L f (x) f(x 0 ) + f (x 0 )(x x 0 ) is also referred to as the linearization, or linear approximation, of f(x) at x 0 Now, suppose that we have a function of two variables, f : D R R, and a point (x 0, y 0 ) D Furthermore, suppose that the first partial derivatives of f, f x and f y, exist at (x 0, y 0 ) Because the graph of this function is a surface, it follows that a linear function that approximates f near (x 0, y 0 ) would have a graph that is a plane Just as the tangent line of f(x) at x 0 passes through the point (x 0, f(x 0 )), and has a slope that is equal to f (x 0 ), the instantaneous rate of change of f(x) with respect to x at x 0, a plane that best approximates f(x, y) at (x 0, y 0 ) must pass through the point (x 0, y 0, f(x 0, y 0 )), and the slope of the plane in the x- and y-directions, respectively, should be equal to the values of f x (x 0, y 0 ) and f y (x 0, y 0 ) Since a general linear function of two variables can be described by the formula L f (x, y) A(x x 0 ) + B(y y 0 ) + C, so that L f (x 0, y 0 ) C, and a simple differentiation yields L f x A, L f y B, 1

we conclude that the linear function that best approximates f(x, y) near (x 0, y 0 ) is the linear approximation L f (x, y) f(x 0, y 0 ) + x (x 0, y 0 )(x x 0 ) + y (x 0, y 0 )(y y 0 ) Furthermore, the graph of this function is called the tangent plane of f(x, y) at (x 0, y 0 ) Its equation is z z 0 x (x 0, y 0 )(x x 0 ) + y (x 0, y 0 )(y y 0 ) Example Let f(x, y) x y + 3y, and let (x 0, y 0 ) (1, 1) Then f(x 0, y 0 ) 5, and the first partial derivatives at (x 0, y 0 ) are f x (1, 1) 4xy x1,y1 4, f y (1, 1) x + 6y x1,y1 8 It follows that the tangent plane at (1, 1) has the equation and the linearization of f at (1, 1) is z 5 4(x 1) + 8(y 1), L f (x, y) 5 + 4(x 1) + 8(y 1) Let (x, y) (11, 11) Then f(x, y) 69, while L f (x, y) 6, for an error of 69 6 009 However, if (x, y) (101, 101), then f(x, y) 51090, while L f (x, y) 51, for an error of 51090 51 000090 That is, moving 10 times as close to (1, 1) decreased the error by a factor of over 100 Another useful application of a linear approximation is to estimate the error in the value of a function, given estimates of error in its inputs Given a function z f(x, y), and its linearization L f (x, y) around a point (x 0, y 0 ), if x 0 and y 0 are measured values and dx x x 0 and dz y y 0 are regarded as errors in x 0 and y 0, then the error in z can be estimated by computing dz z z 0 L f (x, y) f(x 0, y 0 ) [f(x 0, y 0 ) + f x (x 0, y 0 )(x x 0 ) + f y (x 0, y 0 )(y y 0 ) f(x 0, y 0 ) f x (x 0, y 0 ) dx + f y (x 0, y 0 ) dy The variables dx and dy are called differentials, and dz is called the total differential, as it depends on the values of dx and dy The total differential dz is only an estimate of the error in z; the actual error is given by Δz f(x, y) f(x 0, y 0 ), when the actual errors in x and y, Δx x x 0 and Δy y y 0, are known Since this is rarely the case in practice, one instead estimates the error in z from estimates dx and dy of the errors in x and y

Example Recall that the volume of a cylinder with radius r and height h is V πr h Suppose that r 5 cm and h 10 cm Then the volume is V 50π cm 3 If the measurement error in r and h is at most 01 cm, then, to estimate the error in the computed volume, we first compute V r πrh 100π, V h πr 5π It follows that the error in V is approximately dv V r dr + V h dh 01(100π + 5π) 15π cm 3 If we specify Δr 01 and Δh 01, and compute the actual volume using radius r + Δr 51 and height h + Δh 101, we obtain which yields the actual error V + ΔV π(51) (101) 6701π cm 3, ΔV 6701π 50π 1701π cm 3 Therefore, the estimate of the error, dv, is quite accurate Functions of More than Two Variables The concepts of a tangent plane and linear approximation generalize to more than two variables in a straightforward manner Specifically, given f : D R n R and p 0 (x (0) 1, x(0),, x(0) n ) D, we define the tangent space of f(x 1, x,, x n ) at p 0 to be the n-dimensional hyperplane in R n+1 whose points (x 1, x,, x n, y) satisfy the equation y y 0 (x 1 x (0) 1 ) + x (x x (0) ) + + (x n x (0) n ), where y 0 f Similarly, the linearization of f at p 0 is the function L f (x 1, x,, x n ) defined by L f (x 1, x,, x n ) y 0 + (x 1 x (0) 1 ) + x (x x (0) ) + + (x n x (0) n ) The Gradient Vector It can be seen from the above definitions that writing formulas that involve the partial derivatives of functions of n variables can be cumbersome This can be addressed by expressing collections of partial derivatives of functions of several variables using vectors and matrices, especially for vector-valued functions of several variables 3

By convention, a point p 0 (x (0) 1, x(0),, x(0) n ), which can be identified with the position vector p 0 x (0) 1, x(0),, x(0) n, is considered to be a column vector p 0 x (0) 1 x (0) x (0) n Also, by convention, given a function of n variables, f : D R n R, the collection of its partial derivatives with respect to all of its variables is written as a row vector [ f x This vector is called the gradient of f at p 0 Viewing the partial derivatives of f as a vector allows us to use vector operations to describe, much more concisely, the linearization of f Specifically, the linearization of f at p 0, evaluated at a point p (x 1, x,, x n ), can be written as L f (p) f + (x 1 x (0) 1 x (x x (0) 1 x (x n x (0) n f + (x i x (0) i ) x i i1 f + f (p p 0 ), where f (p p 0 ) is the dot product, also known as the inner product, of the vectors f and p p 0 Recall that given two vectors u u 1, u,, u n and v v 1, v,, v n, the dot product of u and v, denoted by u v, is defined by u v n u i v i u 1 v 1 + u v + + u n v n u v cos θ, i1 where θ is the angle between u and v Example Let f : R 3 R be defined by Then Let (x 0, y 0, z 0 ) (1,, 1) Then f(x, y, z) 3x y 3 z 4 f(x, y, z) [ f x f y f z [ 6xy 3 z 4 9x y z 4 1x y 3 z 3 f(x 0, y 0, z 0 ) f(1,, 1) [ f x (1,, 1) f y (1,, 1) f z (1,, 1) [ 48 36 96 n ) 4

It follows that the linearization of f at (x 0, y 0, z 0 ) is L f (x, y, z) f(1,, 1) + f(1,, 1) x 1, y, z + 1 4 + 48, 36, 96 x 1, y, z + 1 4 + 48(x 1) + 36(y ) 96(z + 1) 48x + 36y 96z 19 At the point (11, 19, 11), we have f(11, 19, 11) 365, while L f (11, 19, 11) 348 Because f is changing rapidly in all coordinate directions at (1,, 1), it is not surprising that the linearization of f at this point is not highly accurate The Jacobian Matrix Now, let f : D R n R m be a vector-valued function of n variables, with component functions f 1 (p) f (p) f(p), f m (p) where each f i : D R m Combining the two conventions described above, the partial derivatives of these component functions at a point p 0 D are arranged in an m n matrix 1 1 x 1 J f x m m x m This matrix is called the Jacobian matrix of f at p 0 It is also referred to as the derivative of f at x 0, since it reduces to the scalar f (x 0 ) when f is a scalar-valued function of one variable Note that rows of J f correspond to component functions, and columns correspond to independent variables This allows us to view J f as the following collections of rows or columns: J f f 1 f f m [ x The Jacobian matrix provides a concise way of describing the linearization of a vector-valued function, just the gradient does for a scalar-valued function The linearization of f at p 0 is the 5

function L f (p), defined by L f (p) f 1 f f m f + + n j1 1 (x 1 x (0) 1 ) + + m x j (x j x (0) j ) f + J f (p p 0 ), 1 (x n x (0) m n ) where the expression J f (p p 0 ) involves matrix multiplication of the matrix J f and the vector p p 0 Note the similarity between this definition, and the definition of the linearization of a function of a single variable In general, given a m n matrix A; that is, a matrix A with m rows and n columns, and an n p matrix B, the product AB is the m p matrix C, where the entry in row i and column j of C is obtained by computing the dot product of row i of A and column j of B When computing the linearization of a vector-valued function f at the point p 0 in its domain, the ith component function of the linearization is obtained by adding the value of the ith component function at p 0, f i, to the dot product of f i and the vector p p 0, where p is the vector at which the linearization is to be evaluated Example Let f : R R be defined by [ f1 (x, y) f(x, y) f (x, y) [ e x cos y e x sin y Then the Jacobian matrix, or derivative, of f is the matrix [ [ [ f1 (x, y) (f1 ) J f (x, y) x (f 1 ) y e x cos y f (x, y) (f ) x (f ) y e x sin y e x sin y e x cos y Let (x 0, y 0 ) (0, π/4) Then we have J f (x 0, y 0 ) [, and the linearization of f at (x 0, y 0 ) is [ f1 (x L f (x, y) 0, y 0 ) f (x 0, y 0 ) [ x x0 + J f (x 0, y 0 ) y y 0 6

[ [ + + [ x ( ) y π ( 4 ) y π x + 4 [ x 0 y π 4 At the point (x 1, y 1 ) (01, 08), we have [ 076998 f(x 1, y 1 ) 05873 [ 076749, L f (x 1, y 1 ) 057601 Because of the relatively small partial derivatives at (x 0, y 0 ), the linearization at this point yields a fairly accurate approximation at (x 1, y 1 ) Differentiability Before using a linearization to approximate a function near a point p 0, it is helpful to know whether this linearization is actually an accurate approximation of the function in the first place That is, we need to know if the function is differentiable at p 0, which, informally, means that its instantaneous rate of change at p 0 is well-defined In the single-variable case, a function f(x) is differentiable at x 0 if f (x 0 ) exists; that is, if the limit exists In other words, we must have f (x 0 ) lim x x 0 f(x) f(x 0 ) x x 0 f(x) f(x 0 ) f (x 0 )(x x 0 ) lim 0 x x 0 x x 0 But f(x 0 )+f (x 0 )(x x 0 ) is just the linearization of f at x 0, so we can say that f is differentiable at x 0 if and only if f(x) L f (x) lim 0 x x 0 x x 0 Note that this is a stronger statement than simply requiring that lim f(x) L f (x) 0, x x 0 because as x approaches x 0, 1/(x x 0 ) approaches, so the difference f(x) L f (x) must approach zero particularly rapidly in order for the fraction [f(x) L f (x)/(x x 0 ) to approach zero That is, the linearization must be a sufficiently accurate approximation of f near x 0 for this to be the case, in order for f to be differentiable at x 0 7

This notion of differentiability is readily generalized to functions of several variables Given f : D R n R m, and p 0 D, we say that f is differentiable at p 0 if where L f (p) is the linearization of f at p 0 f(p) L f (p) lim 0, p p 0 p p 0 Example Let f(x, y) x y To verify that this function is differentiable at (x 0, y 0 ) (1, 1), we first compute f x xy and f y x It follows that the linearization of f at (1, 1) is L f (x, y) f(1, 1) + f x (1, 1)(x 1) + f y (1, 1)(y 1) 1 + (x 1) + (y 1) x + y Therefore, f is differentiable at (1, 1) if x y (x + y ) x y (x + y ) lim lim (x,y) (1,1) (x, y) (1, 1) (x,y) (1,1) (x 1) + (y 1) 0 By rewriting this expression as and noting that x y (x + y ) x 1 y(x + 1) (x 1) + (y 1) (x 1) + (y 1), lim y(x + 1) 0, 0 x 1 (x,y) (1,1) (x 1) + (y 1) 1, we conclude that the limit actually is zero, and therefore f is differentiable There are three important conclusions that we can make regarding differentiable functions: If all partial derivatives of f at p 0 exist, and are continuous, then f is differentiable at p 0 Furthermore, if f is differentiable at p 0, then it is continuous at p 0 Note that the converse is not true; for example, f(x) x is continuous at x 0, but it is not differentiable there, because f (x) does not exist there If f is differentiable at p 0, then its first partial derivatives exist at p 0 This statement might seem redundant, because the first partial derivatives are used in the definition of the linearization, but it is important nonetheless, because the converse of this statement is not true That is, if a function s first partial derivatives exist at a point, it is not necessarily differentiable at that point The notion of differentiability is related to not only partial derivatives, which only describe how a function changes as one of its variables changes, but also the instantaneous rate of change of a function as its variables change along any direction If a function is differentiable at a point, that means its rate of change along any direction is well-defined We will explore this idea further in Lecture 7 8

Practice Problems 1 Compute the equation of the tangent plane of f(x, y) e x cos y + e x sin y at (x 0, y 0 ) ( 1, π/) Then, use the linearization of f at this point to approximate the value of f( 11, 16) How accurate is this approximation? Compute the equation of the tangent space of f(x, y, z) 1 ln (x 1) + (y + ) + z at (x 0, y 0, z 0 ) (1, 1, 1) Then, use the linearization of f at this point to approximate the value of f(101, 099, 105) How accurate is this approximation? 3 Let f(x, y) x + y Use the definition of differentiability to show that this function is differentiable at (x 0, y 0 ) (1, 1) 4 Suppose that the coordinates of two points (x 1, y 1 ) (, 3) and (x, y ) (7, 5) are obtained by measurements, for which the maximum error in each is 001 Estimate the maximum error in the distance between the two points Additional Practice Problems Additional practice problems from the recommended textbooks are: Stewart: Section 114, Exercises 1-5 odd, 15-7 odd Marsden/Tromba: Section 3, Exercises 5, 9, 15 9