Section 4.2. Types of Differentiation

Size: px
Start display at page:

Download "Section 4.2. Types of Differentiation"

Transcription

1 42 Types of Differentiation 1 Section 42 Types of Differentiation Note In this section we define differentiation of various structures with respect to a scalar, a vector, and a matrix Definition Let vector y be a function of scalar variable x so that y = y(x) = [y 1,y 2,,y n ] = [y 1 (x),y 2 (x),,y n (x)] The derivative of vector y with respect to scalar x is the vector y = [ y1, y 2,, y ] n = [y 1,y 2,,y n] Let matrix Y be a function of scalar variable x so that Y = Y (x) = [y ij (x)] The derivative of matrix Y with respect to scalar x is the matrix [ ] Y = yij = [y ij ] Note To define differentiation with respect to a vector x of an object Φ which is a function of a vector, we use the usual definition: Φ(x + ty) Φ(x) lim where y is any conformable vector with x (this will give the derivative of Φ in the direction of y) Notice that we can use the usual ε/δ definition of limit, provided we have a metric on the objects Φ We start with scalar valued functions

2 42 Types of Differentiation 2 Definition Let f : R n R be a scalar valued function of a vector The derivative of scalar valued function f with respect to vector x = [x 1,x 2,,x n ] is [ f f =, f,, f ] 1 2 n This derivative is the gradient of f, denoted f (sometimes read dell f ) Note Let A be a given (constant) n n matrix and x R n The quadratic form x T Ax is a scalar and the derivative of the quadratic form in the direction y is (x + ty) T A(x + ty) x T Ax lim (x T + ty T )A(x + ty) x T Ax = lim x T Ax + ty T Ax + tx T Ay + t 2 y T Ay x T Ax = lim = lim(y T Ax + x T AY + ty T Ay) = y T Ax + x T Ay t 0 = y T Ax + (x T Ay) T since x T Ay is a 1 1 matrix and so is symmetric = y T Ax + y T A T x = y T (A + A T )x Notice that x 2 2 = xt x = x T Ix so this example shows that the directional derivative of the Euclidean norm exists in all directions y Note We now consider derivatives of vector valued functions of vectors, f : R n R m

3 42 Types of Differentiation 3 Definition Let f : S R m where S R n With f = f(x 1,x 2,,x n ) = (f 1 (x 1,x 2,,x n ),f 2 (x 1,x 2,,x n ),,f m (x 1,x 2,,x n )), define the matrix gradient f/ as the n m matrix with (i,j) entry f i / j : f 1 f f = f 1 2 f 2 2 f 1 n f 2 n This is also sometimes denoted f (Gentle uses the notation f T /) The m n matrix f) T = ( f) T is the Jacobian of f (denoted by Gentle as f/ T ) f m 1 f m 2 f m n Note You have seen the Jacobian when dealing with substitution in a multiple integral (it sort of plays the role of du in one variable u-substitution) See my online Calculus 3 notes for Substitutions in Multiple Integrals : edu/gardnerr/2110/notes-12e/c15s8pdf Definition Let Y be a p q matrix of functions f ij : R n R, Y = [f ij ] where f ij = f ij (x 1,x 2,,x m ) Define the derivative of Y as Y = [ f ij] That is, Y/ is a three dimensional object (or a matrix of vectors) of dimension p q n where the (i,j,k) entry is f ij / k For fixed k we denote the p q matrix with (i,j) entry as the (i,j,k ) entry of Y/ as Y/ k

4 42 Types of Differentiation 4 Note The linearity and chain rule properties of partial derivatives allow us to establish rules of differentiation in these new settings For example, if Y = [f ij ] where f ij : R n R is a square nonsingular matrix then we show in Exercise 43 that Y 1 = Y 1 ( ) Y Y 1 In Exercise 42A we show for Y = [f ij ] that [tr(y )] = tr ( ) Y Definition Let f : R n R be a scalar valued function of a vector The matrix H with (i,j) entry as is the Hessian of f, denoted H: i j H = n 1 Gentle denotes the Hessian as 2 f and x 2 n 2 2 f T 1 n 2 n 2 n Note In the case that f = f(x,y) then det(h) = det 2 y y y 2 = 2 f 2 y 2 ( 2 ) 2 f y (assuming the mixed partials are equals, which is really an assumption of continuity) and det(h) is related to the curvature of the surface z = f(x, y); see The Gauss Curvature in Detail from my Differential Geometry notes: etsuedu/gardnerr/5310/5310pdf/dg1-6pdf (see the example on pages 3 and

5 42 Types of Differentiation 5 4) Because of this, the Hessian (which is often defined as det(h) instead of H itself) can be used to find extrema and saddle points of z = f(x,y) in a style similar to the use of the Second Derivative Test for functions y = f(x) See Extreme Values and Saddle Points from Calculus 3: notes-12e/c14s7pdf Note Gentle lists f/ for several types of functions f(x) where x is an n- vector Let a be a constant scalar, b a constant conformable vector, and A a constant conformable matrix Then we have (with corrections to Gentle as given in the Errata): f(x) ax b T x x T b x T x xx T b T Ax x T Ab f/ ai b b 2x x I + I x A T b Ab x T Ax (A + A T )x exp(( 1/2)x T Ax) exp(( 1/2)x T Ax)Ax if A = A T x 2 2 2x V (x) 2x/(n 1) The proofs of some of these are to be given in Exercise 42B

6 42 Types of Differentiation 6 Definition Let X be a matrix of independent variables and f a function of X so that f = f(x) = f([x ij ]) The derivative of function f with respect to matrix X is the matrix f [ ] f = ij Theorem 421 Differentiation of scalar valued function f satisfies the following (1) f T = ( f ) T (2) For X square and f(x) = tr(x), f = I (3) For AX a square matrix where A is constant, [tr(ax)] (4) [tr(xt X)] = 2X (5) With a and b constant vectors, [at Xb] = abt (6) [det(x)] = (adj(x)) T = A T Note A quick search of the internet reveals that there are a number of definitions for the derivative of a matrix with respect to a matrix Inspired by Gentle s general approach, we take the following definition Definition Let X be a matrix of independent variables and let Y be a matrix of functions of X, Y = [y ij ] = [y ij (X)] The derivative of matrix Y with respect to matrix X is the matrix Y [ ] Y = ij

7 42 Types of Differentiation 7 Note Gentle claims that that the following hold Given his cryptic statement of equation (415) on page 155, I am not totally confident that all of the following hold under our definition of a matrix with respect to a matrix Note Gentle lists f/ for several types of functions f(x) where X is a matrix Let a and b be a conformable vectors, and A a constant conformable matrix Then we have: f(x) f/ Justification a T Xb ab T Theorem 421(5) tr(ax) A T Theorem 421(3) tr(x T X) 2X Theorem 421(4) BX XC BXC I B C T I C T B If X is square and invertible as required, then: f(x) f/ Justification tr(x) I Theorem 421(2) tr(x k ) kx k 1 Exercise 42C(1) tr(bx 1 C) (X 1 CBX 1 ) T det(x) detx(x 1 ) T Exercise 42C(2) log(det(x)) (X 1 ) T Exercise 42C(3) detx k kdet(x) k (X 1 ) T Exercise 42C(4) BX 1 C (X 1 C) T BX 1 Revised: 4/1/2018

Multivariable Calculus

Multivariable Calculus 2 Multivariable Calculus 2.1 Limits and Continuity Problem 2.1.1 (Fa94) Let the function f : R n R n satisfy the following two conditions: (i) f (K ) is compact whenever K is a compact subset of R n. (ii)

More information

OR MSc Maths Revision Course

OR MSc Maths Revision Course OR MSc Maths Revision Course Tom Byrne School of Mathematics University of Edinburgh t.m.byrne@sms.ed.ac.uk 15 September 2017 General Information Today JCMB Lecture Theatre A, 09:30-12:30 Mathematics revision

More information

Section 4.5. Integration and Expectation

Section 4.5. Integration and Expectation 4.5. Integration and Expectation 1 Section 4.5. Integration and Expectation Note. In this section we consider integrals of scalar valued functions of a vector and matrix valued functions of a scalar. We

More information

MAT 419 Lecture Notes Transcribed by Eowyn Cenek 6/1/2012

MAT 419 Lecture Notes Transcribed by Eowyn Cenek 6/1/2012 (Homework 1: Chapter 1: Exercises 1-7, 9, 11, 19, due Monday June 11th See also the course website for lectures, assignments, etc) Note: today s lecture is primarily about definitions Lots of definitions

More information

The Derivative. Appendix B. B.1 The Derivative of f. Mappings from IR to IR

The Derivative. Appendix B. B.1 The Derivative of f. Mappings from IR to IR Appendix B The Derivative B.1 The Derivative of f In this chapter, we give a short summary of the derivative. Specifically, we want to compare/contrast how the derivative appears for functions whose domain

More information

Optimization and Calculus

Optimization and Calculus Optimization and Calculus To begin, there is a close relationship between finding the roots to a function and optimizing a function. In the former case, we solve for x. In the latter, we solve: g(x) =

More information

REVIEW OF DIFFERENTIAL CALCULUS

REVIEW OF DIFFERENTIAL CALCULUS REVIEW OF DIFFERENTIAL CALCULUS DONU ARAPURA 1. Limits and continuity To simplify the statements, we will often stick to two variables, but everything holds with any number of variables. Let f(x, y) be

More information

Calculus 2502A - Advanced Calculus I Fall : Local minima and maxima

Calculus 2502A - Advanced Calculus I Fall : Local minima and maxima Calculus 50A - Advanced Calculus I Fall 014 14.7: Local minima and maxima Martin Frankland November 17, 014 In these notes, we discuss the problem of finding the local minima and maxima of a function.

More information

D(f/g)(P ) = D(f)(P )g(p ) f(p )D(g)(P ). g 2 (P )

D(f/g)(P ) = D(f)(P )g(p ) f(p )D(g)(P ). g 2 (P ) We first record a very useful: 11. Higher derivatives Theorem 11.1. Let A R n be an open subset. Let f : A R m and g : A R m be two functions and suppose that P A. Let λ A be a scalar. If f and g are differentiable

More information

. D Matrix Calculus D 1

. D Matrix Calculus D 1 D Matrix Calculus D 1 Appendix D: MATRIX CALCULUS D 2 In this Appendix we collect some useful formulas of matrix calculus that often appear in finite element derivations D1 THE DERIVATIVES OF VECTOR FUNCTIONS

More information

Math 600 Day 1: Review of advanced Calculus

Math 600 Day 1: Review of advanced Calculus Math 600 Day 1: Review of advanced Calculus Ryan Blair University of Pennsylvania Thursday September 8, 2010 Ryan Blair (U Penn) Math 600 Day 1: Review of advanced Calculus Thursday September 8, 2010 1

More information

446 CHAP. 8 NUMERICAL OPTIMIZATION. Newton's Search for a Minimum of f(x,y) Newton s Method

446 CHAP. 8 NUMERICAL OPTIMIZATION. Newton's Search for a Minimum of f(x,y) Newton s Method 446 CHAP. 8 NUMERICAL OPTIMIZATION Newton's Search for a Minimum of f(xy) Newton s Method The quadratic approximation method of Section 8.1 generated a sequence of seconddegree Lagrange polynomials. It

More information

MA102: Multivariable Calculus

MA102: Multivariable Calculus MA102: Multivariable Calculus Rupam Barman and Shreemayee Bora Department of Mathematics IIT Guwahati Differentiability of f : U R n R m Definition: Let U R n be open. Then f : U R n R m is differentiable

More information

Notes on Random Vectors and Multivariate Normal

Notes on Random Vectors and Multivariate Normal MATH 590 Spring 06 Notes on Random Vectors and Multivariate Normal Properties of Random Vectors If X,, X n are random variables, then X = X,, X n ) is a random vector, with the cumulative distribution

More information

Variational Methods & Optimal Control

Variational Methods & Optimal Control Variational Methods & Optimal Control lecture 02 Matthew Roughan Discipline of Applied Mathematics School of Mathematical Sciences University of Adelaide April 14, 2016

More information

6.1 Matrices. Definition: A Matrix A is a rectangular array of the form. A 11 A 12 A 1n A 21. A 2n. A m1 A m2 A mn A 22.

6.1 Matrices. Definition: A Matrix A is a rectangular array of the form. A 11 A 12 A 1n A 21. A 2n. A m1 A m2 A mn A 22. 61 Matrices Definition: A Matrix A is a rectangular array of the form A 11 A 12 A 1n A 21 A 22 A 2n A m1 A m2 A mn The size of A is m n, where m is the number of rows and n is the number of columns The

More information

Several variables. x 1 x 2. x n

Several variables. x 1 x 2. x n Several variables Often we have not only one, but several variables in a problem The issues that come up are somewhat more complex than for one variable Let us first start with vector spaces and linear

More information

Functions of Several Variables

Functions of Several Variables Jim Lambers MAT 419/519 Summer Session 2011-12 Lecture 2 Notes These notes correspond to Section 1.2 in the text. Functions of Several Variables We now generalize the results from the previous section,

More information

Lecture 2: Review of Prerequisites. Table of contents

Lecture 2: Review of Prerequisites. Table of contents Math 348 Fall 217 Lecture 2: Review of Prerequisites Disclaimer. As we have a textbook, this lecture note is for guidance and supplement only. It should not be relied on when preparing for exams. In this

More information

8. Conjugate functions

8. Conjugate functions L. Vandenberghe EE236C (Spring 2013-14) 8. Conjugate functions closed functions conjugate function 8-1 Closed set a set C is closed if it contains its boundary: x k C, x k x = x C operations that preserve

More information

g(t) = f(x 1 (t),..., x n (t)).

g(t) = f(x 1 (t),..., x n (t)). Reading: [Simon] p. 313-333, 833-836. 0.1 The Chain Rule Partial derivatives describe how a function changes in directions parallel to the coordinate axes. Now we shall demonstrate how the partial derivatives

More information

Extrema of Functions of Several Variables

Extrema of Functions of Several Variables Extrema of Functions of Several Variables MATH 311, Calculus III J. Robert Buchanan Department of Mathematics Fall 2011 Background (1 of 3) In single-variable calculus there are three important results

More information

Second Order ODEs. Second Order ODEs. In general second order ODEs contain terms involving y, dy But here only consider equations of the form

Second Order ODEs. Second Order ODEs. In general second order ODEs contain terms involving y, dy But here only consider equations of the form Second Order ODEs Second Order ODEs In general second order ODEs contain terms involving y, dy But here only consider equations of the form A d2 y dx 2 + B dy dx + Cy = 0 dx, d2 y dx 2 and F(x). where

More information

FALL 2018 MATH 4211/6211 Optimization Homework 1

FALL 2018 MATH 4211/6211 Optimization Homework 1 FALL 2018 MATH 4211/6211 Optimization Homework 1 This homework assignment is open to textbook, reference books, slides, and online resources, excluding any direct solution to the problem (such as solution

More information

HW2 Solutions

HW2 Solutions 8.024 HW2 Solutions February 4, 200 Apostol.3 4,7,8 4. Show that for all x and y in real Euclidean space: (x, y) 0 x + y 2 x 2 + y 2 Proof. By definition of the norm and the linearity of the inner product,

More information

minimize x x2 2 x 1x 2 x 1 subject to x 1 +2x 2 u 1 x 1 4x 2 u 2, 5x 1 +76x 2 1,

minimize x x2 2 x 1x 2 x 1 subject to x 1 +2x 2 u 1 x 1 4x 2 u 2, 5x 1 +76x 2 1, 4 Duality 4.1 Numerical perturbation analysis example. Consider the quadratic program with variables x 1, x 2, and parameters u 1, u 2. minimize x 2 1 +2x2 2 x 1x 2 x 1 subject to x 1 +2x 2 u 1 x 1 4x

More information

Jim Lambers MAT 419/519 Summer Session Lecture 13 Notes

Jim Lambers MAT 419/519 Summer Session Lecture 13 Notes Jim Lambers MAT 419/519 Summer Session 2011-12 Lecture 13 Notes These notes correspond to Section 4.1 in the text. Least Squares Fit One of the most fundamental problems in science and engineering is data

More information

Basics of Calculus and Algebra

Basics of Calculus and Algebra Monika Department of Economics ISCTE-IUL September 2012 Basics of linear algebra Real valued Functions Differential Calculus Integral Calculus Optimization Introduction I A matrix is a rectangular array

More information

Chapter 2. Vectors and Vector Spaces

Chapter 2. Vectors and Vector Spaces 2.1. Operations on Vectors 1 Chapter 2. Vectors and Vector Spaces Section 2.1. Operations on Vectors Note. In this section, we define several arithmetic operations on vectors (especially, vector addition

More information

Machine Learning Brett Bernstein. Recitation 1: Gradients and Directional Derivatives

Machine Learning Brett Bernstein. Recitation 1: Gradients and Directional Derivatives Machine Learning Brett Bernstein Recitation 1: Gradients and Directional Derivatives Intro Question 1 We are given the data set (x 1, y 1 ),, (x n, y n ) where x i R d and y i R We want to fit a linear

More information

Exercises for Multivariable Differential Calculus XM521

Exercises for Multivariable Differential Calculus XM521 This document lists all the exercises for XM521. The Type I (True/False) exercises will be given, and should be answered, online immediately following each lecture. The Type III exercises are to be done

More information

Stat 206: Linear algebra

Stat 206: Linear algebra Stat 206: Linear algebra James Johndrow (adapted from Iain Johnstone s notes) 2016-11-02 Vectors We have already been working with vectors, but let s review a few more concepts. The inner product of two

More information

A function(al) f is convex if dom f is a convex set, and. f(θx + (1 θ)y) < θf(x) + (1 θ)f(y) f(x) = x 3

A function(al) f is convex if dom f is a convex set, and. f(θx + (1 θ)y) < θf(x) + (1 θ)f(y) f(x) = x 3 Convex functions The domain dom f of a functional f : R N R is the subset of R N where f is well-defined. A function(al) f is convex if dom f is a convex set, and f(θx + (1 θ)y) θf(x) + (1 θ)f(y) for all

More information

1 Overview. 2 A Characterization of Convex Functions. 2.1 First-order Taylor approximation. AM 221: Advanced Optimization Spring 2016

1 Overview. 2 A Characterization of Convex Functions. 2.1 First-order Taylor approximation. AM 221: Advanced Optimization Spring 2016 AM 221: Advanced Optimization Spring 2016 Prof. Yaron Singer Lecture 8 February 22nd 1 Overview In the previous lecture we saw characterizations of optimality in linear optimization, and we reviewed the

More information

Calculus for the Life Sciences II Assignment 6 solutions. f(x, y) = 3π 3 cos 2x + 2 sin 3y

Calculus for the Life Sciences II Assignment 6 solutions. f(x, y) = 3π 3 cos 2x + 2 sin 3y Calculus for the Life Sciences II Assignment 6 solutions Find the tangent plane to the graph of the function at the point (0, π f(x, y = 3π 3 cos 2x + 2 sin 3y Solution: The tangent plane of f at a point

More information

Linear Algebra, part 2 Eigenvalues, eigenvectors and least squares solutions

Linear Algebra, part 2 Eigenvalues, eigenvectors and least squares solutions Linear Algebra, part 2 Eigenvalues, eigenvectors and least squares solutions Anna-Karin Tornberg Mathematical Models, Analysis and Simulation Fall semester, 2013 Main problem of linear algebra 2: Given

More information

Course Summary Math 211

Course Summary Math 211 Course Summary Math 211 table of contents I. Functions of several variables. II. R n. III. Derivatives. IV. Taylor s Theorem. V. Differential Geometry. VI. Applications. 1. Best affine approximations.

More information

Section 3.3. Matrix Rank and the Inverse of a Full Rank Matrix

Section 3.3. Matrix Rank and the Inverse of a Full Rank Matrix 3.3. Matrix Rank and the Inverse of a Full Rank Matrix 1 Section 3.3. Matrix Rank and the Inverse of a Full Rank Matrix Note. The lengthy section (21 pages in the text) gives a thorough study of the rank

More information

CLOSED RANGE POSITIVE OPERATORS ON BANACH SPACES

CLOSED RANGE POSITIVE OPERATORS ON BANACH SPACES Acta Math. Hungar., 142 (2) (2014), 494 501 DOI: 10.1007/s10474-013-0380-2 First published online December 11, 2013 CLOSED RANGE POSITIVE OPERATORS ON BANACH SPACES ZS. TARCSAY Department of Applied Analysis,

More information

Implicit Functions, Curves and Surfaces

Implicit Functions, Curves and Surfaces Chapter 11 Implicit Functions, Curves and Surfaces 11.1 Implicit Function Theorem Motivation. In many problems, objects or quantities of interest can only be described indirectly or implicitly. It is then

More information

MTH4101 CALCULUS II REVISION NOTES. 1. COMPLEX NUMBERS (Thomas Appendix 7 + lecture notes) ax 2 + bx + c = 0. x = b ± b 2 4ac 2a. i = 1.

MTH4101 CALCULUS II REVISION NOTES. 1. COMPLEX NUMBERS (Thomas Appendix 7 + lecture notes) ax 2 + bx + c = 0. x = b ± b 2 4ac 2a. i = 1. MTH4101 CALCULUS II REVISION NOTES 1. COMPLEX NUMBERS (Thomas Appendix 7 + lecture notes) 1.1 Introduction Types of numbers (natural, integers, rationals, reals) The need to solve quadratic equations:

More information

Computational Optimization. Mathematical Programming Fundamentals 1/25 (revised)

Computational Optimization. Mathematical Programming Fundamentals 1/25 (revised) Computational Optimization Mathematical Programming Fundamentals 1/5 (revised) If you don t know where you are going, you probably won t get there. -from some book I read in eight grade If you do get there,

More information

2 Functions of random variables

2 Functions of random variables 2 Functions of random variables A basic statistical model for sample data is a collection of random variables X 1,..., X n. The data are summarised in terms of certain sample statistics, calculated as

More information

CHAPTER 7. An Introduction to Numerical Methods for. Linear and Nonlinear ODE s

CHAPTER 7. An Introduction to Numerical Methods for. Linear and Nonlinear ODE s A SERIES OF CLASS NOTES FOR 2005-2006 TO INTRODUCE LINEAR AND NONLINEAR PROBLEMS TO ENGINEERS, SCIENTISTS, AND APPLIED MATHEMATICIANS DE CLASS NOTES 1 A COLLECTION OF HANDOUTS ON FIRST ORDER ORDINARY DIFFERENTIAL

More information

Matrix Algebra, Class Notes (part 2) by Hrishikesh D. Vinod Copyright 1998 by Prof. H. D. Vinod, Fordham University, New York. All rights reserved.

Matrix Algebra, Class Notes (part 2) by Hrishikesh D. Vinod Copyright 1998 by Prof. H. D. Vinod, Fordham University, New York. All rights reserved. Matrix Algebra, Class Notes (part 2) by Hrishikesh D. Vinod Copyright 1998 by Prof. H. D. Vinod, Fordham University, New York. All rights reserved. 1 Converting Matrices Into (Long) Vectors Convention:

More information

Convex Optimization Theory. Chapter 5 Exercises and Solutions: Extended Version

Convex Optimization Theory. Chapter 5 Exercises and Solutions: Extended Version Convex Optimization Theory Chapter 5 Exercises and Solutions: Extended Version Dimitri P. Bertsekas Massachusetts Institute of Technology Athena Scientific, Belmont, Massachusetts http://www.athenasc.com

More information

3.5 Quadratic Approximation and Convexity/Concavity

3.5 Quadratic Approximation and Convexity/Concavity 3.5 Quadratic Approximation and Convexity/Concavity 55 3.5 Quadratic Approximation and Convexity/Concavity Overview: Second derivatives are useful for understanding how the linear approximation varies

More information

3. Convex functions. basic properties and examples. operations that preserve convexity. the conjugate function. quasiconvex functions

3. Convex functions. basic properties and examples. operations that preserve convexity. the conjugate function. quasiconvex functions 3. Convex functions Convex Optimization Boyd & Vandenberghe basic properties and examples operations that preserve convexity the conjugate function quasiconvex functions log-concave and log-convex functions

More information

Inequality Constraints

Inequality Constraints Chapter 2 Inequality Constraints 2.1 Optimality Conditions Early in multivariate calculus we learn the significance of differentiability in finding minimizers. In this section we begin our study of the

More information

3. Convex functions. basic properties and examples. operations that preserve convexity. the conjugate function. quasiconvex functions

3. Convex functions. basic properties and examples. operations that preserve convexity. the conjugate function. quasiconvex functions 3. Convex functions Convex Optimization Boyd & Vandenberghe basic properties and examples operations that preserve convexity the conjugate function quasiconvex functions log-concave and log-convex functions

More information

Complex Variables. Chapter 2. Analytic Functions Section Harmonic Functions Proofs of Theorems. March 19, 2017

Complex Variables. Chapter 2. Analytic Functions Section Harmonic Functions Proofs of Theorems. March 19, 2017 Complex Variables Chapter 2. Analytic Functions Section 2.26. Harmonic Functions Proofs of Theorems March 19, 2017 () Complex Variables March 19, 2017 1 / 5 Table of contents 1 Theorem 2.26.1. 2 Theorem

More information

Partial Fractions. Calculus 2 Lia Vas

Partial Fractions. Calculus 2 Lia Vas Calculus Lia Vas Partial Fractions rational function is a quotient of two polynomial functions The method of partial fractions is a general method for evaluating integrals of rational function The idea

More information

The Contraction Mapping Theorem and the Implicit and Inverse Function Theorems

The Contraction Mapping Theorem and the Implicit and Inverse Function Theorems The Contraction Mapping Theorem and the Implicit and Inverse Function Theorems The Contraction Mapping Theorem Theorem The Contraction Mapping Theorem) Let B a = { x IR d x < a } denote the open ball of

More information

Tangent spaces, normals and extrema

Tangent spaces, normals and extrema Chapter 3 Tangent spaces, normals and extrema If S is a surface in 3-space, with a point a S where S looks smooth, i.e., without any fold or cusp or self-crossing, we can intuitively define the tangent

More information

MATH529 Fundamentals of Optimization Unconstrained Optimization II

MATH529 Fundamentals of Optimization Unconstrained Optimization II MATH529 Fundamentals of Optimization Unconstrained Optimization II Marco A. Montes de Oca Mathematical Sciences, University of Delaware, USA 1 / 31 Recap 2 / 31 Example Find the local and global minimizers

More information

Math 5630: Iterative Methods for Systems of Equations Hung Phan, UMass Lowell March 22, 2018

Math 5630: Iterative Methods for Systems of Equations Hung Phan, UMass Lowell March 22, 2018 1 Linear Systems Math 5630: Iterative Methods for Systems of Equations Hung Phan, UMass Lowell March, 018 Consider the system 4x y + z = 7 4x 8y + z = 1 x + y + 5z = 15. We then obtain x = 1 4 (7 + y z)

More information

The Contraction Mapping Principle

The Contraction Mapping Principle Chapter 3 The Contraction Mapping Principle The notion of a complete space is introduced in Section 1 where it is shown that every metric space can be enlarged to a complete one without further condition.

More information

2. Linear algebra. matrices and vectors. linear equations. range and nullspace of matrices. function of vectors, gradient and Hessian

2. Linear algebra. matrices and vectors. linear equations. range and nullspace of matrices. function of vectors, gradient and Hessian FE661 - Statistical Methods for Financial Engineering 2. Linear algebra Jitkomut Songsiri matrices and vectors linear equations range and nullspace of matrices function of vectors, gradient and Hessian

More information

The Contraction Mapping Theorem and the Implicit Function Theorem

The Contraction Mapping Theorem and the Implicit Function Theorem The Contraction Mapping Theorem and the Implicit Function Theorem Theorem The Contraction Mapping Theorem Let B a = { x IR d x < a } denote the open ball of radius a centred on the origin in IR d. If the

More information

Mathematical modelling Chapter 2 Nonlinear models and geometric models

Mathematical modelling Chapter 2 Nonlinear models and geometric models Mathematical modelling Chapter 2 Nonlinear models and geometric models Faculty of Computer and Information Science University of Ljubljana 2018/2019 3. Nonlinear models Given is a sample of points {(x

More information

LECTURE NOTES ELEMENTARY NUMERICAL METHODS. Eusebius Doedel

LECTURE NOTES ELEMENTARY NUMERICAL METHODS. Eusebius Doedel LECTURE NOTES on ELEMENTARY NUMERICAL METHODS Eusebius Doedel TABLE OF CONTENTS Vector and Matrix Norms 1 Banach Lemma 20 The Numerical Solution of Linear Systems 25 Gauss Elimination 25 Operation Count

More information

11 a 12 a 21 a 11 a 22 a 12 a 21. (C.11) A = The determinant of a product of two matrices is given by AB = A B 1 1 = (C.13) and similarly.

11 a 12 a 21 a 11 a 22 a 12 a 21. (C.11) A = The determinant of a product of two matrices is given by AB = A B 1 1 = (C.13) and similarly. C PROPERTIES OF MATRICES 697 to whether the permutation i 1 i 2 i N is even or odd, respectively Note that I =1 Thus, for a 2 2 matrix, the determinant takes the form A = a 11 a 12 = a a 21 a 11 a 22 a

More information

Matrix Theory. A.Holst, V.Ufnarovski

Matrix Theory. A.Holst, V.Ufnarovski Matrix Theory AHolst, VUfnarovski 55 HINTS AND ANSWERS 9 55 Hints and answers There are two different approaches In the first one write A as a block of rows and note that in B = E ij A all rows different

More information

MATH 5720: Unconstrained Optimization Hung Phan, UMass Lowell September 13, 2018

MATH 5720: Unconstrained Optimization Hung Phan, UMass Lowell September 13, 2018 MATH 57: Unconstrained Optimization Hung Phan, UMass Lowell September 13, 18 1 Global and Local Optima Let a function f : S R be defined on a set S R n Definition 1 (minimizers and maximizers) (i) x S

More information

Math 207 Honors Calculus III Final Exam Solutions

Math 207 Honors Calculus III Final Exam Solutions Math 207 Honors Calculus III Final Exam Solutions PART I. Problem 1. A particle moves in the 3-dimensional space so that its velocity v(t) and acceleration a(t) satisfy v(0) = 3j and v(t) a(t) = t 3 for

More information

Tangent Planes, Linear Approximations and Differentiability

Tangent Planes, Linear Approximations and Differentiability Jim Lambers MAT 80 Spring Semester 009-10 Lecture 5 Notes These notes correspond to Section 114 in Stewart and Section 3 in Marsden and Tromba Tangent Planes, Linear Approximations and Differentiability

More information

NOTES ON MULTIVARIABLE CALCULUS: DIFFERENTIAL CALCULUS

NOTES ON MULTIVARIABLE CALCULUS: DIFFERENTIAL CALCULUS NOTES ON MULTIVARIABLE CALCULUS: DIFFERENTIAL CALCULUS SAMEER CHAVAN Abstract. This is the first part of Notes on Multivariable Calculus based on the classical texts [6] and [5]. We present here the geometric

More information

Workshop I The R n Vector Space. Linear Combinations

Workshop I The R n Vector Space. Linear Combinations Workshop I Workshop I The R n Vector Space. Linear Combinations Exercise 1. Given the vectors u = (1, 0, 3), v = ( 2, 1, 2), and w = (1, 2, 4) compute a) u + ( v + w) b) 2 u 3 v c) 2 v u + w Exercise 2.

More information

CSCI5654 (Linear Programming, Fall 2013) Lectures Lectures 10,11 Slide# 1

CSCI5654 (Linear Programming, Fall 2013) Lectures Lectures 10,11 Slide# 1 CSCI5654 (Linear Programming, Fall 2013) Lectures 10-12 Lectures 10,11 Slide# 1 Today s Lecture 1. Introduction to norms: L 1,L 2,L. 2. Casting absolute value and max operators. 3. Norm minimization problems.

More information

Many of you got these steps reversed or otherwise out of order.

Many of you got these steps reversed or otherwise out of order. Problem 1. Let (X, d X ) and (Y, d Y ) be metric spaces. Suppose that there is a bijection f : X Y such that for all x 1, x 2 X. 1 10 d X(x 1, x 2 ) d Y (f(x 1 ), f(x 2 )) 10d X (x 1, x 2 ) Show that if

More information

Chapter 7. Extremal Problems. 7.1 Extrema and Local Extrema

Chapter 7. Extremal Problems. 7.1 Extrema and Local Extrema Chapter 7 Extremal Problems No matter in theoretical context or in applications many problems can be formulated as problems of finding the maximum or minimum of a function. Whenever this is the case, advanced

More information

Appendix 5. Derivatives of Vectors and Vector-Valued Functions. Draft Version 24 November 2008

Appendix 5. Derivatives of Vectors and Vector-Valued Functions. Draft Version 24 November 2008 Appendix 5 Derivatives of Vectors and Vector-Valued Functions Draft Version 24 November 2008 Quantitative genetics often deals with vector-valued functions and here we provide a brief review of the calculus

More information

Section 21. The Metric Topology (Continued)

Section 21. The Metric Topology (Continued) 21. The Metric Topology (cont.) 1 Section 21. The Metric Topology (Continued) Note. In this section we give a number of results for metric spaces which are familar from calculus and real analysis. We also

More information

Electromagnetism HW 1 math review

Electromagnetism HW 1 math review Electromagnetism HW math review Problems -5 due Mon 7th Sep, 6- due Mon 4th Sep Exercise. The Levi-Civita symbol, ɛ ijk, also known as the completely antisymmetric rank-3 tensor, has the following properties:

More information

Higher order derivative

Higher order derivative 2 Î 3 á Higher order derivative Î 1 Å Iterated partial derivative Iterated partial derivative Suppose f has f/ x, f/ y and ( f ) = 2 f x x x 2 ( f ) = 2 f x y x y ( f ) = 2 f y x y x ( f ) = 2 f y y y

More information

1. Background: The SVD and the best basis (questions selected from Ch. 6- Can you fill in the exercises?)

1. Background: The SVD and the best basis (questions selected from Ch. 6- Can you fill in the exercises?) Math 35 Exam Review SOLUTIONS Overview In this third of the course we focused on linear learning algorithms to model data. summarize: To. Background: The SVD and the best basis (questions selected from

More information

Recitation 1. Gradients and Directional Derivatives. Brett Bernstein. CDS at NYU. January 21, 2018

Recitation 1. Gradients and Directional Derivatives. Brett Bernstein. CDS at NYU. January 21, 2018 Gradients and Directional Derivatives Brett Bernstein CDS at NYU January 21, 2018 Brett Bernstein (CDS at NYU) Recitation 1 January 21, 2018 1 / 23 Initial Question Intro Question Question We are given

More information

Multiple Integrals and Vector Calculus (Oxford Physics) Synopsis and Problem Sets; Hilary 2015

Multiple Integrals and Vector Calculus (Oxford Physics) Synopsis and Problem Sets; Hilary 2015 Multiple Integrals and Vector Calculus (Oxford Physics) Ramin Golestanian Synopsis and Problem Sets; Hilary 215 The outline of the material, which will be covered in 14 lectures, is as follows: 1. Introduction

More information

Nonlinear Optimization

Nonlinear Optimization Nonlinear Optimization (Com S 477/577 Notes) Yan-Bin Jia Nov 7, 2017 1 Introduction Given a single function f that depends on one or more independent variable, we want to find the values of those variables

More information

MATH H53 : Final exam

MATH H53 : Final exam MATH H53 : Final exam 11 May, 18 Name: You have 18 minutes to answer the questions. Use of calculators or any electronic items is not permitted. Answer the questions in the space provided. If you run out

More information

Solutions to Homework 7

Solutions to Homework 7 Solutions to Homework 7 Exercise #3 in section 5.2: A rectangular box is inscribed in a hemisphere of radius r. Find the dimensions of the box of maximum volume. Solution: The base of the rectangular box

More information

Multivariable Calculus Notes. Faraad Armwood. Fall: Chapter 1: Vectors, Dot Product, Cross Product, Planes, Cylindrical & Spherical Coordinates

Multivariable Calculus Notes. Faraad Armwood. Fall: Chapter 1: Vectors, Dot Product, Cross Product, Planes, Cylindrical & Spherical Coordinates Multivariable Calculus Notes Faraad Armwood Fall: 2017 Chapter 1: Vectors, Dot Product, Cross Product, Planes, Cylindrical & Spherical Coordinates Chapter 2: Vector-Valued Functions, Tangent Vectors, Arc

More information

Problem 1 (Exercise 2.2, Monograph)

Problem 1 (Exercise 2.2, Monograph) MS&E 314/CME 336 Assignment 2 Conic Linear Programming January 3, 215 Prof. Yinyu Ye 6 Pages ASSIGNMENT 2 SOLUTIONS Problem 1 (Exercise 2.2, Monograph) We prove the part ii) of Theorem 2.1 (Farkas Lemma

More information

Linear Ordinary Differential Equations

Linear Ordinary Differential Equations MTH.B402; Sect. 1 20180703) 2 Linear Ordinary Differential Equations Preliminaries: Matrix Norms. Denote by M n R) the set of n n matrix with real components, which can be identified the vector space R

More information

and verify that it satisfies the differential equation:

and verify that it satisfies the differential equation: MOTIVATION: Chapter One: Basic and Review Why study differential equations? Suppose we know how a certain quantity changes with time (for example, the temperature of coffee in a cup, the number of people

More information

Chapter 4. Inverse Function Theorem. 4.1 The Inverse Function Theorem

Chapter 4. Inverse Function Theorem. 4.1 The Inverse Function Theorem Chapter 4 Inverse Function Theorem d d d d d d d d d d d d d d d d d d d d d d d d d d d d d d d d d d d d d d d d d d d d d d d d d d d d d d d d d d d d d d d d d d d d d d d d d d dd d d d d This chapter

More information

THE INVERSE FUNCTION THEOREM

THE INVERSE FUNCTION THEOREM THE INVERSE FUNCTION THEOREM W. PATRICK HOOPER The implicit function theorem is the following result: Theorem 1. Let f be a C 1 function from a neighborhood of a point a R n into R n. Suppose A = Df(a)

More information

Lecture notes for Math Multivariable Calculus

Lecture notes for Math Multivariable Calculus Lecture notes for Math 417-517 Multivariable alculus J. Dimock Dept. of Mathematics SUNY at Buffalo Buffalo, NY 14260 December 4, 2012 ontents 1 multivariable calculus 3 1.1 vectors....................................

More information

Multiplication of Polynomials

Multiplication of Polynomials Summary 391 Chapter 5 SUMMARY Section 5.1 A polynomial in x is defined by a finite sum of terms of the form ax n, where a is a real number and n is a whole number. a is the coefficient of the term. n is

More information

Convex Functions. Daniel P. Palomar. Hong Kong University of Science and Technology (HKUST)

Convex Functions. Daniel P. Palomar. Hong Kong University of Science and Technology (HKUST) Convex Functions Daniel P. Palomar Hong Kong University of Science and Technology (HKUST) ELEC5470 - Convex Optimization Fall 2017-18, HKUST, Hong Kong Outline of Lecture Definition convex function Examples

More information

4.6. Linear Operators on Hilbert Spaces

4.6. Linear Operators on Hilbert Spaces 4.6. Linear Operators on Hilbert Spaces 1 4.6. Linear Operators on Hilbert Spaces Note. This section explores a number of different kinds of bounded linear transformations (or, equivalently, operators

More information

Lecture 2: Convex functions

Lecture 2: Convex functions Lecture 2: Convex functions f : R n R is convex if dom f is convex and for all x, y dom f, θ [0, 1] f is concave if f is convex f(θx + (1 θ)y) θf(x) + (1 θ)f(y) x x convex concave neither x examples (on

More information

3 Applications of partial differentiation

3 Applications of partial differentiation Advanced Calculus Chapter 3 Applications of partial differentiation 37 3 Applications of partial differentiation 3.1 Stationary points Higher derivatives Let U R 2 and f : U R. The partial derivatives

More information

We denote the derivative at x by DF (x) = L. With respect to the standard bases of R n and R m, DF (x) is simply the matrix of partial derivatives,

We denote the derivative at x by DF (x) = L. With respect to the standard bases of R n and R m, DF (x) is simply the matrix of partial derivatives, The derivative Let O be an open subset of R n, and F : O R m a continuous function We say F is differentiable at a point x O, with derivative L, if L : R n R m is a linear transformation such that, for

More information

Symmetric Matrices and Eigendecomposition

Symmetric Matrices and Eigendecomposition Symmetric Matrices and Eigendecomposition Robert M. Freund January, 2014 c 2014 Massachusetts Institute of Technology. All rights reserved. 1 2 1 Symmetric Matrices and Convexity of Quadratic Functions

More information

MATHEMATICAL ECONOMICS: OPTIMIZATION. Contents

MATHEMATICAL ECONOMICS: OPTIMIZATION. Contents MATHEMATICAL ECONOMICS: OPTIMIZATION JOÃO LOPES DIAS Contents 1. Introduction 2 1.1. Preliminaries 2 1.2. Optimal points and values 2 1.3. The optimization problems 3 1.4. Existence of optimal points 4

More information

Lecture 10: Powers of Matrices, Difference Equations

Lecture 10: Powers of Matrices, Difference Equations Lecture 10: Powers of Matrices, Difference Equations Difference Equations A difference equation, also sometimes called a recurrence equation is an equation that defines a sequence recursively, i.e. each

More information

Knowledge Discovery and Data Mining 1 (VO) ( )

Knowledge Discovery and Data Mining 1 (VO) ( ) Knowledge Discovery and Data Mining 1 (VO) (707.003) Review of Linear Algebra Denis Helic KTI, TU Graz Oct 9, 2014 Denis Helic (KTI, TU Graz) KDDM1 Oct 9, 2014 1 / 74 Big picture: KDDM Probability Theory

More information

Chapter 4. The First Fundamental Form (Induced Metric)

Chapter 4. The First Fundamental Form (Induced Metric) Chapter 4. The First Fundamental Form (Induced Metric) We begin with some definitions from linear algebra. Def. Let V be a vector space (over IR). A bilinear form on V is a map of the form B : V V IR which

More information

1. Find the solution of the following uncontrolled linear system. 2 α 1 1

1. Find the solution of the following uncontrolled linear system. 2 α 1 1 Appendix B Revision Problems 1. Find the solution of the following uncontrolled linear system 0 1 1 ẋ = x, x(0) =. 2 3 1 Class test, August 1998 2. Given the linear system described by 2 α 1 1 ẋ = x +

More information