Section 4.2. Types of Differentiation
|
|
- Aubrie Christiana Ray
- 5 years ago
- Views:
Transcription
1 42 Types of Differentiation 1 Section 42 Types of Differentiation Note In this section we define differentiation of various structures with respect to a scalar, a vector, and a matrix Definition Let vector y be a function of scalar variable x so that y = y(x) = [y 1,y 2,,y n ] = [y 1 (x),y 2 (x),,y n (x)] The derivative of vector y with respect to scalar x is the vector y = [ y1, y 2,, y ] n = [y 1,y 2,,y n] Let matrix Y be a function of scalar variable x so that Y = Y (x) = [y ij (x)] The derivative of matrix Y with respect to scalar x is the matrix [ ] Y = yij = [y ij ] Note To define differentiation with respect to a vector x of an object Φ which is a function of a vector, we use the usual definition: Φ(x + ty) Φ(x) lim where y is any conformable vector with x (this will give the derivative of Φ in the direction of y) Notice that we can use the usual ε/δ definition of limit, provided we have a metric on the objects Φ We start with scalar valued functions
2 42 Types of Differentiation 2 Definition Let f : R n R be a scalar valued function of a vector The derivative of scalar valued function f with respect to vector x = [x 1,x 2,,x n ] is [ f f =, f,, f ] 1 2 n This derivative is the gradient of f, denoted f (sometimes read dell f ) Note Let A be a given (constant) n n matrix and x R n The quadratic form x T Ax is a scalar and the derivative of the quadratic form in the direction y is (x + ty) T A(x + ty) x T Ax lim (x T + ty T )A(x + ty) x T Ax = lim x T Ax + ty T Ax + tx T Ay + t 2 y T Ay x T Ax = lim = lim(y T Ax + x T AY + ty T Ay) = y T Ax + x T Ay t 0 = y T Ax + (x T Ay) T since x T Ay is a 1 1 matrix and so is symmetric = y T Ax + y T A T x = y T (A + A T )x Notice that x 2 2 = xt x = x T Ix so this example shows that the directional derivative of the Euclidean norm exists in all directions y Note We now consider derivatives of vector valued functions of vectors, f : R n R m
3 42 Types of Differentiation 3 Definition Let f : S R m where S R n With f = f(x 1,x 2,,x n ) = (f 1 (x 1,x 2,,x n ),f 2 (x 1,x 2,,x n ),,f m (x 1,x 2,,x n )), define the matrix gradient f/ as the n m matrix with (i,j) entry f i / j : f 1 f f = f 1 2 f 2 2 f 1 n f 2 n This is also sometimes denoted f (Gentle uses the notation f T /) The m n matrix f) T = ( f) T is the Jacobian of f (denoted by Gentle as f/ T ) f m 1 f m 2 f m n Note You have seen the Jacobian when dealing with substitution in a multiple integral (it sort of plays the role of du in one variable u-substitution) See my online Calculus 3 notes for Substitutions in Multiple Integrals : edu/gardnerr/2110/notes-12e/c15s8pdf Definition Let Y be a p q matrix of functions f ij : R n R, Y = [f ij ] where f ij = f ij (x 1,x 2,,x m ) Define the derivative of Y as Y = [ f ij] That is, Y/ is a three dimensional object (or a matrix of vectors) of dimension p q n where the (i,j,k) entry is f ij / k For fixed k we denote the p q matrix with (i,j) entry as the (i,j,k ) entry of Y/ as Y/ k
4 42 Types of Differentiation 4 Note The linearity and chain rule properties of partial derivatives allow us to establish rules of differentiation in these new settings For example, if Y = [f ij ] where f ij : R n R is a square nonsingular matrix then we show in Exercise 43 that Y 1 = Y 1 ( ) Y Y 1 In Exercise 42A we show for Y = [f ij ] that [tr(y )] = tr ( ) Y Definition Let f : R n R be a scalar valued function of a vector The matrix H with (i,j) entry as is the Hessian of f, denoted H: i j H = n 1 Gentle denotes the Hessian as 2 f and x 2 n 2 2 f T 1 n 2 n 2 n Note In the case that f = f(x,y) then det(h) = det 2 y y y 2 = 2 f 2 y 2 ( 2 ) 2 f y (assuming the mixed partials are equals, which is really an assumption of continuity) and det(h) is related to the curvature of the surface z = f(x, y); see The Gauss Curvature in Detail from my Differential Geometry notes: etsuedu/gardnerr/5310/5310pdf/dg1-6pdf (see the example on pages 3 and
5 42 Types of Differentiation 5 4) Because of this, the Hessian (which is often defined as det(h) instead of H itself) can be used to find extrema and saddle points of z = f(x,y) in a style similar to the use of the Second Derivative Test for functions y = f(x) See Extreme Values and Saddle Points from Calculus 3: notes-12e/c14s7pdf Note Gentle lists f/ for several types of functions f(x) where x is an n- vector Let a be a constant scalar, b a constant conformable vector, and A a constant conformable matrix Then we have (with corrections to Gentle as given in the Errata): f(x) ax b T x x T b x T x xx T b T Ax x T Ab f/ ai b b 2x x I + I x A T b Ab x T Ax (A + A T )x exp(( 1/2)x T Ax) exp(( 1/2)x T Ax)Ax if A = A T x 2 2 2x V (x) 2x/(n 1) The proofs of some of these are to be given in Exercise 42B
6 42 Types of Differentiation 6 Definition Let X be a matrix of independent variables and f a function of X so that f = f(x) = f([x ij ]) The derivative of function f with respect to matrix X is the matrix f [ ] f = ij Theorem 421 Differentiation of scalar valued function f satisfies the following (1) f T = ( f ) T (2) For X square and f(x) = tr(x), f = I (3) For AX a square matrix where A is constant, [tr(ax)] (4) [tr(xt X)] = 2X (5) With a and b constant vectors, [at Xb] = abt (6) [det(x)] = (adj(x)) T = A T Note A quick search of the internet reveals that there are a number of definitions for the derivative of a matrix with respect to a matrix Inspired by Gentle s general approach, we take the following definition Definition Let X be a matrix of independent variables and let Y be a matrix of functions of X, Y = [y ij ] = [y ij (X)] The derivative of matrix Y with respect to matrix X is the matrix Y [ ] Y = ij
7 42 Types of Differentiation 7 Note Gentle claims that that the following hold Given his cryptic statement of equation (415) on page 155, I am not totally confident that all of the following hold under our definition of a matrix with respect to a matrix Note Gentle lists f/ for several types of functions f(x) where X is a matrix Let a and b be a conformable vectors, and A a constant conformable matrix Then we have: f(x) f/ Justification a T Xb ab T Theorem 421(5) tr(ax) A T Theorem 421(3) tr(x T X) 2X Theorem 421(4) BX XC BXC I B C T I C T B If X is square and invertible as required, then: f(x) f/ Justification tr(x) I Theorem 421(2) tr(x k ) kx k 1 Exercise 42C(1) tr(bx 1 C) (X 1 CBX 1 ) T det(x) detx(x 1 ) T Exercise 42C(2) log(det(x)) (X 1 ) T Exercise 42C(3) detx k kdet(x) k (X 1 ) T Exercise 42C(4) BX 1 C (X 1 C) T BX 1 Revised: 4/1/2018
Multivariable Calculus
2 Multivariable Calculus 2.1 Limits and Continuity Problem 2.1.1 (Fa94) Let the function f : R n R n satisfy the following two conditions: (i) f (K ) is compact whenever K is a compact subset of R n. (ii)
More informationOR MSc Maths Revision Course
OR MSc Maths Revision Course Tom Byrne School of Mathematics University of Edinburgh t.m.byrne@sms.ed.ac.uk 15 September 2017 General Information Today JCMB Lecture Theatre A, 09:30-12:30 Mathematics revision
More informationSection 4.5. Integration and Expectation
4.5. Integration and Expectation 1 Section 4.5. Integration and Expectation Note. In this section we consider integrals of scalar valued functions of a vector and matrix valued functions of a scalar. We
More informationMAT 419 Lecture Notes Transcribed by Eowyn Cenek 6/1/2012
(Homework 1: Chapter 1: Exercises 1-7, 9, 11, 19, due Monday June 11th See also the course website for lectures, assignments, etc) Note: today s lecture is primarily about definitions Lots of definitions
More informationThe Derivative. Appendix B. B.1 The Derivative of f. Mappings from IR to IR
Appendix B The Derivative B.1 The Derivative of f In this chapter, we give a short summary of the derivative. Specifically, we want to compare/contrast how the derivative appears for functions whose domain
More informationOptimization and Calculus
Optimization and Calculus To begin, there is a close relationship between finding the roots to a function and optimizing a function. In the former case, we solve for x. In the latter, we solve: g(x) =
More informationREVIEW OF DIFFERENTIAL CALCULUS
REVIEW OF DIFFERENTIAL CALCULUS DONU ARAPURA 1. Limits and continuity To simplify the statements, we will often stick to two variables, but everything holds with any number of variables. Let f(x, y) be
More informationCalculus 2502A - Advanced Calculus I Fall : Local minima and maxima
Calculus 50A - Advanced Calculus I Fall 014 14.7: Local minima and maxima Martin Frankland November 17, 014 In these notes, we discuss the problem of finding the local minima and maxima of a function.
More informationD(f/g)(P ) = D(f)(P )g(p ) f(p )D(g)(P ). g 2 (P )
We first record a very useful: 11. Higher derivatives Theorem 11.1. Let A R n be an open subset. Let f : A R m and g : A R m be two functions and suppose that P A. Let λ A be a scalar. If f and g are differentiable
More information. D Matrix Calculus D 1
D Matrix Calculus D 1 Appendix D: MATRIX CALCULUS D 2 In this Appendix we collect some useful formulas of matrix calculus that often appear in finite element derivations D1 THE DERIVATIVES OF VECTOR FUNCTIONS
More informationMath 600 Day 1: Review of advanced Calculus
Math 600 Day 1: Review of advanced Calculus Ryan Blair University of Pennsylvania Thursday September 8, 2010 Ryan Blair (U Penn) Math 600 Day 1: Review of advanced Calculus Thursday September 8, 2010 1
More information446 CHAP. 8 NUMERICAL OPTIMIZATION. Newton's Search for a Minimum of f(x,y) Newton s Method
446 CHAP. 8 NUMERICAL OPTIMIZATION Newton's Search for a Minimum of f(xy) Newton s Method The quadratic approximation method of Section 8.1 generated a sequence of seconddegree Lagrange polynomials. It
More informationMA102: Multivariable Calculus
MA102: Multivariable Calculus Rupam Barman and Shreemayee Bora Department of Mathematics IIT Guwahati Differentiability of f : U R n R m Definition: Let U R n be open. Then f : U R n R m is differentiable
More informationNotes on Random Vectors and Multivariate Normal
MATH 590 Spring 06 Notes on Random Vectors and Multivariate Normal Properties of Random Vectors If X,, X n are random variables, then X = X,, X n ) is a random vector, with the cumulative distribution
More informationVariational Methods & Optimal Control
Variational Methods & Optimal Control lecture 02 Matthew Roughan Discipline of Applied Mathematics School of Mathematical Sciences University of Adelaide April 14, 2016
More information6.1 Matrices. Definition: A Matrix A is a rectangular array of the form. A 11 A 12 A 1n A 21. A 2n. A m1 A m2 A mn A 22.
61 Matrices Definition: A Matrix A is a rectangular array of the form A 11 A 12 A 1n A 21 A 22 A 2n A m1 A m2 A mn The size of A is m n, where m is the number of rows and n is the number of columns The
More informationSeveral variables. x 1 x 2. x n
Several variables Often we have not only one, but several variables in a problem The issues that come up are somewhat more complex than for one variable Let us first start with vector spaces and linear
More informationFunctions of Several Variables
Jim Lambers MAT 419/519 Summer Session 2011-12 Lecture 2 Notes These notes correspond to Section 1.2 in the text. Functions of Several Variables We now generalize the results from the previous section,
More informationLecture 2: Review of Prerequisites. Table of contents
Math 348 Fall 217 Lecture 2: Review of Prerequisites Disclaimer. As we have a textbook, this lecture note is for guidance and supplement only. It should not be relied on when preparing for exams. In this
More information8. Conjugate functions
L. Vandenberghe EE236C (Spring 2013-14) 8. Conjugate functions closed functions conjugate function 8-1 Closed set a set C is closed if it contains its boundary: x k C, x k x = x C operations that preserve
More informationg(t) = f(x 1 (t),..., x n (t)).
Reading: [Simon] p. 313-333, 833-836. 0.1 The Chain Rule Partial derivatives describe how a function changes in directions parallel to the coordinate axes. Now we shall demonstrate how the partial derivatives
More informationExtrema of Functions of Several Variables
Extrema of Functions of Several Variables MATH 311, Calculus III J. Robert Buchanan Department of Mathematics Fall 2011 Background (1 of 3) In single-variable calculus there are three important results
More informationSecond Order ODEs. Second Order ODEs. In general second order ODEs contain terms involving y, dy But here only consider equations of the form
Second Order ODEs Second Order ODEs In general second order ODEs contain terms involving y, dy But here only consider equations of the form A d2 y dx 2 + B dy dx + Cy = 0 dx, d2 y dx 2 and F(x). where
More informationFALL 2018 MATH 4211/6211 Optimization Homework 1
FALL 2018 MATH 4211/6211 Optimization Homework 1 This homework assignment is open to textbook, reference books, slides, and online resources, excluding any direct solution to the problem (such as solution
More informationHW2 Solutions
8.024 HW2 Solutions February 4, 200 Apostol.3 4,7,8 4. Show that for all x and y in real Euclidean space: (x, y) 0 x + y 2 x 2 + y 2 Proof. By definition of the norm and the linearity of the inner product,
More informationminimize x x2 2 x 1x 2 x 1 subject to x 1 +2x 2 u 1 x 1 4x 2 u 2, 5x 1 +76x 2 1,
4 Duality 4.1 Numerical perturbation analysis example. Consider the quadratic program with variables x 1, x 2, and parameters u 1, u 2. minimize x 2 1 +2x2 2 x 1x 2 x 1 subject to x 1 +2x 2 u 1 x 1 4x
More informationJim Lambers MAT 419/519 Summer Session Lecture 13 Notes
Jim Lambers MAT 419/519 Summer Session 2011-12 Lecture 13 Notes These notes correspond to Section 4.1 in the text. Least Squares Fit One of the most fundamental problems in science and engineering is data
More informationBasics of Calculus and Algebra
Monika Department of Economics ISCTE-IUL September 2012 Basics of linear algebra Real valued Functions Differential Calculus Integral Calculus Optimization Introduction I A matrix is a rectangular array
More informationChapter 2. Vectors and Vector Spaces
2.1. Operations on Vectors 1 Chapter 2. Vectors and Vector Spaces Section 2.1. Operations on Vectors Note. In this section, we define several arithmetic operations on vectors (especially, vector addition
More informationMachine Learning Brett Bernstein. Recitation 1: Gradients and Directional Derivatives
Machine Learning Brett Bernstein Recitation 1: Gradients and Directional Derivatives Intro Question 1 We are given the data set (x 1, y 1 ),, (x n, y n ) where x i R d and y i R We want to fit a linear
More informationExercises for Multivariable Differential Calculus XM521
This document lists all the exercises for XM521. The Type I (True/False) exercises will be given, and should be answered, online immediately following each lecture. The Type III exercises are to be done
More informationStat 206: Linear algebra
Stat 206: Linear algebra James Johndrow (adapted from Iain Johnstone s notes) 2016-11-02 Vectors We have already been working with vectors, but let s review a few more concepts. The inner product of two
More informationA function(al) f is convex if dom f is a convex set, and. f(θx + (1 θ)y) < θf(x) + (1 θ)f(y) f(x) = x 3
Convex functions The domain dom f of a functional f : R N R is the subset of R N where f is well-defined. A function(al) f is convex if dom f is a convex set, and f(θx + (1 θ)y) θf(x) + (1 θ)f(y) for all
More information1 Overview. 2 A Characterization of Convex Functions. 2.1 First-order Taylor approximation. AM 221: Advanced Optimization Spring 2016
AM 221: Advanced Optimization Spring 2016 Prof. Yaron Singer Lecture 8 February 22nd 1 Overview In the previous lecture we saw characterizations of optimality in linear optimization, and we reviewed the
More informationCalculus for the Life Sciences II Assignment 6 solutions. f(x, y) = 3π 3 cos 2x + 2 sin 3y
Calculus for the Life Sciences II Assignment 6 solutions Find the tangent plane to the graph of the function at the point (0, π f(x, y = 3π 3 cos 2x + 2 sin 3y Solution: The tangent plane of f at a point
More informationLinear Algebra, part 2 Eigenvalues, eigenvectors and least squares solutions
Linear Algebra, part 2 Eigenvalues, eigenvectors and least squares solutions Anna-Karin Tornberg Mathematical Models, Analysis and Simulation Fall semester, 2013 Main problem of linear algebra 2: Given
More informationCourse Summary Math 211
Course Summary Math 211 table of contents I. Functions of several variables. II. R n. III. Derivatives. IV. Taylor s Theorem. V. Differential Geometry. VI. Applications. 1. Best affine approximations.
More informationSection 3.3. Matrix Rank and the Inverse of a Full Rank Matrix
3.3. Matrix Rank and the Inverse of a Full Rank Matrix 1 Section 3.3. Matrix Rank and the Inverse of a Full Rank Matrix Note. The lengthy section (21 pages in the text) gives a thorough study of the rank
More informationCLOSED RANGE POSITIVE OPERATORS ON BANACH SPACES
Acta Math. Hungar., 142 (2) (2014), 494 501 DOI: 10.1007/s10474-013-0380-2 First published online December 11, 2013 CLOSED RANGE POSITIVE OPERATORS ON BANACH SPACES ZS. TARCSAY Department of Applied Analysis,
More informationImplicit Functions, Curves and Surfaces
Chapter 11 Implicit Functions, Curves and Surfaces 11.1 Implicit Function Theorem Motivation. In many problems, objects or quantities of interest can only be described indirectly or implicitly. It is then
More informationMTH4101 CALCULUS II REVISION NOTES. 1. COMPLEX NUMBERS (Thomas Appendix 7 + lecture notes) ax 2 + bx + c = 0. x = b ± b 2 4ac 2a. i = 1.
MTH4101 CALCULUS II REVISION NOTES 1. COMPLEX NUMBERS (Thomas Appendix 7 + lecture notes) 1.1 Introduction Types of numbers (natural, integers, rationals, reals) The need to solve quadratic equations:
More informationComputational Optimization. Mathematical Programming Fundamentals 1/25 (revised)
Computational Optimization Mathematical Programming Fundamentals 1/5 (revised) If you don t know where you are going, you probably won t get there. -from some book I read in eight grade If you do get there,
More information2 Functions of random variables
2 Functions of random variables A basic statistical model for sample data is a collection of random variables X 1,..., X n. The data are summarised in terms of certain sample statistics, calculated as
More informationCHAPTER 7. An Introduction to Numerical Methods for. Linear and Nonlinear ODE s
A SERIES OF CLASS NOTES FOR 2005-2006 TO INTRODUCE LINEAR AND NONLINEAR PROBLEMS TO ENGINEERS, SCIENTISTS, AND APPLIED MATHEMATICIANS DE CLASS NOTES 1 A COLLECTION OF HANDOUTS ON FIRST ORDER ORDINARY DIFFERENTIAL
More informationMatrix Algebra, Class Notes (part 2) by Hrishikesh D. Vinod Copyright 1998 by Prof. H. D. Vinod, Fordham University, New York. All rights reserved.
Matrix Algebra, Class Notes (part 2) by Hrishikesh D. Vinod Copyright 1998 by Prof. H. D. Vinod, Fordham University, New York. All rights reserved. 1 Converting Matrices Into (Long) Vectors Convention:
More informationConvex Optimization Theory. Chapter 5 Exercises and Solutions: Extended Version
Convex Optimization Theory Chapter 5 Exercises and Solutions: Extended Version Dimitri P. Bertsekas Massachusetts Institute of Technology Athena Scientific, Belmont, Massachusetts http://www.athenasc.com
More information3.5 Quadratic Approximation and Convexity/Concavity
3.5 Quadratic Approximation and Convexity/Concavity 55 3.5 Quadratic Approximation and Convexity/Concavity Overview: Second derivatives are useful for understanding how the linear approximation varies
More information3. Convex functions. basic properties and examples. operations that preserve convexity. the conjugate function. quasiconvex functions
3. Convex functions Convex Optimization Boyd & Vandenberghe basic properties and examples operations that preserve convexity the conjugate function quasiconvex functions log-concave and log-convex functions
More informationInequality Constraints
Chapter 2 Inequality Constraints 2.1 Optimality Conditions Early in multivariate calculus we learn the significance of differentiability in finding minimizers. In this section we begin our study of the
More information3. Convex functions. basic properties and examples. operations that preserve convexity. the conjugate function. quasiconvex functions
3. Convex functions Convex Optimization Boyd & Vandenberghe basic properties and examples operations that preserve convexity the conjugate function quasiconvex functions log-concave and log-convex functions
More informationComplex Variables. Chapter 2. Analytic Functions Section Harmonic Functions Proofs of Theorems. March 19, 2017
Complex Variables Chapter 2. Analytic Functions Section 2.26. Harmonic Functions Proofs of Theorems March 19, 2017 () Complex Variables March 19, 2017 1 / 5 Table of contents 1 Theorem 2.26.1. 2 Theorem
More informationPartial Fractions. Calculus 2 Lia Vas
Calculus Lia Vas Partial Fractions rational function is a quotient of two polynomial functions The method of partial fractions is a general method for evaluating integrals of rational function The idea
More informationThe Contraction Mapping Theorem and the Implicit and Inverse Function Theorems
The Contraction Mapping Theorem and the Implicit and Inverse Function Theorems The Contraction Mapping Theorem Theorem The Contraction Mapping Theorem) Let B a = { x IR d x < a } denote the open ball of
More informationTangent spaces, normals and extrema
Chapter 3 Tangent spaces, normals and extrema If S is a surface in 3-space, with a point a S where S looks smooth, i.e., without any fold or cusp or self-crossing, we can intuitively define the tangent
More informationMATH529 Fundamentals of Optimization Unconstrained Optimization II
MATH529 Fundamentals of Optimization Unconstrained Optimization II Marco A. Montes de Oca Mathematical Sciences, University of Delaware, USA 1 / 31 Recap 2 / 31 Example Find the local and global minimizers
More informationMath 5630: Iterative Methods for Systems of Equations Hung Phan, UMass Lowell March 22, 2018
1 Linear Systems Math 5630: Iterative Methods for Systems of Equations Hung Phan, UMass Lowell March, 018 Consider the system 4x y + z = 7 4x 8y + z = 1 x + y + 5z = 15. We then obtain x = 1 4 (7 + y z)
More informationThe Contraction Mapping Principle
Chapter 3 The Contraction Mapping Principle The notion of a complete space is introduced in Section 1 where it is shown that every metric space can be enlarged to a complete one without further condition.
More information2. Linear algebra. matrices and vectors. linear equations. range and nullspace of matrices. function of vectors, gradient and Hessian
FE661 - Statistical Methods for Financial Engineering 2. Linear algebra Jitkomut Songsiri matrices and vectors linear equations range and nullspace of matrices function of vectors, gradient and Hessian
More informationThe Contraction Mapping Theorem and the Implicit Function Theorem
The Contraction Mapping Theorem and the Implicit Function Theorem Theorem The Contraction Mapping Theorem Let B a = { x IR d x < a } denote the open ball of radius a centred on the origin in IR d. If the
More informationMathematical modelling Chapter 2 Nonlinear models and geometric models
Mathematical modelling Chapter 2 Nonlinear models and geometric models Faculty of Computer and Information Science University of Ljubljana 2018/2019 3. Nonlinear models Given is a sample of points {(x
More informationLECTURE NOTES ELEMENTARY NUMERICAL METHODS. Eusebius Doedel
LECTURE NOTES on ELEMENTARY NUMERICAL METHODS Eusebius Doedel TABLE OF CONTENTS Vector and Matrix Norms 1 Banach Lemma 20 The Numerical Solution of Linear Systems 25 Gauss Elimination 25 Operation Count
More information11 a 12 a 21 a 11 a 22 a 12 a 21. (C.11) A = The determinant of a product of two matrices is given by AB = A B 1 1 = (C.13) and similarly.
C PROPERTIES OF MATRICES 697 to whether the permutation i 1 i 2 i N is even or odd, respectively Note that I =1 Thus, for a 2 2 matrix, the determinant takes the form A = a 11 a 12 = a a 21 a 11 a 22 a
More informationMatrix Theory. A.Holst, V.Ufnarovski
Matrix Theory AHolst, VUfnarovski 55 HINTS AND ANSWERS 9 55 Hints and answers There are two different approaches In the first one write A as a block of rows and note that in B = E ij A all rows different
More informationMATH 5720: Unconstrained Optimization Hung Phan, UMass Lowell September 13, 2018
MATH 57: Unconstrained Optimization Hung Phan, UMass Lowell September 13, 18 1 Global and Local Optima Let a function f : S R be defined on a set S R n Definition 1 (minimizers and maximizers) (i) x S
More informationMath 207 Honors Calculus III Final Exam Solutions
Math 207 Honors Calculus III Final Exam Solutions PART I. Problem 1. A particle moves in the 3-dimensional space so that its velocity v(t) and acceleration a(t) satisfy v(0) = 3j and v(t) a(t) = t 3 for
More informationTangent Planes, Linear Approximations and Differentiability
Jim Lambers MAT 80 Spring Semester 009-10 Lecture 5 Notes These notes correspond to Section 114 in Stewart and Section 3 in Marsden and Tromba Tangent Planes, Linear Approximations and Differentiability
More informationNOTES ON MULTIVARIABLE CALCULUS: DIFFERENTIAL CALCULUS
NOTES ON MULTIVARIABLE CALCULUS: DIFFERENTIAL CALCULUS SAMEER CHAVAN Abstract. This is the first part of Notes on Multivariable Calculus based on the classical texts [6] and [5]. We present here the geometric
More informationWorkshop I The R n Vector Space. Linear Combinations
Workshop I Workshop I The R n Vector Space. Linear Combinations Exercise 1. Given the vectors u = (1, 0, 3), v = ( 2, 1, 2), and w = (1, 2, 4) compute a) u + ( v + w) b) 2 u 3 v c) 2 v u + w Exercise 2.
More informationCSCI5654 (Linear Programming, Fall 2013) Lectures Lectures 10,11 Slide# 1
CSCI5654 (Linear Programming, Fall 2013) Lectures 10-12 Lectures 10,11 Slide# 1 Today s Lecture 1. Introduction to norms: L 1,L 2,L. 2. Casting absolute value and max operators. 3. Norm minimization problems.
More informationMany of you got these steps reversed or otherwise out of order.
Problem 1. Let (X, d X ) and (Y, d Y ) be metric spaces. Suppose that there is a bijection f : X Y such that for all x 1, x 2 X. 1 10 d X(x 1, x 2 ) d Y (f(x 1 ), f(x 2 )) 10d X (x 1, x 2 ) Show that if
More informationChapter 7. Extremal Problems. 7.1 Extrema and Local Extrema
Chapter 7 Extremal Problems No matter in theoretical context or in applications many problems can be formulated as problems of finding the maximum or minimum of a function. Whenever this is the case, advanced
More informationAppendix 5. Derivatives of Vectors and Vector-Valued Functions. Draft Version 24 November 2008
Appendix 5 Derivatives of Vectors and Vector-Valued Functions Draft Version 24 November 2008 Quantitative genetics often deals with vector-valued functions and here we provide a brief review of the calculus
More informationSection 21. The Metric Topology (Continued)
21. The Metric Topology (cont.) 1 Section 21. The Metric Topology (Continued) Note. In this section we give a number of results for metric spaces which are familar from calculus and real analysis. We also
More informationElectromagnetism HW 1 math review
Electromagnetism HW math review Problems -5 due Mon 7th Sep, 6- due Mon 4th Sep Exercise. The Levi-Civita symbol, ɛ ijk, also known as the completely antisymmetric rank-3 tensor, has the following properties:
More informationHigher order derivative
2 Î 3 á Higher order derivative Î 1 Å Iterated partial derivative Iterated partial derivative Suppose f has f/ x, f/ y and ( f ) = 2 f x x x 2 ( f ) = 2 f x y x y ( f ) = 2 f y x y x ( f ) = 2 f y y y
More information1. Background: The SVD and the best basis (questions selected from Ch. 6- Can you fill in the exercises?)
Math 35 Exam Review SOLUTIONS Overview In this third of the course we focused on linear learning algorithms to model data. summarize: To. Background: The SVD and the best basis (questions selected from
More informationRecitation 1. Gradients and Directional Derivatives. Brett Bernstein. CDS at NYU. January 21, 2018
Gradients and Directional Derivatives Brett Bernstein CDS at NYU January 21, 2018 Brett Bernstein (CDS at NYU) Recitation 1 January 21, 2018 1 / 23 Initial Question Intro Question Question We are given
More informationMultiple Integrals and Vector Calculus (Oxford Physics) Synopsis and Problem Sets; Hilary 2015
Multiple Integrals and Vector Calculus (Oxford Physics) Ramin Golestanian Synopsis and Problem Sets; Hilary 215 The outline of the material, which will be covered in 14 lectures, is as follows: 1. Introduction
More informationNonlinear Optimization
Nonlinear Optimization (Com S 477/577 Notes) Yan-Bin Jia Nov 7, 2017 1 Introduction Given a single function f that depends on one or more independent variable, we want to find the values of those variables
More informationMATH H53 : Final exam
MATH H53 : Final exam 11 May, 18 Name: You have 18 minutes to answer the questions. Use of calculators or any electronic items is not permitted. Answer the questions in the space provided. If you run out
More informationSolutions to Homework 7
Solutions to Homework 7 Exercise #3 in section 5.2: A rectangular box is inscribed in a hemisphere of radius r. Find the dimensions of the box of maximum volume. Solution: The base of the rectangular box
More informationMultivariable Calculus Notes. Faraad Armwood. Fall: Chapter 1: Vectors, Dot Product, Cross Product, Planes, Cylindrical & Spherical Coordinates
Multivariable Calculus Notes Faraad Armwood Fall: 2017 Chapter 1: Vectors, Dot Product, Cross Product, Planes, Cylindrical & Spherical Coordinates Chapter 2: Vector-Valued Functions, Tangent Vectors, Arc
More informationProblem 1 (Exercise 2.2, Monograph)
MS&E 314/CME 336 Assignment 2 Conic Linear Programming January 3, 215 Prof. Yinyu Ye 6 Pages ASSIGNMENT 2 SOLUTIONS Problem 1 (Exercise 2.2, Monograph) We prove the part ii) of Theorem 2.1 (Farkas Lemma
More informationLinear Ordinary Differential Equations
MTH.B402; Sect. 1 20180703) 2 Linear Ordinary Differential Equations Preliminaries: Matrix Norms. Denote by M n R) the set of n n matrix with real components, which can be identified the vector space R
More informationand verify that it satisfies the differential equation:
MOTIVATION: Chapter One: Basic and Review Why study differential equations? Suppose we know how a certain quantity changes with time (for example, the temperature of coffee in a cup, the number of people
More informationChapter 4. Inverse Function Theorem. 4.1 The Inverse Function Theorem
Chapter 4 Inverse Function Theorem d d d d d d d d d d d d d d d d d d d d d d d d d d d d d d d d d d d d d d d d d d d d d d d d d d d d d d d d d d d d d d d d d d d d d d d d d d dd d d d d This chapter
More informationTHE INVERSE FUNCTION THEOREM
THE INVERSE FUNCTION THEOREM W. PATRICK HOOPER The implicit function theorem is the following result: Theorem 1. Let f be a C 1 function from a neighborhood of a point a R n into R n. Suppose A = Df(a)
More informationLecture notes for Math Multivariable Calculus
Lecture notes for Math 417-517 Multivariable alculus J. Dimock Dept. of Mathematics SUNY at Buffalo Buffalo, NY 14260 December 4, 2012 ontents 1 multivariable calculus 3 1.1 vectors....................................
More informationMultiplication of Polynomials
Summary 391 Chapter 5 SUMMARY Section 5.1 A polynomial in x is defined by a finite sum of terms of the form ax n, where a is a real number and n is a whole number. a is the coefficient of the term. n is
More informationConvex Functions. Daniel P. Palomar. Hong Kong University of Science and Technology (HKUST)
Convex Functions Daniel P. Palomar Hong Kong University of Science and Technology (HKUST) ELEC5470 - Convex Optimization Fall 2017-18, HKUST, Hong Kong Outline of Lecture Definition convex function Examples
More information4.6. Linear Operators on Hilbert Spaces
4.6. Linear Operators on Hilbert Spaces 1 4.6. Linear Operators on Hilbert Spaces Note. This section explores a number of different kinds of bounded linear transformations (or, equivalently, operators
More informationLecture 2: Convex functions
Lecture 2: Convex functions f : R n R is convex if dom f is convex and for all x, y dom f, θ [0, 1] f is concave if f is convex f(θx + (1 θ)y) θf(x) + (1 θ)f(y) x x convex concave neither x examples (on
More information3 Applications of partial differentiation
Advanced Calculus Chapter 3 Applications of partial differentiation 37 3 Applications of partial differentiation 3.1 Stationary points Higher derivatives Let U R 2 and f : U R. The partial derivatives
More informationWe denote the derivative at x by DF (x) = L. With respect to the standard bases of R n and R m, DF (x) is simply the matrix of partial derivatives,
The derivative Let O be an open subset of R n, and F : O R m a continuous function We say F is differentiable at a point x O, with derivative L, if L : R n R m is a linear transformation such that, for
More informationSymmetric Matrices and Eigendecomposition
Symmetric Matrices and Eigendecomposition Robert M. Freund January, 2014 c 2014 Massachusetts Institute of Technology. All rights reserved. 1 2 1 Symmetric Matrices and Convexity of Quadratic Functions
More informationMATHEMATICAL ECONOMICS: OPTIMIZATION. Contents
MATHEMATICAL ECONOMICS: OPTIMIZATION JOÃO LOPES DIAS Contents 1. Introduction 2 1.1. Preliminaries 2 1.2. Optimal points and values 2 1.3. The optimization problems 3 1.4. Existence of optimal points 4
More informationLecture 10: Powers of Matrices, Difference Equations
Lecture 10: Powers of Matrices, Difference Equations Difference Equations A difference equation, also sometimes called a recurrence equation is an equation that defines a sequence recursively, i.e. each
More informationKnowledge Discovery and Data Mining 1 (VO) ( )
Knowledge Discovery and Data Mining 1 (VO) (707.003) Review of Linear Algebra Denis Helic KTI, TU Graz Oct 9, 2014 Denis Helic (KTI, TU Graz) KDDM1 Oct 9, 2014 1 / 74 Big picture: KDDM Probability Theory
More informationChapter 4. The First Fundamental Form (Induced Metric)
Chapter 4. The First Fundamental Form (Induced Metric) We begin with some definitions from linear algebra. Def. Let V be a vector space (over IR). A bilinear form on V is a map of the form B : V V IR which
More information1. Find the solution of the following uncontrolled linear system. 2 α 1 1
Appendix B Revision Problems 1. Find the solution of the following uncontrolled linear system 0 1 1 ẋ = x, x(0) =. 2 3 1 Class test, August 1998 2. Given the linear system described by 2 α 1 1 ẋ = x +
More information