Chapter 3. Differentiable Mappings. 1. Differentiable Mappings

Size: px
Start display at page:

Download "Chapter 3. Differentiable Mappings. 1. Differentiable Mappings"

Transcription

1 Chapter 3 Differentiable Mappings 1 Differentiable Mappings Let V and W be two linear spaces over IR A mapping L from V to W is called a linear mapping if L(u + v) = Lu + Lv for all u, v V and L(λv) = λ(lv) for all λ IR and v V If L is a linear mapping from V to W, then for any c IR, cl is a linear mapping Moreover, if L and M are two linear mappings from V to W, then L + M is a linear mapping Let U, V and W be linear spaces If M is a linear mapping from U to V and L is a linear mapping from V to W, then the composite mapping L M is a linear mapping from U to W Now assume that V = IR k and W = IR m, where k and m are positive integers If L is a linear mapping from IR k to IR m, then there exists a unique matrix B = (b ij ) 1 i m,1 j k such that b 11 b 12 b 1k b L(x) = Bx = 21 b 22 b 2k b m1 b m2 b mk x 1 x 2 x k, x = x 1 x 2 x k IRk The matrix B is the matrix representation of the linear mapping L with respect to the standard bases We call B the standard matrix of L Often we will use B to denote the linear mapping x Bx, x IR k Suppose that B and C are the standard matrices of linear mapping L and M from IR k to IR m, respectively Then B + C is the standard matrix of L + M Moreover, if c is a real number, then cb is the standard matrix of cl Furthermore, if C is the standard matrix of a linear mapping M from IR d to IR k, and if B is the standard matrix of a linear mapping L from IR k to IR m, then BC is the standard matrix of the composite linear mapping L M from IR d to IR m The norm of a linear mapping L from IR k to IR m is defined by L := sup{ L(x) : x IR k, x 1} Thus, L 0 It is easily seen that L = 0 if and only if L(x) = 0 for all x IR k Moreover, for any real number c, cl = c L If M is also a linear mapping from IR k to IR m, then L + M L + M 1

2 Let U be a nonempty open subset of IR k A mapping f from U to IR m is said to be differentiable at a point a in U if there exists a linear mapping L from IR k to IR m such that f(x) f(a) L(x a) lim x a x a The linear mapping L satisfying the above condition is unique This linear mapping is denoted by df a and is called the differential of f at a If f is differentiable at every point = 0 in U, then we call f a differentiable mapping from U to IR m Theorem 11 Let f be a mapping from an open set U in IR k to IR m, and let a be a point in U Suppose that f(x) = (f 1 (x), f 2 (x),, f m (x)) for x U, where f 1, f 2,, f m are real-valued functions on U Then f is differentiable at a if and only if f 1, f 2,, f m are differentiable at a If this is the case, then the standard matrix of the differential df a is D 1 f 1 (a) D 2 f 1 (a) D k f 1 (a) D 1 f 2 (a) D 2 f 2 (a) D k f 2 (a) D 1 f m (a) D 2 f m (a) D k f m (a) Proof Suppose that f is differentiable at a and the standard matrix of the differential df a is B = (b ij ) 1 i m,1 j k For each i {1,, m}, let v i be the ith row vector (b i1,, b ik ) IR k We have It follows that f i (x) f i (a) v i, x a f(x) f(a) L(x a) f i (x) f i (a) v i, x a lim x a x a Hence, for each i {1,, m}, f i is differentiable at a and b ij = D j f i (a) for 1 i m and 1 j k Conversely, suppose that f i is differentiable at a for each i {1,, m} For each i, there exists a vector v i IR k such that f i (x) f i (a) v i, x a lim x a x a Let B be the m k matrix with v 1, v 2,, v m as its rows, and let L be the corresponding linear mapping from IR k to IR m Then we have = 0 = 0 f(x) f(a) L(x a) m f i (x) f i (a) v i, x a i=1 2

3 Consequently, f(x) f(a) L(x a) lim x a x a = 0 This shows that the mapping f is differentiable at a We define the Jacobian matrix of f at a to be Df(a) := ( D j f i (a) ) 1 i m,1 j k An application of the mean value theorem gives the following useful result for differentiable mappings Theorem 12 Let f = (f 1,, f m ) be a differentiable mapping from an open set U in IR k to IR m Let a and b be two distinct points in U such that the closed line segment [a, b] is contained in U If K is a real number such that Df(x) K for all x in the open line segment (a, b), then f(b) f(a) K b a Proof The theorem is obviously true if f(b) = f(a) In what follows we assume that f(b) f(a) For a vector v = (v 1,, v m ) IR m we define h(x) := v, f(x) = v 1 f 1 (x) + + v m f m (x), x U Then v is a differentiable function on U By the mean value theorem (Theorem 41 in Chapter 2), there exists some x (a, b) such that h(b) h(a) = h(x), b a) = v[df(x)](b a), where v is regarded as a 1 m matrix, Df(x) is an m k matrix, and b a is regarded as a k 1 matrix Since Df(x) K, it follows that v, f(b) f(a) v Df(x) b a K v b a Choosing v := [f(b) f(a)]/ f(b) f(a) in the above inequalities, we obtain v = 1 and v, f(b) f(a) = f(b) f(a) Therefore, f(b) f(a) K b a The following theorem gives the chain rule for the composition of two differentiable mappings 3

4 Theorem 13 Let f be a mapping from an open set U in IR k to IR m, and let g be a mapping from an open set V in IR k to IR n Suppose that a U and f(u) V If f is differentiable at a, and if g is differentiable at b := f(a), then the composite mapping g f is differentiable at a and d(g f) a = dg f(a) df a Consequently, D(g f)(a) = Dg(f(a))Df(a) Proof We write S for dg b and write T for df a Since g is differentiable at b = f(a), there exists r > 0 such that y B r (b) implies y V and g(y) g(b) S(y b) ε y b Since f is differentiable at a, there exists δ > 0 such that x B δ (a) implies x U, f(x) B r (b) and f(x) f(a) T (x a) ε x a In what follows we assume that x B δ (a) Then we have f(x) f(a) f(x) f(a) T (x a) + T (x a) (ε + T ) x a Moreover, S(f(x) f(a)) S T (x a) S f(x) f(a) T (x a) ε S x a Since f(x) lies in B r (b), we have g(f(x)) g(f(a)) S(f(x) f(a)) ε f(x) f(a) By using the triangle inequality, we derive from the above inequalities that g f(x) g f(a) S T (x a) ε S x a +ε f(x) f(a) ε( S +ε+ T ) x a This shows that g f is differentiable at a and d(g f) a = S T In other words, d(g f) a = dg f(a) df a In the above theorem, the mapping f can be represented as y s = f s (x 1,, x k ), s = 1,, m, (x 1,, x k ) U, and the mapping g can be represented as z i = g i (y 1,, y m ), i = 1,, n, (y 1,, y m ) V 4

5 If we use the traditional notation, then the chain rule can be expressed as z i x j = m s=1 z i y s y s x j, i = 1,, n, j = 1,, k Example Let f be the mapping from IR 2 to IR 2 given by u = ρ cos θ, v = ρ sin θ, (ρ, θ) IR 2, and let g be the mapping from IR 2 to IR 2 given by We have and x = u 2 v 2, y = 2uv, (u, v) IR 2 [ ] cos θ ρ sin θ Df(ρ, θ) = sin θ ρ cos θ [ 2u 2v Dg(u, v) = 2v 2u By the chain rule we obtain D(g f)(ρ, θ) = Dg(ρ cos θ, ρ sin θ)df(ρ, θ) Consequently, [ ] [ ] [ ] 2ρ cos θ 2ρ sin θ cos θ ρ sin θ 2ρ cos(2θ) 2ρ D(g f)(ρ, θ) = = 2 sin(2θ) 2ρ sin θ 2ρ cos θ sin θ ρ cos θ 2ρ sin(2θ) 2ρ 2 cos(2θ) ] 2 The Jacobian Determinant Let f be a differentiable mapping from an open set U in IR k to IR k For a point a U, the Jacobian determinant of f at a is defined to be J f (a) := det(df(a)) = det ( D j f i (a) ) 1 i,j k Example 1 Let f be the mapping from IR 2 to IR 2 given by u = ρ cos θ, v = ρ sin θ, (ρ, θ) IR 2 Then J f (ρ, θ) = cos θ sin θ ρ sin θ ρ cos θ = ρ Example 2 Let g be the mapping from IR 3 to IR 3 given by x = ρ cos θ sin φ, y = ρ sin θ sin φ, z = ρ cos φ, (ρ, θ, φ) IR 3 5

6 Then cos θ sin φ ρ sin θ sin φ ρ cos θ cos φ J g (ρ, θ, φ) = sin θ sin φ ρ cos θ sin φ ρ sin θ cos φ cos φ 0 ρ sin φ = ρ2 sin φ We are in a position to review basic properties of determinants Let A = (a ij ) 1 i,j n be an n n matrix of real numbers If n = 1, we define det(a 11 ) := a 11 Suppose that n > 1 For a fixed pair (i, j) (1 i, j n) we use A ij to denote the (n 1) (n 1) matrix obtained by deleting the ith row and the jth column from A We define det A := n ( 1) 1+j A 1j In particular, if A = (a ij ) 1 i,j 2 is a 2 2 matrix, then det A = a 11 a 12 a 21 a 22 := a 11a 22 a 12 a 21 j=1 If A = (a ij ) 1 i,j 3 is a 3 3 matrix, then a 11 a 12 a 13 det A = a 21 a 22 a 23 a 31 a 32 a 33 := a 11 a 22 a 23 a 32 a 33 a 12 a 21 a 23 a 31 a 33 + a 13 a 21 a 22 a 31 a 32 For an n n matrix A = (a ij ) 1 i,j n we use A i to denote its ith column (i = 1,, n) Thus A can be written as [A 1, A 2,, A n ] By using an induction argument we can easily verify the following properties of determinants (d1) If A is the identity matrix, that is, a ij = 1 for i = j and a ij = 0 for i j, then det A = 1 (d2) The determinant of a matrix A is a multilinear function of its columns More precisely, if the jth column A j is equal to a sum of two column vectors A i det[a 1,, A i 1, A i + A i, A i+1,, A n ] and A i, then = det[a 1,, A i 1, A i, A i+1,, A n ] + det[a 1,, A i 1, A i, A i+1,, A n ] Furthermore, if c IR, then det[a 1,, A i 1, ca i, A i+1,, A n ] = c det[a 1,, A i 1, A i, A i+1,, A n ] 6

7 (d3) If two adjacent columns of a matrix A are equal, ie, if A i = A i+1 for some i in {1,, n 1}, then det A = 0 The above three conditions (d1), (d2), and (d3) characterize the properties of determinants In other words, all the properties of determinants can be derived from (d1), (d2), and (d3) Let us derive the following property: (d4) If two columns of a matrix are interchanged, then its determinant changes by a sign We establish this property first when two adjacent columns A i and A i+1 are interchanged By (d3) we have det[, A i + A i+1, A i + A i+1, ] = 0 Applying the multilinear property (d2) to the above determinant, we obtain det[, A i, A i, ] + det[, A i, A i+1, ] + det[, A i+1, A i, ] + det[, A i+1, A i+1, ] = 0 By (d3) we have det[, A i, A i, ] = 0 and det[, A i+1, A i+1, ] = 0 Hence, det[, A i+1, A i, ] = det[, A i, A i+1, ] Now we strengthen property (d3) as follows: (d5) If two columns of a matrix A are equal, then det A = 0 Assume that two columns of the matrix A are equal We can change the matrix by a successive interchange of adjacent columns until we obtain a matrix A with equal adjacent columns By what has been proved for (d4) we have det A = det A or det A = det A But det A = 0 by (d3) Hence det A = 0 We can now finish the proof of (d4) Suppose that the ith column and the jth column are interchanged, where 1 i < j n By (d5) we have det[, A i + A j,, A i + A j, ] = 0 Expanding the above determinant as before, we obtain det[, A j,, A i, ] = det[, A i,, A j, ] The following property is also useful (d6) If one adds a scalar multiple of one column to another column, then the value of the determinant does not change 7

8 Suppose that the ith column A i of a matrix A is replaced by A i + ca j, where j i and c IR By (d2) we have det[, A i 1, A i + ca j, A i+1, ] = det[, A i 1, A i, A i+1, ] + c det[, A i 1, A j, A i+1, ] There are two determinants on the right of the above equality The first determinant is just det A, and the second determinant is equal to 0, since two of its columns are equal This verifies (d6) For c IR and i {1,, n}, let Q i (c) be the matrix obtained from the n n identity matrix I by multiplying its ith column by c For 1 i j n, we use P ij to denote the matrix obtained from I by interchanging the ith column and the jth column For α IR and 1 i j n, we use R ij (α) to denote the matrix obtained from I by adding the α multiple of the jth column to the ith column A square matrix is called an elementary matrix if it has one of the forms P ij, Q i (c), or R ij (α) Theorem 21 If A and B are two square matrices of the same size, then det(ab) = (det A)(det B) and det(a T ) = det A Proof Let A and B be two n n matrices Then B can be written as a product of elementary matrices Hence, in order to prove det(ab) = (det A)(det B), it suffices to show that det(ae) = (det A)(det E) for each elementary matrix E If E = P ij, then AP ij is the matrix obtained by interchanging the ith column and the jth column of A; hence det(ap ij ) = det A = (det A)(det P ij ) If E = Q i (c), then AQ i (c) is the matrix obtained from A by multiplying its ith column by c; hence det(aq i (c)) = c det A = (det A)(det Q i (c)) If E = R ij (α), then AR ij (α) is the matrix obtained from A by adding the α multiple of the jth column to the ith column; hence det(ar ij (α)) = det A = (det A)(det R ij (α)) This completes the proof of det(ab) = (det A)(det B) Let us show that det E T = det E for any elementary matrix E Indeed, Pij T = P ij and Q i (c) T = Q i (c) Moreover, R ij (α) T = R ji (α) Hence, det R ij (α) T = 1 = det R ij (α) The 8

9 matrix A can be written as A = E 1 E k, where E 1,, E k are elementary matrices We have det A T = det(e T k E T 1 ) = (det E T k ) (det E T 1 ) = (det E k ) (det E 1 ) = det A This completes the proof of det A T = det A An n n matrix A is said to be invertible if there exists an n n matrix B such that AB = BA = I Such a matrix B is uniquely determined by A This matrix B is called the inverse of A and will be denoted by A 1 Theorem 22 A square matrix A is invertible if and only if det A 0 Proof Let A be an n n matrix If A is invertible, then there exists an n n matrix B such that AB = I It follows that 1 = det I = det(ab) = (det A)(det B) This shows that det A 0 If E is an elementary matrix and det E 0, then E is invertible Indeed, P ij is invertible since P ij P ij = I Moreover, R ij (α) is invertible, because R ij (α)r ij ( α) = I If E = Q i (c) and det E 0, then c = det E 0 In this case Q i (c) is invertible, since Q i (1/c)Q i (c) = Q i (c)q i (1/c) = I Now suppose that det A 0 We write A as A = E 1 E k, where E 1,, E k are elementary matrices Since det A = det(e 1 ) det(e k ) and det A 0, we have det E j 0 for each j {1,, k} By what has been proved, each E j is invertible Consequently, (E 1 k This shows that A is invertible E1 1 )A = A(E 1 k E1 1 ) = I 3 The Inverse Function Theorem The main theorem of this section establishes sufficient conditions for the existence of a local inverse of a continuously differentiable mapping Theorem 31 Let U be an open set in IR k and let f = (f 1,, f k ) be a continuously differentiable mapping from U to IR k Suppose that a is a point in U such that J f (a) 0 Then there exist an open set U 1 with a U 1 U and an open set V 1 with f(a) V 1 f(u) such that f is a one-to-one mapping from U 1 onto V 1 Moreover, the inverse mapping g of f U1 is continuously differentiable on V 1 Proof Let S denote the Jacobian matrix Df(a) = (D j f i (a)) 1 i,j k Since the Jacobian determinant J f (a) 0, the matrix S is invertible Let T := S 1 For given y IR k, 9

10 consider the mapping h from U to IR k defined by h(x) := x T (f(x) y), x U If there exists x U such that h(x ) = x, then T (f(x ) y) = 0 Since T is invertible, it follows that f(x ) = y Thus, the problem of solving the equation f(x) = y is reduced to the problem of finding a fixed point of the mapping h We observe that Dh(x) = I T (Df(x)), x U In particular, Dh(a) = I T (Df(a)) = I T S = 0, where 0 stands for the k k matrix with all entries being 0 Since f is continuously differentiable on U, so is h Hence, there exists some r > 0 such that B r (a) U and Dh(x) < 1/2 for all x B r (a) Consequently, the matrix I Dh(x) is invertible Thus, the Jacobian matrix Df(x) = T 1 (I Dh(x)) is invertible for all x B r (a) Moreover, by Theorem 12 we have h(x ) h(x ) 1 2 x x for all x, x B r (a) It follows that T [f(x ) f(x )] = [x x ] [h(x ) h(x )] x x h(x ) h(x ) 1 2 x x Therefore, f(x ) f(x ) 1 2 T x x In particular, f Br (a) is one-to-one for all x, x B r (a) Let δ := r/(2 T ) and V 1 := B δ (b), where b := f(a) Let y V 1 Our goal is to find x U such that f(x) = y For this purpose, we use the following iteration scheme: x 0 := a and x k+1 := h(x k ) for k = 0, 1, 2, We shall use mathematical induction to prove that the statement P k : x k+1 B r (a) and x k+1 x k < r 2 k+1 is true for all k IN 0 For k = 0 we have x 1 x 0 = T (f(x 0 ) y) It follows that x 1 a = x 1 x 0 T y b < T δ = r 2 This verifies P 0 Suppose that k > 0 and P j is true for all j < k By the induction hypothesis, x k, x k 1 B r (a) Hence, x k+1 x k = h(x k ) h(x k 1 ) 1 2 xk x k 1 < r 2 k = r 2 k+1

11 Moreover, x k+1 a = x k+1 x 0 = (x j+1 x j ) Thus, x k+1 B r (a) j=0 x j+1 x j < j=0 r < r 2j+1 This verifies P k and thereby completes the induction procedure Since x k+1 x k < r/2 k+1 for all k IN 0, the sequence (x k ) k=0,1, converges to some x in IR k Letting k on the both sides of the equation x k+1 = h(x k ), we obtain x = h(x ) Therefore, f(x ) = y Furthermore, x a = lim k xk+1 a x j+1 x j = x 1 x 0 + j=0 j=0 x j+1 x j < r Let U 1 := B r (a) f 1 (V 1 ) Then U 1 is an open set and a U 1 U By what has been proved, for any y V 1, there exists a unique x B r (a) such that f(x ) = y Clearly, x B r (a) f 1 (V 1 ) = U 1 This shows that f is a one-to-one mapping from U 1 onto V 1 Moreover, f(a) = b V 1 = f(u 1 ) f(u) Let g = (g 1,, g k ) be the inverse mapping of f U1 For v V 1, we wish to show that g is differentiable at v Let y V 1, x := g(y) U 1 and u := g(v) U 1 Then y = f(x) and v = f(u) Let S u denote the Jacobian matrix Df(u) Then S u is invertible We have It follows that g(y) g(v) S 1 u g(y) g(v) S 1 u j=1 (y v) = Su 1 [ f(x) f(u) Su (x u) ] (y v) Su 1 f(x) f(u) S u (x u) Note that y v = f(x) f(u) x u /(2 T ) Since f is differentiable at u, we have Consequently, f(x) f(u) S u (x u) lim x u x u lim y v g(y) g(v) Su 1 (y v) y v Therefore, g is differentiable at v, and Dg(v) = Su 1 on U 1, we conclude that Dg is continuous on V 1 = 0 = 0 Example Let f = (f 1, f 2 ) be the mapping from IR 2 to IR 2 given by = (Df(u)) 1 Since Df is continuous f 1 (x 1, x 2 ) := x 2 1 x 2 2, f 2 (x 1, x 2 ) := 2x 1 x 2, (x 1, x 2 ) IR 2 11

12 Given (y 1, y 2 ) IR 2, we wish to solve the system of equations f 1 (x 1, x 2 ) = y 1, f 2 (x 1, x 2 ) = y 2 We have y 2 1 = (x 2 1 x 2 2) 2 and y 2 2 = 4x 2 1x 2 2 It follows that y y 2 2 = (x x 2 2) 2 Hence, x x 2 2 = y y2 2 Thus, for (y 1, y 2 ) = (0, 0), the only solution is (x 1, x 2 ) = (0, 0) For (y 1, y 2 ) (0, 0), we derive from x x 2 2 = y y2 2 and x2 1 x 2 2 = y 1 that (x 1, x 2 ) = ±( [y1 + [ y1 ) y ] y2 /2, + y1 2 + ] y2 2 /2 for y 2 0 or (x 1, x 2 ) = ±( [y1 + [ y1 ) y ] y2 /2, + y1 2 + ] y2 2 /2 for y 2 < 0 Consequently, f maps IR 2 \ {(0, 0)} two-to-one onto IR 2 \ {(0, 0)} Let us compute the Jacobian of f We have J f (x 1, x 2 ) = D 1f 1 (x 1, x 2 ) D 2 f 1 (x 1, x 2 ) D 1 f 2 (x 1, x 2 ) D 2 f 2 (x 1, x 2 ) = 2x 1 2x 2 2x 2 2x 1 = 4(x2 1 + x 2 2) By Theorem 31, for (a 1, a 2 ) (0, 0), there exist an open set U 1 in IR 2 containing (a 1, a 2 ) and an open set V 1 in IR 2 containing f(a 1, a 2 ) such that f is a one-to-one mapping from U 1 onto V 1 Indeed, if we choose r := a a2 2 > 0, then f B r (a 1,a 2 ) is one-to-one Let g = (g 1, g 2 ) be the inverse of f Br (a 1,a 2 ) By Theorem 31 and the chain rule we have Dg(y 1, y 2 ) = where (y 1, y 2 ) = f(x 1, x 2 ) for (x 1, x 2 ) B r (a 1, a 2 ) [ ] 1 x1 x 2 2(x x2 2 ), x 2 x 1 In the above example, the inverse mapping could be found in an explicit form This is not possible in general But the Inverse Mapping Theorem is still applicable It gives us a powerful tool to analyze mappings and curvilinear coordinates 4 The Implicit Function Theorem Theorem 41 Let f = (f 1,, f m ) be a continuously differentiable mapping from an open set U in IR k+m to IR m Each f i (i = 1,, m) is a function of (x 1,, x k, y 1,, y m ) Suppose that (a, b) = (a 1,, a k, b 1,, b m ) is a point in U such that f i (a, b) = 0 for i = 1,, m If ( fi ) det (a, b) y 0, j 1 i,j m 12

13 then there exist an open set V in IR k containing a = (a 1,, a k ) and a continuously differentiable mapping g = (g 1,, g m ) from V to IR m such that g(a) = b = (b 1,, b m ) and that for i = 1,, m, f i ( x1,, x k, g 1 (x 1,, x k ),, g m (x 1,, x k ) ) = 0 (x 1,, x k ) V Proof Let F = (F 1,, F k, F k+1,, F m ) be the mapping from U to IR k+m given by F i (x, y) = x i for i = 1,, k and F k+j (x, y) = f j (x, y) for j = 1,, m, where x := (x 1,, x k ) and y := (y 1,, y m ) Clearly, F (a, b) = (a 1,, a k, 0,, 0) The Jacobian matrix of F at (a, b) is [ I 0 S T where I is the k k identity matrix, 0 is the k m zero matrix, ( fi ) ( fi ) S = (a, b) and T = (a, b) x j 1 i m,1 j k y j ], 1 i m,1 j m By our assumption, det T 0 Hence, the Jacobian determinant J F (a, b) 0 By the Inverse Function Theorem, there exist an open set U 1 in IR k+m with (a, b) U 1 U and an open set V 1 in IR k+m with F (a, b) V 1 F (U) such that F is an one-to-one mapping from U 1 onto V 1 Let G := (G 1, G 2,, G k+m ) be the inverse mapping of F U1 Then G is continuously differentiable on V 1 Set V := {(x 1,, x k ) IR k : (x 1,, x k, 0,, 0) V 1 } Then V is an open set in IR k Moreover, since (a 1,, a k, 0,, 0) = F (a, b) V 1, we have a = (a 1,, a k ) V For j = 1,, m, let g j (x 1,, x k ) := G k+j (x 1,, x k, 0,, 0), (x 1,, x k ) V Then g := (g 1,, g m ) is a continuously differentiable mapping from V to IR m Since F (a, b) = (a 1,, a k, 0,, 0) = (a, 0), we have G(a, 0) = (a, b) In light of the definition of g we see that b = g(a) Moreover, since F G is the identity mapping on V, we have F (G(x, 0)) = (x, 0) for all x = (x 1,, x k ) V Consequently, we obtain G i (x, 0) = x i for i = 1,, k and F k+j (G(x, 0)) = 0 for j = 1,, m Therefore, for j = 1,, m we have f j ( x1,, x k, g 1 (x 1,, x k ),, g m (x 1,, x k ) ) = 0 (x 1,, x k ) V This completes the proof of the theorem 13

14 Let Z := {(x 1,, x k, y 1,, y m ) U 1 : f j (x 1,, x k, y 1,, y m ) = 0 for j = 1,, m} Let ϕ be the mapping given by ϕ(x 1,, x k ) := ( x 1,, x k, g 1 (x 1,, x k ),, g m (x 1,, x k ) ), (x 1,, x k ) V From the above proof we see that ϕ is a one-to-one mapping from V onto Z Example 1 Let S 1 and S 2 be two surfaces in IR 3 represented by S 1 := {(x, y, z) IR 3 : x 2 (y 2 + z 2 ) = 5} and S 2 := {(x, y, z) IR 3 : (x z) 2 + y 2 = 2} Clearly, (1, 1, 2) S 1 S 2 Let F and G be the functions given by F (x, y, z) := x 2 (y 2 + z 2 ) 5 and G(x, y, z) := (x z) 2 + y 2 2, (x, y, z) IR 3 At the point (1, 1, 2) we have F y G y F z G z = 2x 2 y 2x 2 z 2y 2(x z) = = 4 0 By Theorem 41 we can find an interval V in IR containing 1, an open set U in IR 3 containing (1, 1, 2), and functions f and g from V to IR such that f(1) = 1, g(1) = 2, and (f, g) maps V one-to-one and onto U (S 1 S 2 ) In particular, F (x, f(x), g(x)) = 0 and G(x, f(x), g(x)) = 0 for all x V By using the chain rule it follows that Consequently, F x + F y f (x) + F z g (x) = 0 and G x + G y f (x) + G z g (x) = 0 2x(y 2 + z 2 ) + 2x 2 yf (x) + 2x 2 zg (x) = 0 and 2(x z) + 2yf (x) 2(x z)g (x) = 0 Solving the above system of equations for f (x) and g (x) we obtain f (x) = y2 z + z 3 xy 2 x 2 z x 2 y and g (x) = x2 xz y 2 z 2 x 2, x V 14

15 Example 2 Consider the system of equations F (x, y, u, v) = u 2 + v 2 x 2 y = 0 and G(x, y, u, v) = u + v xy + 1 = 0 Clearly, F (2, 1, 1, 2) = 0 and G(2, 1, 1, 2) = 0 At the point (2, 1, 1, 2) we have F F u v = 2u 2v 1 1 = = 6 0 G u G v By Theorem 41 we can find an open set V in IR 2 containing (2, 1), and functions f and g from V to IR such that f(2, 1) = 1, g(2, 1) = 2, and F (x, y, f(x, y), g(x, y)) = 0 and G(x, y, f(x, y), g(x, y)) = 0 for all (x, y) V Differentiating both sides of the above equations with respect to x, we obtain F x + F f u x + F v g x = 0 and G x + G f u x + G g v x = 0 Consequently, f x = x yv g and u + v x = x + yu, (x, y) V u + v Similarly, differentiating with respect to y yields Hence we obtain F y + F f u y + F v f y = 1 2xv 2(u + v) g y = 0 and G y + G f u y + G g v y = 0 and g y = 1 + 2xu, (x, y) V 2(u + v) 5 Constrained Optimization In this section, as an application of the inverse function theorem, we study the Lagrange multiplier method for constrained optimization Theorem 51 Let f, g 1,, g k be real-valued continuously differentiable functions of (x 1,, x n ) defined on an open set U in IR n with n > k Let Z := {z U : g 1 (z) = = g k (z) = 0} Suppose that there exist a point a in Z and an open ball B r (a) U such that f(z) f(a) (or f(z) f(a)) for all z Z B r (a) Suppose also that the rows of the Jacobian matrix ( gi ) (a) x j 1 i k,1 j n 15

16 are linearly independent Then there exist real numbers λ 1,, λ k such that f(a) + λ 1 g 1 (a) + + λ k g k (a) = 0 Proof By a permutation of the index set {1,, n} if necessary we may assume that ( gi ) det (a) x 0 j 1 i,j k Consequently, we can find real numbers λ 1,, λ k such that the equality f x j (a) + i=1 λ i g i x j (a) = 0 holds for j = 1,, k Our proof will be complete if we can show that the above equality also holds for j = k + 1,, n Suppose that a = (a 1,, a n ) We write a = (a, a ), where a := (a 1,, a k ) and a := (a k+1,, a n ) By the Implicit Function Theorem, there exist an open set V in IR n k containing a = (a k+1,, a n ) and a continuously differentiable mapping φ = (φ 1,, φ k ) from V to IR k such that φ(a k+1,, a n ) = (a 1,, a k ) and that for i = 1,, k, g i ( φ1 (x k+1,, x n ),, φ k (x k+1,, x n ), x k+1,, x n ) = 0 (xk+1,, x n ) V For (x k+1,, x n ) V, define h(x k+1,, x n ) := f ( φ 1 (x k+1,, x n ),, φ k (x k+1,, x n ), x k+1,, x n ) Then h is a continuously differentiable function on V and it attains a local minimum (or a local maximum) at the point a = (a k+1,, a n ) Hence, for m = k + 1,, n we have 0 = h x m (a ) = j=1 f x j (a) φ j x m (a ) + f x m (a), where the chain rule has been used to derive the second equality Furthermore, we have j=1 Consequently, j=1 g i x j (a) φ j x m (a ) + g i x m (a) = 0, i = 1,, k, m = k + 1,, n f x j (a) φ j x m (a ) + f x m (a) + i=1 λ i( j=1 16 g i (a) φ j (a ) + g ) i (a) = 0 x j x m x m

17 It follows that f x m (a) + i=1 λ i g i x m (a) = 0, m = k + 1,, n This completes the proof The above theorem gives the following method to find local minima or maxima for a continuous differentiable function f subject to the constraint g 1 = = g k = 0 Set the Lagrange function L(x 1,, x n ) := f(x 1,, x n ) + λ i g i (x 1,, x n ), (x 1,, x n ) U, where λ 1,, λ k are Lagrange multipliers Solve the system of n + k equations i=1 { L x j (x 1,, x n ) = 0 g i (x 1,, x n ) = 0 for j = 1,, n for i = 1,, k for (x 1,, x n ) and (λ 1,, λ k ) Example Let us find the extreme values (maximum and minimum) of the function f(x 1, x 2, x 3 ) = x x x 3 3 subject to the constraints x x x 2 3 = 4 and x 1 + x 2 + x 3 = 1 Note that the set E := {(x 1, x 2, x 3 ) IR 3 : x x x 2 3 = 4, x 1 + x 2 + x 3 = 1} is a compact set So f attains its maximum and minimum on E We use the Lagrange multiplier method to find the maximum and minimum of f on E Let g 1 (x 1, x 2, x 3 ) := x x x and g 2 (x 1, x 2, x 3 ) := x 1 + x 2 + x 3 1 The Lagrange function is L(x 1, x 2, x 3, λ 1, λ 2 ) = (x x x 3 3) + λ 1 (x x x 2 3 4) + λ 2 (x 1 + x 2 + x 3 1) Setting L x j = 0 for j = 1, 2, 3, we obtain 3x λ 1 2x 1 + λ 2 = 0, 3x λ 1 2x 2 + λ 2 = 0, 3x λ 1 2x 3 + λ 2 = 0 17

18 It follows that Consequently, 3x 2 1 2x 1 1 3x 2 2 2x 2 1 3x 2 3 2x 3 1 = 0 (x 1 x 2 )(x 1 x 3 )(x 2 x 3 ) = 0 Thus, one and only one of the cases x 1 = x 2, x 1 = x 3, and x 2 = x 3 must occur Suppose that x 1 = x 2 This together with x 1 + x 2 + x 3 = 1 yields x 3 = 1 2x 1 Substituting x 2 = x 1 and x 3 = 1 2x 1 into the equation x x x 2 3 = 4, we obtain It has two solutions: x x (1 2x 1 ) 2 = 1 x 1 = and x 1 = Hence, the optimization problem has the following solutions: ( , , ) 3 and ( , , ) Other solutions are obtained from the above solutions by permutations of {x 1, x 2, x 3 } Note that the rows of the Jacobian matrix [ g1 g 1 g 1 ] x 1 x 2 x 3 = g 2 g 2 g 2 x 1 x 2 x 3 are linearly independent at any of these points [ 2x1 2x 2 2x ] It is easily seen that the first set of solutions correspond to the minimum value, and the second set of solutions correspond to the maximum value 18

Differentiation. f(x + h) f(x) Lh = L.

Differentiation. f(x + h) f(x) Lh = L. Analysis in R n Math 204, Section 30 Winter Quarter 2008 Paul Sally, e-mail: sally@math.uchicago.edu John Boller, e-mail: boller@math.uchicago.edu website: http://www.math.uchicago.edu/ boller/m203 Differentiation

More information

We denote the derivative at x by DF (x) = L. With respect to the standard bases of R n and R m, DF (x) is simply the matrix of partial derivatives,

We denote the derivative at x by DF (x) = L. With respect to the standard bases of R n and R m, DF (x) is simply the matrix of partial derivatives, The derivative Let O be an open subset of R n, and F : O R m a continuous function We say F is differentiable at a point x O, with derivative L, if L : R n R m is a linear transformation such that, for

More information

THE INVERSE FUNCTION THEOREM

THE INVERSE FUNCTION THEOREM THE INVERSE FUNCTION THEOREM W. PATRICK HOOPER The implicit function theorem is the following result: Theorem 1. Let f be a C 1 function from a neighborhood of a point a R n into R n. Suppose A = Df(a)

More information

Math113: Linear Algebra. Beifang Chen

Math113: Linear Algebra. Beifang Chen Math3: Linear Algebra Beifang Chen Spring 26 Contents Systems of Linear Equations 3 Systems of Linear Equations 3 Linear Systems 3 2 Geometric Interpretation 3 3 Matrices of Linear Systems 4 4 Elementary

More information

Math 421, Homework #9 Solutions

Math 421, Homework #9 Solutions Math 41, Homework #9 Solutions (1) (a) A set E R n is said to be path connected if for any pair of points x E and y E there exists a continuous function γ : [0, 1] R n satisfying γ(0) = x, γ(1) = y, and

More information

COMPLETE METRIC SPACES AND THE CONTRACTION MAPPING THEOREM

COMPLETE METRIC SPACES AND THE CONTRACTION MAPPING THEOREM COMPLETE METRIC SPACES AND THE CONTRACTION MAPPING THEOREM A metric space (M, d) is a set M with a metric d(x, y), x, y M that has the properties d(x, y) = d(y, x), x, y M d(x, y) d(x, z) + d(z, y), x,

More information

MATH Topics in Applied Mathematics Lecture 12: Evaluation of determinants. Cross product.

MATH Topics in Applied Mathematics Lecture 12: Evaluation of determinants. Cross product. MATH 311-504 Topics in Applied Mathematics Lecture 12: Evaluation of determinants. Cross product. Determinant is a scalar assigned to each square matrix. Notation. The determinant of a matrix A = (a ij

More information

INVERSE FUNCTION THEOREM and SURFACES IN R n

INVERSE FUNCTION THEOREM and SURFACES IN R n INVERSE FUNCTION THEOREM and SURFACES IN R n Let f C k (U; R n ), with U R n open. Assume df(a) GL(R n ), where a U. The Inverse Function Theorem says there is an open neighborhood V U of a in R n so that

More information

Definition 2.3. We define addition and multiplication of matrices as follows.

Definition 2.3. We define addition and multiplication of matrices as follows. 14 Chapter 2 Matrices In this chapter, we review matrix algebra from Linear Algebra I, consider row and column operations on matrices, and define the rank of a matrix. Along the way prove that the row

More information

Equality: Two matrices A and B are equal, i.e., A = B if A and B have the same order and the entries of A and B are the same.

Equality: Two matrices A and B are equal, i.e., A = B if A and B have the same order and the entries of A and B are the same. Introduction Matrix Operations Matrix: An m n matrix A is an m-by-n array of scalars from a field (for example real numbers) of the form a a a n a a a n A a m a m a mn The order (or size) of A is m n (read

More information

1. Bounded linear maps. A linear map T : E F of real Banach

1. Bounded linear maps. A linear map T : E F of real Banach DIFFERENTIABLE MAPS 1. Bounded linear maps. A linear map T : E F of real Banach spaces E, F is bounded if M > 0 so that for all v E: T v M v. If v r T v C for some positive constants r, C, then T is bounded:

More information

II. Determinant Functions

II. Determinant Functions Supplemental Materials for EE203001 Students II Determinant Functions Chung-Chin Lu Department of Electrical Engineering National Tsing Hua University May 22, 2003 1 Three Axioms for a Determinant Function

More information

2. Linear algebra. matrices and vectors. linear equations. range and nullspace of matrices. function of vectors, gradient and Hessian

2. Linear algebra. matrices and vectors. linear equations. range and nullspace of matrices. function of vectors, gradient and Hessian FE661 - Statistical Methods for Financial Engineering 2. Linear algebra Jitkomut Songsiri matrices and vectors linear equations range and nullspace of matrices function of vectors, gradient and Hessian

More information

Implicit Functions, Curves and Surfaces

Implicit Functions, Curves and Surfaces Chapter 11 Implicit Functions, Curves and Surfaces 11.1 Implicit Function Theorem Motivation. In many problems, objects or quantities of interest can only be described indirectly or implicitly. It is then

More information

Chapter 4 - MATRIX ALGEBRA. ... a 2j... a 2n. a i1 a i2... a ij... a in

Chapter 4 - MATRIX ALGEBRA. ... a 2j... a 2n. a i1 a i2... a ij... a in Chapter 4 - MATRIX ALGEBRA 4.1. Matrix Operations A a 11 a 12... a 1j... a 1n a 21. a 22.... a 2j... a 2n. a i1 a i2... a ij... a in... a m1 a m2... a mj... a mn The entry in the ith row and the jth column

More information

Section 1.6. M N = [a ij b ij ], (1.6.2)

Section 1.6. M N = [a ij b ij ], (1.6.2) The Calculus of Functions of Several Variables Section 16 Operations with Matrices In the previous section we saw the important connection between linear functions and matrices In this section we will

More information

Several variables. x 1 x 2. x n

Several variables. x 1 x 2. x n Several variables Often we have not only one, but several variables in a problem The issues that come up are somewhat more complex than for one variable Let us first start with vector spaces and linear

More information

ALGEBRAIC TOPOLOGY N. P. STRICKLAND

ALGEBRAIC TOPOLOGY N. P. STRICKLAND ALGEBRAIC TOPOLOGY N. P. STRICKLAND 1. Introduction In this course, we will study metric spaces (which will often be subspaces of R n for some n) with interesting topological structure. Here are some examples

More information

Chapter 2. Square matrices

Chapter 2. Square matrices Chapter 2. Square matrices Lecture notes for MA1111 P. Karageorgis pete@maths.tcd.ie 1/18 Invertible matrices Definition 2.1 Invertible matrices An n n matrix A is said to be invertible, if there is a

More information

Differentiation in higher dimensions

Differentiation in higher dimensions Capter 2 Differentiation in iger dimensions 2.1 Te Total Derivative Recall tat if f : R R is a 1-variable function, and a R, we say tat f is differentiable at x = a if and only if te ratio f(a+) f(a) tends

More information

f(x, y) = 1 2 x y2 xy 3

f(x, y) = 1 2 x y2 xy 3 Problem. Find the critical points of the function and determine their nature. We have We find the critical points We calculate f(x, y) = 2 x2 + 3 2 y2 xy 3 f x = x y 3 = 0 = x = y 3 f y = 3y 3xy 2 = 0

More information

Introduction to Matrix Algebra

Introduction to Matrix Algebra Introduction to Matrix Algebra August 18, 2010 1 Vectors 1.1 Notations A p-dimensional vector is p numbers put together. Written as x 1 x =. x p. When p = 1, this represents a point in the line. When p

More information

Mathematical Analysis Outline. William G. Faris

Mathematical Analysis Outline. William G. Faris Mathematical Analysis Outline William G. Faris January 8, 2007 2 Chapter 1 Metric spaces and continuous maps 1.1 Metric spaces A metric space is a set X together with a real distance function (x, x ) d(x,

More information

The Contraction Mapping Theorem and the Implicit Function Theorem

The Contraction Mapping Theorem and the Implicit Function Theorem The Contraction Mapping Theorem and the Implicit Function Theorem Theorem The Contraction Mapping Theorem Let B a = { x IR d x < a } denote the open ball of radius a centred on the origin in IR d. If the

More information

MA2AA1 (ODE s): The inverse and implicit function theorem

MA2AA1 (ODE s): The inverse and implicit function theorem MA2AA1 (ODE s): The inverse and implicit function theorem Sebastian van Strien (Imperial College) February 3, 2013 Differential Equations MA2AA1 Sebastian van Strien (Imperial College) 0 Some of you did

More information

THE INVERSE FUNCTION THEOREM FOR LIPSCHITZ MAPS

THE INVERSE FUNCTION THEOREM FOR LIPSCHITZ MAPS THE INVERSE FUNCTION THEOREM FOR LIPSCHITZ MAPS RALPH HOWARD DEPARTMENT OF MATHEMATICS UNIVERSITY OF SOUTH CAROLINA COLUMBIA, S.C. 29208, USA HOWARD@MATH.SC.EDU Abstract. This is an edited version of a

More information

1 Directional Derivatives and Differentiability

1 Directional Derivatives and Differentiability Wednesday, January 18, 2012 1 Directional Derivatives and Differentiability Let E R N, let f : E R and let x 0 E. Given a direction v R N, let L be the line through x 0 in the direction v, that is, L :=

More information

Math 291-2: Final Exam Solutions Northwestern University, Winter 2016

Math 291-2: Final Exam Solutions Northwestern University, Winter 2016 Math 29-2: Final Exam Solutions Northwestern University, Winter 206 Determine whether each of the following statements is true or false f it is true, explain why; if it is false, give a counterexample

More information

MATH 323 Linear Algebra Lecture 6: Matrix algebra (continued). Determinants.

MATH 323 Linear Algebra Lecture 6: Matrix algebra (continued). Determinants. MATH 323 Linear Algebra Lecture 6: Matrix algebra (continued). Determinants. Elementary matrices Theorem 1 Any elementary row operation σ on matrices with n rows can be simulated as left multiplication

More information

SPRING OF 2008 D. DETERMINANTS

SPRING OF 2008 D. DETERMINANTS 18024 SPRING OF 2008 D DETERMINANTS In many applications of linear algebra to calculus and geometry, the concept of a determinant plays an important role This chapter studies the basic properties of determinants

More information

MATHEMATICS. IMPORTANT FORMULAE AND CONCEPTS for. Final Revision CLASS XII CHAPTER WISE CONCEPTS, FORMULAS FOR QUICK REVISION.

MATHEMATICS. IMPORTANT FORMULAE AND CONCEPTS for. Final Revision CLASS XII CHAPTER WISE CONCEPTS, FORMULAS FOR QUICK REVISION. MATHEMATICS IMPORTANT FORMULAE AND CONCEPTS for Final Revision CLASS XII 2016 17 CHAPTER WISE CONCEPTS, FORMULAS FOR QUICK REVISION Prepared by M. S. KUMARSWAMY, TGT(MATHS) M. Sc. Gold Medallist (Elect.),

More information

The Contraction Mapping Theorem and the Implicit and Inverse Function Theorems

The Contraction Mapping Theorem and the Implicit and Inverse Function Theorems The Contraction Mapping Theorem and the Implicit and Inverse Function Theorems The Contraction Mapping Theorem Theorem The Contraction Mapping Theorem) Let B a = { x IR d x < a } denote the open ball of

More information

Functions of Several Variables (Rudin)

Functions of Several Variables (Rudin) Functions of Several Variables (Rudin) Definition: Let X and Y be finite-dimensional real vector spaces. Then L(X, Y ) denotes the set of all linear transformations from X to Y and L(X) denotes the set

More information

3. Vector spaces 3.1 Linear dependence and independence 3.2 Basis and dimension. 5. Extreme points and basic feasible solutions

3. Vector spaces 3.1 Linear dependence and independence 3.2 Basis and dimension. 5. Extreme points and basic feasible solutions A. LINEAR ALGEBRA. CONVEX SETS 1. Matrices and vectors 1.1 Matrix operations 1.2 The rank of a matrix 2. Systems of linear equations 2.1 Basic solutions 3. Vector spaces 3.1 Linear dependence and independence

More information

Review of Multi-Calculus (Study Guide for Spivak s CHAPTER ONE TO THREE)

Review of Multi-Calculus (Study Guide for Spivak s CHAPTER ONE TO THREE) Review of Multi-Calculus (Study Guide for Spivak s CHPTER ONE TO THREE) This material is for June 9 to 16 (Monday to Monday) Chapter I: Functions on R n Dot product and norm for vectors in R n : Let X

More information

ELEMENTARY LINEAR ALGEBRA WITH APPLICATIONS. 1. Linear Equations and Matrices

ELEMENTARY LINEAR ALGEBRA WITH APPLICATIONS. 1. Linear Equations and Matrices ELEMENTARY LINEAR ALGEBRA WITH APPLICATIONS KOLMAN & HILL NOTES BY OTTO MUTZBAUER 11 Systems of Linear Equations 1 Linear Equations and Matrices Numbers in our context are either real numbers or complex

More information

Math 61CM - Quick answer key to section problems Fall 2018

Math 61CM - Quick answer key to section problems Fall 2018 Math 6CM - Quick answer key to section problems Fall 08 Cédric De Groote These are NOT complete solutions! You are always expected to write down all the details and justify everything. This document is

More information

Prepared by: M. S. KumarSwamy, TGT(Maths) Page

Prepared by: M. S. KumarSwamy, TGT(Maths) Page Prepared by: M. S. KumarSwamy, TGT(Maths) Page - 50 - CHAPTER 3: MATRICES QUICK REVISION (Important Concepts & Formulae) MARKS WEIGHTAGE 03 marks Matrix A matrix is an ordered rectangular array of numbers

More information

Multivariable Calculus

Multivariable Calculus 2 Multivariable Calculus 2.1 Limits and Continuity Problem 2.1.1 (Fa94) Let the function f : R n R n satisfy the following two conditions: (i) f (K ) is compact whenever K is a compact subset of R n. (ii)

More information

Math Advanced Calculus II

Math Advanced Calculus II Math 452 - Advanced Calculus II Manifolds and Lagrange Multipliers In this section, we will investigate the structure of critical points of differentiable functions. In practice, one often is trying to

More information

2 b 3 b 4. c c 2 c 3 c 4

2 b 3 b 4. c c 2 c 3 c 4 OHSx XM511 Linear Algebra: Multiple Choice Questions for Chapter 4 a a 2 a 3 a 4 b b 1. What is the determinant of 2 b 3 b 4 c c 2 c 3 c 4? d d 2 d 3 d 4 (a) abcd (b) abcd(a b)(b c)(c d)(d a) (c) abcd(a

More information

f(x) f(z) c x z > 0 1

f(x) f(z) c x z > 0 1 INVERSE AND IMPLICIT FUNCTION THEOREMS I use df x for the linear transformation that is the differential of f at x.. INVERSE FUNCTION THEOREM Definition. Suppose S R n is open, a S, and f : S R n is a

More information

Determinants Chapter 3 of Lay

Determinants Chapter 3 of Lay Determinants Chapter of Lay Dr. Doreen De Leon Math 152, Fall 201 1 Introduction to Determinants Section.1 of Lay Given a square matrix A = [a ij, the determinant of A is denoted by det A or a 11 a 1j

More information

Chapter 4. Inverse Function Theorem. 4.1 The Inverse Function Theorem

Chapter 4. Inverse Function Theorem. 4.1 The Inverse Function Theorem Chapter 4 Inverse Function Theorem d d d d d d d d d d d d d d d d d d d d d d d d d d d d d d d d d d d d d d d d d d d d d d d d d d d d d d d d d d d d d d d d d d d d d d d d d d dd d d d d This chapter

More information

Mathematics Course 111: Algebra I Part I: Algebraic Structures, Sets and Permutations

Mathematics Course 111: Algebra I Part I: Algebraic Structures, Sets and Permutations Mathematics Course 111: Algebra I Part I: Algebraic Structures, Sets and Permutations D. R. Wilkins Academic Year 1996-7 1 Number Systems and Matrix Algebra Integers The whole numbers 0, ±1, ±2, ±3, ±4,...

More information

Nonlinear equations. Norms for R n. Convergence orders for iterative methods

Nonlinear equations. Norms for R n. Convergence orders for iterative methods Nonlinear equations Norms for R n Assume that X is a vector space. A norm is a mapping X R with x such that for all x, y X, α R x = = x = αx = α x x + y x + y We define the following norms on the vector

More information

CLASS 12 ALGEBRA OF MATRICES

CLASS 12 ALGEBRA OF MATRICES CLASS 12 ALGEBRA OF MATRICES Deepak Sir 9811291604 SHRI SAI MASTERS TUITION CENTER CLASS 12 A matrix is an ordered rectangular array of numbers or functions. The numbers or functions are called the elements

More information

Course Summary Math 211

Course Summary Math 211 Course Summary Math 211 table of contents I. Functions of several variables. II. R n. III. Derivatives. IV. Taylor s Theorem. V. Differential Geometry. VI. Applications. 1. Best affine approximations.

More information

MA102: Multivariable Calculus

MA102: Multivariable Calculus MA102: Multivariable Calculus Rupam Barman and Shreemayee Bora Department of Mathematics IIT Guwahati Differentiability of f : U R n R m Definition: Let U R n be open. Then f : U R n R m is differentiable

More information

Analysis-3 lecture schemes

Analysis-3 lecture schemes Analysis-3 lecture schemes (with Homeworks) 1 Csörgő István November, 2015 1 A jegyzet az ELTE Informatikai Kar 2015. évi Jegyzetpályázatának támogatásával készült Contents 1. Lesson 1 4 1.1. The Space

More information

Economics 204 Summer/Fall 2011 Lecture 11 Monday August 8, f(x + h) (f(x) + ah) = a = 0 = 0. f(x + h) (f(x) + T x (h))

Economics 204 Summer/Fall 2011 Lecture 11 Monday August 8, f(x + h) (f(x) + ah) = a = 0 = 0. f(x + h) (f(x) + T x (h)) Economics 204 Summer/Fall 2011 Lecture 11 Monday August 8, 2011 Sections 41-43 (Unified) Definition 1 Let f : I R, where I R is an open interval f is differentiable at x I if for some a R f(x + h) f lim

More information

The Inverse Function Theorem 1

The Inverse Function Theorem 1 John Nachbar Washington University April 11, 2014 1 Overview. The Inverse Function Theorem 1 If a function f : R R is C 1 and if its derivative is strictly positive at some x R, then, by continuity of

More information

Algebra II. Paulius Drungilas and Jonas Jankauskas

Algebra II. Paulius Drungilas and Jonas Jankauskas Algebra II Paulius Drungilas and Jonas Jankauskas Contents 1. Quadratic forms 3 What is quadratic form? 3 Change of variables. 3 Equivalence of quadratic forms. 4 Canonical form. 4 Normal form. 7 Positive

More information

Linear Algebra. Matrices Operations. Consider, for example, a system of equations such as x + 2y z + 4w = 0, 3x 4y + 2z 6w = 0, x 3y 2z + w = 0.

Linear Algebra. Matrices Operations. Consider, for example, a system of equations such as x + 2y z + 4w = 0, 3x 4y + 2z 6w = 0, x 3y 2z + w = 0. Matrices Operations Linear Algebra Consider, for example, a system of equations such as x + 2y z + 4w = 0, 3x 4y + 2z 6w = 0, x 3y 2z + w = 0 The rectangular array 1 2 1 4 3 4 2 6 1 3 2 1 in which the

More information

SYLLABUS. 1 Linear maps and matrices

SYLLABUS. 1 Linear maps and matrices Dr. K. Bellová Mathematics 2 (10-PHY-BIPMA2) SYLLABUS 1 Linear maps and matrices Operations with linear maps. Prop 1.1.1: 1) sum, scalar multiple, composition of linear maps are linear maps; 2) L(U, V

More information

Lecture 7. Econ August 18

Lecture 7. Econ August 18 Lecture 7 Econ 2001 2015 August 18 Lecture 7 Outline First, the theorem of the maximum, an amazing result about continuity in optimization problems. Then, we start linear algebra, mostly looking at familiar

More information

2 Sequences, Continuity, and Limits

2 Sequences, Continuity, and Limits 2 Sequences, Continuity, and Limits In this chapter, we introduce the fundamental notions of continuity and limit of a real-valued function of two variables. As in ACICARA, the definitions as well as proofs

More information

Linear Algebra II. 2 Matrices. Notes 2 21st October Matrix algebra

Linear Algebra II. 2 Matrices. Notes 2 21st October Matrix algebra MTH6140 Linear Algebra II Notes 2 21st October 2010 2 Matrices You have certainly seen matrices before; indeed, we met some in the first chapter of the notes Here we revise matrix algebra, consider row

More information

Basics of Calculus and Algebra

Basics of Calculus and Algebra Monika Department of Economics ISCTE-IUL September 2012 Basics of linear algebra Real valued Functions Differential Calculus Integral Calculus Optimization Introduction I A matrix is a rectangular array

More information

Linear Algebra Notes. Lecture Notes, University of Toronto, Fall 2016

Linear Algebra Notes. Lecture Notes, University of Toronto, Fall 2016 Linear Algebra Notes Lecture Notes, University of Toronto, Fall 2016 (Ctd ) 11 Isomorphisms 1 Linear maps Definition 11 An invertible linear map T : V W is called a linear isomorphism from V to W Etymology:

More information

2. Matrix Algebra and Random Vectors

2. Matrix Algebra and Random Vectors 2. Matrix Algebra and Random Vectors 2.1 Introduction Multivariate data can be conveniently display as array of numbers. In general, a rectangular array of numbers with, for instance, n rows and p columns

More information

k=1 ( 1)k+j M kj detm kj. detm = ad bc. = 1 ( ) 2 ( )+3 ( ) = = 0

k=1 ( 1)k+j M kj detm kj. detm = ad bc. = 1 ( ) 2 ( )+3 ( ) = = 0 4 Determinants The determinant of a square matrix is a scalar (i.e. an element of the field from which the matrix entries are drawn which can be associated to it, and which contains a surprisingly large

More information

Differential Topology Solution Set #2

Differential Topology Solution Set #2 Differential Topology Solution Set #2 Select Solutions 1. Show that X compact implies that any smooth map f : X Y is proper. Recall that a space is called compact if, for every cover {U } by open sets

More information

MATH 213 Linear Algebra and ODEs Spring 2015 Study Sheet for Midterm Exam. Topics

MATH 213 Linear Algebra and ODEs Spring 2015 Study Sheet for Midterm Exam. Topics MATH 213 Linear Algebra and ODEs Spring 2015 Study Sheet for Midterm Exam This study sheet will not be allowed during the test Books and notes will not be allowed during the test Calculators and cell phones

More information

LINEAR ALGEBRA REVIEW

LINEAR ALGEBRA REVIEW LINEAR ALGEBRA REVIEW SPENCER BECKER-KAHN Basic Definitions Domain and Codomain. Let f : X Y be any function. This notation means that X is the domain of f and Y is the codomain of f. This means that for

More information

NOTES ON LINEAR ALGEBRA. 1. Determinants

NOTES ON LINEAR ALGEBRA. 1. Determinants NOTES ON LINEAR ALGEBRA 1 Determinants In this section, we study determinants of matrices We study their properties, methods of computation and some applications We are familiar with the following formulas

More information

M311 Functions of Several Variables. CHAPTER 1. Continuity CHAPTER 2. The Bolzano Weierstrass Theorem and Compact Sets CHAPTER 3.

M311 Functions of Several Variables. CHAPTER 1. Continuity CHAPTER 2. The Bolzano Weierstrass Theorem and Compact Sets CHAPTER 3. M311 Functions of Several Variables 2006 CHAPTER 1. Continuity CHAPTER 2. The Bolzano Weierstrass Theorem and Compact Sets CHAPTER 3. Differentiability 1 2 CHAPTER 1. Continuity If (a, b) R 2 then we write

More information

Undergraduate Mathematical Economics Lecture 1

Undergraduate Mathematical Economics Lecture 1 Undergraduate Mathematical Economics Lecture 1 Yu Ren WISE, Xiamen University September 15, 2014 Outline 1 Courses Description and Requirement 2 Course Outline ematical techniques used in economics courses

More information

1.1 Limits and Continuity. Precise definition of a limit and limit laws. Squeeze Theorem. Intermediate Value Theorem. Extreme Value Theorem.

1.1 Limits and Continuity. Precise definition of a limit and limit laws. Squeeze Theorem. Intermediate Value Theorem. Extreme Value Theorem. STATE EXAM MATHEMATICS Variant A ANSWERS AND SOLUTIONS 1 1.1 Limits and Continuity. Precise definition of a limit and limit laws. Squeeze Theorem. Intermediate Value Theorem. Extreme Value Theorem. Definition

More information

Evaluating Determinants by Row Reduction

Evaluating Determinants by Row Reduction Evaluating Determinants by Row Reduction MATH 322, Linear Algebra I J. Robert Buchanan Department of Mathematics Spring 2015 Objectives Reduce a matrix to row echelon form and evaluate its determinant.

More information

Analysis and Linear Algebra. Lectures 1-3 on the mathematical tools that will be used in C103

Analysis and Linear Algebra. Lectures 1-3 on the mathematical tools that will be used in C103 Analysis and Linear Algebra Lectures 1-3 on the mathematical tools that will be used in C103 Set Notation A, B sets AcB union A1B intersection A\B the set of objects in A that are not in B N. Empty set

More information

Analysis Finite and Infinite Sets The Real Numbers The Cantor Set

Analysis Finite and Infinite Sets The Real Numbers The Cantor Set Analysis Finite and Infinite Sets Definition. An initial segment is {n N n n 0 }. Definition. A finite set can be put into one-to-one correspondence with an initial segment. The empty set is also considered

More information

Math Bootcamp An p-dimensional vector is p numbers put together. Written as. x 1 x =. x p

Math Bootcamp An p-dimensional vector is p numbers put together. Written as. x 1 x =. x p Math Bootcamp 2012 1 Review of matrix algebra 1.1 Vectors and rules of operations An p-dimensional vector is p numbers put together. Written as x 1 x =. x p. When p = 1, this represents a point in the

More information

Metric Spaces and Topology

Metric Spaces and Topology Chapter 2 Metric Spaces and Topology From an engineering perspective, the most important way to construct a topology on a set is to define the topology in terms of a metric on the set. This approach underlies

More information

MATH2210 Notebook 2 Spring 2018

MATH2210 Notebook 2 Spring 2018 MATH2210 Notebook 2 Spring 2018 prepared by Professor Jenny Baglivo c Copyright 2009 2018 by Jenny A. Baglivo. All Rights Reserved. 2 MATH2210 Notebook 2 3 2.1 Matrices and Their Operations................................

More information

Section Summary. Definition of a Function.

Section Summary. Definition of a Function. Section 2.3 Section Summary Definition of a Function. Domain, Codomain Image, Preimage Injection, Surjection, Bijection Inverse Function Function Composition Graphing Functions Floor, Ceiling, Factorial

More information

A fixed point theorem for weakly Zamfirescu mappings

A fixed point theorem for weakly Zamfirescu mappings A fixed point theorem for weakly Zamfirescu mappings David Ariza-Ruiz Dept. Análisis Matemático, Fac. Matemáticas, Universidad de Sevilla, Apdo. 1160, 41080-Sevilla, Spain Antonio Jiménez-Melado Dept.

More information

ACI-matrices all of whose completions have the same rank

ACI-matrices all of whose completions have the same rank ACI-matrices all of whose completions have the same rank Zejun Huang, Xingzhi Zhan Department of Mathematics East China Normal University Shanghai 200241, China Abstract We characterize the ACI-matrices

More information

MATH 240 Spring, Chapter 1: Linear Equations and Matrices

MATH 240 Spring, Chapter 1: Linear Equations and Matrices MATH 240 Spring, 2006 Chapter Summaries for Kolman / Hill, Elementary Linear Algebra, 8th Ed. Sections 1.1 1.6, 2.1 2.2, 3.2 3.8, 4.3 4.5, 5.1 5.3, 5.5, 6.1 6.5, 7.1 7.2, 7.4 DEFINITIONS Chapter 1: Linear

More information

= 1 and 2 1. T =, and so det A b d

= 1 and 2 1. T =, and so det A b d Chapter 8 Determinants The founder of the theory of determinants is usually taken to be Gottfried Wilhelm Leibniz (1646 1716, who also shares the credit for inventing calculus with Sir Isaac Newton (1643

More information

Solution to Homework 1

Solution to Homework 1 Solution to Homework Sec 2 (a) Yes It is condition (VS 3) (b) No If x, y are both zero vectors Then by condition (VS 3) x = x + y = y (c) No Let e be the zero vector We have e = 2e (d) No It will be false

More information

4. Determinants.

4. Determinants. 4. Determinants 4.1. Determinants; Cofactor Expansion Determinants of 2 2 and 3 3 Matrices 2 2 determinant 4.1. Determinants; Cofactor Expansion Determinants of 2 2 and 3 3 Matrices 3 3 determinant 4.1.

More information

Math 3108: Linear Algebra

Math 3108: Linear Algebra Math 3108: Linear Algebra Instructor: Jason Murphy Department of Mathematics and Statistics Missouri University of Science and Technology 1 / 323 Contents. Chapter 1. Slides 3 70 Chapter 2. Slides 71 118

More information

Chapter 8: Taylor s theorem and L Hospital s rule

Chapter 8: Taylor s theorem and L Hospital s rule Chapter 8: Taylor s theorem and L Hospital s rule Theorem: [Inverse Mapping Theorem] Suppose that a < b and f : [a, b] R. Given that f (x) > 0 for all x (a, b) then f 1 is differentiable on (f(a), f(b))

More information

Linear Algebra March 16, 2019

Linear Algebra March 16, 2019 Linear Algebra March 16, 2019 2 Contents 0.1 Notation................................ 4 1 Systems of linear equations, and matrices 5 1.1 Systems of linear equations..................... 5 1.2 Augmented

More information

SMSTC (2017/18) Geometry and Topology 2.

SMSTC (2017/18) Geometry and Topology 2. SMSTC (2017/18) Geometry and Topology 2 Lecture 1: Differentiable Functions and Manifolds in R n Lecturer: Diletta Martinelli (Notes by Bernd Schroers) a wwwsmstcacuk 11 General remarks In this lecture

More information

Homework Set #8 Solutions

Homework Set #8 Solutions Exercises.2 (p. 19) Homework Set #8 Solutions Assignment: Do #6, 8, 12, 14, 2, 24, 26, 29, 0, 2, 4, 5, 6, 9, 40, 42 6. Reducing the matrix to echelon form: 1 5 2 1 R2 R2 R1 1 5 0 18 12 2 1 R R 2R1 1 5

More information

Lecture 2: Review of Prerequisites. Table of contents

Lecture 2: Review of Prerequisites. Table of contents Math 348 Fall 217 Lecture 2: Review of Prerequisites Disclaimer. As we have a textbook, this lecture note is for guidance and supplement only. It should not be relied on when preparing for exams. In this

More information

APPLICATIONS OF DIFFERENTIABILITY IN R n.

APPLICATIONS OF DIFFERENTIABILITY IN R n. APPLICATIONS OF DIFFERENTIABILITY IN R n. MATANIA BEN-ARTZI April 2015 Functions here are defined on a subset T R n and take values in R m, where m can be smaller, equal or greater than n. The (open) ball

More information

International Competition in Mathematics for Universtiy Students in Plovdiv, Bulgaria 1994

International Competition in Mathematics for Universtiy Students in Plovdiv, Bulgaria 1994 International Competition in Mathematics for Universtiy Students in Plovdiv, Bulgaria 1994 1 PROBLEMS AND SOLUTIONS First day July 29, 1994 Problem 1. 13 points a Let A be a n n, n 2, symmetric, invertible

More information

MATH 106 LINEAR ALGEBRA LECTURE NOTES

MATH 106 LINEAR ALGEBRA LECTURE NOTES MATH 6 LINEAR ALGEBRA LECTURE NOTES FALL - These Lecture Notes are not in a final form being still subject of improvement Contents Systems of linear equations and matrices 5 Introduction to systems of

More information

Elementary maths for GMT

Elementary maths for GMT Elementary maths for GMT Linear Algebra Part 2: Matrices, Elimination and Determinant m n matrices The system of m linear equations in n variables x 1, x 2,, x n a 11 x 1 + a 12 x 2 + + a 1n x n = b 1

More information

Fundamentals of Engineering Analysis (650163)

Fundamentals of Engineering Analysis (650163) Philadelphia University Faculty of Engineering Communications and Electronics Engineering Fundamentals of Engineering Analysis (6563) Part Dr. Omar R Daoud Matrices: Introduction DEFINITION A matrix is

More information

Elementary Matrices. which is obtained by multiplying the first row of I 3 by -1, and

Elementary Matrices. which is obtained by multiplying the first row of I 3 by -1, and Elementary Matrices In this special handout of material not contained in the text, we introduce the concept of elementary matrix. Elementary matrices are useful in several ways that will be shown in this

More information

A matrix A is invertible i det(a) 6= 0.

A matrix A is invertible i det(a) 6= 0. Chapter 4 Determinants 4.1 Definition Using Expansion by Minors Every square matrix A has a number associated to it and called its determinant, denotedbydet(a). One of the most important properties of

More information

Transformations from R m to R n.

Transformations from R m to R n. Transformations from R m to R n 1 Differentiablity First of all because of an unfortunate combination of traditions (the fact that we read from left to right and the way we define matrix multiplication

More information

UNIVERSAL IDENTITIES. b = det

UNIVERSAL IDENTITIES. b = det UNIVERSAL IDENTITIES KEITH CONRAD 1 Introduction We want to describe an idea which reduces the verification of algebraic identities valid over all commutative rings to the verification over the complex

More information

Optimization Theory. A Concise Introduction. Jiongmin Yong

Optimization Theory. A Concise Introduction. Jiongmin Yong October 11, 017 16:5 ws-book9x6 Book Title Optimization Theory 017-08-Lecture Notes page 1 1 Optimization Theory A Concise Introduction Jiongmin Yong Optimization Theory 017-08-Lecture Notes page Optimization

More information

Linear Systems and Matrices

Linear Systems and Matrices Department of Mathematics The Chinese University of Hong Kong 1 System of m linear equations in n unknowns (linear system) a 11 x 1 + a 12 x 2 + + a 1n x n = b 1 a 21 x 1 + a 22 x 2 + + a 2n x n = b 2.......

More information

NOTES ON MULTIVARIABLE CALCULUS: DIFFERENTIAL CALCULUS

NOTES ON MULTIVARIABLE CALCULUS: DIFFERENTIAL CALCULUS NOTES ON MULTIVARIABLE CALCULUS: DIFFERENTIAL CALCULUS SAMEER CHAVAN Abstract. This is the first part of Notes on Multivariable Calculus based on the classical texts [6] and [5]. We present here the geometric

More information