Chapter 2: Matrix Algebra

Similar documents
MATH 2331 Linear Algebra. Section 2.1 Matrix Operations. Definition: A : m n, B : n p. Example: Compute AB, if possible.

Matrix operations Linear Algebra with Computer Science Application

Chapter 4 - MATRIX ALGEBRA. ... a 2j... a 2n. a i1 a i2... a ij... a in

All of my class notes can be found at

Chapter 1: Linear Equations

Chapter 1: Linear Equations

Review Notes for Linear Algebra True or False Last Updated: February 22, 2010

Lecture Notes in Linear Algebra

Equality: Two matrices A and B are equal, i.e., A = B if A and B have the same order and the entries of A and B are the same.

1 Last time: inverses

Linear Algebra Summary. Based on Linear Algebra and its applications by David C. Lay

MAT 242 CHAPTER 4: SUBSPACES OF R n

Review Let A, B, and C be matrices of the same size, and let r and s be scalars. Then

Chapter 1: Systems of linear equations and matrices. Section 1.1: Introduction to systems of linear equations

Linear Algebra March 16, 2019

1. Determine by inspection which of the following sets of vectors is linearly independent. 3 3.

1 Last time: determinants

Study Guide for Linear Algebra Exam 2

Chapter 3. Directions: For questions 1-11 mark each statement True or False. Justify each answer.

Math 1553 Introduction to Linear Algebra

MAT 2037 LINEAR ALGEBRA I web:

Math 4A Notes. Written by Victoria Kala Last updated June 11, 2017

Math 54 HW 4 solutions

IMPORTANT DEFINITIONS AND THEOREMS REFERENCE SHEET

Linear Algebra: Lecture notes from Kolman and Hill 9th edition.

Fall Inverse of a matrix. Institute: UC San Diego. Authors: Alexander Knop

IMPORTANT DEFINITIONS AND THEOREMS REFERENCE SHEET

MTH501- Linear Algebra MCQS MIDTERM EXAMINATION ~ LIBRIANSMINE ~

MATH10212 Linear Algebra Lecture Notes

LINEAR ALGEBRA SUMMARY SHEET.

Homework Set #8 Solutions

Row Space and Column Space of a Matrix

Math Camp II. Basic Linear Algebra. Yiqing Xu. Aug 26, 2014 MIT

Announcements Monday, October 29

Math113: Linear Algebra. Beifang Chen

MATH10212 Linear Algebra Lecture Notes

Matrices and Matrix Algebra.

Elementary Linear Algebra Review for Exam 2 Exam is Monday, November 16th.

Lecture Summaries for Linear Algebra M51A

Matrices and systems of linear equations

BASIC NOTIONS. x + y = 1 3, 3x 5y + z = A + 3B,C + 2D, DC are not defined. A + C =

CSL361 Problem set 4: Basic linear algebra

MTH 464: Computational Linear Algebra

Matrix Algebra. Matrix Algebra. Chapter 8 - S&B

Linear Equations in Linear Algebra

Math 4377/6308 Advanced Linear Algebra

EXERCISE SET 5.1. = (kx + kx + k, ky + ky + k ) = (kx + kx + 1, ky + ky + 1) = ((k + )x + 1, (k + )y + 1)

Chapter 2 Notes, Linear Algebra 5e Lay

Math Linear Algebra Final Exam Review Sheet

Glossary of Linear Algebra Terms. Prepared by Vince Zaccone For Campus Learning Assistance Services at UCSB

MATRIX ALGEBRA AND SYSTEMS OF EQUATIONS. + + x 1 x 2. x n 8 (4) 3 4 2

Applied Matrix Algebra Lecture Notes Section 2.2. Gerald Höhn Department of Mathematics, Kansas State University

Math 3108: Linear Algebra

Linear Algebra. Linear Equations and Matrices. Copyright 2005, W.R. Winfrey

LINEAR ALGEBRA REVIEW

University of Ottawa

March 27 Math 3260 sec. 56 Spring 2018

Math 140, c Benjamin Aurispa. 2.1 Matrices

Solution to Homework 1

Linear equations in linear algebra

MATH 2360 REVIEW PROBLEMS

2. Every linear system with the same number of equations as unknowns has a unique solution.

Dimension. Eigenvalue and eigenvector

A FIRST COURSE IN LINEAR ALGEBRA. An Open Text by Ken Kuttler. Matrix Arithmetic

Vector Space Basics. 1 Abstract Vector Spaces. 1. (commutativity of vector addition) u + v = v + u. 2. (associativity of vector addition)

MATH 221: SOLUTIONS TO SELECTED HOMEWORK PROBLEMS

MATH 2331 Linear Algebra. Section 1.1 Systems of Linear Equations. Finding the solution to a set of two equations in two variables: Example 1: Solve:

February 20 Math 3260 sec. 56 Spring 2018

can only hit 3 points in the codomain. Hence, f is not surjective. For another example, if n = 4

MATH 20F: LINEAR ALGEBRA LECTURE B00 (T. KEMP)

Chapter 6: Orthogonality

Solving a system by back-substitution, checking consistency of a system (no rows of the form

MTH 464: Computational Linear Algebra

MATH 240 Spring, Chapter 1: Linear Equations and Matrices

ICS 6N Computational Linear Algebra Matrix Algebra

MATH 323 Linear Algebra Lecture 12: Basis of a vector space (continued). Rank and nullity of a matrix.

Numerical Linear Algebra Homework Assignment - Week 2

Systems of Linear Equations and Matrices

MATH SOLUTIONS TO PRACTICE PROBLEMS - MIDTERM I. 1. We carry out row reduction. We begin with the row operations

MAT Linear Algebra Collection of sample exams

Working with Block Structured Matrices

Math 123, Week 5: Linear Independence, Basis, and Matrix Spaces. Section 1: Linear Independence

Math Linear algebra, Spring Semester Dan Abramovich

Matrix & Linear Algebra

Linear Algebra Highlights

Solutions to Math 51 First Exam April 21, 2011

We showed that adding a vector to a basis produces a linearly dependent set of vectors; more is true.

Systems of Linear Equations and Matrices

EK102 Linear Algebra PRACTICE PROBLEMS for Final Exam Spring 2016

Linear Systems and Matrices

Math 369 Exam #2 Practice Problem Solutions

DS-GA 1002 Lecture notes 0 Fall Linear Algebra. These notes provide a review of basic concepts in linear algebra.

Math 321: Linear Algebra

Solutions to Exam I MATH 304, section 6

Section 4.5. Matrix Inverses

Chapter SSM: Linear Algebra. 5. Find all x such that A x = , so that x 1 = x 2 = 0.

Gaussian elimination

A matrix over a field F is a rectangular array of elements from F. The symbol

APPENDIX: MATHEMATICAL INDUCTION AND OTHER FORMS OF PROOF

Elementary Linear Algebra

Transcription:

Chapter 2: Matrix Algebra (Last Updated: October 12, 2016) These notes are derived primarily from Linear Algebra and its applications by David Lay (4ed). Write A = 1. Matrix operations [a 1 a n. Then entry a ij is the ith entry of the jth column (ith row and jth column). The diagonal entries of A are those a ij with i = j. A diagonal matrix is one such that a ij = 0 if i j. The zero matrix 0 is a matrix with a ij = 0 for all i, j. We ll discuss operations on matrices and how these correspond to operations on linear transformations. Let A = (a ij ) and = (b ij ) be matrices of the same size m n and c R. Matrix addition: A + is defined as the m n matrix whose i, j entry is a ij + b ij. Scalar multiplication: ca is defined as the m n matrix whose i, j entry if ca ij. 3 2 1 5 Example 1. Let A = 1 5, = 2 3, and c = 3. Compute A + c. 1 3 1 2 Theorem 2. Let A,, C be matrices of the same size and r, s R scalars. (1) A + = + A (2) (A + ) + C = A + ( + C) (3) A + 0 = A (4) r(a + ) = ra + r (5) (r + s)a = ra + sa (6) r(sa) = (rs)a It is worth observing the similarity between these properties and those of vectors. Later in the course we will study vector spaces (a generalization of R n and C n ). The space of m n is another example of a vector space because it obeys the linearity properties of those spaces. One can prove the above properties directly and without much difficulty. On the other hand, one could prove these by recognizing matrices as linear transformations and using those properties. If T and S are linear transformations with standard matrices A and, respectively, then T + S is a linear transformation with standard matrix A +. Similarly, if c is a scalar then the standard matrix of (ct ) is just ca. What is the standard matrix associated to the composition of two standard matrices? If T : R n R m and S : R p R q, then T S only makes sense of q = n. Let A and again be the standard matrices of T and S, respectively, so A is m n and is n p. Then (T S)(x) = T (S(x)) = T (x) = A(x) = A(x) for all x R n.

Hence, the matrix product A is the standard matrix of T S. S T R p R n A R m A T S If A is an m n matrix and is an n p matrix with columns b 1,..., b p, then the matrix product A is the m p matrix whose columns are Ab 1,..., Ab p. That is, A = A [b 1 b p = [Ab 1 Ab p. 2 1 1 2 4 Example 3. Let A =, =. Compute A. 1 5 3 1 6 In Example 3, the product A is not defined. In general, A A even if both are defined. Also, cancellation does not work. That is, it is possible for A = 0 even if A, 0. Another way of phrasing this definition is that each column of A is a linear combination of the columns of A using weights from the corresponding columns of. Let A = (a ij ) be m n and = (b ij ) n p. Let C = (c ij ) = A. Using the row-column rule, n c ij = a ik b kj = a i1 b 1j + a i2 b 2j + + a in b nj. k=1 Example 4. Use the row-column rule to compute A with A, as in Example 3. The n n identity matrix is a diagonal matrix I n with 1 for every diagonal entry. Theorem 5. Let A,, C be matrices of appropriate sizes such that each sum/product is defined and let r R be a scalar. (1) A(C) = (A)C (2) A( + C) = A + AC (3) ( + C)A = A + CA (4) r(a) = (ra) = A(r) (5) I m A = A = AI n If A is an n n matrix and k a positive integer, then the matrix power A k denotes the product of k copies of A. y definition, A 0 = I n. If A = (a ij ) is an m n matrix, the transpose of A, denoted A T, is an n m matrix whose columns are the rows of A. That is, the i, j entry of A T is a ji. Example 6. Let A, be as in Example 3. Compute A T, T, and (A) T. Theorem 7. Let A and be matrices of appropriate sizes. (1) (A T ) T = A (2) (A + ) T = A T + T (3) (ra) T = r(a T ) (4) (A) T = A T T

2. The inverse of a matrix In this section we only consider square matrices (m = n). Recall that the n n identity matrix I n is the diagonal matrix with 1s along the diagonal. For any n n matrix A, we have I n A = AI n = A. Definition 1. An n n matrix A is said to be invertible if there exists an n n matrix C such that AC = I n and CA = I n. We call C the inverse of A. A matrix which is not invertible is said to be singular. 2 5 7 5 Example 8. Let A = and C =. Verify that C is the inverse of A. 3 7 3 2 The next proposition explains why it is proper to refer to C as the inverse of A and not an inverse. Proposition 9. Let A be an invertible n n matrix. The inverse of A is unique. Proof. Suppose, C are inverses of A. Then = I n = (AC) = (A)C = I n C = C. Hence, = C. If A is invertible, we denote the inverse of A as A 1. Theorem 10. If A is an n n invertible matrix, then for each b R n, the equation Ax = b has a unique solution x = A 1 b. Proof. If A is invertible, then clearly, Ax = A(A 1 b) = (AA 1 )b = I n b = b, so A 1 b is a solution. We claim it is unique. Suppose u is another solution, then Au = b = A(A 1 b). Multiplying both sides by A 1 on the left gives, A 1 (Au) = A 1 (A(A 1 b)) (A 1 A)u = (A 1 A)(A 1 b) I n u = I n (A 1 b) u = A 1 b. Hence, the solution is unique. a b d b Theorem 11. Let A =. If ad bc 0, then A is invertible and A 1 = 1 ad bc. c d c a If ad bc = 0, then A is singular. Proof. Suppose ad bc 0. We claim that A 1 satisfies the definition of the inverse of A. a b AA 1 1 d b 1 ad bc 0 = = = I 2, c d ad bc c a ad bc 0 ad bc Similarly, A 1 A = I n. Hence, A 1 is the inverse of A. Exercise. Prove that if ad bc = 0, then A is singular. (Hint: Use Theorem 10. You might first consider the case that a = 0.)

Example 12. Use matrix inversion to solve the system 3x 1 + 4x 2 = 3 5x 1 + 6x 2 = 7. Theorem 13. Let A and be invertible n n matrices. (1) A 1 is invertible and (A 1 ) 1 = A. (2) A is invertible and (A) 1 = 1 A 1. (3) A T is invertible and (A T ) 1 = (A 1 ) T. Proof. (1) follows from the symmetry in the definition of A 1. For (2), it suffices to show that 1 A 1 satisfies the definition of the inverse for A by Proposition 9. y associativity, (A)( 1 A 1 ) = (A( 1 ))A 1 = (AI n )A 1 = AA 1 = I n. Similarly, ( 1 A 1 )(A) = I n. (3) Follows from Proposition 9 and Theorem 7, (A T )(A 1 ) T = (A 1 A) T = In T = I n. Similarly, (A 1 ) T A T = I n. Next we want to show how one can obtain the inverse of a matrix in general. The method is quite basic, but to explain why it works takes a bit more muscle, so we ll hold off on that momentarily. Theorem 14. An n n matrix A is invertible if and only if A is row equivalent to I n, and in this case, any sequence of elementary row operations that reduces A to I n also transforms A n into A 1. In other words, we form the augmented matrix [A I n and row reduce A to I n (if possible). The result will be the augmented matrix [I n A 1. Example 15. Find the inverse of the matrix 0 1 2 A = 1 0 3. 4 3 8

Elementary Matrices An elementary matrix is one obtained by performing a single row operation to an identity matrix. Example 16. The following are elementary matrices. performed on I 3 to obtain each one. 1 0 0 0 1 0 1 0 0 0 1 0, 1 0 0, 0 1 0. 4 0 1 0 0 1 Identify which row operation has been 0 0 5 Lemma 17. If an elementary row operation is performed on an m n matrix A, the resulting matrix can be written EA, where the m m matrix E is created by performing the same row operation on I m. Proof. Let E be the elementary matrix obtained by multiplying row k of I n by m 0. That is, E = [e 1 me k e n. Let A be any n n matrix. Then, EA = E [a 1 a n = [Ea 1 Ea n. It is clear that the result of the multiplication Ea i is the multiply the kth entry of a i by m. Thus, the resulting matrix EA is the same as A, except each entry in the kth row is multiplied by m. The proofs for the other two row operations are left as an exercise. Lemma 18. Each elementary matrix is invertible. The inverse of E is the elementary matrix of the same type that transforms E back into I. Proof. Let E be an elementary matrix and F the elementary matrix that transforms E back into I. Clearly such an F exists because every row operation is reversible. Moreover, E must reverse F. y Lemma 17, F E = I and EF = I, so F is the inverse of E. We are now ready to prove Theorem 14. Proof. Suppose that A is invertible. Then the equation Ax = b has a solution for all b R n. Hence, A has a pivot in each row. ecause A is square, A has n pivots and so the reduced row echelon form of A is I n. That is, A I n. Conversely, suppose A I n. Then there exists a series of row operations which transform A into I n. Denote the corresponding elementary matrices by E 1,..., E p. That is, E p (E p 1 E 1 A)) = I n. That is E p E 1 A = I n. Since the product of invertible matrices is invertible, then A = (E p E 1 ) 1. Since the inverse of an invertible matrix is invertible, it follows that A is invertible and the inverse is E p E 1.

3. Characterizations of invertible matrices Theorem 19 (The Invertible Matrix Theorem). Let A be an n n matrix. The following are equivalent. (1) A is invertible. (2) A is row equivalent to I n. (3) A has n pivot positions. (4) The equation Ax = 0 has only the trivial solution. (5) The columns of A form a linearly independent set. (6) The linear transformation x Ax is 1-1. (7) The equation Ax = b has at least one solution for each b R n. (8) The columns of A span R n. (9) The linear transformation x Ax is onto. (10) There exists an n n matrix C such that CA = I n. (11) There exists an n n matrix D such that AD = I n. (12) A T is invertible. Proof. (1) (2) by Theorem 14. (2) (3) is clear because I n has n pivots and row operations do not change the number of pivots. (2) (4) because row operations do not change the solution set and the only solution to x = I n x = 0 is 0. (4) (6) by Theorem 35 in Chapter 1. (5) (6) and (8) (9) by Theorem 36 in Chapter 1. (3) (7) (8) by Theorem 17 in Chapter 1. Thus, (1) (9) are all equivalent. (1) (12) by Theorem 13. Clearly, (1) implies (10) and (11). If (10) holds, then Ax = 0 has only the trivial solution. To see this, left multiply on both sides by C (the left inverse of A). Then x = I n x = (CA)x = C(Ax) = C0 = 0. Thus, (10) (4). If (11) holds, then D T is a left inverse of A T. Hence, by the above this implies that A T is invertible and so A is invertible. A linear transformation T : R n R n is said to be invertible if there exists S : R n R n such that for all x R n, S(T (x)) = x and T (S(x)) = x. We call S = T 1 the inverse of T. Theorem 20. Let T : R n R n be a linear transformation with standard matrix A. Then T is invertible if and only if A is an invertible matrix. In that case, the linear transformation S : R n R n given by S(x) = A 1 x is the unique function such that S(T (x)) = x and T (S(x)) = x for all x R n.

Proof. Suppose T is invertible with inverse S. We claim A is invertible. Let b R n and set x = S(b). Then T (x) = T (S(b)) = b, so T is onto and hence by the Invertible Matrix Theorem, A is invertible. Conversely, suppose A is invertible. Define a function S : R n R n by S(x) = A 1 x. We claim S is the inverse of T. S(T (x)) = S(Ax) = A 1 (Ax) = (A 1 A)x = I n x = x T (S(x)) = T (A 1 x) = A(A 1 x) = (AA 1 )x = I n x = x. Hence, the claim holds. The proof of the uniqueness of S is left as an exercise.

4. Partitioned Matrices Definition 2. A partition of a matrix A is a decomposition of A into rectangular submatrices A 1,..., A n such that each entry in A lies in some submatrix. Example 21. The matrix can be partitioned as 1 3 1 4 A = 2 1 0 7 A = 1 3 1 4 where A 11 =, A 12 =, A 21 = 2 1 0 7 4 5 2 1 [ A11 A 12 A 12 A 22 4 5 2, and A 22 = 1. Note that there are many other partitions of this matrix and in general a matrix will have several partitions. I leave it as an exercise to write another partition of this matrix. If A and are matrices of the same sense and partition in the same way, then one may obtain A+ by adding corresponding partitions. Similarly, we can multiple A and as partitioned matrices so long as multiplication between the blocks makes sense. Example 22. Let A be as in Example 21 and 1 3 1 = where 1 = 0 7 and 2 = 2 2 0 [ 1 3. Then A11 1 + A 12 2 A =. A 21 1 + A 22 2 One way to partition a matrix is with only vertical or only horizontal lines. In this way we can denote A = [ col 1 (A) col 2 (A) col n () row 1 (A) or = row 2 (A). row n () Theorem 23 (Column-row expansion of A). If A is m n and is n p then A = col 1 (A) row 1 (A) + + col n (A) row n ().

We say a matrix A is block diagonal if A 1 0 0 0 A 2 0 A =...... 0 A n where each A i is a square matrix. If each A i is invertible, then A is invertible and A 1 1 0 0 0 A 1 2 0 A =...... 0 A 1 n 1 2 0 Example 24. Let A = 1 0 0. Find A 1 without using row reduction. 0 0 3 A11 A 12 We say a matrix A is block upper triangular if it is of the form A =. Suppose A 11 is 0 A 21 p p and A 22 is q q. If A is invertible, then we can compute the inverse by setting A11 A 12 11 12 Ip 0 =. 0 A 22 21 22 0 I q Multiplying on the left gives A11 11 + A 12 21 A 11 12 + A 12 22 Ip 0 =. A 22 21 A 22 22 0 I q Since A 22 21 = 0 and A 22 is invertible, then 21 = 0. Hence, the (1,1)-entry becomes A 11 11 = I p, so 11 = A 1 11. Moreover, since A 22 22 = I q, then 22 = A 1 22. Thus, the above reduces to Ip A 11 12 + A 12 A 1 22 Ip 0 =. 0 I q 0 I q Since A 11 12 + A 12 A 1 22 = 0, then 12 = A 1 11 A 12A 1 22 so [ A = A 1 1 11 A 1 11 = A 12A 1 22 0 A 1. 22

6. The Leontief Input-Output Model Suppose we divide the economy of the US into n sectors that produce or output goods and services. We can then denote the output of each sector as a production vector x R n. Another part of the economy only consumes. We denote this by a final demand vector d that lists the values of the goods and services demanded by the nonproductive part of the economy. In between these two are the goods and services needed by the producers to create the products that are in demand. Optimally, one would balance all aspects of the production, but this is very complex. A model for finding this balance is due to Wassily Leontief. Ideally, one would find a production level x such that the amounts produced will balance the total demand, so that { amt produced x } = { int demand } + { final demand d }. The model assumes that for each sector, there is a unit consumption vector in R n that lists the inputs needed per unit of output in the sector (in millions of dollars). Example 25. Suppose the economy consists of three sectors: manufacturing, agriculture, and services, with unit consumption vectors c 1, c 2, and c 3, respectively. Inputs consumed per unit of output Purchased from: Manufacturing Agriculture Services Manufacturing.50.40.20 Agriculture.20.30.10 Services.10.10.30 We have 100c 2 = c 1 c 2 c 3 50 20. Hence, to produce 100 units of output, manufacturing will order (or 10 demand) 50 units from other parts of manufacturing, 20 units from agriculture, and 10 units from services. If manufacturing produces x 1 units of output, then x 1 c 1 represents the intermediate demands of manufacturing, because the amounts in x 1 c 1 will be consumed in the process of creating x 1 units of output. Similarly, x 2 c 2 and x 3 c 3 list the corresponding intermediate demands of agriculture and services, respectively. Hence, { int demand } = x 1 c 1 + x 2 c 2 + x 3 c 3 = Cx where C is the consumption matrix [c 1 c 2 c 3. In this example,.50.40.20 C =.20.30.10.10.10.30

The Leontief Input-Output Model x = Cx + d amt produced int demand final demand This can be rewritten as (I C)x = d where I is the n n identity matrix. Example 26. In the economy from the previous example, suppose the final demand is 50 units for manufacturing, 30 units for agriculture, and 20 units for services. Find the production level x that will satisfy this demand. This amounts to row reducing the agumented matrix [I C d where d is given in the problem. We can compute I C directly and we find that.50.40.20 50 1 0 0 226 C =.20.70.10 30 0 1 0 119..10.10.70 20 0 0 1 78 How else might we solve this problem? Theorem 27. Let C be the consumption matrix for an economy, and let d be the final demand. If C and d have nonnegative entries and if each column sum of C is less than 1, then (I C) 1 exists and the production vector x = (I C) 1 d has nonnegative entries and is the unique solution of x = Cx + d. In a 3 3 system, finding the inverse is not difficult, though a little tedious and we d repeat the row reduction above. However, in the context of this theorem there is another way to get the inverse, or at least a close approximation of it. Suppose the initial demand is set at d. This creates an intermediate demand of Cd. To meet this intermediate demand, industries will need more input. This creates additional intermediate demands of C(Cd) = C 2 d. This process repeats (forever?) and in the next round we have C(C 2 d) = C 3 d. Hence, the production level that will meet this demand is d = d + Cd + C 2 d + C 3 d + = (I + C + C 2 + C 3 + )d. y our hypotheses on C, C m 0 as m. Hence, (I C) 1 I + C + C 2 + + C m where the approximation may be made as close to (I C) 1 as desired by taking m sufficiently large. Under what conditions will this ever be exact? The i, j entry in (I C) 1 is the increased amounts that sector i will have to produce to satisfy an increase of 1 unit in the final demand from sector j.

8. Subspaces of R n So far we have focused on linear transformations whose domain and codomain are exclusively R n. Now we explore subspaces and the concept of a basis. You should think of a basis as an efficient way of writing the span of a set of vectors. Definition 3. A subspace of R n is any set H in R n such that (1) 0 H; (2) If u, v H, then u + v H; (3) If u H and c R, then cu H. {0} and R n are always subspaces of R n. Any other subspace is called proper. 1 Example 28. The set H = k 1 : k R is a subspace of R3. More generally, the span of any 1 set of vectors in R n is a subspace of R n. Example 29. Let L be a line in R n not through the origin. Show that L is not a subspace of R n. Example 30. Define a set H in R 2 by the following property: v H if v has exactly one nonzero entry. Is H a subspace of R 2? Next we ll define two subspaces related to linear transformations. different ways, we ll see next time how they are connected. Though they are defined in Definition 4. Let A be an m n matrix. The column space of a matrix A is the set Col(A) of all linear combinations of the columns of A. The null space is the set Nul(A) of all solutions of the homogeneous equation Ax = 0. If A = [a 1 a n with a i R m, then Col(A) = Span{a 1,..., a n } so Col(A) is a subspace of R m. We will show momentarily that Nul(A) is a subspace of R n. 1 3 4 3 Example 31. Let A = 4 6 2 and b = 3. Is b Col(A)? What is Nul(A)? 3 7 6 Theorem 32. The null space of an m n matrix A is a subspace of R n. Proof. y definition, Nul(A) is a set in R n and clearly A0 = 0, so 0 Nul(A). Let x, y Nul(A) so Ax = 0 and Ay = 0. Then A(x + y) = Ax + Ay = 0 + 0 = 0. Thus, x + y Nul(A). Finally, suppose x Nul(A) and c R. Then A(cx) = c(ax) = c0 = 0, so cx Nul(A). Thus, Nul(A) is a subspace of R n. 4

Definition 5. A basis for a subspace of H of R n is a linearly independent set in H that spans H. If A is n n and invertible then the columns of A are a basis for R n because they are linearly independent and span R n. Recall that e 1,..., e n are called the standard basis vectors for R n (now you know why). Proposition 33. Any basis of R n has size n. Proof. Let {a 1,..., a p } be a basis for R n. We proved in Chapter 1 that if p > n, then the set is linearly dependent and hence not a basis of R n. Hence, p n. Assume all the a i are linearly independent. Then there are exactly p pivots in the RREF of A = [a 1 a p. Hence, if p < n, then there exist b R n for which the equation Ax = b is inconsistent (take b to be all zeros and 1 in the row with no pivots). In this case, the a i do not span R n. Hence, we must have p = n. Example 34. Find a basis for Nul(A) in Example 31. Example 35. Find a basis for the null space of 3 6 1 1 7 A = 1 2 2 3 1 2 4 5 8 4 Example 36. Find a basis for the column space of 1 0 3 5 0 = 0 1 2 1 0 0 0 0 0 1 0 0 0 0 0 Theorem 37. The pivot columns of a matrix A form a basis for the column space of A. Proof. Row reduction does not change the linear relations between the columns. Each pivot column corresponds to a standard basis vector in RREF(A) and hence the pivot columns are linearly independent. Let b be a non-pivot column of A. Since b does not have a pivot, then all of its nonzero entries lie to the right of the pivots. Consequently, it is a linear combination of the pivot columns to its left. Thus, the pivot columns span Col(A).

9. Dimension and Rank Definition 6. The dimension of a nonzero subspace H of R n, denoted dim H, is the number of vectors in any basis of H. The dimension of the zero subspace is defined to be zero. The rank of a matrix A is the dimension of the column space of A (ranka = dim Col A). The nullity of a matrix A is the dimension of the null space of A (nul A = dim Nul A). At the end of this section we will prove that dimension is well-defined (that is, all bases have the same dimension). We have already proved this for R n. 1 3 2 6 Example 38. Find the rank and nullity of A = 3 9 1 5 2 6 1 9. 5 15 0 14 A row reduces to the following matrix. 1 3 0 0 0 0 1 0 0 0 0 1. 0 0 0 0 There are three pivot colums, so ranka = 3. On the other hand, there is one free variable (x 2 ), and so nul A = 1. Theorem 39 (Rank-Nullity Theorem). If a matrix A has n columns, then ranka + nul A = n. Proof. We know that ranka is the number of pivot columns in A and nul A is the number of columns corresponding to free variables. Since each columns is exactly one of these, the result follows. Theorem 40 (The Invertible Matrix Theorem). Let A be an n n matrix. The following are equivalent to the statement that A is invertible. (13) The columns of A form a basis of R n. (14) Col A = R n. (15) dim Col A = n. (16) ranka = n. (17) Nul A = {0}. (18) nul A = 0. Proposition 41. Let H be a subspace of R n with basis = {b 1,..., b p }. If x H, then x can be written in exactly one way as a linear combination of basis vectors in. Proof. Suppose x can be written in two ways: x = c 1 b 1 + + c p b p = d 1 b 1 + + d p b p. Taking differences we find 0 = (c 1 d 1 )b 1 + + (c p d p )b p. y linear independence, c i d i = 0 for all i = 1,..., p. That is, c i = d i for i = 1,..., p. We now want to set up the theorem that says the notion of dimension is well-defined.

Definition 7. Suppose = {b 1,..., b p } is a basis for a subspace H of R n. For each x H, the coordinates of x relative to are the weights c 1,..., c p such that x = c 1 b 1 + + c p b p. The coordinate vector of x relative to (or the -coordinate vector) is [x = 1 c.. The map T : H R p given by x [x is called a coordinate mapping of. 3 1 3 Example 42. Let = 6, 0 and x = 12. Find [x. 2 1 7 Proposition 43. Let = {b 1,..., b p } be a basis for a subspace H of R n. The coordinate mapping T is a 1-1 and onto (bijective) linear transformation. Proof. y Proposition 41, is T is well-defined. Linearity is left as an exercise. Let x = c 1 b 1 + + c p b p H such that T (x) = 0. Then [x = (c i ) = 0 so c i = 0 for all i. Hence, x = 0 and T is 1-1. To prove onto, assume a = (a i ) R p. Then y = a 1 b 1 + + a p b p H because spans H and T = [y = a. Theorem 44. Let H be a subspace of R n with basis = {b 1,..., b p }. Then every basis of H has p elements. Proof. Let {u 1,..., u m } be a set in H with m > p. We claim the set is linearly dependent (and hence cannot be a basis). Note that the set {[u 1,..., [u m } is linearly dependent in R p because m > p. Hence, there exist scalars c 1,..., c m (not all zero) such that c p c 1 [u 1 + + c m [u m = 0. ecause the coordinate mapping is a linear transformation, this implies [c 1 u 1 + + c m u m = 0. Moreover, because the coordinate mapping is 1-1 this implies c 1 u 1 + + c m u m = 0, so the set is linearly dependent. Hence, any basis of H can have at most p elements. Now suppose is another basis with k p elements. y the above argument, can have no more elements than, so p k. This implies p = k so every basis has p elements. Theorem 45 (The asis Theorem). Let H be a p-dimensional subspace of R n. independent set of exactly p elements in H is automatically a basis for H. elements of H that spans H is automatically a basis for H. Any linearly Also, any set of p

Proof. Since H has dimension p, then there exists a basis = {b 1,..., b p }. Let U = {u 1,..., u p } be a set of p linearly independent elements in H. If U spans H, then U is a basis. Otherwise, there exists some element u p+1 / Span U. Hence, {u 1,..., u p, u p+1 } is linearly independent. We can continue this process, but not indefinitely. The number of linearly independent elements in the set cannot exceed n (the dimension of R n ). Thus, our set will eventually span all of H, and hence will be a basis. ut a basis for H cannot have more than p elements by Theorem 44. Hence, U must have been a basis to begin with. Now suppose V = {v 1,..., v p } is a set of p elements in H that span H. If the set is not linearly independent, then one of the vectors is a linear combination of the others. Hence, we can remove that element, say v 1 and the span of {v 2,..., v p } is the same as the span of V. Continue in this way until the remaining elements are linearly independent. ut then the set is linearly independent and spans H, hence is a basis. ut by Theorem 45, every basis has exactly p elements, a contradiction unless V was a basis to begin with. Another way of viewing the previous theorem is the following. The first says that any linearly independent set in H can be extended to a basis. The second says that any spanning set contains a basis.

Change of basis (in R n ) Let be a basis for R n and denote by E the standard basis of R n. The coordinate mapping T is as a map from the standard basis to. { } 1 1 3 Example 46. Let =, and x =. Compute the standard matrix of T. Use 5 4 1 this to find the -coordinates of x. We need to compute T (e 1 ) and T (e 2 ). We know how to do this problem in general, Here we ll just observe that e 1 = 4b 1 + 5b 2 and e 2 = e 1 e 2. Thus, 4 1 T (e 1 ) = [e 1 = and T (e 2 ) = [e 2 =. 5 1 Thus, the standard matrix of T, denoted P, is [ 4 1 13 P = T (e 1 ) T (e 2 ) =, so T (x) = P x =. 5 1 16 4 We can check this, x = 13b 1 + 17b 2 =. 1 In the last section we proved that T is a 1-1 and onto (bijective) linear map and hence it is invertible with inverse T 1. The standard matrix of T 1 1 is P. It s easy to see that T (b i ) = e i (see for yourself in the example above then verify this fact in general). Hence, T 1 [ P 1 = T 1 (e 1) T (e n ) = [b 1 b n. This is just the matrix whose columns are the vectors in the basis! Example 47. Use the above trick to compute P in Example 46. 1 1 4 1 Since P 1 =. Then P = (P 1 5 4 ) 1 =. 5 1 (e i) = b i. Thus, Now say we have two bases of R n, and C. We have the tools we need to change from the basis to the basis C. Consider the following diagram. R n R n C T T C R n E

If we want to go from to C, that is, convert coordinates in to coordinates in C, we need to first go from to E and then from E to C. R n T C T 1 R n C T T C T 1 R n E Definition 8. Suppose we have bases = {b 1,..., b n } and C = {c 1,..., c n } of R n. Then the change of basis transformation is T C : R n R n [x [x C is linear and is the composition T C T 1 where T and T C are the coordinate mappings of and C, respectively. Its standard matrix is P C = P C P 1 and is called the change of basis matrix. { } { } 1 1 1 2 Example 48. Let =, be as before and C =,. Find the change of 5 4 3 5 1 basis matrix P C. Given a vector x R n with [x =, use P C to compute [x C. 2 We previously computed P 1 so it is left to compute P C. We know that 1 2 5 2 P 1 C = so P C = (P 1 C 3 5 ) 1 =. 3 1 Hence, P C = P C P 1 = 5 2 1 1 15 13 =. 3 1 5 4 8 7 To find the coordinates [x C, we apply P C to [x, 15 13 1 11 P C [x = =. 8 7 2 6 We can check by converting both to the standard basis, 1 1 x = (1)b 1 + ( 2)b 2 = and x = ( 11)c 1 + (6)c 2 =. 3 3

Example 49. Let 1 0 2 5 0 7 = 1, 3, 0 and C = 2, 0, 3. 2 1 5 1 1 2 oth and C are bases for R 3 (check this!). Let T and T C be their respective coordinate mappings with standard matrices P and P C, respectively. Then 1 0 2 15 2 6 5 0 7 3 7 0 P 1 = 1 3 0, P = 5 1 2, P 1 C = 2 0 3, P C = 1 3 1. 2 1 5 7 1 3 1 1 2 2 5 0 Thus, Write x R 3 with [x = 3 2 10 21 6 P C = P C P 1 = 6 8 7 7 15 4. Applying our map above gives [x C = P C [x = 1 check by writing both in terms of the standard basis, 1 x = 3b 1 + 2b 2 b 1 = 3 = 18c 1 5c 2 + 13c 3. 3 18 5 13. We The above generalizes somewhat. Let T : R n R n be any linear transformation and, C bases of R n. The map [T,C : R n R n [x [T (x) C is linear. In fact, it is just the composition T C T T 1. If A T is the standard matrix of T, then the standard matrix of [T,C is P C A T P 1. Hence, we have the commutative diagrams. R n T R n R n P R n T T,C A T P,C R n T C R n R n P C R n When T = id R n, then this reduces to the above case.