Vector Space Basics. 1 Abstract Vector Spaces. 1. (commutativity of vector addition) u + v = v + u. 2. (associativity of vector addition)


 Imogene Underwood
 1 years ago
 Views:
Transcription
1 Vector Space Basics (Remark: these notes are highly formal and may be a useful reference to some students however I am also posting Ray Heitmann's notes to Canvas for students interested in a direct computational approach. Chapter 3 of Braun also covers most of this material. I will assume basic knowledge about matrices, matrix multiplication, matrix inversion, row reduction etc.) 1 Abstract Vector Spaces Denition 1. A (real) vector space is a set V with two binary operations + : V V V and : R V V, called vector addition and scalar multiplication respectively, such that all the following properties hold: 1. (commutativity of vector addition) u, v V, u + v = v + u 2. (associativity of vector addition) u, v, w V, u + (v + w) = (u + v) + w 3. (existence of additive identity) z V such that y V, z + y = y + z = y 1
2 (a) (corollary to properties (13): there is only one such z, call it 0 V ) 4. (existence of additive inverse) u V w V such that u + w = w + u = 0 V (a) (we say that w is an additive inverse to u if w + u = u + w = 0 V ; property (4) says that every u V has at least one additive inverse) 5. (scalar multiplication distributes over vector addition) u, v V, c R, c (u + v) = (c u) + (c v) 6. (scalar multiplication distributes over addition in R) u V, a, b R, (a + b) u = (a u) + (b u) 7. (compatibility of scalar multiplication with multiplication in R) u V, a, b R, (ab) u = a (b u) 8. (identity law for scalar multiplication) u V, 1 u = u 2
3 The above list of assumptions is essentially minimal to obtain the full power of linear algebra; the price for this generality is that several obvious results are not actually completely obvious. We will show now that all the usual behavior we desire does follow from the above assumptions. Lemma 2. Let u V ; then, 0 u = 0 V. Also, if c R then c 0 V = 0 V. Proof. By property (6) and the fact that = 0 holds in R, we have: 0 u = (0 + 0) u = (0 u) + (0 u) Let w V be an additive inverse of 0 u; such an element of V does exist by property (4). In particular we have (0 u) + w = w + (0 u) = 0 V Hence 0 V = w + (0 u) = w + [(0 u) + (0 u)] Using the associativity of vector addition, this implies 0 V = [w + (0 u)] + (0 u) But w is an additive inverse of 0 u so this implies (using property (3)) 0 V = 0 V + (0 u) = 0 u 3
4 In particular we have 0 u = 0 V. Taking u = 0 V yields 0 0 V = 0 V, therefore using property (7) we have c 0 V = c (0 0 V ) = (c0) 0 V = 0 0 V = 0 V Lemma 3. Let u V and let v be an additive inverse of u; then, v = ( 1) u. Proof. First, by the previous lemma, ( 1) u is itself an additive inverse of u, because for instance we have: u + (( 1) u) = (1 u) + (( 1) u) = (1 + ( 1)) u = 0 u = 0 V (which properties have been used?) On the other hand since v is an additive inverse of u we have v + u = u + v = 0 V therefore v = v + 0 V = v + [u + (( 1) u)] = = (v + u) + (( 1) u) = 0 V + (( 1) u) = ( 1) u so v = ( 1) u as desired. 4
5 We can conclude from the above lemmas that each u V has a unique additive inverse, and moreover it is equal to ( 1) u. This allows us to dene vector subtraction the following way: if u, v V, u v = u + (( 1) v) In particular vector subtraction : V V V is now its own binary operation, distinct from vector addition, and we can check that the usual rules of subtraction are obeyed. Here are a number of examples of vector spaces, ranging from very concrete to highly abstract: Example 4. R n is a vector space if, for every x = (x 1, x 2,..., x n ) R n and y = (y 1, y 2,..., y n ) R n, and c R, we dene x + y = (x 1 + y 1, x 2 + y 2,..., x n + y n ) c x = (cx 1, cx 2,..., cx n ) Example 5. The set of all m n real matrices, M mn, is a vector space under entrywise addition and entrywise scalar multiplication (the usual addition and scalar multiplication operations for matrices). 5
6 Example 6. Let A M mn be a xed matrix. Then the set of all x R n such that Ax = 0 R m is a vector space under the usual vector addition and scalar multiplication operations of R n. Example 7. Let A M mn be a xed matrix and let 0 R m b R m be a nonzero vector such that the inhomogeneous equation Ax = b has at least one solution, say x 0. Then the set of all x R n such that Ax = b is not a vector space under the usual operations of R n. However, it is a vector space under the following operations, call them and, where x, y R n and c R: x y = (x x 0 ) + (y x 0 ) + x 0 c x = c (x x 0 ) + x 0 (It is not at all obvious that these operations dene a vector space it has to be checked carefully.) Note that a dierent choice of x 0 would result in dierent vector space operations on the set of solutions of the inhomogeneous system, because for instance x x = x if and only if x = x 0. Example 8. The set of all polynomials with degree at most 5, call it P 5, is a vector space under the following operations: for any two polynomials p, q, 6
7 and c R, (p + q) (t) = p (t) + q (t) (c p) (t) = cp (t) Note that P 5 is in some sense the same as R 6, because the addition and scalar multiplication of fthdegree polynomials is the same as addition and scalar multiplication of their coecients, which are six in number. Example 9. The set of all polynomials (in one variable) with degree at most n, written P n, is a vector space under the same operations as in Example 8. (note that P n is somehow the same as R n+1 ). Example 10. The set P of all polynomials (in one variable) is a vector space under the same operations as in Example 8. These can be expressed as p (t) = n=0 a nt n where all but nitely many a n 's are zero. Example 11. The set of all functions f : R R is a vector space under the following operations: for two functions f, g : R R and c R, (f + g) (t) = f (t) + g (t) (c f) (t) = cf (t) 7
8 Example 12. The set of all power series with nonzero radius of convergence is a vector space under the usual addition and scalar multiplication. (This requires us to prove that the sum of two power series with nonzero radius of convergence again has nonzero radius of convergence.) Example 13. The set of all formal power series (which may have radius of convergence equal to zero) is a vector space under the following operations: a n X n + n=0 b n X n = n=0 ( ) c a n X n = n=0 (a n + b n ) X n n=0 (ca n ) X n Note that in this denition we cannot generally substitute any real value for X (except perhaps X = 0) because we are not guaranteed any convergence. For example n=0 n!xn is a valid formal power series but it does not converge for any real X 0. The formal power series is a useful construction when you want to prove abstract results which would hold for any convergent series, without actually considering the detailed convergence process in your proof. n=0 8
9 1.1 Subspaces. Checking that you have a dened a vector space is usually a two step process. First you have to show that the vector and scalar multiplication operations actually map into V (and not some bigger set or nowhere at all), and second you have to verify all eight properties stated in the denition of a vector space. This is extremely painful and tedious in general which is why we introduce the following denition: Denition 14. Let V be a vector space, with vector addition + and scalar multiplication, and let S be a subset of V such that 0 V S. Suppose that the following two properties hold: x, y S, x + y S x S, c R, c x S Then it follows from the denition of vector spaces that S is itself a vector space with operations + and inherited from V. We say that the set S, equipped with the operations + and from V, is a subspace of V. Remark 15. Note that usually we only know that x + y V and c x V ; thus only very special subsets of V can be subspaces. The convenience of this denition is the following: we already have a number of mathematical objects which we know are vector spaces (such as R n, or the set of all polynomials in one variable). Hence if we have a set 9
10 S which is embedded in some vector space V, and we can show that S is closed under the operations of V, then S is automatically a vector space under those same operations. (If we want to put dierent operations on S then this trick does not work!) Remark 16. Clearly an equivalent denition is obtained if we let S be a nonempty subset of V which is closed under addition and scalar multiplication (then automatically it follows that 0 V S). The empty set is never a vector space since it does not contain a zero element (because it does not contain any elements!). On the other hand if V is any vector space then the singleton set {0 V } is actually a subspace of V, and a vector space in its own right. Most theorems we will prove will either hold trivially or fail trivially for the trivial vector space {0 V }. When theorems fail for the trivial vector space we will try to say for any nontrivial vector space V... When theorems hold trivially for the trivial vector space we will not provide a separate proof for this case (since it is trivial). From now on (unless stated otherwise) we will write cx instead of c x for scalar multiplication, and + will be understood as vector addition as above; additionally, 0 V will simply be written 0. Example 17. Let A M mn, then the following set S = {x R n such that Ax = 0} is a subspace of R n (under the usual addition and scalar multiplication in 10
11 R n ). To prove this, note that if Ax = 0 and Ay = 0 and c R, then A (x + y) = 0 and A (cx) = 0; and, clearly 0 S. 2 Linear Transformations Denition 18. Let V, W be vector spaces and suppose T : V W is a map. (This means that for every v V, there is an assignment T (v) W ; the word map is interchangeable with the word function.) We say that T is a linear map, or a linear transformation, if both of the following properties hold: u, v V, T (u + v) = T (u) + T (v) u V, c R, T (cu) = ct (u) (Note that in either formula, the operations on the left hand side occur in V whereas the operations on the right hand side occur in W.) We sometimes abbreviate T u in place of T (u) when T is a linear transformation. Proposition 19. Let U, V, W be vector spaces and suppose T : U V and S : V W are linear transformations. Then the composition S T : U W is also a linear transformation. Denition 20. Let V be a vector space; then we dene the identity transformation Id V : V V by Id V (v) = v. Lemma 21. The identity transformation on any vector space V is a linear transformation. 11
12 Lemma 22. Let V, W be vector spaces and let T : V W be a linear transformation. Then T Id V = T and Id W T = T. Example 23. Let A M mn be any m n real matrix, and dene the map T : R n R m by T x = Ax In other words T x is what we get if we view x R n as a column vector and multiply A from the left. That this denes a linear transformation follows from the properties of matrix multiplication. Example 24. Let P n denote the space of polynomials with degree at most n. Dene the map T : P n P n 1 by T p = p where p (t) = n k=0 ka kt k 1 is the derivative of p (t) = n k=0 a kt k. By the properties of dierentiation, this denes a linear transformation. Example 25. Dene the map T : M mn M nm by T A = A T 12
13 that is T takes A to the transpose of A. Then T is a linear transformation by the properties of transpose. Example 26. Fix any nonsingular matrix B M nn and dene the map T : M nn M nn by T A = B 1 AB Then T is a linear transformation by the properties of matrix multiplication. Sometimes we are interested in linear transformations that completely identify two spaces, so that (at least as far as vector space structure is concerned) the two spaces are the same. Denition 27. Let V, W be vector spaces and let T : V W be a linear map. We say that T is one to one if the following property holds: u, v V, T (u) = T (v) = u = v Equivalently (by contrapositive) T is one to one if u v implies T (u) T (v); that is, distinct points map to distinct points. Denition 28. Let V, W be vector spaces and let T : V W be a linear 13
14 map. We say that T is onto if w W v V such that T (v) = w In other words T is onto if its range is all of W. Denition 29. Let V, W be vector spaces and let T : V W be a linear map. We say that T is a linear isomorphism if it is onetoone and onto. If there exists a linear isomorphism T : V W then we say that V and W are linearly isomorphic. These denitions are highly abstract so we will try to make them more concrete with some examples. Example 30. Let P 5 denote the space of polynomials with degree at most ve. Dene the map T : R 6 P 5 by T ((a 0,..., a 5 )) = p (a0,...,a 5 ) where p (a0,...,a 5 ) (t) = 5 a k t k k=0 Then T : R 6 P 5 is a linear isomorphism. Denition 31. Let V, W be vector spaces and let T : V W be a linear transformation. We say that T is invertible if there exists a linear trans 14
15 formation S : W V such that S T = Id V and T S = Id W. Such a transformation S is called an inverse of T. Remark 32. If T : V W is a linear transformation of vector spaces V, W and there exists a map S : W V (not assumed linear) such that S T = Id V and T S = Id W, then it automatically follows that S is a linear transformation. The proof is two lines: Sw 1 + Sw 2 = S (T (Sw 1 + Sw 2 )) = S (T Sw 1 + T Sw 2 ) = S (w 1 + w 2 ) cs (w) = S (T (cs (w))) = S (ct Sw) = S (cw) Proposition 33. A linear transformation has at most one inverse. Proof. Let T : V W be a linear transformation of vector spaces V, W ; furthermore, suppose that T has two inverses, S 1 : W V and S 2 : W V. Then we have S 1 = S 1 Id W = S 1 (T S 2 ) = (S 1 T ) S 2 = Id V S 2 = S 2 hence S 1 = S 2. Since a linear transformation T can have at most one inverse, when it has one we call it the inverse of T and we write it T 1. Example 34. Let A M nn be a nonsingular matrix with inverse matrix 15
16 A 1. Dene the linear transformations S, T : R n R n by Sx = Ax T x = A 1 x Then S, T are both invertible linear transformations; furthermore, S 1 = T and T 1 = S. Theorem 35. Let V, W be vector spaces and let T : V W be a linear transformation. Then T is invertible if and only if T is a linear isomorphism. Proof. Assume T is invertible, with inverse T 1 : W V. We see that T is onetoone because v 1 v 2 = T 1 (T (v 1 v 2 )) = T 1 (T v 1 T v 2 ) so if T v 1 = T v 2 then v 1 = v 2. Additionally for any w W we have w = T ( T 1 w ) so w = T v where v = T 1 w; hence, T is onto. Since T is onetoone and onto, T is a linear isomorphism. Now suppose instead that T is a linear isomorphism. Since T is onto, for any w W there is some v V such that T v = w; morover, since T is onetoone, there can be at most one such v. Therefore we can dene a map 16
17 S : W V so that Sw is the unique vector v V such that T v = w. We have w W, T (Sw) = w by denition of S, and therefore T S = Id W. Additionally, v V, S (T v) = v and again this follows from the denition of S, so S T = Id V. We easily show that S is a linear transformation; altogether we can conclude that S is an inverse of T, and in particular T is invertible. Remark 36. If T : V W is a linear isomorphism then T 1 : W V is also a linear isomorphism. 3 Linear Independence, Bases, Dimension Denition 37. Let V be a vector space. A subset E V is said to be linearly dependent if there exists a nite collection of distinct elements v 1, v 2,..., v N E, and scalars c 1, c 2,..., c N R, such that at least one c i 0 and c 1 v 1 + c 2 v c N v N = 0 Denition 38. Let V be a vector space. A subset E V is said to be 17
18 linearly independent if it is not linearly dependent. Remark 39. The empty set is linearly independent. Denition 40. Let V be a vector space; we say that v V is a linear combination (or nite linear combination) of the vectors v 1, v 2,..., v N V if there exist scalars c 1, c 2,..., c N R such that v = c 1 v 1 + c 2 v c N v N Lemma 41. If V is a vector space and E V is a subset, then E is linearly dependent if and only if there exists a vector v E which is a linear combination of other elements v 1, v 2,..., v N E. Denition 42. Let V be a vector space and let E V be a subset. Then we dene span E to be the set of all (nite) linear combinations of elements of E. We also dene span = {0} V. Lemma 43. If V is a vector space and E V then span E is a subspace of V. Moreover if E W V and W is a subspace of V, then span E W. Thus span E is the smallest subspace of V containing every element of E. Denition 44. Let V be a vector space and let B V be a subset. We say that B is a basis of V if B is linearly independent and span B = V. 18
19 Denition 45. Let V be a vector space; if there exists a subset B V such that B is a basis of V and B is a nite set, then we say that V is nite dimensional. If V is not nite dimensional then we say that V is innite dimensional. Example 46. The space P n of polynomials (in one variable) with degree at most n is nite dimensional, because the set {1, t, t 2,..., t n } is a basis of P n. The space P of all polynomials (in one variable) is innite dimensional. Lemma 47. Let V, W be vector spaces and let T : V W be a linear transformation. If T is a linear isomorphism and V is nite dimensional then W is also nite dimensional. Lemma 48. Let V be an innite dimensional vector space; then, for every n N there exists a linearly independent subset E V such that E has exactly n elements. Proof. Use induction. For n = 1 this is trivial: let E = {x 0 } for any 0 x 0 V (such an x 0 exists because the trivial vector space is nite dimensional). Suppose now that for some n N there exist a linearly independent subset E V such that E has exactly n elements. We claim that span E V ; indeed, if this were not the case then V would be nite dimensional. Therefore, there exists a vector z V such that z / span E. Then E {z} is a linearly independent subset of V having exactly n + 1 elements. 19
20 Lemma 49. Let V be a nite dimensional vector space, with a nite basis B having exactly N elements. Then every linearly independent subset of V has at most N elements. Proof. Let E V be a linearly independent subset; we will assume E has at least N + 1 elements to reach a contradiction, hence proving the lemma. Let v 1, v 2,..., v N+1 E be N + 1 distinct elements of E. Since B is a basis of V, each v V is a linear combination of elements of B. Denoting the elements of B as w 1, w 2,..., w N we have numbers c i,j R such that v 1 = c 1,1 w 1 + c 2,1 w c N,1 w N v 2 = c 1,2 w 1 + c 2,2 w c N,2 w N. v N+1 = c 1,N+1 w 1 + c 2,N+1 w c N,N+1 w N Arrange the numbers c i,j as the following N (N + 1) matrix: C = c 1,1 c 1,2... c 1,N+1 c 2,1 c 2,2... c 2,N c N,1 c N,2 c N,N+1 20
21 We can solve the equation Cx = 0 (with x R N+1 ) by row reduction. Now in reduced row echelon form (RREF) each row can have at most one pivot; since there are N rows, there can be at most N pivots in the RREF. Hence there is at least one free variable, which can take on any real value. Therefore there are innitely many solutions to the equation Cx = 0, and this certainly implies that there exists some x 0 such that Cx = 0. Call this vector x = ( x 1, x 2,..., x N+1 ) 0. Now consider the following vector: ṽ = x 1 v 1 + x 2 v x N+1 v N+1 Using the more compact summation notation, this can be written But v k = N j=1 c j,kw j, therefore Rearranging, this says ṽ = ṽ = N+1 k=1 N+1 k=1 x k v k N c j,k x k w j j=1 ṽ = N j=1 ( N+1 k=1 c j,k x k ) w j But N+1 k=1 c j,k x k is just the jth entry of the vector Cx, and by construction 21
22 Cx = 0. Therefore N+1 k=1 c j,k x k for each j {1, 2,..., N} and we have ṽ = N 0w j = 0 j=1 Hence ṽ = 0. Then again we have ṽ = N+1 k=1 x kv k, hence x 1 v 1 + x 2 v x N+1 v N+1 = 0 Since v 1, v 2,..., v N+1 are distinct elements of E, and the numbers x k are not all zero, this implies that the set E is not linearly independent, so we have a contradiction. Theorem 50. Let V be a vector space and let W V be a subspace. If V is nite dimensional then W is nite dimensional. Proof. Suppose W is innite dimensional. By Lemma 48, for each natural number n there is a linearly independent subset E of W having exactly n elements. But a linearly independent subset of W is also a linearly independent subset of V. Therefore, for each natural number n there is a linearly independent subset E of V having exactly n elements. On the other hand, V is nite dimensional so it has a nite basis B. Let N be the number of elements of B. Then by Lemma 49, any linearly independent subset of V has at most N elements. But we just said that, for every n N, V has a linearly independent subset E having exactly n elements; choosing n = N + 1 yields the contradiction. 22
23 Remark 51. Note that in Theorem 50, we have proven that if W is a subspace of the nite dimensional space V then W is nite dimensional; in particular, W has a nite basis. However we did not actually construct any particular basis for W ; indeed, given a basis B of V, it is entirely possible that B W =. Due to Theorem 50, we do not always have to exhibit a nite basis to show that a vector space W is nite dimensional; it is sucient to show that W is linearly isomorphic to a subspace of a nite dimensional space. Example 52. Let V be the set of all smooth functions f : R R such that t R, f (t) f (t) = 0 Now V is a vector space, and it is a subspace of the space of all smooth functions on R, but that larger vector space is not nite dimensional. To show that V is nite dimensional, we can dene the following map T : R 2 V T c 1 c 2 = c 1 f 1 + c 2 f 2 where f 1 (t) = e t and f 2 (t) = e t. Then T is onetoone because the Wronskian W [f 1, f 2 ] = 2 0; also, T is onto because all solutions of the ODE are of the form c 1 f 1 + c 2 f 2 for some constants c 1, c 2. Hence V is linearly isomorphic to the nite dimensional space R 2, so V is itself nite dimensional. (Note that we could equally well observe that {f 1, f 2 } is a basis of V in order to conclude that V is nite dimensional.) 23
24 Theorem 53. Let V be a nite dimensional vector space; furthermore, let B 1 be a basis of V, and suppose B 2 is also a basis of V. Then B 1 and B 2 are both nite sets and they have the same number of elements. Proof. Since V is nite dimensional, there is a nite basis of V, call it B 0. Let N 0 N {0} be the number of elements of B 0. By Lemma 49, since B 0 is a nite basis of V and B 1 is a linearly independent subset of V, we nd that B 1 has at most N 0 elements. Moreover, again by Lemma 49, since B 0 is a nite basis of V and B 2 is a linearly independent subset of V, we nd that B 2 has at most N 0 elements. In particular, both B 1 and B 2 are nite bases. Now since B 1, B 2 are nite sets, let N 1, N 2 N {0} denote (respectively) the size of B 1, B 2. By Lemma 49, since B 1 is a nite basis of V and B 2 is a linearly independent subset of V, we have that N 2 N 1. Then again, since B 2 is a nite basis of V and B 1 is a linearly indepdendent subset of V, we have that N 1 N 2. Therefore, N 1 = N 2. Denition 54. Let V be a nite dimensional vector space (then V has a nite basis because that is what it means to be nite dimensional). The dimension of V is dened to be the number of elements in a basis of V ; by Theorem 53, it does not matter which basis we choose. We write the dimension of V as dim V. If V is innite dimensional we may write dim V = as a convenient (but nonrigorous) shorthand. Theorem 55. Let V, W be vector spaces and let T : V W be a linear transformation. If T is a linear isomorphism, and either V or W is nite 24
25 dimensional, then both V and W are nite dimensional and dim V = dim W. Proof. Simply observe, if V is nite dimensional, that the image of any basis of V under T is a basis of W. Similarly, if W is nite dimensional, then the image of any basis of W under T 1 is a basis of V. Example 56. (Euclidean space) dim R n = n Example 57. (polynomials of degree at most n) dim P n = n + 1 Example 58. (all m n real matrices) dim M mn = mn Example 59. (all polynomials) dim P = Theorem 60. Let V be a nitedimensional vector space with dim V = n. Then V is linearly isomorphic to R n. Proof. Let B = {v 1, v 2,..., v n } be a nite basis of V. Dene a map T : R n V as follows: T c 1 c 2. c n = c 1 v 1 + c 2 v c n v n 25
26 It is trivial to check that T is a linear map. Clearly T is onto, since any v V can be written as a linear combination of the vectors v 1, v 2,..., v n (since B is a basis). So it only remains to show that T is onetoone. Suppose there are numbers c 1, c 2,..., c n and c 1, c 2,..., c n such that T c 1 c 2. c n = T c 1 c 2. c n Then by the denition of T we have c 1 v 1 + c 2 v c n v n = c 1 v 1 + c 2 v c n v n Therefore (c 1 c 1 ) v 1 + (c 2 c 2 ) v (c n c n ) v n = 0 But B is a basis, hence linearly independent, so we conclude c 1 c 1 = 0, c 2 c 2 = 0,..., c n c n = 0. Hence c 1 c 2. c n = c 1 c 2. c n 26
27 so T is onetoone. 4 Matrix Representation of Linear Transformations We have seen in Theorem 60 that, just by choosing a basis, any nitedimensional vector space can be regarded as equivalent (in the sense of linear isomorphism) to a copy of R n. (Note carefully that extra structures, such as dot products, are not necessarily preserved even for linear isomorphisms from R n to itself.) We have also seen that if A M mn is a matrix then A denes a linear transformation R n R m by leftmultiplication of any (column) vector x R n. What we are going to show is that any linear transformation R n R m arises as leftmultiplication by some m n matrix. Though we will not go into all the details (which you can nd in any linear algebra textbook), by combining this result with Theorem 60, any linear transformation of nitedimensional vector spaces V and W can be represented by a matrix. Of course the matrix will depend on your choice of bases for V and W ; there is a standard rule (written in any linear algebra textbook) for transforming the matrix of a linear transformation from one pair of bases to another pair. We will not discuss those details. Theorem 61. Let T : R n R m be a linear transformation (where elements 27
28 of R n and R m are regarded as column vectors). Then there exists a unique m n matrix A M mn such that x R n, T x = Ax where Ax is the usual matrixvector product. Proof. Let us rst prove the uniqueness. Suppose that A, B M mn are two matrices that both coincide with T ; in that case, we clearly have x R n, Ax = Bx Therefore, taking x = e j (with 1 j n) and dotting both sides against e i (with 1 i m) we have 1 j n, 1 i m, e T i Ae j = e T i Be j But this is equivalent to the following statement: 1 j n, 1 i m, a ij = b ij in particular A = B. Now we turn to the existence. Let us write y j = T e j R m for 1 j n; furthermore, let us dene the numbers a ij, with 1 i m and 1 j n, 28
29 by the following formula: a ij = e T i y j Dene the matrix A as follows: A = a 11 a a 1n a 21 a a 2n a m1 a m2 a mn Clearly Ae j = y j. Let x R n ; then we can write x = c 1 e 1 + c 2 e c n e n = n k=1 c ke k. Therefore, On the other hand, T x = Ax = n c k T e k = k=1 n c k Ae k = k=1 n c k y k k=1 n c k y k Since both T x and Ax are equal to n k=1 c ky k, it follows that T x = Ax. But x R n was arbitrary so we conclude that k=1 x R n, T x = Ax 29
Math Linear Algebra Final Exam Review Sheet
Math 151 Linear Algebra Final Exam Review Sheet Vector Operations Vector addition is a componentwise operation. Two vectors v and w may be added together as long as they contain the same number n of
More informationELEMENTARY LINEAR ALGEBRA WITH APPLICATIONS. 1. Linear Equations and Matrices
ELEMENTARY LINEAR ALGEBRA WITH APPLICATIONS KOLMAN & HILL NOTES BY OTTO MUTZBAUER 11 Systems of Linear Equations 1 Linear Equations and Matrices Numbers in our context are either real numbers or complex
More informationInverses and Elementary Matrices
Inverses and Elementary Matrices 1122013 Matrix inversion gives a method for solving some systems of equations Suppose a 11 x 1 +a 12 x 2 + +a 1n x n = b 1 a 21 x 1 +a 22 x 2 + +a 2n x n = b 2 a n1 x
More informationEXERCISE SET 5.1. = (kx + kx + k, ky + ky + k ) = (kx + kx + 1, ky + ky + 1) = ((k + )x + 1, (k + )y + 1)
EXERCISE SET 5. 6. The pair (, 2) is in the set but the pair ( )(, 2) = (, 2) is not because the first component is negative; hence Axiom 6 fails. Axiom 5 also fails. 8. Axioms, 2, 3, 6, 9, and are easily
More informationMAT 242 CHAPTER 4: SUBSPACES OF R n
MAT 242 CHAPTER 4: SUBSPACES OF R n JOHN QUIGG 1. Subspaces Recall that R n is the set of n 1 matrices, also called vectors, and satisfies the following properties: x + y = y + x x + (y + z) = (x + y)
More informationMATH 304 Linear Algebra Lecture 10: Linear independence. Wronskian.
MATH 304 Linear Algebra Lecture 10: Linear independence. Wronskian. Spanning set Let S be a subset of a vector space V. Definition. The span of the set S is the smallest subspace W V that contains S. If
More information4.3  Linear Combinations and Independence of Vectors
 Linear Combinations and Independence of Vectors De nitions, Theorems, and Examples De nition 1 A vector v in a vector space V is called a linear combination of the vectors u 1, u,,u k in V if v can be
More informationLinear Algebra Massoud Malek
CSUEB Linear Algebra Massoud Malek Inner Product and Normed Space In all that follows, the n n identity matrix is denoted by I n, the n n zero matrix by Z n, and the zero vector by θ n An inner product
More informationMath 110, Spring 2015: Midterm Solutions
Math 11, Spring 215: Midterm Solutions These are not intended as model answers ; in many cases far more explanation is provided than would be necessary to receive full credit. The goal here is to make
More informationNOTES (1) FOR MATH 375, FALL 2012
NOTES 1) FOR MATH 375, FALL 2012 1 Vector Spaces 11 Axioms Linear algebra grows out of the problem of solving simultaneous systems of linear equations such as 3x + 2y = 5, 111) x 3y = 9, or 2x + 3y z =
More informationAbstract Vector Spaces and Concrete Examples
LECTURE 18 Abstract Vector Spaces and Concrete Examples Our discussion of linear algebra so far has been devoted to discussing the relations between systems of linear equations, matrices, and vectors.
More informationNONCOMMUTATIVE POLYNOMIAL EQUATIONS. Edward S. Letzter. Introduction
NONCOMMUTATIVE POLYNOMIAL EQUATIONS Edward S Letzter Introduction My aim in these notes is twofold: First, to briefly review some linear algebra Second, to provide you with some new tools and techniques
More informationA matrix is a rectangular array of. objects arranged in rows and columns. The objects are called the entries. is called the size of the matrix, and
Section 5.5. Matrices and Vectors A matrix is a rectangular array of objects arranged in rows and columns. The objects are called the entries. A matrix with m rows and n columns is called an m n matrix.
More information290 J.M. Carnicer, J.M. Pe~na basis (u 1 ; : : : ; u n ) consisting of minimally supported elements, yet also has a basis (v 1 ; : : : ; v n ) which f
Numer. Math. 67: 289{301 (1994) Numerische Mathematik c SpringerVerlag 1994 Electronic Edition Least supported bases and local linear independence J.M. Carnicer, J.M. Pe~na? Departamento de Matematica
More informationAPPENDIX: MATHEMATICAL INDUCTION AND OTHER FORMS OF PROOF
ELEMENTARY LINEAR ALGEBRA WORKBOOK/FOR USE WITH RON LARSON S TEXTBOOK ELEMENTARY LINEAR ALGEBRA CREATED BY SHANNON MARTIN MYERS APPENDIX: MATHEMATICAL INDUCTION AND OTHER FORMS OF PROOF When you are done
More informationMath Linear Algebra II. 1. Inner Products and Norms
Math 342  Linear Algebra II Notes 1. Inner Products and Norms One knows from a basic introduction to vectors in R n Math 254 at OSU) that the length of a vector x = x 1 x 2... x n ) T R n, denoted x,
More informationMATH10212 Linear Algebra B Homework Week 4
MATH22 Linear Algebra B Homework Week 4 Students are strongly advised to acquire a copy of the Textbook: D. C. Lay Linear Algebra and its Applications. Pearson, 26. ISBN 5228734. Normally, homework
More informationMath Camp Lecture 4: Linear Algebra. Xiao Yu Wang. Aug 2010 MIT. Xiao Yu Wang (MIT) Math Camp /10 1 / 88
Math Camp 2010 Lecture 4: Linear Algebra Xiao Yu Wang MIT Aug 2010 Xiao Yu Wang (MIT) Math Camp 2010 08/10 1 / 88 Linear Algebra Game Plan Vector Spaces Linear Transformations and Matrices Determinant
More informationLECTURE VI: SELFADJOINT AND UNITARY OPERATORS MAT FALL 2006 PRINCETON UNIVERSITY
LECTURE VI: SELFADJOINT AND UNITARY OPERATORS MAT 204  FALL 2006 PRINCETON UNIVERSITY ALFONSO SORRENTINO 1 Adjoint of a linear operator Note: In these notes, V will denote a ndimensional euclidean vector
More information1. General Vector Spaces
1.1. Vector space axioms. 1. General Vector Spaces Definition 1.1. Let V be a nonempty set of objects on which the operations of addition and scalar multiplication are defined. By addition we mean a rule
More informationChapter 7. Linear Algebra: Matrices, Vectors,
Chapter 7. Linear Algebra: Matrices, Vectors, Determinants. Linear Systems Linear algebra includes the theory and application of linear systems of equations, linear transformations, and eigenvalue problems.
More informationMATRIX ALGEBRA AND SYSTEMS OF EQUATIONS. + + x 1 x 2. x n 8 (4) 3 4 2
MATRIX ALGEBRA AND SYSTEMS OF EQUATIONS SYSTEMS OF EQUATIONS AND MATRICES Representation of a linear system The general system of m equations in n unknowns can be written a x + a 2 x 2 + + a n x n b a
More information1 Linear transformations; the basics
Linear Algebra Fall 2013 Linear Transformations 1 Linear transformations; the basics Definition 1 Let V, W be vector spaces over the same field F. A linear transformation (also known as linear map, or
More informationLinear Algebra Notes. Lecture Notes, University of Toronto, Fall 2016
Linear Algebra Notes Lecture Notes, University of Toronto, Fall 2016 (Ctd ) 11 Isomorphisms 1 Linear maps Definition 11 An invertible linear map T : V W is called a linear isomorphism from V to W Etymology:
More informationDot Products, Transposes, and Orthogonal Projections
Dot Products, Transposes, and Orthogonal Projections David Jekel November 13, 2015 Properties of Dot Products Recall that the dot product or standard inner product on R n is given by x y = x 1 y 1 + +
More information4 Vector Spaces. 4.1 Basic Definition and Examples. Lecture 10
Lecture 10 4 Vector Spaces 4.1 Basic Definition and Examples Throughout mathematics we come across many types objects which can be added and multiplied by scalars to arrive at similar types of objects.
More informationa11 a A = : a 21 a 22
Matrices The study of linear systems is facilitated by introducing matrices. Matrix theory provides a convenient language and notation to express many of the ideas concisely, and complicated formulas are
More informationMATH 2030: MATRICES. Example 0.2. Q:Define A 1 =, A. 3 4 A: We wish to find c 1, c 2, and c 3 such that. c 1 + c c
MATH 2030: MATRICES Matrix Algebra As with vectors, we may use the algebra of matrices to simplify calculations. However, matrices have operations that vectors do not possess, and so it will be of interest
More information2 so Q[ 2] is closed under both additive and multiplicative inverses. a 2 2b 2 + b
. FINITEDIMENSIONAL VECTOR SPACES.. Fields By now you ll have acquired a fair knowledge of matrices. These are a concrete embodiment of something rather more abstract. Sometimes it is easier to use matrices,
More informationOur goal is to solve a general constant coecient linear second order. this way but that will not always happen). Once we have y 1, it will always
October 5 Relevant reading: Section 2.1, 2.2, 2.3 and 2.4 Our goal is to solve a general constant coecient linear second order ODE a d2 y dt + bdy + cy = g (t) 2 dt where a, b, c are constants and a 0.
More information3  Vector Spaces Definition vector space linear space u, v,
3  Vector Spaces Vectors in R and R 3 are essentially matrices. They can be vieed either as column vectors (matrices of size and 3, respectively) or ro vectors ( and 3 matrices). The addition and scalar
More informationChapter Two Elements of Linear Algebra
Chapter Two Elements of Linear Algebra Previously, in chapter one, we have considered single first order differential equations involving a single unknown function. In the next chapter we will begin to
More informationIMPORTANT DEFINITIONS AND THEOREMS REFERENCE SHEET
IMPORTANT DEFINITIONS AND THEOREMS REFERENCE SHEET This is a (not quite comprehensive) list of definitions and theorems given in Math 1553. Pay particular attention to the ones in red. Study Tip For each
More informationIMPORTANT DEFINITIONS AND THEOREMS REFERENCE SHEET
IMPORTANT DEFINITIONS AND THEOREMS REFERENCE SHEET This is a (not quite comprehensive) list of definitions and theorems given in Math 1553. Pay particular attention to the ones in red. Study Tip For each
More informationSolutions to Math 51 First Exam April 21, 2011
Solutions to Math 5 First Exam April,. ( points) (a) Give the precise definition of a (linear) subspace V of R n. (4 points) A linear subspace V of R n is a subset V R n which satisfies V. If x, y V then
More informationLecture 8: A Crash Course in Linear Algebra
Math/CS 120: Intro. to Math Professor: Padraic Bartlett Lecture 8: A Crash Course in Linear Algebra Week 9 UCSB 2014 Qué sed de saber cuánto! Pablo Neruda, Oda a los Números 1 Linear Algebra In the past
More information7 Matrix Operations. 7.0 Matrix Multiplication + 3 = 3 = 4
7 Matrix Operations Copyright 017, Gregory G. Smith 9 October 017 The product of two matrices is a sophisticated operations with a wide range of applications. In this chapter, we defined this binary operation,
More information(1.) For any subset P S we denote by L(P ) the abelian group of integral relations between elements of P, i.e. L(P ) := ker Z P! span Z P S S : For ea
Torsion of dierentials on toric varieties Klaus Altmann Institut fur reine Mathematik, HumboldtUniversitat zu Berlin Ziegelstr. 13a, D10099 Berlin, Germany. Email: altmann@mathematik.huberlin.de Abstract
More informationMath 24 Spring 2012 Questions (mostly) from the Textbook
Math 24 Spring 2012 Questions (mostly) from the Textbook 1. TRUE OR FALSE? (a) The zero vector space has no basis. (F) (b) Every vector space that is generated by a finite set has a basis. (c) Every vector
More information0.2 Vector spaces. J.A.Beachy 1
J.A.Beachy 1 0.2 Vector spaces I m going to begin this section at a rather basic level, giving the definitions of a field and of a vector space in much that same detail as you would have met them in a
More informationReview of Linear Algebra
Review of Linear Algebra Throughout these notes, F denotes a field (often called the scalars in this context). 1 Definition of a vector space Definition 1.1. A F vector space or simply a vector space
More informationChapter 4 & 5: Vector Spaces & Linear Transformations
Chapter 4 & 5: Vector Spaces & Linear Transformations Philip Gressman University of Pennsylvania Philip Gressman Math 240 002 2014C: Chapters 4 & 5 1 / 40 Objective The purpose of Chapter 4 is to think
More informationMatrices and Vectors. Definition of Matrix. An MxN matrix A is a twodimensional array of numbers A =
30 MATHEMATICS REVIEW G A.1.1 Matrices and Vectors Definition of Matrix. An MxN matrix A is a twodimensional array of numbers A = a 11 a 12... a 1N a 21 a 22... a 2N...... a M1 a M2... a MN A matrix can
More informationAnswers in blue. If you have questions or spot an error, let me know. 1. Find all matrices that commute with A =. 4 3
Answers in blue. If you have questions or spot an error, let me know. 3 4. Find all matrices that commute with A =. 4 3 a b If we set B = and set AB = BA, we see that 3a + 4b = 3a 4c, 4a + 3b = 3b 4d,
More informationTopics. Vectors (column matrices): Vector addition and scalar multiplication The matrix of a linear function y Ax The elements of a matrix A : A ij
Topics Vectors (column matrices): Vector addition and scalar multiplication The matrix of a linear function y Ax The elements of a matrix A : A ij or a ij lives in row i and column j Definition of a matrix
More informationMA 0540 fall 2013, Row operations on matrices
MA 0540 fall 2013, Row operations on matrices December 2, 2013 This is all about m by n matrices (of real or complex numbers). If A is such a matrix then A corresponds to a linear map F n F m, which we
More informationVector Spaces. (1) Every vector space V has a zero vector 0 V
Vector Spaces 1. Vector Spaces A (real) vector space V is a set which has two operations: 1. An association of x, y V to an element x+y V. This operation is called vector addition. 2. The association of
More informationT ((x 1, x 2,..., x n )) = + x x 3. , x 1. x 3. Each of the four coordinates in the range is a linear combination of the three variables x 1
MATH 37 Linear Transformations from Rn to Rm Dr. Neal, WKU Let T : R n R m be a function which maps vectors from R n to R m. Then T is called a linear transformation if the following two properties are
More informationMath Camp Notes: Linear Algebra I
Math Camp Notes: Linear Algebra I Basic Matrix Operations and Properties Consider two n m matrices: a a m A = a n a nm Then the basic matrix operations are as follows: a + b a m + b m A + B = a n + b n
More informationChapter 2. Error Correcting Codes. 2.1 Basic Notions
Chapter 2 Error Correcting Codes The identification number schemes we discussed in the previous chapter give us the ability to determine if an error has been made in recording or transmitting information.
More informationBasic Concepts in Linear Algebra
Basic Concepts in Linear Algebra Grady B Wright Department of Mathematics Boise State University February 2, 2015 Grady B Wright Linear Algebra Basics February 2, 2015 1 / 39 Numerical Linear Algebra Linear
More informationA FIRST COURSE IN LINEAR ALGEBRA. An Open Text by Ken Kuttler. Matrix Arithmetic
A FIRST COURSE IN LINEAR ALGEBRA An Open Text by Ken Kuttler Matrix Arithmetic Lecture Notes by Karen Seyffarth Adapted by LYRYX SERVICE COURSE SOLUTION AttributionNonCommercialShareAlike (CC BYNCSA)
More informationMatrix Arithmetic. a 11 a. A + B = + a m1 a mn. + b. a 11 + b 11 a 1n + b 1n = a m1. b m1 b mn. and scalar multiplication for matrices via.
Matrix Arithmetic There is an arithmetic for matrices that can be viewed as extending the arithmetic we have developed for vectors to the more general setting of rectangular arrays: if A and B are m n
More informationISOMETRIES OF R n KEITH CONRAD
ISOMETRIES OF R n KEITH CONRAD 1. Introduction An isometry of R n is a function h: R n R n that preserves the distance between vectors: h(v) h(w) = v w for all v and w in R n, where (x 1,..., x n ) = x
More informationChapter 2. General Vector Spaces. 2.1 Real Vector Spaces
Chapter 2 General Vector Spaces Outline : Real vector spaces Subspaces Linear independence Basis and dimension Row Space, Column Space, and Nullspace 2 Real Vector Spaces 2 Example () Let u and v be vectors
More informationMathematics Department Stanford University Math 61CM/DM Vector spaces and linear maps
Mathematics Department Stanford University Math 61CM/DM Vector spaces and linear maps We start with the definition of a vector space; you can find this in Section A.8 of the text (over R, but it works
More informationDetermining a span. λ + µ + ν = x 2λ + 2µ 10ν = y λ + 3µ 9ν = z.
Determining a span Set V = R 3 and v 1 = (1, 2, 1), v 2 := (1, 2, 3), v 3 := (1 10, 9). We want to determine the span of these vectors. In other words, given (x, y, z) R 3, when is (x, y, z) span(v 1,
More information4 ORTHOGONALITY ORTHOGONALITY OF THE FOUR SUBSPACES 4.1
4 ORTHOGONALITY ORTHOGONALITY OF THE FOUR SUBSPACES 4.1 Two vectors are orthogonal when their dot product is zero: v w = orv T w =. This chapter moves up a level, from orthogonal vectors to orthogonal
More informationLinear Independence x
Linear Independence A consistent system of linear equations with matrix equation Ax = b, where A is an m n matrix, has a solution set whose graph in R n is a linear object, that is, has one of only n +
More information1 Invariant subspaces
MATH 2040 Linear Algebra II Lecture Notes by Martin Li Lecture 8 Eigenvalues, eigenvectors and invariant subspaces 1 In previous lectures we have studied linear maps T : V W from a vector space V to another
More informationFormal power series rings, inverse limits, and Iadic completions of rings
Formal power series rings, inverse limits, and Iadic completions of rings Formal semigroup rings and formal power series rings We next want to explore the notion of a (formal) power series ring in finitely
More information1 Vectors. Notes for Bindel, Spring 2017 Numerical Analysis (CS 4220)
Notes for 20170130 Most of mathematics is best learned by doing. Linear algebra is no exception. You have had a previous class in which you learned the basics of linear algebra, and you will have plenty
More informationMATH 20F: LINEAR ALGEBRA LECTURE B00 (T. KEMP)
MATH 20F: LINEAR ALGEBRA LECTURE B00 (T KEMP) Definition 01 If T (x) = Ax is a linear transformation from R n to R m then Nul (T ) = {x R n : T (x) = 0} = Nul (A) Ran (T ) = {Ax R m : x R n } = {b R m
More informationCalculating determinants for larger matrices
Day 26 Calculating determinants for larger matrices We now proceed to define det A for n n matrices A As before, we are looking for a function of A that satisfies the product formula det(ab) = det A det
More informationStat 159/259: Linear Algebra Notes
Stat 159/259: Linear Algebra Notes Jarrod Millman November 16, 2015 Abstract These notes assume you ve taken a semester of undergraduate linear algebra. In particular, I assume you are familiar with the
More informationExtra Problems for Math 2050 Linear Algebra I
Extra Problems for Math 5 Linear Algebra I Find the vector AB and illustrate with a picture if A = (,) and B = (,4) Find B, given A = (,4) and [ AB = A = (,4) and [ AB = 8 If possible, express x = 7 as
More informationFundamentals of Linear Algebra. Marcel B. Finan Arkansas Tech University c All Rights Reserved
Fundamentals of Linear Algebra Marcel B. Finan Arkansas Tech University c All Rights Reserved 2 PREFACE Linear algebra has evolved as a branch of mathematics with wide range of applications to the natural
More informationELEMENTARY LINEAR ALGEBRA
ELEMENTARY LINEAR ALGEBRA K. R. MATTHEWS DEPARTMENT OF MATHEMATICS UNIVERSITY OF QUEENSLAND Corrected Version, 7th April 013 Comments to the author at keithmatt@gmail.com Chapter 1 LINEAR EQUATIONS 1.1
More informationINSTITIÚID TEICNEOLAÍOCHTA CHEATHARLACH INSTITUTE OF TECHNOLOGY CARLOW MATRICES
1 CHAPTER 4 MATRICES 1 INSTITIÚID TEICNEOLAÍOCHTA CHEATHARLACH INSTITUTE OF TECHNOLOGY CARLOW MATRICES 1 Matrices Matrices are of fundamental importance in 2dimensional and 3dimensional graphics programming
More informationMATH 323 Linear Algebra Lecture 6: Matrix algebra (continued). Determinants.
MATH 323 Linear Algebra Lecture 6: Matrix algebra (continued). Determinants. Elementary matrices Theorem 1 Any elementary row operation σ on matrices with n rows can be simulated as left multiplication
More informationIntroduction to Determinants
Introduction to Determinants For any square matrix of order 2, we have found a necessary and sufficient condition for invertibility. Indeed, consider the matrix The matrix A is invertible if and only if.
More information2: LINEAR TRANSFORMATIONS AND MATRICES
2: LINEAR TRANSFORMATIONS AND MATRICES STEVEN HEILMAN Contents 1. Review 1 2. Linear Transformations 1 3. Null spaces, range, coordinate bases 2 4. Linear Transformations and Bases 4 5. Matrix Representation,
More informationSystems of Linear Equations. By: Tri Atmojo Kusmayadi and Mardiyana Mathematics Education Sebelas Maret University
Systems of Linear Equations By: Tri Atmojo Kusmayadi and Mardiyana Mathematics Education Sebelas Maret University Standard of Competency: Understanding the properties of systems of linear equations, matrices,
More informationSupplementary Material for MTH 299 Online Edition
Supplementary Material for MTH 299 Online Edition Abstract This document contains supplementary material, such as definitions, explanations, examples, etc., to complement that of the text, How to Think
More informationElementary Linear Algebra Review for Exam 2 Exam is Monday, November 16th.
Elementary Linear Algebra Review for Exam Exam is Monday, November 6th. The exam will cover sections:.4,..4, 5. 5., 7., the class notes on Markov Models. You must be able to do each of the following. Section.4
More informationMatrix & Linear Algebra
Matrix & Linear Algebra Jamie Monogan University of Georgia For more information: http://monogan.myweb.uga.edu/teaching/mm/ Jamie Monogan (UGA) Matrix & Linear Algebra 1 / 84 Vectors Vectors Vector: A
More informationFall Inverse of a matrix. Institute: UC San Diego. Authors: Alexander Knop
Fall 2017 Inverse of a matrix Authors: Alexander Knop Institute: UC San Diego RowColumn Rule If the product AB is defined, then the entry in row i and column j of AB is the sum of the products of corresponding
More information4. Linear transformations as a vector space 17
4 Linear transformations as a vector space 17 d) 1 2 0 0 1 2 0 0 1 0 0 0 1 2 3 4 32 Let a linear transformation in R 2 be the reflection in the line = x 2 Find its matrix 33 For each linear transformation
More information4. Matrix inverses. left and right inverse. linear independence. nonsingular matrices. matrices with linearly independent columns
L. Vandenberghe ECE133A (Winter 2018) 4. Matrix inverses left and right inverse linear independence nonsingular matrices matrices with linearly independent columns matrices with linearly independent rows
More informationMath 115A: Linear Algebra
Math 115A: Linear Algebra Michael Andrews UCLA Mathematics Department February 9, 218 Contents 1 January 8: a little about sets 4 2 January 9 (discussion) 5 2.1 Some definitions: union, intersection, set
More informationAppendix C Vector and matrix algebra
Appendix C Vector and matrix algebra Concepts Scalars Vectors, rows and columns, matrices Adding and subtracting vectors and matrices Multiplying them by scalars Products of vectors and matrices, scalar
More informationChapter 3: Theory Review: Solutions Math 308 F Spring 2015
Chapter : Theory Review: Solutions Math 08 F Spring 05. What two properties must a function T : R m R n satisfy to be a linear transformation? (a) For all vectors u and v in R m, T (u + v) T (u) + T (v)
More informationTHE MINIMAL POLYNOMIAL AND SOME APPLICATIONS
THE MINIMAL POLYNOMIAL AND SOME APPLICATIONS KEITH CONRAD. Introduction The easiest matrices to compute with are the diagonal ones. The sum and product of diagonal matrices can be computed componentwise
More informationThe GaussJordan Elimination Algorithm
The GaussJordan Elimination Algorithm Solving Systems of Real Linear Equations A. Havens Department of Mathematics University of Massachusetts, Amherst January 24, 2018 Outline 1 Definitions Echelon Forms
More informationChapter 1 Matrices and Systems of Equations
Chapter 1 Matrices and Systems of Equations System of Linear Equations 1. A linear equation in n unknowns is an equation of the form n i=1 a i x i = b where a 1,..., a n, b R and x 1,..., x n are variables.
More informationLinear algebra 2. Yoav Zemel. March 1, 2012
Linear algebra 2 Yoav Zemel March 1, 2012 These notes were written by Yoav Zemel. The lecturer, Shmuel Berger, should not be held responsible for any mistake. Any comments are welcome at zamsh7@gmail.com.
More informationFact: Every matrix transformation is a linear transformation, and vice versa.
Linear Transformations Definition: A transformation (or mapping) T is linear if: (i) T (u + v) = T (u) + T (v) for all u, v in the domain of T ; (ii) T (cu) = ct (u) for all scalars c and all u in the
More information2.1 Matrices. 3 5 Solve for the variables in the following matrix equation.
2.1 Matrices Reminder: A matrix with m rows and n columns has size m x n. (This is also sometimes referred to as the order of the matrix.) The entry in the ith row and jth column of a matrix A is denoted
More informationMath 24 Winter 2010 Sample Solutions to the Midterm
Math 4 Winter Sample Solutions to the Midterm (.) (a.) Find a basis {v, v } for the plane P in R with equation x + y z =. We can take any two noncollinear vectors in the plane, for instance v = (,, )
More informationMATH 2030: MATRICES ,, a m1 a m2 a mn If the columns of A are the vectors a 1, a 2,...,a n ; A is represented as A 1. .
MATH 030: MATRICES Matrix Operations We have seen how matrices and the operations on them originated from our study of linear equations In this chapter we study matrices explicitely Definition 01 A matrix
More informationGENERATING SETS KEITH CONRAD
GENERATING SETS KEITH CONRAD 1 Introduction In R n, every vector can be written as a unique linear combination of the standard basis e 1,, e n A notion weaker than a basis is a spanning set: a set of vectors
More informationIntroduction to Matrix Algebra
Introduction to Matrix Algebra August 18, 2010 1 Vectors 1.1 Notations A pdimensional vector is p numbers put together. Written as x 1 x =. x p. When p = 1, this represents a point in the line. When p
More informationRecall the convention that, for us, all vectors are column vectors.
Some linear algebra Recall the convention that, for us, all vectors are column vectors. 1. Symmetric matrices Let A be a real matrix. Recall that a complex number λ is an eigenvalue of A if there exists
More informationA linear algebra proof of the fundamental theorem of algebra
A linear algebra proof of the fundamental theorem of algebra Andrés E. Caicedo May 18, 2010 Abstract We present a recent proof due to Harm Derksen, that any linear operator in a complex finite dimensional
More informationLecture 17: Section 4.2
Lecture 17: Section 4.2 Shuanglin Shao November 4, 2013 Subspaces We will discuss subspaces of vector spaces. Subspaces Definition. A subset W is a vector space V is called a subspace of V if W is itself
More informationINVERSE OF A MATRIX [2.2]
INVERSE OF A MATRIX [2.2] The inverse of a matrix: Introduction We have a mapping from R n to R n represented by a matrix A. Can we invert this mapping? i.e. can we find a matrix (call it B for now) such
More informationSAMPLE OF THE STUDY MATERIAL PART OF CHAPTER 1 Introduction to Linear Algebra
1.1. Introduction SAMPLE OF THE STUDY MATERIAL PART OF CHAPTER 1 Introduction to Linear algebra is a specific branch of mathematics dealing with the study of vectors, vector spaces with functions that
More informationCHEVALLEY S THEOREM AND COMPLETE VARIETIES
CHEVALLEY S THEOREM AND COMPLETE VARIETIES BRIAN OSSERMAN In this note, we introduce the concept which plays the role of compactness for varieties completeness. We prove that completeness can be characterized
More informationRings. Chapter 1. Definition 1.2. A commutative ring R is a ring in which multiplication is commutative. That is, ab = ba for all a, b R.
Chapter 1 Rings We have spent the term studying groups. A group is a set with a binary operation that satisfies certain properties. But many algebraic structures such as R, Z, and Z n come with two binary
More informationMATH PRACTICE EXAM 1 SOLUTIONS
MATH 2359 PRACTICE EXAM SOLUTIONS SPRING 205 Throughout this exam, V and W will denote vector spaces over R Part I: True/False () For each of the following statements, determine whether the statement is
More information