Duality of finite-dimensional vector spaces
|
|
- Tabitha Melton
- 6 years ago
- Views:
Transcription
1 CHAPTER I Duality of finite-dimensional vector spaces 1 Dual space Let E be a finite-dimensional vector space over a field K The vector space of linear maps E K is denoted by E, so E = L(E, K) This vector space is called the dual space of E Its elements are called linear forms on E For ϕ E and u E, let ϕ, u = ϕ(u) K 11 Proposition For ϕ, ψ E, u, v E and α, β K, 1 αϕ + βψ, u = α ϕ, u + β ψ, u 2 ϕ, uα + vβ = ϕ, u α + ϕ, v β 3 If ϕ, u = 0 for all u E, then ϕ = 0 4 If ϕ, u = 0 for all ϕ E, then u = 0 Proof 1 This follows from the definition of the sum and the scalar multiplication in E 2 This follows from the fact that ϕ is a linear map 3 This is the definition of the zero map 4 If u 0, then u is the first element in some basis of E Let (u, e 2,, e n ) be such a basis From the construction principle for linear maps, it follows that there is a linear form ϕ E which maps u to 1 and the other basis vectors to 0 This proves the contrapositive of implication 4 11 Dual basis Let e = (e 1,, e n ) be a basis of E By the construction principle for linear maps, there exists for all i = 1,, n a linear form e i E which maps e i to 1 and the other basis vectors to 0, ie, { e i, e 1 if i = j, j = δ ij = 0 if i j Thus, for u = n j=1 e ju j E we have n e i, u = e i, e j u j = j=1 n δ ij u j = u i, which shows that the linear form e i maps every vector of E to its i-th coordinate with respect to the basis e Therefore, for all u E, n (1) u = e i e i, u n, 12 Theorem The sequence e = (e 1,, e n) is a basis of E In particular, dime = dime Proof The elements e 1,, e n are linearly independent: Let n α ie i n 0 = α i e i, e j = j=1 n α i e i, e j = 1 n α i δ ij = α j = 0 For j = 1,,
2 2 ORTHOGONALITY 2 The sequence e spans E : Let us show that for all ϕ E, n (2) ϕ = ϕ, e i e i It suffices to show that both sides take the same value on every vector u E From (1), it follows that n n ϕ, u = ϕ, e j e j, u = ϕ, e j e j, u On the other hand, j=1 n ϕ, e i e i, u = j=1 n ϕ, e i e i, u 12 Bidual space Every vector u E defines a linear form ev u : E K by Thus, ev u (E ) Moreover, the map ev u (ϕ) = ϕ(u) ev: E E ; is linear since for u, v E, α, β K and ϕ E, u ev u ev uα+vβ (ϕ) = ϕ(uα + vβ) = ϕ(u)α + ϕ(v)β = ev u (ϕ)α + ev v (ϕ)β 13 Theorem The map ev: E E is a vector space isomorphism Proof First, we show that ev is injective: if u E is such that ev u = 0, then ϕ, u = 0 for all ϕ E, hence u = 0 by Proposition 11 Therefore, Kerev = 0 and ev is injective On the other hand, applying Theorem 12 twice, we find Since ev is injective, it is therefore also surjective dime = dime = dim E The isomorphism ev is natural, inasmuch as it does not depend on the choice of bases in E and E It is used to identify these two vector spaces Thus, for u E, it is agreed that u = ev u E Therefore, u, ϕ is defined for u E and ϕ E, and we have u, ϕ = ev u, ϕ = ϕ(u) = ϕ, u 2 Orthogonality For every subspace V E, the orthogonal subspace V 0 E is defined by It is easily checked that V 0 is a subspace of E 21 Proposition dimv 0 = dime dimv V 0 = {ϕ E ϕ, u = 0 for all u V } Proof Let (e 1,, e r ) be a basis of V, which we extend into a basis e = (e 1,, e r, e r+1,, e n ) of E Consider the last n r elements of the dual basis e = (e 1,, e r,, e n) We shall show: V 0 = span e r+1,, e n The proposition follows, since the right side is a subspace of dimension n r, as e r+1,, e n are linearly independent Equation (2) shows that every ϕ V 0 is a linear combination of e r+1,, e n Therefore, V 0 span e r+1,, e n
3 3 TRANSPOSITION 3 On the other hand, every v V is a linear combination of (e 1,, e r ), hence equation (1) shows that e i, v = 0 for i = r + 1,, n Therefore, e r+1,, e n V 0, hence span e r+1,, e n V 0 22 Corollary For every subspace V E, V 00 = V (under the identification of E and E, since V 00 E ) Proof By definition, V 00 = {u E u, ϕ = 0 for all ϕ V 0 } Since ϕ, u = 0 for u V and ϕ V 0, we have V V 00 Now, the preceding proposition and Theorem 12 yield dimv 00 = dime dimv 0 = dime dimv 0 = dimv Therefore, the inclusion V V 00 is an equality 3 Transposition Let A: E F be a linear map between finite-dimensional vector spaces The transpose map is defined by A t : F E A t (ϕ) = ϕ A for ϕ F 31 Proposition A t is the only linear map from F to E such that A t (ϕ), u = ϕ, A(u) for all ϕ F and all u E Proof For ϕ F and u E we have A t (ϕ), u = ϕ A, u = ϕ A(u) = ϕ, A(u), hence A t satisfies the property Suppose B: F E is another linear map with the same property Then for all ϕ F and all u E, B(ϕ), u = ϕ, A(u) = A t (ϕ), u, hence B(ϕ) = A t (ϕ) and therefore B = A t Now, let e and f be arbitrary bases of E and F respectively, and let e, f be the dual bases 32 Proposition The matrices of A t and A are related as follows: Proof Let f (A) e = (a ij ) 1 i m, ie, m (1) A(e j ) = f i a ij and e (A t ) f = (a ij ) 1 i n, ie, 1 j m A t (f j ) = e (At ) f = f (A) t e for j = 1,, n n e i a ij for j = 1,, m
4 3 TRANSPOSITION 4 We have to prove a ij = a ji for i = 1,, n and j = 1,, m Substituting A t (fj ) for ϕ in equation (2) of section 11, we obtain A t (f j ) = n A t (fj ), e i e i for j = 1,, m, hence a ij = A t (f j ), e i By Proposition 31, it follows that a ij = f j, A(e i ) On the other hand, we may compute the right side of this last equation by using equation (1), which yields f j, A(e i ) = f n j, n f k a ki = fj, f k a ki = a ji k=1 Thus, a ij = a ji for i = 1,, n and j = 1,, m k=1 33 Proposition 1 For A, B: E F and α, β K, 2 For A: E F and B: F G, 3 For A: E F, (αa + βb) t = αa t + βb t (B A) t = A t B t (A t ) t = A (under the identifications E = E and F = F, since (A t ) t : E F ) Proof These properties can be proved either by reduction to the corresponding properties of matrices (by using arbitrary bases of the relevant spaces), or by using the characterization of transpose maps in Proposition 31
5 CHAPTER II Quotient spaces 1 Definition Let E be an arbitrary vector space over a field K and let V E be an arbitrary subspace For x E let x + V = {x + v v V } E 11 Lemma For x, y E, the equation x + V = y + V is equivalent to x y V Proof Suppose first x+v = y+v Since x = x+0 x+v, we have x y+v, hence x = y+v for some v V and therefore x y = v V Conversely, suppose x y = v V For all w V we then have x + w = y + (v + w) y + V, hence x + V y + V We also have y + w = x + (w v) x + V, hence y + V x + V, and therefore x + V = y + V Let E/V = {x + V x E} An addition and a scalar multiplication are defined on this set by (x + V ) + (y + V ) = (x + y) + V for x, y E, (x + V )α = xα + V for x E and α K It is easily checked that these operations are well-defined and endow the set E/V with a K-vector space structure In view of the definitions above, the map defined by ε V : E E/V ε V (x) = x + V is a surjective linear map It is called the canonical epimorphism of E onto E/V Its kernel is the set of x E such that x + V = 0 + V The lemma shows that this condition holds if and only if x V ; therefore Kerε V = V 12 Proposition Suppose E is finite-dimensional; then Moreover, dime/v = dime dim V (a) if (e 1,, e n ) is a basis of E which extends a basis (e 1,, e r ) of V, then (e r+1 +V,, e n +V ) is a basis of E/V ; (b) if (e 1,, e r ) is a basis of V and u r+1,, u n are vectors of E such that (u r+1 +V,, u n +V ) is a basis of E/V, then (e 1,, e r, u r+1,, u n ) is a basis of E 5
6 2 INDUCED LINEAR MAPS 6 Proof Since Kerε V = V and Imε V = E/V, the relation between the dimensions of the kernel and the image of a linear map yields dime = dimv + dime/v (a) Since dime/v = dime dimv = n r, it suffices to prove that the sequence (e r+1 +V,, e n +V ) spans E/V Every x E is a linear combination of e 1,, e n If then x = e 1 x e n x n, x + V = ε V (x) = ε V (e 1 )x ε V (e n )x n Now, ε V (e i ) = 0 for i = 1,, r since e i V, hence x + V = (e r+1 + V )x r (e n + V )x n (b) Since dime = dimv + dim E/V, it suffices to prove that e 1,, e r, u r+1,, u n are linearly independent Suppose (1) e 1 α e r α r + u r+1 β r u n β n = 0 Taking the image of each side under ε V, we obtain (u r+1 + V )β r (u n + V )β n = 0, hence β r+1 = = β n = 0 Therefore, by (1) it follows that α 1 = = α r = 0, since (e 1,, e r ) is a basis of V 2 Induced linear maps Let A: E F be a linear map and let V E be a subspace A linear map A: E/V F is said to be induced by A if In other words, A(x + V ) = A(x) for all x E A ε V = A This condition is also expressed as follows: the diagram E ε V E/V is commutative or commutes A A F 21 Proposition A linear map A: E F induces a linear map if and only if V KerA A: E/V F Proof If A ε V = A, then A(v) = 0 for all v V since V = Kerε V For the converse, suppose V KerA and consider x, x E If x + V = x + V, then x x KerA, hence A(x x ) = 0 and therefore We may then define a map A: E/V F by A(x) = A(x ) A(x + V ) = A(x) for all x E It is easily checked that the map A thus defined is linear
7 2 INDUCED LINEAR MAPS 7 Now, consider a linear map A: E F and subspaces V E and W F A linear map is said to be induced by A if A: E/V F/W A(x + V ) = A(x) + W for all x E or, alternatively, if A ε V = ε W A This condition can also be expressed by saying that the following diagram is commutative: E A F ε V E/V A ε W F/W or that A makes the above diagram commute 22 Corollary The linear map A: E F induces a linear map A: E/V F/W if and only if A(V ) W Proof Consider the linear map A = ε W A: E F/W The preceding proposition shows that there exists a linear map A such that A = A ε V if and only if V KerA In view of the definition of A, this condition is equivalent to A(V ) Kerε W = W If A(V ) W, we may restrict the linear map A to the subspace V and get a linear map A V : V W We aim to compare the matrices of A, of A V and of A: E/V F/W Consider a basis e = (e 1,, e r ) of V, a basis e = (e r+1 + V,, e n + V ) of E/V and the sequence e = (e 1,, e r, e r+1,, e n ) which, according to Proposition 12, is a basis of E Similarly, let f = (f 1,, f s ) be a basis of W, f = (f s+1 + W,, f m + W) a basis of F/W and consider the basis f = (f 1,, f s, f s+1,, f m ) of F 23 Proposition f(a) e = ( f (A V ) e 0 f (A) e ) Proof Let A(e j ) = m f ia ij for j = 1,, n, so that f(a) e = (a ij ) 1 i m For j = 1,, r we have e j V, hence A(e j ) W Therefore, A(e j ) is a linear combination of f 1,, f s, hence a ij = 0 for j = 1,, r and i = s + 1,, m Moreover, A(e j ) = A V (e j ) for j = 1,, n, hence A V (e j ) = s f ia ij and therefore On the other hand, for j = r + 1,, n, f (A V ) e A(e j + V ) = A(e j ) + W = = (a ij ) 1 i s 1 j r m (f i + W)a ij For i = 1,, s, we have f i W, hence f i +W = 0 The preceding equation thus yields A(e j +V ) = m i=s+1 (f i + W)a ij for j = r + 1,, n, hence f (A) e = (a ij ) s+1 i m r+
8 3 TRIANGULATION 8 The proof is thus complete In particular, if F = E and W = V, e(a) e = ( e (A V ) e 0 e(a) e and since the determinant of a block-triangular matrix is the product of the determinants of the blocks the following corollary follows: 24 Corollary deta = deta V deta and Pc A = Pc A V Pc A 3 Triangulation 31 Proposition Let A: E E be a linear operator on a finite-dimensional vector space E There exists a basis e of E such that the matrix e (A) e is upper triangular if and only if the characteristic polynomial Pc A decomposes into a product of factors of degree 1 Proof If the matrix e (A) e is upper triangular, let e(a) e = λ 1 0 λ n then Pc A (X) = (X λ 1 )(X λ n ) Conversely, suppose Pc A decomposes into a product of factors of degree 1 We argue by induction on the dimension of E If dim E = 1, there is nothing to prove since every matrix of order 1 is upper triangular Let dim E = n By hypothesis, the characteristic polynomial of A has a root λ Let e 1 E be an eigenvector of A with eigenvalue λ and let V = e 1 K be the vector space spanned by e 1 We have hence A induces a linear operator, A(V ) = A(e 1 )K = e 1 λk V, A: E/V E/V By Corollary 24, we have Pc A (X) = (X λ)pc A (X), hence the characteristic polynomial of A decomposes into a product of factors of degree 1 By induction, there exists a basis e = (e 2 + V,, e n + V ) of E/V such that the matrix e (A) e is upper triangular Proposition 12 shows that the sequence e = (e 1, e 2,, e n ) is a basis of E Moreover, by Proposition 23, ( ) λ e(a) e = 0 e(a) e Since the matrix e (A) e is upper triangular, so is e (A) e )
9 CHAPTER III Tensor product 1 Tensor product of vector spaces Let E 1,, E n, F be vector spaces over a field K Recall that a map is called n-linear or multilinear if f(x 1,, x i 1, x i α + x i α, x i+1,, x n ) = f : E 1 E n F f(x 1,, x i 1, x i, x i+1,, x n )α + f(x 1,, x i 1, x i, x i+1,, x n )α for each i = 1,, n and for all x 1 E 1,, x i 1 E i 1, x i, x i E i, x i+1 E i+1,, x n E n and α, α K In other words, for each i = 1,, n and for all x 1 E 1,, x i 1 E i 1, x i+1 E i+1,, x n E n fixed, the map f(x 1,, x i 1,, x i+1,, x n ): E i F which carries x E i to f(x 1,, x i 1, x, x i+1,, x n ) is linear For example, all the products for which the distributive law holds are bilinear maps Clearly, if f : E 1 E n F is multilinear and A: F G is linear, then the composite map A f : E 1 E n G is multilinear It turns out that, for given E 1,, E n, there exists a universal linear map : E 1 E n E 1 E n, called the tensor product of E 1,, E n, from which all the multilinear maps originating in E 1 E n can be derived The goal of this section is to define this universal map 11 Universal property Let E 1,, E n be vector spaces over a field K A pair (f, F) consisting of a K-vector space F and a multilinear map f : E 1 E n F is called a universal product of E 1,, E n if the following property (called the universal property) holds: for every K-vector space G and every multilinear map g: E 1 E n G there exists a unique linear map A: F G such that g = A f or, in other words, which makes the following diagram commute: E 1 E n f F g A G The existence of a universal product is far from obvious By contrast, the uniqueness (up to isomorphism) easily follows from the definition by abstract nonsense 11 Proposition Suppose (f, F) and (f, F ) are two universal products of E 1,, E n Then there is one and only one vector space isomorphism ϕ: F F such that the following diagram 9
10 1 TENSOR PRODUCT OF VECTOR SPACES 10 commutes: E 1 E n f F f ϕ F Proof The existence of a unique linear map ϕ: F F such that f = ϕ f follows from the universal property of (f, F) We have to show that ϕ is bijective To achieve this goal, observe that the universal property of (f, F ) yields a linear map ϕ : F F such that f = ϕ f Now, the following diagrams commute: E 1 E n f F E 1 E n f F f Id F F f ϕ ϕ F In view of the uniqueness condition in the universal property of (f, F), it follows that ϕ ϕ = Id F Similarly, the universal property of (f, F ) shows that ϕ ϕ = Id F Therefore, ϕ and ϕ are reciprocal bijections This proposition shows that a universal product of E 1,, E n, if it exists, can be considered as unique We shall refer to this universal product as the tensor product of E 1,, E n and denote it as (, E 1 E n ) The image of an n-tuple (x 1,, x n ) E 1 E n under is denoted x 1 x n Our next goal is to prove the existence of the tensor product We consider first the case where n = 2 and then derive the general case by associativity 12 The tensor product of two vector spaces 12 Proposition Let µ: K m K n K m n be defined by µ ( ) (x i ) 1 i m, (y j ) = (xi y j ) 1 i m The pair (µ, K m n ) is a universal product of K m and K n Proof Denoting the elements of K m and K n as column vectors, we have µ(x, y) = x y t for x K m and y K n, where denotes the matrix multiplication K m 1 K 1 n K m n Therefore, the bilinearity of µ follows from the properties of the matrix multiplication Let (c i ) 1 i m (resp (c j ) ) be the canonical basis of K m (resp K n ) Let also e ij = µ(c i, c j ) for i = 1,, m and j = 1,, n Thus, e ij K m n is the matrix whose only non-zero entry is a 1 at the i-th row and j-th column The family (e ij ) 1 i m is a basis of K m n Therefore, given any bilinear map B: K m K n F, we may use the construction principle to obtain a linear map ϕ: K m n F such that ϕ(e ij ) = B(c i, c j ) For all x = hence x 1 x m K m and y = µ(x, y) = 1 i m y 1 y n K n, we have x = m c ix i, y = n j=1 c j y j, x i y j µ(c i, c j ) = 1 i m x i y j e ij
11 1 TENSOR PRODUCT OF VECTOR SPACES 11 and B(x, y) = 1 i m x i y j B(c i, c j ) = 1 i m ( x i y j ϕ(e ij ) = ϕ 1 i m x i y j e ij ) Therefore, B = ϕ µ Uniqueness of the linear map ϕ: K m n F such that B = ϕ µ is clear, since this equation yields B(c i, c j) = ϕ ( µ(c i, c j) ) = ϕ(e ij ), and a linear map is uniquely determined by the image of a basis Thus, (µ, K m n ) is a universal product of K m and K n 13 Theorem Any two vector spaces over a field K have a tensor product Proof We give the proof for finite-dimensional vector spaces only If dime = m and dimf = n, we may identify E to K m and F to K n by the choice of arbitrary bases of E and of F (by mapping every vector in E (resp F) to the m-tuple (resp n-tuple) of its coordinates with respect to the chosen basis) The preceding proposition then shows that (µ, K m n ) is a universal tensor product of E and F Note that, even though bases of E and F were chosen in the proof of the theorem, the tensor product itself does not depend on the choice of bases (but its identification with (µ, K m n ) does) It is a canonical construction To simplify the notation, we suppress the bilinear map in the tensor product notation (, E F), and refer to the tensor product of E and F simply as E F 14 Corollary If e = (e i ) 1 i m is a basis of E and f = (f j ) is a basis of F, then (e i f j ) 1 i m is a basis of E F In particular, dim(e F) = dim E dimf Proof It was observed in the proof of Proposition 12 that ( µ(c i, c j )) 1 i m is a basis of K m n If E is identified with K m and F with K n by means of the bases e and f, then e and f correspond to the canonical bases of K m and K n respectively, hence µ(c i, c j ) = e i f j The corollary shows that every element in E F has the form 1 i m α ij e i f j for some α ij K Grouping terms, we obtain n ( m ) α ij e i f j = α ij e i f j 1 i m j=1 hence every element in E F can be written in the form x 1 y x r y r for suitable x 1,, x r E and y 1,, y r F (and r N) Note that this expression is not unique: for instance, (e 1 + e 2 ) f 1 + (e 1 e 2 ) f 2 = e 1 (f 1 + f 2 ) + e 2 (f 1 f 2 ) 13 The tensor product of three or more vector spaces Let E 1, E 2, E 3 be vector spaces over K We may consider the tensor products E 1 E 2 and (E 1 E 2 ) E 3 The map : E 1 E 2 E 3 (E 1 E 2 ) E 3 defined by (x 1, x 2, x 3 ) = (x 1 x 2 ) x 3 is 3-linear 15 Proposition The pair (, (E 1 E 2 ) E 3 ) is a universal product of E1, E 2 and E 3
12 1 TENSOR PRODUCT OF VECTOR SPACES 12 Proof Let T : E 1 E 2 E 3 F be a 3-linear map For all x 3 E 3, the map T(,, x 3 ): E 1 E 2 F is bilinear, hence there is a unique linear map ϕ x3 : E 1 E 2 F such that ϕ x3 (x 1 x 2 ) = T(x 1, x 2, x 3 ) for all x 1 E 1, x 2 E 2 For x 1 E 1, x 2 E 2, x 3, x 3 E 3 and α, α K, we have hence T(x 1, x 2, αx 3 + α x 3 ) = αt(x 1, x 2, x 3 ) + α T(x 1, x 2, x 3 ), ϕ αx3+α x 3 (x 1 x 2 ) = αϕ x3 (x 1 x 2 ) + α ϕ x 3 (x 1 x 2 ) Since E 1 E 2 has a basis consisting of products x 1 x 2, by Corollary 14, and since a linear map is uniquely determined by the image of a basis, it follows that (1) ϕ αx3+α x 3 = αϕ x 3 + α ϕ x 3 Consider the map Φ: (E 1 E 2 ) E 3 F defined by Φ(ξ, x 3 ) = ϕ x3 (ξ) for ξ E 1 E 2 and x 3 E 3 By (1) and since each map ϕ x3 is linear, the map Φ is bilinear Therefore, the universal property of tensor products yields a linear map such that In particular, ϕ: (E 1 E 2 ) E 3 F ϕ(ξ x 3 ) = Φ(ξ, x 3 ) for all ξ E 1 E 2, x 3 E 3 ϕ ( (x 1 x 2 ) x 3 ) = ϕx3 (x 1 x 2 ) = T(x 1, x 2, x 3 ), hence T = ϕ Uniqueness of the map ϕ satisfying this equation is clear, since every element in (E 1 E 2 ) E 3 can be written as i (x i y i ) z i = i (x i, y i, z i ) for some x i E 1, y i E 2 and z i E 3 If ϕ and ϕ satisfy T = ϕ = ϕ, we must have ( ) ϕ (x i y i ) z i = T(x i, y i, z i ) = ϕ ( (x i y i ) z i ), i i hence ϕ = ϕ i The proposition above proves the existence of the tensor product E 1 E 2 E 3, and shows that there is a canonical isomorphism E 1 E 2 E 3 = (E 1 E 2 ) E 3 Obviously, we also have E 1 E 2 E 3 = E 1 (E 2 E 3 ) By induction, it follows that the tensor product of arbitrarily many vector spaces exists, and may be defined by associativity The following corollary also follows by induction on the number of factors: 16 Corollary Let (e iji ) 1 ji n i be a basis of E i, for i = 1,, r Then is a basis of E 1 E r In particular, (e 1j1 e 2j2 e rjr ) 1 j1 n 1 1 j r n r dime 1 E r = dime 1 dime r
13 1 TENSOR PRODUCT OF VECTOR SPACES Example Let E 1,, E n be finite-dimensional K-vector spaces and let F be an arbitrary K-vector space Denote by M(E 1 E n, F) the set of all multilinear maps f : E 1 E n F It is easily checked that M(E 1 E n, F) is a subspace of the space F(E 1 E n, F) of all mappings E 1 E n F Let E i be the dual space of E i for i = 1,, n For y F and ϕ 1 E 1,, ϕ n E n, define by Clearly, the map θ y,ϕ1,,ϕ n is multilinear, hence Moreover, it is also clear that θ y,ϕ1,,ϕ n : E 1 E n F θ y,ϕ1,,ϕ n (x 1,, x n ) = y ϕ 1, x 1 ϕ n, x n θ y,ϕ1,,ϕ n M(E 1 E n, F) (y, ϕ 1,, ϕ n ) θ y,ϕ1,,ϕ n is a multilinear map F E 1 E n M(E 1 E n, F) We denote it by θ By the universal property of tensor products, θ induces a linear map Θ: F E 1 E n M(E 1 E n, F) 17 Theorem The map Θ is an isomorphism of vector spaces Proof We have to show that Θ is one-to-one and onto To achieve this, we construct a reciprocal map For i = 1,, n, let (e iji ) 1 ji m i be a basis of E i, and let (e ij i ) 1 ji m i be the dual basis For every multilinear map f M(E 1 E n, F), define Ψ(f) = f(e 1j1,, e njn ) e 1j 1 e nj n F E1 E n 1 j 1 m 1 1 j n m n The map Ψ: M(E 1 E n, F) F E1 E n is linear To prove that Ψ and Θ are reciprocal bijections, we shall show 1 that Ψ Θ is the identity on F E1 En and Θ Ψ is the identity on M(E 1 E n, F) Every element in F E1 E n can be written in the form ξ = y j1 j n e 1j 1 e nj n For this ξ, we have 1 j 1 m 1 1 j n m n Θ(ξ) = θ yj1 jn,e 1j 1,,e njn hence Θ(ξ)(e 1k1,, e nkn ) = 1 j 1 m 1 1 j n m n y j1 j n e 1j 1, e 1k1 e nj n, e nkn 1 j 1 m 1 1 j n m n = y k1 k n for all k 1 = 1,, m 1,, k n = 1,, m n 1 The notation makes the following proof hardly readable The reader is warmly encouraged to check the statement by him/herself!
14 2 FORMALISM 14 Now, by definition of Ψ we have Ψ Θ(ξ) = Θ(ξ)(e 1k1,, e nkn ) e 1k 1 e nk n 1 j 1 m 1 1 j n m n = 1 j 1 m 1 1 j n m n = ξ y k1 k n e 1k 1 e nk n Therefore, Ψ Θ is the identity on F E1 En Now, let f M(E 1 E n, F) We have Θ Ψ(f) = θ f(e1j1,,e njn ),e 1j,,e, 1 njn hence Θ Ψ(f)(e 1k1,, e nkn ) = 1 j 1 m 1 1 j n m n f(e 1j1,, e njn ) e 1j 1, e 1k1 e nj n, e nkn 1 j 1 m 1 1 j n m n = f(e 1k1,, e nkn ) for all k 1 = 1,, n 1,, k n = 1,, m n Since multilinear maps are uniquely determined by their values on basis elements, it follows that Θ Ψ(f) = f, hence Θ Ψ is the identity on M(E 1 E n, F) Note that the isomorphism Θ is canonical, even though its reciprocal isomorphism Ψ is defined above through the choice of bases of E 1,, E n We may therefore use Θ to identify F E 1 E n = M(E 1 E n, F) This provides an alternative definition for the tensor product of vector spaces, at least when all but one of the factors are finite-dimensional 21 Covariant and contravariant behaviour 2 Formalism Let E be a finite-dimensional vector space over a field K, and let (e i ) 1 i n be a basis of E It is sometimes convenient to write the coordinates of vectors x E with upper indices instead of lower indices 2 We thus write n x = e i x i This notation is simplified by the Einstein convention: on every index appearing twice in a formula, the summation is understood: x = e i x i (= x i e i ) For notational coherence, the entries in any change-of-basis matrix receive an upper index and a lower index: if (e α) 1 α n is another basis of E we write (1) e α = e i a i α (instead of e α = n e ia iα ) So, (a i α ) 1 i n are the coordinates of e α with respect to (e i) 1 i n The upper index is the row index and the lower index is the column index If (b α i ) 1 i,α n is the inverse matrix, we have (by definition of the inverse matrix) a i α bβ i = δβ α and a i α bα j = δi j 2 Of course, upper indices should not be confused with exponents There should not be any confusion, since in (multi)linear algebra exponents appear very exceptionally
15 2 FORMALISM 15 where δj i (resp δβ α) is the Kronecker symbol, ie, δj i = 1 if i = j and δi j = 0 if i j We also have (2) e i = e α bα i Now, consider a vector x = e i x i = e αx α Substituting e i a i α for e α, we obtain x = e i a i αx α, hence (3) Similarly, (4) x i = a i α x α x α = b α i x i It is a crucial observation that, although (3) and (4) look like (1) and (2), these equations are essentially different: multiplying by the matrix a i α transforms e i into e α while it has the reverse effect on coordinates, transforming x α into x i Now, consider the dual bases (e i ) 1 i n and (e α) 1 α n (These are bases of the dual space E ) Equation (2) in section 11 of Chapter I yields hence, taking into account (1) and (2), (5) e α = e i e α, e i and e i = e α e i, e α e α = e i bα i and e i = e α ai α In view of these equations, notational coherence dictates to denote e i for e i and e α for e α ; then (5) can be rewritten as (6) e α = e i b α i and e i = e α a i α These equations are exactly like (4) and (3) To complete the picture, consider a linear form ϕ E Equation (2) of section 11 of Chapter I yields ϕ = e i ϕ, e i = e α ϕ, e α Therefore, the coordinates (ϕ i ) 1 i n (resp (ϕ α) 1 α n ) of ϕ with respect to the basis (e i ) 1 i n (resp (e α ) 1 α n ) of E are given by Using (1) and (2), we get (7) ϕ i = ϕ, e i, ϕ α = ϕ, e α ϕ α = ϕ, e i a i α = ϕ i a i α and ϕ i = ϕ, e α b α i = ϕ αb α i, which should be compared to (3) and (4) Thus, coordinates of linear forms behave under a change of basis exactly like basis vectors: this is the covariant behaviour By contrast, coordinates of vectors and dual basis vectors behave in the opposite way to basis vectors: they display the so-called contravariant behaviour 22 Covariant and contravariant tensors For p and q any natural numbers, let Tp(E) q = } E {{ E } E } {{ E } p q if p, q are not both zero, and T0 0(E) = K The elements in T p q (E) are called p-covariant and q-contravariant tensors on E This is a shameless abuse of terminology, since the elements in Tp(E) q do not vary at all It is motivated by the fact that the coordinates of elements in Tp q (E) have p covariant (ie, lower) indices and q contravariant (ie, upper) indices, as we proceed to show Let (e i ) 1 i n be a basis of E, and let (e i ) 1 i n be the dual basis, which is a basis of E Corollary 16 shows that (e i1 e ip e j1 e jq ) 1 i1,,j q n is a basis of Tp(E), q hence every element in Tp(E) q can be written as (8) t = t j1 jq e i1 e ip e j1 e jq
16 2 FORMALISM 16 Let (e α) 1 α n be another basis of E and (e α ) 1 α n be its dual basis, and suppose the bases are related by (9) hence e α = e ia i α and e i = e α bα i e α = e i b α i and e i = e α a i α, as in the preceding section Substituting in (8), we get t = t j1 jq (e α 1 a i1 α 1 ) (e α p a ip α p ) (e β 1 b β1 j 1 ) (e β q b βq j q ) = ( t j1 jq a i1 α 1 a ip α p b β1 j 1 b βq j q ) e α 1 e α p e β 1 e β q Therefore, the coordinates t β 1 β q α 1 α p of t T q p(e) with respect to the basis (e α 1 e α p e β 1 e β q ) 1 α1,,β q n are given by (10) t β 1 β q α 1 α p = t j1 jq a i1 α 1 a ip α p b β1 j 1 b βq j q This shows that the lower indices are indeed covariant while the upper indices are contravariant (as they should) Similarly, we have t j1 jq = t β 1 β q α 1 α p b α1 i 1 b αp i p a j1 β 1 a jq β q From a formal point of view, the change of basis formula (10) is the only condition for an indexed array to be the coordinates of a p-covariant and q-contravariant tensor, as the following tensoriality criterion shows 21 Theorem Suppose that to each basis e = (e i ) 1 i n of E is attached an array of scalars ( j u(e) 1 j q ) 1 i 1,,j Knp+q q n There exists a tensor in Tp q (E) whose coordinates with respect to each basis e are u(e) j1 jq if and only if the arrays attached to any two bases e, e related by (9) satisfy u(e ) β1 βq α 1 α p = u(e) j1 jq a i1 α 1 a ip α p b β1 j 1 b βq j q Proof The if part readily follows from the computation leading to equation (10) To prove the only if part, pick an arbitrary basis ẽ of E and define a tensor u Tp q (E) by u = u(ẽ) j1 jq ẽ i1 ẽ ip ẽ j1 ẽ jq By definition, ( u(ẽ) j1 jq )1 i 1,,j q n is the array of coordinates of u with respect to the basis (ẽi1 ẽ ip ẽ j1 ẽ jq ) 1 i1,,j q n We have to show that for any basis ê = (ê α ) 1 α n, the coordinates of u with respect to (ê α1 ê αp ê β1 ê βq ) 1 α1,,β q n are ( u(ê) α β1 βq 1 α p )1 α 1,,β Suppose q n ê α = ẽ i a i α and ẽ i = ê α b α i, hence êα = ẽ i b α i and ẽ i = ê α a i α Then by (10) the coordinates of u with respect to (ê α1 ê αp ê β1 ê βq ) 1 α1,,β q n are ( u(ẽ) j1 jq a i1 α 1 a ip α p b β1 j 1 b βq j q )1 α 1,,β q n By hypothesis, this array is ( u(ê) β1 βq α 1 α p )1 α 1,,β q n 23 Tensors on euclidean spaces In this section, we assume that the vector space E is finite-dimensional and endowed with a scalar product ( ): E E K, ie, a symmetric bilinear form which is non-degenerate in the sense that (x y) = 0 for all y E implies x = 0 Using the scalar product, we define a linear map Φ: E E
17 2 FORMALISM 17 by mapping every vector x E to the linear form y (x y), so Φ(x), y = (x y) (= (y x)) for x, y E The non-degeneracy condition on the scalar product means that Φ is injective, hence it is bijective since dim E = dime, by Theorem 12 of Chapter I We may thus use Φ to identify E and E, hence also T q p (E) with T p+q(e) or T p+q (E) In the rest of this section, we show how this identification works at the level of coordinates Let (e i ) 1 i n be an arbitrary basis of E For i, j = 1,, n, let g ij = (e i e j ) K The matrix (g ij ) 1 i,j n is the Gram matrix of the scalar product with respect to the basis (e i ) 1 i n Its entries are the coordinates of the 2-covariant tensor g = (e i e j )e i e j E E which corresponds to the scalar product (or to the map Φ) under the identification of Theorem 17: M(E E, K) = E E (= M(E, E )) The tensor g is called the metric tensor on the euclidean space E For i = 1,, n, the basis vector e i E is identified with Φ(e i ) E By equation (2) of section 11 of Chapter I, we have hence Φ(e i ) = e j Φ(e i ), e j = e j (e i e j ) = g ij e j, (11) e i = g ij e j under the identification E = E Consequently, any p-covariant and q-contravariant tensor is identified to t = t j1 jq e i1 e ip e j1 e jq T q p(e) t j1 jq e i1 e ip (g j1k 1 e k1 ) (g jqk q e kq ) = ( t j1 jq g j1k 1 g jqk q ) e i 1 e ip e k1 e kq We may therefore define the (p + q)-covariant coordinates of t as follows: t i1 i pj 1 j q = t k1 kq g k1j 1 g kqj q To define the (p + q)-contravariant coordinates of t, we use the inverse of the Gram matrix (g ij ) 1 i,j n (The Gram matrix is invertible because the scalar product is non-degenerate) Let (g ij ) 1 i,j n be the inverse matrix From equation (11), we derive hence e i = g ij e j, t = t j1 jq (g i1k1 e k1 ) (g ipkp e kp ) e j1 e jq Therefore, the (p + q)-contravariant coordinates of t are = ( t j1 jq g i1k1 g ipkp) e k1 e kp e j1 e jq t i1 ipj1 jq = t j1 jq k 1 k p g k1i1 g kpip
18 3 TENSOR PRODUCT OF LINEAR MAPS 18 3 Tensor product of linear maps 31 Theorem For i = 1,, n, let A i : E i F i be a linear map There is a linear map such that (1) for all x 1 E 1,, x n E n A 1 A n : E 1 E n F 1 F n (A 1 A n )(x 1 x n ) = A 1 (x 1 ) A n (x n ) Proof This follows from the universal property of tensor products, applied to the multilinear map E 1 E n F 1 F n which carries (x 1,, x n ) E 1 E n to A 1 (x 1 ) A n (x n ) Clearly, the linear map A 1 A n is uniquely determined by condition (1), since every element in E 1 E n is a sum of products x 1 x n 32 Proposition 1 If A i : E i F i and B i : F i G i are linear maps, for i = 1,, n, then (B 1 A 1 ) (B n A n ) = (B 1 B n ) (A 1 A n ) 2 Id E1 Id En = Id E1 E n 3 If A i is injective (resp surjective) for all i = 1,, n, then A 1 A n is injective (resp surjective) Proof and 1 For x 1 E 1,, x n E n, we have (B 1 A 1 ) (B n A n )(x 1 x n ) = B 1 A 1 (x 1 ) B n A n (x n ) (B 1 B n ) (A 1 A n )(x 1 x n ) = (B 1 B n ) ( A 1 (x 1 ) A n (x n ) ) 2 For x 1 E 1,, x n E n we have = B 1 A 1 (x 1 ) B n A n (x n ) (Id E1 Id En )(x 1 x n ) = Id E1 (x 1 ) Id En (x n ) = x 1 x n = Id E1 E n (x 1 x n ) 3 If A i is injective (resp surjective), then there is a linear map B i : F i E i such that B i A i = Id Ei (resp A i B i = Id Fi ), hence (B 1 B n ) (A 1 A n ) = Id E1 E n (resp (A 1 A n ) (B 1 B n ) = Id F1 F n ) by the first two parts of the proposition It follows that A 1 A n is injective (resp surjective)
Boolean Inner-Product Spaces and Boolean Matrices
Boolean Inner-Product Spaces and Boolean Matrices Stan Gudder Department of Mathematics, University of Denver, Denver CO 80208 Frédéric Latrémolière Department of Mathematics, University of Denver, Denver
More informationMath 396. Quotient spaces
Math 396. Quotient spaces. Definition Let F be a field, V a vector space over F and W V a subspace of V. For v, v V, we say that v v mod W if and only if v v W. One can readily verify that with this definition
More informationMath 443 Differential Geometry Spring Handout 3: Bilinear and Quadratic Forms This handout should be read just before Chapter 4 of the textbook.
Math 443 Differential Geometry Spring 2013 Handout 3: Bilinear and Quadratic Forms This handout should be read just before Chapter 4 of the textbook. Endomorphisms of a Vector Space This handout discusses
More informationA Brief Introduction to Tensors
A Brief Introduction to Tensors Jay R Walton Fall 2013 1 Preliminaries In general, a tensor is a multilinear transformation defined over an underlying finite dimensional vector space In this brief introduction,
More informationLinear Algebra M1 - FIB. Contents: 5. Matrices, systems of linear equations and determinants 6. Vector space 7. Linear maps 8.
Linear Algebra M1 - FIB Contents: 5 Matrices, systems of linear equations and determinants 6 Vector space 7 Linear maps 8 Diagonalization Anna de Mier Montserrat Maureso Dept Matemàtica Aplicada II Translation:
More information1.8 Dual Spaces (non-examinable)
2 Theorem 1715 is just a restatement in terms of linear morphisms of a fact that you might have come across before: every m n matrix can be row-reduced to reduced echelon form using row operations Moreover,
More informationLinear algebra 2. Yoav Zemel. March 1, 2012
Linear algebra 2 Yoav Zemel March 1, 2012 These notes were written by Yoav Zemel. The lecturer, Shmuel Berger, should not be held responsible for any mistake. Any comments are welcome at zamsh7@gmail.com.
More informationLinear Algebra Notes. Lecture Notes, University of Toronto, Fall 2016
Linear Algebra Notes Lecture Notes, University of Toronto, Fall 2016 (Ctd ) 11 Isomorphisms 1 Linear maps Definition 11 An invertible linear map T : V W is called a linear isomorphism from V to W Etymology:
More information4.2. ORTHOGONALITY 161
4.2. ORTHOGONALITY 161 Definition 4.2.9 An affine space (E, E ) is a Euclidean affine space iff its underlying vector space E is a Euclidean vector space. Given any two points a, b E, we define the distance
More informationLinear Vector Spaces
CHAPTER 1 Linear Vector Spaces Definition 1.0.1. A linear vector space over a field F is a triple (V, +, ), where V is a set, + : V V V and : F V V are maps with the properties : (i) ( x, y V ), x + y
More informationMatrix representation of a linear map
Matrix representation of a linear map As before, let e i = (0,..., 0, 1, 0,..., 0) T, with 1 in the i th place and 0 elsewhere, be standard basis vectors. Given linear map f : R n R m we get n column vectors
More informationNOTES on LINEAR ALGEBRA 1
School of Economics, Management and Statistics University of Bologna Academic Year 207/8 NOTES on LINEAR ALGEBRA for the students of Stats and Maths This is a modified version of the notes by Prof Laura
More informationDS-GA 1002 Lecture notes 0 Fall Linear Algebra. These notes provide a review of basic concepts in linear algebra.
DS-GA 1002 Lecture notes 0 Fall 2016 Linear Algebra These notes provide a review of basic concepts in linear algebra. 1 Vector spaces You are no doubt familiar with vectors in R 2 or R 3, i.e. [ ] 1.1
More informationFirst we introduce the sets that are going to serve as the generalizations of the scalars.
Contents 1 Fields...................................... 2 2 Vector spaces.................................. 4 3 Matrices..................................... 7 4 Linear systems and matrices..........................
More informationVectors To begin, let us describe an element of the state space as a point with numerical coordinates, that is x 1. x 2. x =
Linear Algebra Review Vectors To begin, let us describe an element of the state space as a point with numerical coordinates, that is x 1 x x = 2. x n Vectors of up to three dimensions are easy to diagram.
More information1. General Vector Spaces
1.1. Vector space axioms. 1. General Vector Spaces Definition 1.1. Let V be a nonempty set of objects on which the operations of addition and scalar multiplication are defined. By addition we mean a rule
More informationVector spaces, duals and endomorphisms
Vector spaces, duals and endomorphisms A real vector space V is a set equipped with an additive operation which is commutative and associative, has a zero element 0 and has an additive inverse v for any
More informationFoundations of Matrix Analysis
1 Foundations of Matrix Analysis In this chapter we recall the basic elements of linear algebra which will be employed in the remainder of the text For most of the proofs as well as for the details, the
More informationMath113: Linear Algebra. Beifang Chen
Math3: Linear Algebra Beifang Chen Spring 26 Contents Systems of Linear Equations 3 Systems of Linear Equations 3 Linear Systems 3 2 Geometric Interpretation 3 3 Matrices of Linear Systems 4 4 Elementary
More information5 Linear Transformations
Lecture 13 5 Linear Transformations 5.1 Basic Definitions and Examples We have already come across with the notion of linear transformations on euclidean spaces. We shall now see that this notion readily
More informationTopics in linear algebra
Chapter 6 Topics in linear algebra 6.1 Change of basis I want to remind you of one of the basic ideas in linear algebra: change of basis. Let F be a field, V and W be finite dimensional vector spaces over
More information1 Linear transformations; the basics
Linear Algebra Fall 2013 Linear Transformations 1 Linear transformations; the basics Definition 1 Let V, W be vector spaces over the same field F. A linear transformation (also known as linear map, or
More informationLinear Algebra March 16, 2019
Linear Algebra March 16, 2019 2 Contents 0.1 Notation................................ 4 1 Systems of linear equations, and matrices 5 1.1 Systems of linear equations..................... 5 1.2 Augmented
More informationDIVISORS ON NONSINGULAR CURVES
DIVISORS ON NONSINGULAR CURVES BRIAN OSSERMAN We now begin a closer study of the behavior of projective nonsingular curves, and morphisms between them, as well as to projective space. To this end, we introduce
More informationMath 52H: Multilinear algebra, differential forms and Stokes theorem. Yakov Eliashberg
Math 52H: Multilinear algebra, differential forms and Stokes theorem Yakov Eliashberg March 202 2 Contents I Multilinear Algebra 7 Linear and multilinear functions 9. Dual space.........................................
More informationReview of linear algebra
Review of linear algebra 1 Vectors and matrices We will just touch very briefly on certain aspects of linear algebra, most of which should be familiar. Recall that we deal with vectors, i.e. elements of
More informationMatrix representation of a linear map
Matrix representation of a linear map As before, let e i = (0,..., 0, 1, 0,..., 0) T, with 1 in the i th place and 0 elsewhere, be standard basis vectors. Given linear map f : R n R m we get n column vectors
More informationMATH 101A: ALGEBRA I PART C: TENSOR PRODUCT AND MULTILINEAR ALGEBRA. This is the title page for the notes on tensor products and multilinear algebra.
MATH 101A: ALGEBRA I PART C: TENSOR PRODUCT AND MULTILINEAR ALGEBRA This is the title page for the notes on tensor products and multilinear algebra. Contents 1. Bilinear forms and quadratic forms 1 1.1.
More informationNONCOMMUTATIVE POLYNOMIAL EQUATIONS. Edward S. Letzter. Introduction
NONCOMMUTATIVE POLYNOMIAL EQUATIONS Edward S Letzter Introduction My aim in these notes is twofold: First, to briefly review some linear algebra Second, to provide you with some new tools and techniques
More informationLINEAR ALGEBRA REVIEW
LINEAR ALGEBRA REVIEW JC Stuff you should know for the exam. 1. Basics on vector spaces (1) F n is the set of all n-tuples (a 1,... a n ) with a i F. It forms a VS with the operations of + and scalar multiplication
More informationNOTES ON LINEAR ALGEBRA. 1. Determinants
NOTES ON LINEAR ALGEBRA 1 Determinants In this section, we study determinants of matrices We study their properties, methods of computation and some applications We are familiar with the following formulas
More information5 Quiver Representations
5 Quiver Representations 5. Problems Problem 5.. Field embeddings. Recall that k(y,..., y m ) denotes the field of rational functions of y,..., y m over a field k. Let f : k[x,..., x n ] k(y,..., y m )
More informationMath 3108: Linear Algebra
Math 3108: Linear Algebra Instructor: Jason Murphy Department of Mathematics and Statistics Missouri University of Science and Technology 1 / 323 Contents. Chapter 1. Slides 3 70 Chapter 2. Slides 71 118
More informationMath 350 Fall 2011 Notes about inner product spaces. In this notes we state and prove some important properties of inner product spaces.
Math 350 Fall 2011 Notes about inner product spaces In this notes we state and prove some important properties of inner product spaces. First, recall the dot product on R n : if x, y R n, say x = (x 1,...,
More informationChapter 1: Systems of Linear Equations
Chapter : Systems of Linear Equations February, 9 Systems of linear equations Linear systems Lecture A linear equation in variables x, x,, x n is an equation of the form a x + a x + + a n x n = b, where
More informationMath Topology II: Smooth Manifolds. Spring Homework 2 Solution Submit solutions to the following problems:
Math 132 - Topology II: Smooth Manifolds. Spring 2017. Homework 2 Solution Submit solutions to the following problems: 1. Let H = {a + bi + cj + dk (a, b, c, d) R 4 }, where i 2 = j 2 = k 2 = 1, ij = k,
More informationMath 396. Linear algebra operations on vector bundles
Math 396. Linear algebra operations on vector bundles 1. Motivation Let (X, O) be a C p premanifold with corners, 0 p. We have developed the notion of a C p vector bundle over X as a certain kind of C
More informationChapter 2 Linear Transformations
Chapter 2 Linear Transformations Linear Transformations Loosely speaking, a linear transformation is a function from one vector space to another that preserves the vector space operations. Let us be more
More informationA PRIMER ON SESQUILINEAR FORMS
A PRIMER ON SESQUILINEAR FORMS BRIAN OSSERMAN This is an alternative presentation of most of the material from 8., 8.2, 8.3, 8.4, 8.5 and 8.8 of Artin s book. Any terminology (such as sesquilinear form
More informationLECTURE 3: REPRESENTATION THEORY OF SL 2 (C) AND sl 2 (C)
LECTURE 3: REPRESENTATION THEORY OF SL 2 (C) AND sl 2 (C) IVAN LOSEV Introduction We proceed to studying the representation theory of algebraic groups and Lie algebras. Algebraic groups are the groups
More informationReview 1 Math 321: Linear Algebra Spring 2010
Department of Mathematics and Statistics University of New Mexico Review 1 Math 321: Linear Algebra Spring 2010 This is a review for Midterm 1 that will be on Thursday March 11th, 2010. The main topics
More informationKnowledge Discovery and Data Mining 1 (VO) ( )
Knowledge Discovery and Data Mining 1 (VO) (707.003) Review of Linear Algebra Denis Helic KTI, TU Graz Oct 9, 2014 Denis Helic (KTI, TU Graz) KDDM1 Oct 9, 2014 1 / 74 Big picture: KDDM Probability Theory
More informationSupplementary Notes on Linear Algebra
Supplementary Notes on Linear Algebra Mariusz Wodzicki May 3, 2015 1 Vector spaces 1.1 Coordinatization of a vector space 1.1.1 Given a basis B = {b 1,..., b n } in a vector space V, any vector v V can
More informationWhere is matrix multiplication locally open?
Linear Algebra and its Applications 517 (2017) 167 176 Contents lists available at ScienceDirect Linear Algebra and its Applications www.elsevier.com/locate/laa Where is matrix multiplication locally open?
More informationMTH 309 Supplemental Lecture Notes Based on Robert Messer, Linear Algebra Gateway to Mathematics
MTH 309 Supplemental Lecture Notes Based on Robert Messer, Linear Algebra Gateway to Mathematics Ulrich Meierfrankenfeld Department of Mathematics Michigan State University East Lansing MI 48824 meier@math.msu.edu
More informationSome notes on linear algebra
Some notes on linear algebra Throughout these notes, k denotes a field (often called the scalars in this context). Recall that this means that there are two binary operations on k, denoted + and, that
More informationREPRESENTATION THEORY, LECTURE 0. BASICS
REPRESENTATION THEORY, LECTURE 0. BASICS IVAN LOSEV Introduction The aim of this lecture is to recall some standard basic things about the representation theory of finite dimensional algebras and finite
More informationAlgebra II. Paulius Drungilas and Jonas Jankauskas
Algebra II Paulius Drungilas and Jonas Jankauskas Contents 1. Quadratic forms 3 What is quadratic form? 3 Change of variables. 3 Equivalence of quadratic forms. 4 Canonical form. 4 Normal form. 7 Positive
More informationPhys 201. Matrices and Determinants
Phys 201 Matrices and Determinants 1 1.1 Matrices 1.2 Operations of matrices 1.3 Types of matrices 1.4 Properties of matrices 1.5 Determinants 1.6 Inverse of a 3 3 matrix 2 1.1 Matrices A 2 3 7 =! " 1
More informationQuantum Computing Lecture 2. Review of Linear Algebra
Quantum Computing Lecture 2 Review of Linear Algebra Maris Ozols Linear algebra States of a quantum system form a vector space and their transformations are described by linear operators Vector spaces
More informationProblem set 2. Math 212b February 8, 2001 due Feb. 27
Problem set 2 Math 212b February 8, 2001 due Feb. 27 Contents 1 The L 2 Euler operator 1 2 Symplectic vector spaces. 2 2.1 Special kinds of subspaces....................... 3 2.2 Normal forms..............................
More informationMATHEMATICS 217 NOTES
MATHEMATICS 27 NOTES PART I THE JORDAN CANONICAL FORM The characteristic polynomial of an n n matrix A is the polynomial χ A (λ) = det(λi A), a monic polynomial of degree n; a monic polynomial in the variable
More informationLecture Summaries for Linear Algebra M51A
These lecture summaries may also be viewed online by clicking the L icon at the top right of any lecture screen. Lecture Summaries for Linear Algebra M51A refers to the section in the textbook. Lecture
More informationLinear Algebra. Min Yan
Linear Algebra Min Yan January 2, 2018 2 Contents 1 Vector Space 7 1.1 Definition................................. 7 1.1.1 Axioms of Vector Space..................... 7 1.1.2 Consequence of Axiom......................
More informationSolution to Homework 1
Solution to Homework Sec 2 (a) Yes It is condition (VS 3) (b) No If x, y are both zero vectors Then by condition (VS 3) x = x + y = y (c) No Let e be the zero vector We have e = 2e (d) No It will be false
More information5 Dedekind extensions
18.785 Number theory I Fall 2016 Lecture #5 09/22/2016 5 Dedekind extensions In this lecture we prove that the integral closure of a Dedekind domain in a finite extension of its fraction field is also
More informationLinear Algebra Lecture Notes-II
Linear Algebra Lecture Notes-II Vikas Bist Department of Mathematics Panjab University, Chandigarh-64 email: bistvikas@gmail.com Last revised on March 5, 8 This text is based on the lectures delivered
More information2.2. Show that U 0 is a vector space. For each α 0 in F, show by example that U α does not satisfy closure.
Hints for Exercises 1.3. This diagram says that f α = β g. I will prove f injective g injective. You should show g injective f injective. Assume f is injective. Now suppose g(x) = g(y) for some x, y A.
More informationLinear Algebra. Matrices Operations. Consider, for example, a system of equations such as x + 2y z + 4w = 0, 3x 4y + 2z 6w = 0, x 3y 2z + w = 0.
Matrices Operations Linear Algebra Consider, for example, a system of equations such as x + 2y z + 4w = 0, 3x 4y + 2z 6w = 0, x 3y 2z + w = 0 The rectangular array 1 2 1 4 3 4 2 6 1 3 2 1 in which the
More informationDiagonalizing Matrices
Diagonalizing Matrices Massoud Malek A A Let A = A k be an n n non-singular matrix and let B = A = [B, B,, B k,, B n ] Then A n A B = A A 0 0 A k [B, B,, B k,, B n ] = 0 0 = I n 0 A n Notice that A i B
More information(v, w) = arccos( < v, w >
MA322 Sathaye Notes on Inner Products Notes on Chapter 6 Inner product. Given a real vector space V, an inner product is defined to be a bilinear map F : V V R such that the following holds: For all v
More informationIntroduction to Geometry
Introduction to Geometry it is a draft of lecture notes of H.M. Khudaverdian. Manchester, 18 May 211 Contents 1 Euclidean space 3 1.1 Vector space............................ 3 1.2 Basic example of n-dimensional
More informationLinear and Bilinear Algebra (2WF04) Jan Draisma
Linear and Bilinear Algebra (2WF04) Jan Draisma CHAPTER 10 Symmetric bilinear forms We already know what a bilinear form β on a K-vector space V is: it is a function β : V V K that satisfies β(u + v,
More informationHomework 5 M 373K Mark Lindberg and Travis Schedler
Homework 5 M 373K Mark Lindberg and Travis Schedler 1. Artin, Chapter 3, Exercise.1. Prove that the numbers of the form a + b, where a and b are rational numbers, form a subfield of C. Let F be the numbers
More informationNotes on basis changes and matrix diagonalization
Notes on basis changes and matrix diagonalization Howard E Haber Santa Cruz Institute for Particle Physics, University of California, Santa Cruz, CA 95064 April 17, 2017 1 Coordinates of vectors and matrix
More informationFinal Review Sheet. B = (1, 1 + 3x, 1 + x 2 ) then 2 + 3x + 6x 2
Final Review Sheet The final will cover Sections Chapters 1,2,3 and 4, as well as sections 5.1-5.4, 6.1-6.2 and 7.1-7.3 from chapters 5,6 and 7. This is essentially all material covered this term. Watch
More informationNotes on Linear Algebra
1 Notes on Linear Algebra Jean Walrand August 2005 I INTRODUCTION Linear Algebra is the theory of linear transformations Applications abound in estimation control and Markov chains You should be familiar
More informationMathematics 1. Part II: Linear Algebra. Exercises and problems
Bachelor Degree in Informatics Engineering Barcelona School of Informatics Mathematics Part II: Linear Algebra Eercises and problems February 5 Departament de Matemàtica Aplicada Universitat Politècnica
More informationEquality: Two matrices A and B are equal, i.e., A = B if A and B have the same order and the entries of A and B are the same.
Introduction Matrix Operations Matrix: An m n matrix A is an m-by-n array of scalars from a field (for example real numbers) of the form a a a n a a a n A a m a m a mn The order (or size) of A is m n (read
More informationA matrix over a field F is a rectangular array of elements from F. The symbol
Chapter MATRICES Matrix arithmetic A matrix over a field F is a rectangular array of elements from F The symbol M m n (F ) denotes the collection of all m n matrices over F Matrices will usually be denoted
More informationTopic 1: Matrix diagonalization
Topic : Matrix diagonalization Review of Matrices and Determinants Definition A matrix is a rectangular array of real numbers a a a m a A = a a m a n a n a nm The matrix is said to be of order n m if it
More informationALGEBRA 8: Linear algebra: characteristic polynomial
ALGEBRA 8: Linear algebra: characteristic polynomial Characteristic polynomial Definition 8.1. Consider a linear operator A End V over a vector space V. Consider a vector v V such that A(v) = λv. This
More informationa(b + c) = ab + ac a, b, c k
Lecture 2. The Categories of Vector Spaces PCMI Summer 2015 Undergraduate Lectures on Flag Varieties Lecture 2. We discuss the categories of vector spaces and linear maps. Since vector spaces are always
More informationMatrices and Linear Algebra
Contents Quantitative methods for Economics and Business University of Ferrara Academic year 2017-2018 Contents 1 Basics 2 3 4 5 Contents 1 Basics 2 3 4 5 Contents 1 Basics 2 3 4 5 Contents 1 Basics 2
More informationThe following definition is fundamental.
1. Some Basics from Linear Algebra With these notes, I will try and clarify certain topics that I only quickly mention in class. First and foremost, I will assume that you are familiar with many basic
More informationDot Products, Transposes, and Orthogonal Projections
Dot Products, Transposes, and Orthogonal Projections David Jekel November 13, 2015 Properties of Dot Products Recall that the dot product or standard inner product on R n is given by x y = x 1 y 1 + +
More informationMATH 23a, FALL 2002 THEORETICAL LINEAR ALGEBRA AND MULTIVARIABLE CALCULUS Solutions to Final Exam (in-class portion) January 22, 2003
MATH 23a, FALL 2002 THEORETICAL LINEAR ALGEBRA AND MULTIVARIABLE CALCULUS Solutions to Final Exam (in-class portion) January 22, 2003 1. True or False (28 points, 2 each) T or F If V is a vector space
More informationVector Spaces and Linear Transformations
Vector Spaces and Linear Transformations Wei Shi, Jinan University 2017.11.1 1 / 18 Definition (Field) A field F = {F, +, } is an algebraic structure formed by a set F, and closed under binary operations
More informationChap. 3. Controlled Systems, Controllability
Chap. 3. Controlled Systems, Controllability 1. Controllability of Linear Systems 1.1. Kalman s Criterion Consider the linear system ẋ = Ax + Bu where x R n : state vector and u R m : input vector. A :
More informationCombining Association Schemes
Chapter 4 Combining Association Schemes 4.1 Tensor products In this chapter we take an association scheme on Ω 1 and an association scheme on Ω, and combine them to obtain an association scheme on Ω 1
More information235 Final exam review questions
5 Final exam review questions Paul Hacking December 4, 0 () Let A be an n n matrix and T : R n R n, T (x) = Ax the linear transformation with matrix A. What does it mean to say that a vector v R n is an
More informationMath 113 Midterm Exam Solutions
Math 113 Midterm Exam Solutions Held Thursday, May 7, 2013, 7-9 pm. 1. (10 points) Let V be a vector space over F and T : V V be a linear operator. Suppose that there is a non-zero vector v V such that
More informationUpper triangular matrices and Billiard Arrays
Linear Algebra and its Applications 493 (2016) 508 536 Contents lists available at ScienceDirect Linear Algebra and its Applications www.elsevier.com/locate/laa Upper triangular matrices and Billiard Arrays
More informationAlgebra Workshops 10 and 11
Algebra Workshops 1 and 11 Suggestion: For Workshop 1 please do questions 2,3 and 14. For the other questions, it s best to wait till the material is covered in lectures. Bilinear and Quadratic Forms on
More informationj=1 x j p, if 1 p <, x i ξ : x i < ξ} 0 as p.
LINEAR ALGEBRA Fall 203 The final exam Almost all of the problems solved Exercise Let (V, ) be a normed vector space. Prove x y x y for all x, y V. Everybody knows how to do this! Exercise 2 If V is a
More informationThe Proj Construction
The Proj Construction Daniel Murfet May 16, 2006 Contents 1 Basic Properties 1 2 Functorial Properties 2 3 Products 6 4 Linear Morphisms 9 5 Projective Morphisms 9 6 Dimensions of Schemes 11 7 Points of
More informationAN ELEMENTARY PROOF OF THE SPECTRAL RADIUS FORMULA FOR MATRICES
AN ELEMENTARY PROOF OF THE SPECTRAL RADIUS FORMULA FOR MATRICES JOEL A. TROPP Abstract. We present an elementary proof that the spectral radius of a matrix A may be obtained using the formula ρ(a) lim
More informationLINEAR ALGEBRA MICHAEL PENKAVA
LINEAR ALGEBRA MICHAEL PENKAVA 1. Linear Maps Definition 1.1. If V and W are vector spaces over the same field K, then a map λ : V W is called a linear map if it satisfies the two conditions below: (1)
More information2: LINEAR TRANSFORMATIONS AND MATRICES
2: LINEAR TRANSFORMATIONS AND MATRICES STEVEN HEILMAN Contents 1. Review 1 2. Linear Transformations 1 3. Null spaces, range, coordinate bases 2 4. Linear Transformations and Bases 4 5. Matrix Representation,
More informationINSTITIÚID TEICNEOLAÍOCHTA CHEATHARLACH INSTITUTE OF TECHNOLOGY CARLOW MATRICES
1 CHAPTER 4 MATRICES 1 INSTITIÚID TEICNEOLAÍOCHTA CHEATHARLACH INSTITUTE OF TECHNOLOGY CARLOW MATRICES 1 Matrices Matrices are of fundamental importance in 2-dimensional and 3-dimensional graphics programming
More informationLinear Algebra. Chapter 5
Chapter 5 Linear Algebra The guiding theme in linear algebra is the interplay between algebraic manipulations and geometric interpretations. This dual representation is what makes linear algebra a fruitful
More informationVector Space Basics. 1 Abstract Vector Spaces. 1. (commutativity of vector addition) u + v = v + u. 2. (associativity of vector addition)
Vector Space Basics (Remark: these notes are highly formal and may be a useful reference to some students however I am also posting Ray Heitmann's notes to Canvas for students interested in a direct computational
More informationElementary linear algebra
Chapter 1 Elementary linear algebra 1.1 Vector spaces Vector spaces owe their importance to the fact that so many models arising in the solutions of specific problems turn out to be vector spaces. The
More informationELEMENTARY LINEAR ALGEBRA WITH APPLICATIONS. 1. Linear Equations and Matrices
ELEMENTARY LINEAR ALGEBRA WITH APPLICATIONS KOLMAN & HILL NOTES BY OTTO MUTZBAUER 11 Systems of Linear Equations 1 Linear Equations and Matrices Numbers in our context are either real numbers or complex
More information0.1 Rational Canonical Forms
We have already seen that it is useful and simpler to study linear systems using matrices. But matrices are themselves cumbersome, as they are stuffed with many entries, and it turns out that it s best
More informationChapter 4 Euclid Space
Chapter 4 Euclid Space Inner Product Spaces Definition.. Let V be a real vector space over IR. A real inner product on V is a real valued function on V V, denoted by (, ), which satisfies () (x, y) = (y,
More informationQUATERNIONS AND ROTATIONS
QUATERNIONS AND ROTATIONS SVANTE JANSON 1. Introduction The purpose of this note is to show some well-known relations between quaternions and the Lie groups SO(3) and SO(4) (rotations in R 3 and R 4 )
More informationALGEBRA QUALIFYING EXAM PROBLEMS LINEAR ALGEBRA
ALGEBRA QUALIFYING EXAM PROBLEMS LINEAR ALGEBRA Kent State University Department of Mathematical Sciences Compiled and Maintained by Donald L. White Version: August 29, 2017 CONTENTS LINEAR ALGEBRA AND
More informationChapter 7. Linear Algebra: Matrices, Vectors,
Chapter 7. Linear Algebra: Matrices, Vectors, Determinants. Linear Systems Linear algebra includes the theory and application of linear systems of equations, linear transformations, and eigenvalue problems.
More informationMATH 320: PRACTICE PROBLEMS FOR THE FINAL AND SOLUTIONS
MATH 320: PRACTICE PROBLEMS FOR THE FINAL AND SOLUTIONS There will be eight problems on the final. The following are sample problems. Problem 1. Let F be the vector space of all real valued functions on
More information