Vector Spaces and Linear Maps

Similar documents
MATH SOLUTIONS TO PRACTICE MIDTERM LECTURE 1, SUMMER Given vector spaces V and W, V W is the vector space given by

Homework 11 Solutions. Math 110, Fall 2013.

1 Invariant subspaces

Math 4153 Exam 1 Review

Abstract Vector Spaces

( 9x + 3y. y 3y = (λ 9)x 3x + y = λy 9x + 3y = 3λy 9x + (λ 9)x = λ(λ 9)x. (λ 2 10λ)x = 0

Eigenvalues and Eigenvectors. Prepared by Vince Zaccone For Campus Learning Assistance Services at UCSB

Definition (T -invariant subspace) Example. Example

Review of Linear Algebra

NONCOMMUTATIVE POLYNOMIAL EQUATIONS. Edward S. Letzter. Introduction

MATH 115A: SAMPLE FINAL SOLUTIONS

The converse is clear, since

Lecture 11: Finish Gaussian elimination and applications; intro to eigenvalues and eigenvectors (1)

Math 550 Notes. Chapter 2. Jesse Crawford. Department of Mathematics Tarleton State University. Fall 2010

MATH Linear Algebra

Chapter 1. Vectors, Matrices, and Linear Spaces

Honors Algebra II MATH251 Course Notes by Dr. Eyal Goren McGill University Winter 2007

Lecture 9: Vector Algebra

What is on this week. 1 Vector spaces (continued) 1.1 Null space and Column Space of a matrix

Vector Spaces and Linear Transformations

Math 4153 Exam 3 Review. The syllabus for Exam 3 is Chapter 6 (pages ), Chapter 7 through page 137, and Chapter 8 through page 182 in Axler.

Lecture 6: Corrections; Dimension; Linear maps

A Little Beyond: Linear Algebra

A linear algebra proof of the fundamental theorem of algebra

A linear algebra proof of the fundamental theorem of algebra

ALGEBRA QUALIFYING EXAM PROBLEMS LINEAR ALGEBRA

Lecture Summaries for Linear Algebra M51A

Lecture 19: Polar and singular value decompositions; generalized eigenspaces; the decomposition theorem (1)

Practice Midterm Solutions, MATH 110, Linear Algebra, Fall 2013

Lecture notes - Math 110 Lec 002, Summer The reference [LADR] stands for Axler s Linear Algebra Done Right, 3rd edition.

Definition 3 A Hamel basis (often just called a basis) of a vector space X is a linearly independent set of vectors in X that spans X.

Math 4A Notes. Written by Victoria Kala Last updated June 11, 2017

LINEAR ALGEBRA REVIEW

MATH 1120 (LINEAR ALGEBRA 1), FINAL EXAM FALL 2011 SOLUTIONS TO PRACTICE VERSION

Family Feud Review. Linear Algebra. October 22, 2013

NAME MATH 304 Examination 2 Page 1

Study Guide for Linear Algebra Exam 2

Math 121 Homework 4: Notes on Selected Problems

Lecture 19: Polar and singular value decompositions; generalized eigenspaces; the decomposition theorem (1)

Math 115A: Homework 5

Infinite-Dimensional Triangularization

Definitions for Quizzes

NOTES on LINEAR ALGEBRA 1

MATH 304 Linear Algebra Lecture 34: Review for Test 2.

YORK UNIVERSITY. Faculty of Science Department of Mathematics and Statistics MATH M Test #1. July 11, 2013 Solutions

Math 369 Exam #2 Practice Problem Solutions

A PRIMER ON SESQUILINEAR FORMS

Chapter 4 Vector Spaces And Modules

1 Last time: inverses

Online Exercises for Linear Algebra XM511

08a. Operators on Hilbert spaces. 1. Boundedness, continuity, operator norms

LINEAR ALGEBRA: THEORY. Version: August 12,

MATH 304 Linear Algebra Lecture 23: Diagonalization. Review for Test 2.

Linear Algebra. Rekha Santhanam. April 3, Johns Hopkins Univ. Rekha Santhanam (Johns Hopkins Univ.) Linear Algebra April 3, / 7

Linear Algebra Review

Math 308 Practice Test for Final Exam Winter 2015

MATH 583A REVIEW SESSION #1

IMPORTANT DEFINITIONS AND THEOREMS REFERENCE SHEET

Mathematics Department Stanford University Math 61CM/DM Vector spaces and linear maps

A supplement to Treil

Generalized eigenspaces

80 min. 65 points in total. The raw score will be normalized according to the course policy to count into the final score.

Math 314/ Exam 2 Blue Exam Solutions December 4, 2008 Instructor: Dr. S. Cooper. Name:

LINEAR ALGEBRA REVIEW

Then x 1,..., x n is a basis as desired. Indeed, it suffices to verify that it spans V, since n = dim(v ). We may write any v V as r

Linear Algebra Practice Problems

2. Every linear system with the same number of equations as unknowns has a unique solution.

Linear Algebra 2 Spectral Notes

The following definition is fundamental.

Math 113 Winter 2013 Prof. Church Midterm Solutions

Eigenspaces and Diagonalizable Transformations

Math 113 Midterm Exam Solutions

Remark By definition, an eigenvector must be a nonzero vector, but eigenvalue could be zero.

OHSx XM511 Linear Algebra: Solutions to Online True/False Exercises

Econ Lecture 8. Outline. 1. Bases 2. Linear Transformations 3. Isomorphisms

MATH 304 Linear Algebra Lecture 20: The Gram-Schmidt process (continued). Eigenvalues and eigenvectors.

Linear Algebra Final Exam Solutions, December 13, 2008

MATH 240 Spring, Chapter 1: Linear Equations and Matrices

IMPORTANT DEFINITIONS AND THEOREMS REFERENCE SHEET

University of Colorado Denver Department of Mathematical and Statistical Sciences Applied Linear Algebra Ph.D. Preliminary Exam May 25th, 2018

MAS4107 Linear Algebra 2

Lecture 21: The decomposition theorem into generalized eigenspaces; multiplicity of eigenvalues and upper-triangular matrices (1)

Math 3191 Applied Linear Algebra

(a) II and III (b) I (c) I and III (d) I and II and III (e) None are true.

Definition Suppose S R n, V R m are subspaces. A map U : S V is linear if

Math 113 Practice Final Solutions

MATRIX LIE GROUPS AND LIE GROUPS

Math 594. Solutions 5

Math 113 Homework 5. Bowei Liu, Chao Li. Fall 2013

Math Linear algebra, Spring Semester Dan Abramovich

Math Final December 2006 C. Robinson

MAT224 Practice Exercises - Week 7.5 Fall Comments

Math 110, Spring 2015: Midterm Solutions

1.8 Dual Spaces (non-examinable)

Linear Algebra. Preliminary Lecture Notes

4. Linear transformations as a vector space 17

THE MINIMAL POLYNOMIAL AND SOME APPLICATIONS

Vector Spaces. (1) Every vector space V has a zero vector 0 V

MATH 304 Linear Algebra Lecture 33: Bases of eigenvectors. Diagonalization.

Math 115A: Linear Algebra

Transcription:

Vector Spaces and Linear Maps Garrett Thomas August 14, 2018 1 About This document is part of a series of notes about math and machine learning. You are free to distribute it as you wish. The latest version can be found at http://gwthomas.github.io/notes. Please report any errors to gwthomas@stanford.edu. 2 Vector spaces Vector spaces are the basic setting in which linear algebra happens. A vector space over a field F consists of a set V (the elements of which are called vectors) along with an addition operation + : V V V and a scalar multiplication operation F V V satisfying (i) There exists an additive identity, denoted 0, in V such that v + 0 = v for all v V (ii) For each v V, there exists an additive inverse, denoted v, such that v + ( v) = 0 (iii) There exists a multiplicative identity, denoted 1, in F such that 1v = v for all v V (iv) Commutativity: u + v = v + u for all u, v V (v) Associativity: (u + v) + w = u + (v + w) and α(βv) = (αβ)v for all u, v, w V and α, β F (vi) Distributivity: α(u + v) = αu + αv and (α + β)v = αv + βv for all u, v V and α, β F We will assume F = R or F = C as these are the most common cases by far, although in general it could be any field. To justify the notations 0 for the additive identity and v for the additive inverse of v we must demonstrate their uniqueness. Proof. If w, w V satisfy v + w = v and v + w = v for all v V, then w = w + w = w + w = w This proves the uniqueness of the additive identity. Similarly, if w, w V satisfy v + w = 0 and v + w = 0, then w = w + 0 = w + v + w = 0 + w = w thus showing the uniqueness of the additive inverse. 1

We have 0v = 0 for every v V and α0 = 0 for every α F. Proof. If v V, then v = 1v = (1 + 0)v = 1v + 0v = v + 0v which implies 0v = 0. If α F, then αv = α(v + 0) = αv + α0 which implies α0 = 0. Moreover, ( 1)v = v for every v V, since v + ( 1)v = (1 + ( 1))v = 0v = 0 and we have shown that additive inverses are unique. 2.1 Subspaces Vector spaces can contain other vector spaces. If V is a vector space, then S V is said to be a subspace of V if (i) 0 S (ii) S is closed under addition: u, v S implies u + v S (iii) S is closed under scalar multiplication: v S, α F implies αv S Note that V is always a subspace of V, as is the trivial vector space which contains only 0. Proposition 1. Suppose U and W are subspaces of some vector space. Then U W is a subspace of U and a subspace of W. Proof. We only show that U W is a subspace of U; the same result follows for W since U W = W U. (i) Since 0 U and 0 W by virtue of their being subspaces, we have 0 U W. (ii) If x, y U W, then x, y U so x + y U, and x, y W so x + y W ; hence x + y U W. (iii) If v U W and α F, then v U so αv U, and v W so αv W ; hence αv U W. Thus U W is a subspace of U. 2.1.1 Sums of subspaces If U and W are subspaces of V, then their sum is defined as U + W {u + w : u U, w W } Proposition 2. Let U and W be subspaces of V. Then U + W is a subspace of V. Proof. (i) Since 0 U and 0 V, 0 = 0 + 0 U + W. 2

(ii) If v 1, v 2 U + W, then there exist u 1, u 2 U and w 1, w 2 W such that v 1 = u 1 + w 1 and v 2 = u 2 + w 2, so by the closure of U and W under addition we have v 1 + v 2 = u 1 + w 1 + u 2 + w 2 = u 1 + u }{{} 2 + w 1 + w 2 U + W }{{} U W (iii) If v U + W, α F, then there exists u U and w W such that v = u + w, so by the closure of U and W under scalar multiplication we have Thus U + W is a subspace of V. αv = α(u + w) = }{{} αu + }{{} αw U + W U W If U W = {0}, the sum is said to be a direct sum and written U W. Proposition 3. Suppose U and W are subspaces of some vector space. The following are equivalent: (i) The sum U + W is direct, i.e. U W = {0}. (ii) The only way to write 0 as a sum u + w where u U, w W is to take u = w = 0. (iii) Every vector in U + W can be written uniquely as u + w for some u U and w W. Proof. (i) = (ii): Assume (i), and suppose 0 = u + w where u U, w W. Then u = w, so u W as well, and thus u U W. By assumption this implies u = 0, so w = 0 also. (ii) = (iii): Assume (ii), and suppose v U + W can be written both as v = u + w and v = ũ + w where u, ũ U and w, w W. Then 0 = v v = u + w (ũ + w) = u ũ + w w }{{}}{{} U W so by assumption we must have u ũ = w w = 0, i.e. u = ũ and w = w. (iii) = (i): Assume (iii), and suppose v U W, so that v U and v W, nothing that v W also. Now 0 = v + ( v) and 0 = 0 + 0 both write 0 in the form u + w where u U, w W, so by the assumed uniqueness of this decomposition we conclude that v = 0. Thus U W {0}. It is also clear that {0} U W since U and W are both subspaces, so U W = {0}. 2.2 Span A linear combination of vectors v 1,..., v n V is an expression of the form α 1 v 1 + + α n v n where α 1,..., α n F. The span of a set of vectors is the set of all possible linear combinations of these: span{v 1,..., v n } {α 1 v 1 + + α n v n : α 1,..., α n F} We also define the special case span{} = {0}. Spans of sets of vectors form subspaces. Proposition 4. If v 1,..., v n V, then span{v 1,..., v n } is a subspace of V. 3

Proof. Let S = span{v 1,..., v n }. If n = 0, we have S = span{} = {0}, which is a subspace of V, so assume hereafter that n > 0. (i) Since {v 1,..., v n } is non-empty (as n 1), we have 0 = 0v 1 + + 0v n S. (ii) If x, y S, we can write x = α 1 v 1 + + α n v n and y = β 1 v 1 + + β n v n for some α 1,..., α n, β 1,..., β n F. Then x + y = α 1 v 1 + + α n v n + β 1 v 1 + + β n v n = (α 1 + β 1 )v 1 + + (α n + β n )v n S (iii) If x S, β F, we can write x = α 1 v 1 + + α n v n for some α 1,..., α n F. Then βx = β(α 1 v 1 + + α n v n ) = (βα 1 )v 1 + + (βα n )v n S Thus S is a subspace of V. Adding a vector which lies in the span of a set of vectors to that set does not expand the span in any way. Proposition 5. If w span{v 1,..., v n }, then span{v 1,..., v n, w} = span{v 1,..., v n }. Proof. Suppose w span{v 1,..., v n }. It is clear that span{v 1,..., v n } span{v 1,..., v n, w}: if u span{v 1,..., v n } then it can be written u = α 1 v 1 + +α n v n, but then u = α 1 v 1 + +α n v n +0w, so u span{v 1,..., v n, w}. To prove the other direction, write w = β 1 v 1 + + β n v n. Then for any u span{v 1,..., v m, w}, there exist α 1,..., α n, α n+1 such that u = α 1 v 1 + + α n v n + α n+1 w, and u = α 1 v 1 + + α n v n + α n+1 (β 1 v 1 + + β n v n ) = (α 1 + α n+1 β 1 )v 1 + + (α n + α n+1 β n )v n so u span{v 1,..., v m }. Hence span{v 1,..., v n, w} span{v 1,..., v n } and the claim follows. 2.3 Linear independence A set of vectors v 1,..., v n V is said to be linearly independent if α 1 v 1 + + α n v n = 0 implies α 1 = = α n = 0. Otherwise, there exist α 1,..., α n F, not all zero, such that α 1 v 1 + + α n v n = 0, and v 1,..., v n are said to be linearly dependent. Note that in the case of linear dependence, (at least) one vector can be written as a linear combination of the other vectors in the set: if α j 0, then v j = 1 α j n k j This implies that v j span{v 1,..., v j 1, v j+1,..., v n }. α k v k 4

2.4 Basis If a set of vectors is linearly independent and its span is the whole of V, those vectors are said to be a basis for V. One of the most important properties of bases is that they provide unique representations for every vector in the space they span. Proposition 6. A set of vectors v 1,..., v n V is a basis for V if and only if every vector in V can be written uniquely as a linear combination of v 1,..., v n. Proof. Suppose v 1,..., v n is a basis for V. Then these span V, so if w V then there exist α 1,..., α n such that w = α 1 v 1 + + α n v n. Furthermore, if w = α 1 v 1 + + α n v n also, then 0 = w w = α 1 v 1 + + α n v n ( α 1 v 1 + + α n v n ) = (α 1 α 1 )v n + + (α n α n )v n Since v 1,..., v n are linearly independent, this implies α j α j = 0, i.e. α j = α j, for all j = 1,..., n. Conversely, suppose that every vector in V can be written uniquely as a linear combination of v 1,..., v n. The existence part of this assumption clearly implies that v 1,..., v n span V. To show that v 1,..., v n are linearly independent, suppose and observe that since α 1 v 1 + + α n v n = 0 0v 1 + + 0v n = 0 the uniqueness of the representation of 0 as a linear combination of v 1,..., v n implies that α 1 = = α n = 0. If a vector space is spanned by a finite number of vectors, it is said to be finite-dimensional. Otherwise it is infinite-dimensional. The number of vectors in a basis for a finite-dimensional vector space V is called the dimension of V and denoted dim V. Proposition 7. If v 1,..., v n V are not all zero, then there exists a subset of {v 1,..., v n } which is linearly independent and whose span equals span{v 1,..., v n }. Proof. If v 1,..., v n are linearly independent, then we are done. Otherwise, pick v j which is in span{v 1,..., v j 1, v j+1,..., v n }. Since span{v 1,..., v j 1, v j+1,..., v n } = span{v 1,..., v n }, we can drop v j from this list without changing the span. Repeat this process until the resulting set is linearly independent. As a corollary, we can establish an important fact: every finite-dimensional vector space has a basis. Proof. Let V be a finite-dimensional vector space. By definition, this means that there exist v 1,..., v n such that V = span{v 1,..., v n }. By the previous result, there exists a linearly independent subset of {v 1,..., v n } whose span is V ; this subset is a basis for V. Proposition 8. A linearly independent set of vectors in V can be extended to a basis for V. Proof. Suppose v 1,..., v n V are linearly independent. Let w 1,..., w m be a basis for V. Then v 1,..., v n, w 1,..., w m spans V, so there exists a subset of this set which is linearly independent and also spans V. Moreover, by inspecting the process by which this linearly independent subset 5

is produced, we see that none of v 1,..., v n need be dropped because they are linearly independent. Hence the resulting linearly independent set, which is basis for V, can be chosen to contain v 1,..., v n. The dimensions of sums of subspaces obey a friendly relationship: Proposition 9. If U and W are subspaces of some vector space, then dim(u + W ) = dim U + dim W dim(u W ) Proof. Let v 1,..., v k be a basis for U W, and extend this to bases v 1,..., v k, u 1,..., u m for U and v 1,..., v k, w 1,..., w n for W. We claim that v 1,..., v k, u 1,..., u m, w 1,..., w n is a basis for U + W. If v U + W, then v = u + w for some u U, w W. Then since u = α 1 v 1 + + α k v k + α k+1 u 1 + α k+m u m for some α 1,..., α k+m F and w = β 1 v 1 + + β k v k + β k+1 w 1 + β k+n w n for some β 1,..., β k+n F, we have v = u + w = α 1 v 1 + + α k v k + α k+1 u 1 + α k+m u m + β 1 v 1 + + β k v k + β k+1 w 1 + β k+n w n = (α 1 + β 1 )v 1 + + (α k + β k )v k + α k+1 u 1 + α k+m u m + β k+1 w 1 + β k+n w n so v span{v 1,..., v k, u 1,..., u m, w 1,..., w n }. Moreover, the uniqueness of the expansion of v in this set follows from the uniqueness of the expansions of u and w. Thus v 1,..., v k, u 1,..., u m, w 1,..., w n is a basis for U + W as claimed, so dim U + dim W dim(u W ) = (k + m) + (k + n) k = k + m + n = dim(u + W ) as we set out to show. It follows immediately that dim(u W ) = dim U + dim W since dim(u W ) = dim{0} = 0 if the sum is direct. 3 Linear maps A linear map is a function T : V W, where V and W are vector spaces over F, that satisfies (i) T (u + v) = T (u) + T (v) for all u, v V (ii) T (αv) = αt (v) for all v V, α F The standard notational convention for linear maps (which we will follow hereafter) is to drop unnecessary parentheses, writing T v rather than T (v) if there is no risk of ambiguity, and denote composition of linear maps by ST rather than the usual S T. The identity map on V, which sends each v V to v, is denoted I V, or just I if the vector space V is clear from context. Note that all linear maps (not just the identity) send zero to zero. Proof. For any v V we have T v = T (v + 0 V ) = T v + T 0 V so by subtracting T v from both sides we obtain T 0 V = 0 W. 6

One useful fact regarding linear maps (both conceptually and for proofs) is that they are uniquely determined by their action on a basis. Proposition 10. Let v 1,..., v n be a basis for a vector space V, and w 1,..., w n be arbitrary vectors in a vector space W. Then there exists a unique linear map T : V W such that for all j = 1,..., n. Proof. Define T v j = w j T : V W α 1 v 1 + + α n v n α 1 w 1 + + α n w n This function is well-defined because every vector in V can be expressed uniquely as a linear combination of v 1,..., v n. Then for any x = α 1 v 1 + + α n v n and y = β 1 v 1 + + β n v n in V, T (x + y) = T (α 1 v 1 + + α n v n + β 1 v 1 + + β n v n ) = T ((α 1 + β 1 )v 1 + + (α n + β n )v n ) = (α 1 + β 1 )w 1 + + (α n + β n )w n and if γ F then = α 1 w 1 + + α n w n + β 1 w 1 + + β n w n = T x + T y T (γx) = T (γ(α 1 v 1 + + α n v n )) = T ((γα 1 )v 1 + + (γα n )v n ) = (γα 1 )w 1 + + (γα n )w n = γ(α 1 w 1 + + α n w n ) = γt x so T is a linear map. Now observe that, by construction, T v j = T (0v 1 + + 0v j 1 + 1v j + 0v j+1 + + 0v n ) = 0w 1 + + 0w j 1 + 1w j + 0w j+1 + + 0w n = w j as desired. Towards uniqueness, suppose T is a linear map which satisfies T v j = w j for j = 1,..., n. Any v V can be written uniquely as v = α 1 v 1 + + α n v n, and T (α 1 v 1 + + α n v n ) = α 1 T v1 + + α n T vn = α 1 w 1 + + α n w n by the linearity of T. Hence T = T. 3.1 Isomorphisms Observe that the definition of a linear map is suited to reflect the structure of vector spaces, since it preserves vector spaces two main operations, addition and scalar multiplication. In algebraic terms, a linear map is said to be a homomorphism of vector spaces. An invertible homomorphism where the inverse is also a homomorphism is called an isomorphism. If there exists an isomorphism from V to W, then V and W are said to be isomorphic, and we write V = W. Isomorphic vector spaces are essentially the same in terms of their algebraic structure. It is an interesting fact that finite-dimensional vector spaces of the same dimension over the same field are always isomorphic. 7

Proof. Let V and W be finite-dimensional vector spaces over F, with n dim V = dim W. We will show that V = W by exhibiting an isomorphism from V to W. To this end, let v 1,..., v n and w 1,..., w n be bases for V and W, respectively. Then there exist (unique) linear maps T : V W and S : W V satisfying T v j = w j and Sw j = v j for all j = 1,..., n. Now clearly ST v j = Sw j = v j for each j, so ST = I V. Similarly, T Sw j = T v j = w j for each j, so T S = I W. Hence S = T 1, so T is an isomorphism, and hence V = W. 3.2 Algebra of linear maps Linear maps can be added and scaled to produce new linear maps. That is, if S and T are linear maps from V into W, and α F, it is straightforward to verify that S + T and αt are linear maps from V into W. Since addition and scaling of functions satisfy the usual commutativity/associativity/distributivity rules, the set of linear maps from V into W is also a vector space over F; we denote this space by L(V, W ). The additive identity here is the zero map which sends every v V to 0 W. The composition of linear maps is also a linear map. Proposition 11. Let U, V, and W be vector spaces over a common field F, and suppose S : V W and T : U V are linear maps. Then the composition ST : U W is also a linear map. Proof. For any u, v U, and furthermore if α F then so ST is linear. (ST )(u + v) = S(T (u + v)) = S(T u + T v) = ST u + ST v (ST )(αv) = S(T (αv)) = S(αT v) = αst v The identity map is a multiplicative identity here, as T I V = I W T = T for any T : V W. 3.3 Nullspace If T : V W is a linear map, the nullspace 1 of T is the set of all vectors in V that get mapped to zero: null(t ) {v V : T v = 0 W } Proposition 12. If T : V W is a linear map, then null(t ) is a subspace of V. Proof. (i) We have already seen that T (0 V ) = 0 W, so 0 V null(t ). (ii) If u, v null(t ), then so u + v null(t ). T (u + v) = T u + T v = 0 W + 0 W = 0 W 1 It is sometimes called the kernel by algebraists, but we eschew this terminology because the word kernel has another meaning in machine learning. 8

(iii) If v null(t ), α F, then so αv null(t ). Thus null(t ) is a subspace of V. T (αv) = αt v = α0 W = 0 W Proposition 13. A linear map T : V W is injective if and only if null(t ) = {0 V }. Proof. Assume T is injective. We have seen that T 0 V = 0 W, so 0 V null(t ), and moreover if T v = 0 W the injectivity of T implies v = 0 V, so null(t ) = {0 V }. Conversely, assume null(t ) = {0 V }, and suppose T u = T v for some u, v V. Then 0 = T u T v = T (u v) so u v null(t ), which by assumption implies u v = 0, i.e. u = v. Hence T is injective. 3.4 Range The range of T is (as for any function) the set of all possible outputs of T : range(t ) {T v : v V } Proposition 14. If T : V W is a linear map, then range(t ) is a subspace of W. Proof. (i) We have already seen that T (0 V ) = 0 W, so 0 W range(t ). (ii) If w, z range(t ), there exist u, v V such that w = T u and z = T v. Then T (u + v) = T u + T v = w + z so w + z range(t ). (iii) If w range(t ), α F, there exists v V such that w = T v. Then T (αv) = αt v = αw so αw range(t ). Thus range(t ) is a subspace of W. 4 Eigenvalues and eigenvectors In the special case where the domain and codomain of a linear map are the same, certain vectors may have the special property that the map simply scales them. If T : V V is a linear map and v V is nonzero, then v is said to be an eigenvector of T with corresponding eigenvalue λ F if or equivalently, if T v = λv v null(t λi) This second definition makes it clear that the set of eigenvectors corresponding to a given eigenvalue (along with 0) is a vector space. This vector space is called the eigenspace of T associated with λ. 9