EIGENVALUES AND EIGENVECTORS 3

Similar documents
October 4, 2017 EIGENVALUES AND EIGENVECTORS. APPLICATIONS

September 26, 2017 EIGENVALUES AND EIGENVECTORS. APPLICATIONS

m We can similarly replace any pair of complex conjugate eigenvalues with 2 2 real blocks. = R

DETERMINANTS. , x 2 = a 11b 2 a 21 b 1

Remark By definition, an eigenvector must be a nonzero vector, but eigenvalue could be zero.

Study Guide for Linear Algebra Exam 2

Remark 1 By definition, an eigenvector must be a nonzero vector, but eigenvalue could be zero.

Math 3191 Applied Linear Algebra

Topic 1: Matrix diagonalization

October 25, 2013 INNER PRODUCT SPACES

a 11 a 12 a 11 a 12 a 13 a 21 a 22 a 23 . a 31 a 32 a 33 a 12 a 21 a 23 a 31 a = = = = 12

Econ Slides from Lecture 7

1. Linear systems of equations. Chapters 7-8: Linear Algebra. Solution(s) of a linear system of equations (continued)

Math Matrix Algebra

and let s calculate the image of some vectors under the transformation T.

Numerical Linear Algebra Homework Assignment - Week 2

Chapter Two Elements of Linear Algebra

1. General Vector Spaces

MATH 583A REVIEW SESSION #1

Chapter 5 Eigenvalues and Eigenvectors

1. What is the determinant of the following matrix? a 1 a 2 4a 3 2a 2 b 1 b 2 4b 3 2b c 1. = 4, then det

Linear Algebra, part 2 Eigenvalues, eigenvectors and least squares solutions

Solutions to Final Exam

c c c c c c c c c c a 3x3 matrix C= has a determinant determined by

MATH 240 Spring, Chapter 1: Linear Equations and Matrices

8. Diagonalization.

Lecture Summaries for Linear Algebra M51A

Announcements Wednesday, November 7

Question: Given an n x n matrix A, how do we find its eigenvalues? Idea: Suppose c is an eigenvalue of A, then what is the determinant of A-cI?

Systems of Algebraic Equations and Systems of Differential Equations

MATH 31 - ADDITIONAL PRACTICE PROBLEMS FOR FINAL

Computationally, diagonal matrices are the easiest to work with. With this idea in mind, we introduce similarity:

MTH 464: Computational Linear Algebra

12. Perturbed Matrices

Problem Set 9 Due: In class Tuesday, Nov. 27 Late papers will be accepted until 12:00 on Thursday (at the beginning of class).

Eigenvalues and Eigenvectors. Review: Invertibility. Eigenvalues and Eigenvectors. The Finite Dimensional Case. January 18, 2018

Linear Algebra M1 - FIB. Contents: 5. Matrices, systems of linear equations and determinants 6. Vector space 7. Linear maps 8.

MAT Linear Algebra Collection of sample exams

Chapter 2: Linear Independence and Bases

Linear algebra I Homework #1 due Thursday, Oct Show that the diagonals of a square are orthogonal to one another.

Systems of Second Order Differential Equations Cayley-Hamilton-Ziebur

Equality: Two matrices A and B are equal, i.e., A = B if A and B have the same order and the entries of A and B are the same.

Math Linear Algebra Final Exam Review Sheet

Recall : Eigenvalues and Eigenvectors

Lecture 7: Positive Semidefinite Matrices

A = 3 1. We conclude that the algebraic multiplicity of the eigenvalues are both one, that is,

Eigenvalues and Eigenvectors

Notes on the Matrix-Tree theorem and Cayley s tree enumerator

Review problems for MA 54, Fall 2004.

Elementary maths for GMT

Announcements Wednesday, November 7

Lecture 11: Eigenvalues and Eigenvectors

A matrix is a rectangular array of. objects arranged in rows and columns. The objects are called the entries. is called the size of the matrix, and

GRE Subject test preparation Spring 2016 Topic: Abstract Algebra, Linear Algebra, Number Theory.

MATRIX ALGEBRA AND SYSTEMS OF EQUATIONS. + + x 1 x 2. x n 8 (4) 3 4 2

systems of linear di erential If the homogeneous linear di erential system is diagonalizable,

(b) If a multiple of one row of A is added to another row to produce B then det(b) =det(a).

Chapter 1: Systems of Linear Equations

A matrix is a rectangular array of. objects arranged in rows and columns. The objects are called the entries. is called the size of the matrix, and

Announcements Wednesday, November 01

2 Eigenvectors and Eigenvalues in abstract spaces.

Previously Monte Carlo Integration

Lecture 12: Diagonalization

Calculating determinants for larger matrices

4.1 Eigenvalues, Eigenvectors, and The Characteristic Polynomial

MATH 20F: LINEAR ALGEBRA LECTURE B00 (T. KEMP)

ANALYTICAL MATHEMATICS FOR APPLICATIONS 2018 LECTURE NOTES 3

4. Linear transformations as a vector space 17

Eigenspaces and Diagonalizable Transformations

LU Factorization. A m x n matrix A admits an LU factorization if it can be written in the form of A = LU

1 Invariant subspaces

Problem Set (T) If A is an m n matrix, B is an n p matrix and D is a p s matrix, then show

Announcements Wednesday, November 01

Cayley-Hamilton Theorem

Properties of Linear Transformations from R n to R m

3 Matrix Algebra. 3.1 Operations on matrices

1. In this problem, if the statement is always true, circle T; otherwise, circle F.

33AH, WINTER 2018: STUDY GUIDE FOR FINAL EXAM

MAT1302F Mathematical Methods II Lecture 19

Linear Algebra. Min Yan

Eigenvalues and Eigenvectors: An Introduction

MATH 1120 (LINEAR ALGEBRA 1), FINAL EXAM FALL 2011 SOLUTIONS TO PRACTICE VERSION

Linear Algebra Primer

2 b 3 b 4. c c 2 c 3 c 4

Chapter 7. Linear Algebra: Matrices, Vectors,

Linear Algebra. Matrices Operations. Consider, for example, a system of equations such as x + 2y z + 4w = 0, 3x 4y + 2z 6w = 0, x 3y 2z + w = 0.

2. Every linear system with the same number of equations as unknowns has a unique solution.

Linear Algebra Primer

Diagonalization of Matrix

Linear Algebra- Final Exam Review

Chapter 3. Determinants and Eigenvalues

Homework For each of the following matrices, find the minimal polynomial and determine whether the matrix is diagonalizable.

Review of Linear Algebra

LINEAR ALGEBRA 1, 2012-I PARTIAL EXAM 3 SOLUTIONS TO PRACTICE PROBLEMS

Eigenvalues and Eigenvectors

Diagonalisierung. Eigenwerte, Eigenvektoren, Mathematische Methoden der Physik I. Vorlesungsnotizen zu

18.S34 linear algebra problems (2007)

9.1 Eigenvectors and Eigenvalues of a Linear Map

Econ 204 Supplement to Section 3.6 Diagonalization and Quadratic Forms. 1 Diagonalization and Change of Basis

Midterm for Introduction to Numerical Analysis I, AMSC/CMSC 466, on 10/29/2015

Transcription:

EIGENVALUES AND EIGENVECTORS 3 1. Motivation 1.1. Diagonal matrices. Perhaps the simplest type of linear transformations are those whose matrix is diagonal (in some basis). Consider for example the matrices a1 0 b1 0 (1) M =, N = 0 a 2 0 b 2 It can be easily checked that a1 + b M + N = 1 0 0 a 2 + b 2 and M 1 = 1 a 1 0 1, M k a k = 1 0 0 a 2 0 a k 2 a1 b, MN = 1 0 0 a 2 b 2 Diagonal matrices behave like the bunch of numbers on their diagonal! The linear transformation consisting of multiplication by the matrix M in (1) dialates a 1 times vectors in the direction of e 1 and a 2 times vectors in the direction of e 2. In this chapter we will see that most linear transformations do have diagonal matrices in a special basis, whose elements are called the eigenvectors of the transformation. We will learn how to find these bases. Along the special directions of the eigenvectors the transformation just dialation by a factor, called eigenvalue. 1.2. Example: solving linear di erential equations. Consider the simple equation du dt = u which is linear, homogeneous, with constant coe cients, and unknown function u(t) 2 R (or in C). Its general solution is, as it is well known, u(t) =Ce t. Consider now a similar equation, but where the unknown u(t) is a vector valued function: du (2) dt = Mu, u(t) 2 Rn, M is an n n constant matrix Inspired by the one dimensional case we look for exponential solutions. Substituting in (2) u(t) =e t v (where is a number and v is a constant vector, both be determined) and dividing by e t, we obtain that the scalar and the vector v must satisfy (3) v = Mv or (4) (M I)v = 0

4 RODICA D. COSTIN If the null space of the matrix M I is zero, then the only solution of (4) is v = 0 which gives the (trivial!) solution u(t) 0. If however, we can find special values of for which N (M I) is not null, then we found a nontrivial solution of (2). Such values of are called eigenvalues of the matrix M, and vectors v 2N(M I), v 6= 0, are called eigenvectors corresponding to the eigenvalue. Of course, the necessary and su cient condition for N (M I) 6= {0} is that (5) det(m I) =0 Example. Let us calculate the exponential solutions for " # 1 3 (6) M = 0 2 Looking for eigenvalues of M we solve equation (5), which for (6) is " # " #! 1 3 1 0 1 3 det = =( 1 )(2 ) 0 2 0 1 0 2 with solutions 1 = 1 and 2 = 2. We next determine an eigenvector corresponding to the eigenvalue = 1 = 1: looking for a nozero vector v 1 such that (M 1 I)v 1 = 0 we solve " #" # " # 0 3 x1 0 = 0 3 0 giving x 2 = 0 and x 1 arbitrary; therefore the first eigenvector is any scalar multiple of v 1 =(0, 1) T. Similarly, for the eigenvalue = 2 = 2 we solve (M 2 I)v 2 = 0: " #" # " # 3 3 y1 0 = 0 0 0 which gives y 2 = y 1 and y 1 arbitrary, and the second eigenvector is any scalar multiple of v 2 =(1, 1) T. We found two particular solutions of (2), (6), namely u 1 (t) =e t (0, 1) T and u 2 (t) =e 2t (1, 1) T. These are functions belonging to the null space of the linear operator Lu = du dx M u, therefore any linear combination of these two solutions also belongs to the null space: any C 1 u 1 (t)+c 2 u 2 (t) is also a solution, for and constants C 1,C 2. A bit later we will show that these are all the solutions. x 2 2. Eigenvalues and eigenvectors: definition and calculation 2.1. Definitions. Denote the set of n n (square) matrices with entries in F (= R or C) M n (F )={M M =[M ij ] i,j=1,...n, M ij 2 F } y 2

EIGENVALUES AND EIGENVECTORS 5 A matrix M 2M n (F ) defines an endomorphism the vector space F n (over the scalars F ) by usual multiplication x 7! Mx. Note that a matrix with real entries can also act on C n, since for any x 2 C n also Mx 2 C n. But a matrix with complex non real entries cannot act on R n, since for x 2 R n the image Mx may not belong to R n (while certainly Mx 2 C n ). Definition 1. Let M be an n n matrix acting on the vector space V = F n. Ascalar 2 F is an eigenvalue of M if for some nonzero vector v 2 V, v 6= 0 we have (7) Mv = v The vector v is called eigenvector corresponding to the eigenvalue. Of course, if v is an eigenvector corresponding to multiple cv (for c 6= 0)., then so is any scalar 2.2. The characteristic equation. Equation (7) can be rewritten as M v v = 0, or (M I)v = 0, which means that the nonzero vector v belongs to the null space of the matrix M I, and in particular this matrix is not invertible. Using the theory of matrices, we know that this is equivalent to det(m I) =0 The determinant has the form M 11 M 12... M 1n det(m I) = M 21 M 22... M 2n... M n1 M n2... M nn This is a polynomial in, having degree n. To understand why this is the case, consider first n = 2 and n = 3. For n = 2 the characteristic polynomial is M 11 M 12 M 21 M 22 =(M 11 )(M 22 ) M 12 M 21 = 2 (M 11 + M 22 ) +(M 11 M 22 M 12 M 21 ) which is a quadratic polynomial in ; the dominant coe cient is 1. For n = 3 the characteristic polynomial is and expanding along say, row 1, M 11 M 12 M 13 M 21 M 22 M 23 M 13 M 23 M 33 =( 1) 1+1 (M 11 ) M 22 M 23 M 23 M 33 +( 1) 1+2 M 12 M 21 M 23 M 13 M 33 +( 1) 1+3 M 13 M 21 M 22 M 13 M 23

6 RODICA D. COSTIN = 3 +(M 11 + M 22 + M 33 ) 2 +... which is a cubic polynomial in ; the dominant coe cient is 1. It is easy to show by induction that det(m I) is polynomial in, having degree n, and that the coe cient of n is ( 1) n. Definition 2. The polynomial det(m I) is called the characteristic polynomial of the matrix M, and the equation det(m I) =0is called the characteristic equation of M. Remark. Some authors refer to the characteristic polynomial as det( I M); the two polynomial are either equal or a 1 multiple of each other, since det( I M) =( 1) n det(m I). 2.3. Geometric interpretation of eigenvalues and eigenvectors. Let M be an n n matrix, and T : R n! R n defined by T (x) =Mx be the corresponding linear transformation. If v is an eigenvector corresponding to an eigenvalue of M: Mv = v, then T expands or contracts v (and any vector in its direction) times (and it does not change its direction!). If the eigenvalue/vector are not real, a similar fact is true, only that multiplication by a complex (not real) scalar cannot be easily called an expansion or a contraction (there is no ordering in complex numbers), see the example of rotations, 2.13.1. The special directions of the eigenvectors are called principal axes of the linear transformation (or of the matrix). 2.4. Digression: the fundamental theorem of algebra. 2.4.1. Polynomials of degree two: roots and factorization. Consider polynomials of degree two, with real coe cients: p(x) =ax 2 + bx + c. It is well known that p(x) has real solutions x 1,2 = b ± p b 2 4ac if b 2 4ac 0 2a (where x 1 = x 2 when b 2 4ac = 0), and p(x) has no real solutions if b 2 4ac < 0. When the solutions are real, then the polynomial factors as ax 2 + bx + c = a(x x 1 )(x x 2 ) In particular, if x 1 = x 2 then p(x) =a(x x 1 ) 2 and x 1 is called a double root; x 1 is said to have multiplicity two. It is convenient to say that also in this case p(x) has two roots. If, on the other hand, if p(x) has no real roots, then p cannot be factored within real numbers, and it is called irreducible (over the real numbers). 2.4.2. Complex numbers and factorization of polynomials of degree two. If p(x) =ax 2 +bx+c is irreducible this means that b 2 4ac < 0 and we cannot take the square root of this quantity in order to calculate the two roots of

EIGENVALUES AND EIGENVECTORS 7 p(x). However, writing b 2 4ac =( 1) ( b 2 +4ac) and introducing the symbol i for p 1 we can write the zeroes of p(x) as p x 1,2 = b ± ip b 2 +4ac 2a = b 2a ± i b 2 +4ac 2a 2 R + ir = C Considering the two roots x 1,x 2 complex, we can still factor ax 2 +bx+c = a(x x 1 )(x x 2 ), only now the factors have complex coe cients. Within complex numbers every polynomial of degree two is reducible! Note that the two roots of a quadratic polynomial with real coe cients are complex conjugate: if a, b, c 2 R and x 1,2 62 R then x 2 = x 1. 2.4.3. The fundamental theorem of algebra. It is absolutely remarkable that any polynomial can be completely factored using complex numbers: Theorem 3. The fundamental theorem of algebra Any polynomial p(x) =a n x n +a n 1 x n 1 +...+a 0 with coe cients a j 2 C can be factored (8) a n x n + a n 1 x n 1 +...+ a 0 = a n (x x 1 )(x x 2 )...(x x n ) for a unique set of complex numbers x 1,x 2,...,x n (not necessarily distinct), called the roots of the polynomial p(x). Remark. With probability one, the zeroes x 1,...,x n of polynomials p(x) are distinct. Indeed, if x 1 is a double root (or has higher multiplicity) then both relations p(x 1 ) = 0 and p 0 (x 1 ) = 0 must hold. This means that there is a relation between the coe cients a 0,...a n of p(x) (themultiplet(a 0,...a n ) belongs to an n dimensional surface in C n+1 ). 2.4.4. Factorization within real numbers. If we want to restrict ourselves only within real numbers then we can factor any polynomial into factors of degree one or two: Theorem 4. Factorization within real numbers Any polynomial of degree n with real coe cients can be factored into factors of degree one or two with real coe cients. Theorem 4 is an easy consequence of the (deep) Theorem 3. Indeed, first, factor the polynomial in complex numbers (8). Then note that the zeroes x 1,x 2,...,x n come in pairs of complex conjugate numbers, since if z satisfies p(z) = 0, then also its complex conjugate z satisfies p(z) = 0. Then each pair of factors (x z)(x z) must be replaced in (8) by its expanded value: (x z)(x z) =x 2 (z + z) x + z 2 which is an irreducible polynomial of degree 2, with real coe cients. 2

8 RODICA D. COSTIN 2.5. Diagonal matrices. Let D be a diagonal matrix: 2 3 d 1 0... 0 0 d 2... 0 (9) D = 6 7 4... 5 0 0... d n To find its eigenvalues, calculate det(d I) = d 1 0... 0 0 d 2... 0... 0 0... d n =(d 1 1 )(d 2 2 )...(d n n ) The eigenvalues are precisely the diagonal elements, and the eigenvector corresponding to d j is e j (as it is easy to check). The principal axes of diagonal matrices the coordinate axes. Vectors in the direction of one of these axes preserve their direction and are stretched or compressed: if x = ce k then Dx = d k x. Diagonal matrices are easy to work with: what was noted for the 2 2 matrices in 1 is true in general, and one can easily check that any power D k is the diagonal matrix having d k j on the diagonal. If p(x) is a polynomial p(t) =a k t k + a k 1 t k 1 +...+ a 1 t + a 0 then for any square matrix M one can define p(m) as p(m) =a k M k + a k 1 M k 1 +...+ a 1 M + a 0 I If D is a diagonal matrix (9) then p(d) is the diagonal matrix having p(d j ) on the diagonal. (Check!) Diagonal matrices can be viewed as the collection of their eigenvalues! Exercise. Show that the eigenvalues of an upper (or lower) triangular matrix are the elements on the diagonal. 2.6. Similar matrices have the same eigenvalues. It is very easy to work with diagonal matrices and a natural question arises: which linear transformations have a diagonal matrix in a well chosen basis? This is the main topic we will be exploring for many sections to come. Recall that if the matrix M represents the linear transformation L : V! V in some basis B V of V, and the matrix M represents the same linear transformation L, only in a di erent basis B V, then the two matrices are similar: M = S 1 MS (where S the the matrix of change of basis). Eigenvalues are associated to the linear transformation (rather than its matrix representation):

EIGENVALUES AND EIGENVECTORS 9 Proposition 5. Two similar matrices have the same eigenvalues: if M, M,S are n n matrices, and M = S 1 MS then the eigenvalues of M and of M are the same. This is very easy to check, since det( M I) =det(s 1 MS I) =det S 1 (M I)S =dets 1 det(m I) dets =det(m I) so M and M have the same characteristic equation. 2 2.7. Projections. Recall that projections do satisfy P 2 = P (we saw this for projections in dimension two, and we will prove it in general). Proposition 6. Let P be a square matrix satisfying P 2 = P. Then the eigenvalues of P can only be 0 or 1. Proof. Let be an eigenvalue; this means that there is a nonzero vector v so that P v = v. Applying P to both sides of the equality we obtain P 2 v = P ( v) = P v = 2 v. Using the fact that P 2 v = P v = v it follows that v = 2 v so ( 2 )v = 0 and since v 6= 0 then 2 =0so 2{0, 1}. 2 Example. Consider the projection of R 3 onto the x 1 x 2 plane. Its matrix 2 P = 4 1 0 0 3 0 1 0 5 0 0 0 is diagonal, with eigenvalues 1, 1, 0. 2.8. Trace, determinant and eigenvalues. Definition 7. Let M be an n n matrix, M =[M ij ] i,j=1,...,n. The trace of M is the sum of its elements on the principal diagonal: nx Tr M = The following theorem shows that what we noticed in 2.2 for n =2is true for any n: Theorem 8. Let M be an n n matrix, an let 1,..., n be its n eigenvalues (complex, not necessarily distinct). Then j=1 M jj (10) det M = 1 2... n and (11) Tr M = 1 + 2 +...+ n In particular, the traces of similar matrices are equal, and so are their determinants.

10 RODICA D. COSTIN Sketch of the proof. The coe cients of any polynomial can be written in terms of its roots 1, and, in particular, it is not hard to see that (12) p(x) (x 1)(x 2)...(x n) = x n ( 1 + 2 +...+ n )x n 1 +...+( 1) n ( 1 2... n) In particular, p(0) = ( 1) n 1 2... n. The characteristic polynomial factors as det(m I) =( 1) n ( 1)...( n) ( 1) n p( ) (recall that the dominant coe cient of the characteristic polynomial is ( 1) n ) and (10) follows. To show (11) we expand the determinant det(m I) using minors and cofactors keeping track of the coe cient of n 1. As seen on the examples in 2.2, only the first term in the expansion contains the power n 1, and continuing to expand to lower and lower dimensional determinants, we see that the only term containing n 1 is (M 11 )(M 22 )...(M nn ) =( 1) n n ( 1) n (M 11 + M 22 +...+ M nn ) n 1 + lower powers of which compared to (12) gives (11). 2 2.9. The eigenvalue zero. As an immediate consequence of Theorem 8, we can recognize invertible matrices by looking at their eigenvalues: Corollary 9. AmatrixM is invertible if and only if all its eigenvalues are nonzero. Note that a matrix M has an eigenvalue equal to zero if and only if its null space N (M) is nontrivial. Moreover, the matrix M has dim N (M) eigenvectors linearly independent which correspond to the eigenvalue zero. 2.10. Eigenvectors corresponding to di erent eigenvalues are independent. Theorem 10. Let M be an n n matrix. Let 1,... k a set of distinct eigenvalues of M and v 1,...,v k be corresponding eigenvectors. Then the set v 1,...,v k is linearly independent. In particular, if M has entries in F = R or C, and all the eigenvalues of M are in F and are distinct, then the set of corresponding eigenvectors form a basis for F n. 1 These are called Vieta s formulas.

EIGENVALUES AND EIGENVECTORS 11 Proof. Assume, to obtain a contradiction, that the eigenvectors are linearly dependent: there are c 1,...,c k 2 F not all zero such that (13) c 1 v 1 +...+ c k v k = 0 Step I. We can assume that all c j are not zero, otherwise we just remove those v j from (13) and we have a similar equation with a smaller k. If after this procedure we are left with k = 1, then this implies c 1 v 1 = 0 which contradicts the fact that not all c j are zero or the fact that v 1 6= 0. Otherwise, for k 2 we continue as follows. Step II. Then we can solve (13) for v k : (14) v k = c 0 1v 1 +...+ c 0 k 1 v k 1 where c 0 j = c j/c k. Applying M to both sides of (14) we obtain (15) kv k = c 0 1 1v 1 +...+ c 0 k 1 k 1v k 1 Multiplying (14) by k and subtracting from (15) we obtain (16) 0 = c 0 1( 1 k )v 1 +...+ c 0 k 1 ( k 1 k)v k 1 Note that all c 0 j ( j k) are non-zero (since all c 0 1 are non-zero, and j 6= k ). If k=2, then this implies v 1 = 0 which is a contradiction. If k>2 we go to Step I. with a lower k. The procedure decreases k, therefore it must end, and we have a contradiction. 2 2.11. Diagonalization of matrices with linearly independent eigenvectors. Suppose that the M be an n n matrix has n independent eigenvectors v 1,...,v n. Note that, by Theorem 10, this is the case if we work in F = C and all the eigenvalues are distinct (recall that this happens with probability one). Also this is the case if we work in F = R and all the eigenvalues are real and distinct. Let S be the matrix with columns v 1,...,v n : S =[v 1,...,v n ] which is invertible, since v 1,...,v n are linearly independent. Since Mv j = j v j then (17) M[v 1,...,v n ]=[ 1v 1,..., nv n ]

12 RODICA D. COSTIN The left side of (17) equals MS. To identify the matrix on the right side of (17) note that since Se j = v j then S( j e j )= j v j so 2 3 1 0... 0 0 2... 0 [ 1v 1,..., nv n ]=S[ 1e 1,..., ne n ]=S, where = 6 7 4... 5 0 0... n Relation (17) is therefore MS = S, or S 1 MS = = diagonal Note that the matrix S which diagonalizes a matrix is not unique. For example, we can replace any eigenvector by a scalar multiple of it. Also, we can use di erent orders for the eigenvectors (this will result on a diagonal matrix with the same values on the diagonal, but in di erent positions). Example 1. Consider the matrix (6) for which we found the eigenvalues 1 = 1 and 2 = 2 and the corresponding eigenvectors v 1 = (0, 1) T. v 2 =(1, we have 1) T. Taking S = 0 1 1 1 S 1 MS = 1 0 0 2 Not all matrices are diagonalizable, only those with distinct eigenvalues, and some matrices with multiple eigenvalues. Example 2. The matrix 0 1 (18) N = 0 0 has eigenvalues 1 = 2 = 0 but only one (up to a scalar multiple) eigenvector v 1 = e 1. Multiple eigenvalues are not guaranteed to have an equal number of independent eigenvectors! N is not diagonalizable. Indeed, assume the contrary, to arrive at a contradiction. Suppose there exists an invertible matrix S so that S 1 NS = where is diagonal, hence it has the eigenvalues of N on its diagonal, and therefore it is the zero matrix: S 1 NS = 0, which multiplied by S to the left and S 1 to the right gives N = 0, which is a contradiction. Note that matrices with multiple eigenvalues may still be diagonalized (we will see when this is the case). For example, the identity matrix has all eigenvalues equal (to 1), and is diagonal! A plethora of other examples can be found by just conjugating by an invertible matrix S a diagonal matrix which has some equal elements on the diagonal, namely by taking M = S S 1.