Generalized Eigenvectors and Jordan Form

Similar documents
Remark By definition, an eigenvector must be a nonzero vector, but eigenvalue could be zero.

Math 240 Calculus III

Chapter 7. Canonical Forms. 7.1 Eigenvalues and Eigenvectors

Remark 1 By definition, an eigenvector must be a nonzero vector, but eigenvalue could be zero.

1. In this problem, if the statement is always true, circle T; otherwise, circle F.

235 Final exam review questions

Definition (T -invariant subspace) Example. Example

MATH 1553-C MIDTERM EXAMINATION 3

Jordan Canonical Form Homework Solutions

Chapter 5. Eigenvalues and Eigenvectors

Eigenvalues, Eigenvectors, and Diagonalization

Definitions for Quizzes

Eigenvalues and Eigenvectors

Math Matrix Algebra

MATH 1553 PRACTICE MIDTERM 3 (VERSION B)

and let s calculate the image of some vectors under the transformation T.

Linear Algebra II Lecture 13

Math 3191 Applied Linear Algebra

Diagonalization. Hung-yi Lee

Section 8.2 : Homogeneous Linear Systems

MATH 304 Linear Algebra Lecture 23: Diagonalization. Review for Test 2.

Computationally, diagonal matrices are the easiest to work with. With this idea in mind, we introduce similarity:

Math 2331 Linear Algebra

DIAGONALIZATION. In order to see the implications of this definition, let us consider the following example Example 1. Consider the matrix

MATH 221, Spring Homework 10 Solutions

Linear algebra II Tutorial solutions #1 A = x 1

Diagonalization of Matrix

MAC Module 12 Eigenvalues and Eigenvectors. Learning Objectives. Upon completing this module, you should be able to:

MAC Module 12 Eigenvalues and Eigenvectors

Math Linear Algebra Final Exam Review Sheet

Diagonalization of Matrices

Symmetric and anti symmetric matrices

Math 205, Summer I, Week 4b: Continued. Chapter 5, Section 8

Linear Algebra - Part II

Definition: An n x n matrix, "A", is said to be diagonalizable if there exists a nonsingular matrix "X" and a diagonal matrix "D" such that X 1 A X

Systems of Linear Differential Equations Chapter 7

Linear Algebra Practice Final

Eigenspaces and Diagonalizable Transformations

Linear Algebra review Powers of a diagonalizable matrix Spectral decomposition

Study Guide for Linear Algebra Exam 2

Announcements Monday, November 06

MATH 304 Linear Algebra Lecture 33: Bases of eigenvectors. Diagonalization.

Announcements Wednesday, November 7

Final Exam Practice Problems Answers Math 24 Winter 2012

Math Final December 2006 C. Robinson

Math 250B Midterm III Information Fall 2018 SOLUTIONS TO PRACTICE PROBLEMS

Announcements Wednesday, November 7

4. Linear transformations as a vector space 17

Solutions to Final Exam

Dimension. Eigenvalue and eigenvector

LINEAR ALGEBRA 1, 2012-I PARTIAL EXAM 3 SOLUTIONS TO PRACTICE PROBLEMS

IMPORTANT DEFINITIONS AND THEOREMS REFERENCE SHEET

Lecture 15, 16: Diagonalization

MATH 304 Linear Algebra Lecture 34: Review for Test 2.

Eigenvalues, Eigenvectors. Eigenvalues and eigenvector will be fundamentally related to the nature of the solutions of state space systems.

Homework sheet 4: EIGENVALUES AND EIGENVECTORS. DIAGONALIZATION (with solutions) Year ? Why or why not? 6 9

Linear Algebra. Rekha Santhanam. April 3, Johns Hopkins Univ. Rekha Santhanam (Johns Hopkins Univ.) Linear Algebra April 3, / 7

Linear Algebra review Powers of a diagonalizable matrix Spectral decomposition

Therefore, A and B have the same characteristic polynomial and hence, the same eigenvalues.

MATH SOLUTIONS TO PRACTICE MIDTERM LECTURE 1, SUMMER Given vector spaces V and W, V W is the vector space given by

Calculating determinants for larger matrices

Spectral radius, symmetric and positive matrices

Econ Slides from Lecture 7

2. Every linear system with the same number of equations as unknowns has a unique solution.

Eigenvalues and Eigenvectors

Question: Given an n x n matrix A, how do we find its eigenvalues? Idea: Suppose c is an eigenvalue of A, then what is the determinant of A-cI?

MATH 423 Linear Algebra II Lecture 20: Geometry of linear transformations. Eigenvalues and eigenvectors. Characteristic polynomial.

IMPORTANT DEFINITIONS AND THEOREMS REFERENCE SHEET

NONCOMMUTATIVE POLYNOMIAL EQUATIONS. Edward S. Letzter. Introduction

ft-uiowa-math2550 Assignment NOTRequiredJustHWformatOfQuizReviewForExam3part2 due 12/31/2014 at 07:10pm CST

DM554 Linear and Integer Programming. Lecture 9. Diagonalization. Marco Chiarandini

After we have found an eigenvalue λ of an n n matrix A, we have to find the vectors v in R n such that

Math Ordinary Differential Equations

Math 205, Summer I, Week 4b:

Jordan Normal Form and Singular Decomposition

Linear algebra II Homework #1 solutions A = This means that every eigenvector with eigenvalue λ = 1 must have the form

City Suburbs. : population distribution after m years

Linear Algebra II Lecture 22

Announcements Monday, October 29

Eigenvalues for Triangular Matrices. ENGI 7825: Linear Algebra Review Finding Eigenvalues and Diagonalization

PROBLEM SET. Problems on Eigenvalues and Diagonalization. Math 3351, Fall Oct. 20, 2010 ANSWERS

EIGENVALUES AND EIGENVECTORS

MAT1302F Mathematical Methods II Lecture 19

33A Linear Algebra and Applications: Practice Final Exam - Solutions

MATH 115A: SAMPLE FINAL SOLUTIONS

EXAM. Exam #3. Math 2360, Spring April 24, 2001 ANSWERS

Eigenvalues and Eigenvectors 7.2 Diagonalization

Eigenvalue and Eigenvector Homework

Modeling and Simulation with ODE for MSE

Chapter 5. Linear Algebra. A linear (algebraic) equation in. unknowns, x 1, x 2,..., x n, is. an equation of the form

MAT 1302B Mathematical Methods II

Agenda: Understand the action of A by seeing how it acts on eigenvectors.

Matrices related to linear transformations

Lecture 1: Review of linear algebra

A matrix is a rectangular array of. objects arranged in rows and columns. The objects are called the entries. is called the size of the matrix, and

av 1 x 2 + 4y 2 + xy + 4z 2 = 16.

Chapter 5 Eigenvalues and Eigenvectors

c Igor Zelenko, Fall

LINEAR ALGEBRA BOOT CAMP WEEK 2: LINEAR OPERATORS

A matrix is a rectangular array of. objects arranged in rows and columns. The objects are called the entries. is called the size of the matrix, and

Transcription:

Generalized Eigenvectors and Jordan Form We have seen that an n n matrix A is diagonalizable precisely when the dimensions of its eigenspaces sum to n. So if A is not diagonalizable, there is at least one eigenvalue with a geometric multiplicity (dimension of its eigenspace) which is strictly less than its algebraic multiplicity. In this handout, we will discuss how one can make up for this deficiency of eigenvectors by finding what are called generalized eigenvectors, which can in turn be used to find the Jordan form of the matrix A. First consider the following non-diagonalizable system. Example. The matrix A = [ ] 3 3 has characteristic polynomial (λ 3) ([ 2, so ]) it has only one eigenvalue λ = 3, and the corresponding eigenspace is E 3 = span. Since dim(e 3 ) = < 3, the matrix A is not diagonalizable. Nevertheless, it is still possible to solve the system dy = Ay without much difficulty. Writing out the two equations dy = 3y + y 2 dy 2 = 3y 2 we see that the second equation has general solution y 2 = c e 3t. Plugging this into the first equation gives dy = 3y + c e 3t This is a first order linear equation that can be solved by the method of integrating factors. Its solution is y = c te 3t + c 2 e 3t. Thus the general solution of the system is [ ] [ ] ( [ ] [ ]) [ ] y c te y = = 3t + c 2 e 3t c 2 e 3t = c e 3t t + + c 2 e 3t y 2 The matrix A in the previous example is said to be in Jordan form. Definition. A 2 2 matrix of the form J = [ ] λ λ for some number λ is called a Jordan form matrix.

The following theorem generalizes the calculations of the previous example. Theorem. The general solution of the system dy = Jy, J = [ ] λ λ is y = c e λt (t [ ] + [ ]) [ ] + c 2 e λt In general, suppose A is a 2 2 matrix with a single repeated eigenvalue λ with dim(e λ ) =. Then, as we have seen before, making the change of variables y = Cx transforms the system dy = Ay into the system dx = C ACx. We now want to find a matrix C such that the coefficient matrix C AC for the new system is a Jordan form matrix J. That is, we want AC = CJ. Writing we have Therefore the columns of C must satisfy C = v v 2, J = [ ] λ, λ AC = Av Av 2, CJ = λv λv 2 + v Av = λv Av 2 = λv 2 + v Thus the vector v is an eigenvector with eigenvalue λ. Rewriting these equations (A λi)v = (A λi)v 2 = v it follows that (A λi) 2 v 2 = (A λi)v =. Thus the vector v 2 must be in ker(a λi) 2. 2

Definition 2. A nonzero vector v which satisfies (A λi) p v = for some positive integer p is called a generalized eigenvector of A with eigenvalue λ. The vectors v and v 2 form a generalized eigenvector chain, as the following diagram illustrates: v 2 v A λi A λi Therefore, to find the columns of the matrix C that puts A in Jordan form, we must find a chain of generalized eigenvectors, as follows: Find a nonzero vector v 2 in ker(a λi) 2 that is not in ker(a λi). Set v = (A λi)v 2. Having found such vectors, we can then write down the solution of the system dy From above, we know that the solution of the system dx = Jx is [ ] [ ]) [ ] x = c e (t λt + + c 2 e λt so multiplying this by C we get In summary, we have the following result. y = Cx = c e λt (tv + v 2 ) + c 2 e λt v. = Ay. Theorem 2. Let A be a 2 2 matrix and suppose v 2 v is a chain of generalized eigenvectors of A with eigenvalue λ. Then the general solution of the system dy = Ay is y(t) = c e λt (tv + v 2 ) + c 2 e λt v. Example 2. Let A = [ 2 ] 4 The characteristic polynomial of A is λ 2 6λ + 9 = (λ 3) 2, so λ = 3 is the only eigenvalue of A. Next, we compute [ ] [ ] A 3I =, (A 3I) 2 = Now we choose[ v] 2 to be any vector in ker(a 3I) 2 that is not in ker(a [ ] 3I). One such vector is v 2 =. With this choice, we than have v = (A 3I)v 2 =. The solution of the system dy = Ay is therefore [ ] [ ]) [ ] y = c e (t 3t + + c 2 e 3t. 3

For larger systems it turns out that any chain of generalized eigenvectors of the coefficient matrix can be used to write down a linearly independent set of solutions. Theorem 3. Suppose {v, v 2,..., v k } is a chain of generalized eigenvectors of A with eigenvalue λ. That is, v k v k v k 2 v 2 v. A λi A λi A λi A λi A λi A λi Then y (t) = e λt v y 2 (t) = e λt (v 2 + tv ) [ ] y 3 (t) = e λt v 3 + tv 2 + t2 2! v. [ ] y k (t) = e λt v k + tv k + t2 2! v k 2 + + tk 2 (k 2)! v 2 + tk (k )! v is a linearly independent set of solutions of dy = Ay. Proof. We show that y k is a solution. The proof that the others are solutions is similar. Differentiating gives dy k = e λt [ (λv k + v k ) + t(λv k + v k 2 ) + t2 + tk 2 (k 2)! (λv 2 + v ) + tk (k )! (λv ) 2! (λv k 2 + v k 3 ) + ]. Multiplying by A gives [ ] Ay k = e λt (Av k ) + t(av k ) + t2 2! (Av k 2) + + tk 2 (k 2)! (Av 2) + tk (k )! (Av ). The two expressions above are equal, since Av = λv, Av 2 = λv 2 + v, and so on. Next we show that the chain {v, v 2,..., v k } is necessarily linearly independent. Suppose c v + c 2 v 2 + + c k v k =. () The multiplying by (A λi) k sends v through v k to zero, and v k to v, so we are left with c k v =. Since v is nonzero, this implies c k =. Similar reasoning shows that the remaining coefficients must also be zero. Thus the chain of generalized eigenvectors is linearly independent. Now suppose c y (t) + c 2 y 2 (t) + + c n y n (t) = for all t. Then since y j () = v j, at t = this equation is precisely equation () above, which we just showed implies c = c 2 = = c k =. 4

The next theorem guarantees that we can always find enough solutions of this form to generate a fundamental set of solutions. Theorem 4. Let A be an n n matrix, and suppose λ is an eigenvalue of A with algebraic multiplicity m. Then there is some integer p m such that dim(ker(a λi) p ) = m. Example 3. Let 5 4 A = 4 3 5 3 2 The characteristic polynomial of A is p A (λ) = (λ 2) 3. Since 3 4 A 2I = 4 5, rref 3 4 the eigenspace E 2 = ker(a 2I) has dimension. Next, since (A 2I) 2 =, rref the space ker(a 2I) 2 has dimension 2. Finally, since (A 2I) 3 =, the space ker(a 2I) 3 has dimension 3, which matches the algebraic multiplicity of the eigenvalue λ = 2. We can now form a chain of 3 generalized eigenvectors by choosing a vector v 3 in ker(a 2I) 3 and defining v 2 = (A 2I)v 3 and v = (A 2I)v 2 = (A 2I) 2 v 3. To ensure that v 2 and v are both non-zero, we need v 3 to not be in ker(a 2I) 2 (which in turn implies that v 3 is not in ker(a 2I)). Since ker(a 2I) 3 = R 3, we can choose v 3 to be any vector not in ker(a 2I) 2. Since the first column of (A 2I) 2 is non-zero, we may choose 3 v 3 = = v 2 = (A 2I)v 3 = 4 = v = (A 2I)v 2 =. 3 5

Then y (t) = e 2t 3 y 2 (t) = e 2t t + 4 3 3 y 3 (t) = e 2t t2 + t 4 + 2! 3 is a fundamental set of solutions of dy = Ay. There may in general be more than one chain of generalized eigenvectors corresponding to a given eigenvalue. Since the last vector in each chain is an eigenvector, the number of chains corresponding to an eigenvalue λ is equal to the dimension of the eigenspace E λ. Example 4. Let 4 A = 2 2. 4 The characteristic polynomial of A is p A (λ) = (λ 3) 3. Since A 3I = 2 2 2, rref the eigenspace E 3 = ker(a 3I) has dimension 2, so there will be two chains. Next, since (A 3I) 2 =, the space ker(a 3I) 2 has dimension 3, which matches the algebraic multiplicity of λ = 3. Thus one of the chains will have length 2, so the other must have length. We now form a chain of 2 generalized eigenvectors by choosing v 2 in ker(a 3I) 2 such that v 2 is not in ker(a 3I). Since every vector is in ker(a 3I) 2 and the first column of A 3I is non-zero,we may again choose v 2 = = v = (A 3I)v 2 = To form a basis for R 3, we now need one additional chain of one generalized eigenvector. This vector must be an eigenvector independent from v. Since E 3 = span,, 6 2

and neither of these spanning vectors is itself a scalar multiple of v, we may choose either one of them. So let w = Then we have the two chains v 2 v w. The first chain generates the two solutions y (t) = e 3t 2 y 2 (t) = e 3t t 2 + and the second chain generates a third solution y 3 (t) = e 3t and together y, y 2 and y 3 form a fundamental set of solutions of dy = Ay. Example 5. Let 5 3 2 A = 5 3 5 5 The characteristic polynomial of A is (λ 5) 4. Since 3 2 A 5I = 3 rref 3, the eigenspace E 5 = ker(a 5I) has dimension 2, and there are two chains. Since (A 5I) 2 = the dimension of the space ker(a 5I) 2 is 4, matching the algebraic multiplicity of λ = 5. 7

To form a chain of length 2, we first choose a vector v 2 in ker(a 5I) 2 which is not in ker(a 5I). Let v 2 = = v = (A 5I)v 2 =. We now need one more chain of length 2. To find a second chain, we need to choose another vector w 2 in ker(a 5I) 2 which is not in ker(a 5I), in such a way that w 2 is independent of v 2 and w = (A 5I)w 2 is independent of v. One such choice is w 2 = = w = (A 5I)w 2 = 2 3. Thus y (t) = e 5t y 2 (t) = e 5t t + y 3 (t) = e 5t 2 3 y 4 (t) = e 5t t 2 3 + is a fundamental set of solutions of dy = Ay. Jordan Form for n n Matrices Definition 3. A matrix of the form λ λ......... λ consisting of λ in each entry along the main diagonal, in each entry directly above the main diagonal, and elsewhere, is called an elementary Jordan block. A matrix of the form J J 2... J k 8

where each J i is an elementary Jordan block (with possibly different values of λ) is called a Jordan form matrix. Example 6. The following are examples of Jordan form matrices: 2 2 3 2 2, 2, 3, 3 2 3 3 Theorem 5. For any n n matrix A, there is a matrix C and a Jordan form matrix J such that C AC = J. The Jordan matrix J is determined by the number and length of the generalized eigenvector chains for A. Example 7. Let A be the matrix in Example 3. Since Av = 2v Av 2 = v + 2v 2 Av 2 = v 2 + 2v 3, letting implies C = v v 2 v 3 2 AC = 2v v + 2v 2 v 2 + 2v 3 = v v 2 v 3 2 = CJ, 2 so C AC = J consists of a single elementary Jordan block. Example 8. Let A be the matrix in Example 4. Since Av = 3v Av 2 = v + 3v 2 Aw = 3w, letting implies C = v v 2 w 3 AC = 3v v + 3v 2 w = v v 2 w 3 = CJ, 3 so C AC = J consists of 2 elementary Jordan blocks, one 2 2 and one. 9