Lecture Note on Linear Algebra 16. Eigenvalues and Eigenvectors

Similar documents
Linear Algebra 18 Orthogonality

Remark 1 By definition, an eigenvector must be a nonzero vector, but eigenvalue could be zero.

Remark By definition, an eigenvector must be a nonzero vector, but eigenvalue could be zero.

Lecture Note on Linear Algebra 14. Linear Independence, Bases and Coordinates

Therefore, A and B have the same characteristic polynomial and hence, the same eigenvalues.

Math 3191 Applied Linear Algebra

and let s calculate the image of some vectors under the transformation T.

DIAGONALIZATION. In order to see the implications of this definition, let us consider the following example Example 1. Consider the matrix

Linear Algebra 20 Symmetric Matrices and Quadratic Forms

Diagonalization. MATH 322, Linear Algebra I. J. Robert Buchanan. Spring Department of Mathematics

Dimension. Eigenvalue and eigenvector

Eigenvalues, Eigenvectors, and Diagonalization

Diagonalization of Matrix

Recall : Eigenvalues and Eigenvectors

A = 3 1. We conclude that the algebraic multiplicity of the eigenvalues are both one, that is,

2 Eigenvectors and Eigenvalues in abstract spaces.

Lecture 12: Diagonalization

Chapter 5 Eigenvalues and Eigenvectors

LINEAR ALGEBRA 1, 2012-I PARTIAL EXAM 3 SOLUTIONS TO PRACTICE PROBLEMS

Eigenvalues and Eigenvectors 7.2 Diagonalization

MATH 221, Spring Homework 10 Solutions

Study Guide for Linear Algebra Exam 2

Eigenvalues and Eigenvectors

LINEAR ALGEBRA REVIEW

Eigenvalues and Eigenvectors

MAT 1302B Mathematical Methods II

MATH 1120 (LINEAR ALGEBRA 1), FINAL EXAM FALL 2011 SOLUTIONS TO PRACTICE VERSION

Computationally, diagonal matrices are the easiest to work with. With this idea in mind, we introduce similarity:

Definition (T -invariant subspace) Example. Example

c c c c c c c c c c a 3x3 matrix C= has a determinant determined by

Linear Algebra II Lecture 13

1 Invariant subspaces

Unit 5: Matrix diagonalization

Lecture 11: Eigenvalues and Eigenvectors

Eigenvalues and Eigenvectors

Lecture Note on Linear Algebra 17. Standard Inner Product

ANSWERS. E k E 2 E 1 A = B

Solutions to Final Practice Problems Written by Victoria Kala Last updated 12/5/2015

Eigenvalues and Eigenvectors

PRACTICE PROBLEMS FOR THE FINAL

City Suburbs. : population distribution after m years

The converse is clear, since

Diagonalization. Hung-yi Lee

Linear Algebra 1. M.T.Nair Department of Mathematics, IIT Madras. and in that case x is called an eigenvector of T corresponding to the eigenvalue λ.

Eigenvalue and Eigenvector Homework

Eigenvalues and Eigenvectors A =

MATH 20F: LINEAR ALGEBRA LECTURE B00 (T. KEMP)

Solutions to Final Exam

Riemann s Hypothesis and Conjecture of Birch and Swinnerton-Dyer are False

Chapter 5. Eigenvalues and Eigenvectors

Eigenvalues and Eigenvectors 7.1 Eigenvalues and Eigenvecto

MAT1302F Mathematical Methods II Lecture 19

MATH 304 Linear Algebra Lecture 33: Bases of eigenvectors. Diagonalization.

Final Review Written by Victoria Kala SH 6432u Office Hours R 12:30 1:30pm Last Updated 11/30/2015

LINEAR ALGEBRA SUMMARY SHEET.

Announcements Monday, November 06

Diagonalization. P. Danziger. u B = A 1. B u S.

(a) II and III (b) I (c) I and III (d) I and II and III (e) None are true.

DM554 Linear and Integer Programming. Lecture 9. Diagonalization. Marco Chiarandini

Announcements Wednesday, November 01

MAC Module 12 Eigenvalues and Eigenvectors. Learning Objectives. Upon completing this module, you should be able to:

MAC Module 12 Eigenvalues and Eigenvectors

Linear Algebra review Powers of a diagonalizable matrix Spectral decomposition

MTH 464: Computational Linear Algebra

Calculating determinants for larger matrices

Econ Slides from Lecture 7

From Lay, 5.4. If we always treat a matrix as defining a linear transformation, what role does diagonalisation play?

1.1 Limits and Continuity. Precise definition of a limit and limit laws. Squeeze Theorem. Intermediate Value Theorem. Extreme Value Theorem.

Math 20F Practice Final Solutions. Jor-el Briones

Online Exercises for Linear Algebra XM511

Question: Given an n x n matrix A, how do we find its eigenvalues? Idea: Suppose c is an eigenvalue of A, then what is the determinant of A-cI?

Math 205, Summer I, Week 4b:

MA 265 FINAL EXAM Fall 2012

OHSx XM511 Linear Algebra: Solutions to Online True/False Exercises

Announcements Monday, October 29

5.3.5 The eigenvalues are 3, 2, 3 (i.e., the diagonal entries of D) with corresponding eigenvalues. Null(A 3I) = Null( ), 0 0

Lecture 15, 16: Diagonalization

TMA Calculus 3. Lecture 21, April 3. Toke Meier Carlsen Norwegian University of Science and Technology Spring 2013

Lecture Summaries for Linear Algebra M51A

Unit 5. Matrix diagonaliza1on

Summer Session Practice Final Exam

The eigenvalues are the roots of the characteristic polynomial, det(a λi). We can compute

EIGENVALUES AND EIGENVECTORS 3

EIGENVALUES AND EIGENVECTORS

(b) If a multiple of one row of A is added to another row to produce B then det(b) =det(a).

Math 315: Linear Algebra Solutions to Assignment 7

Eigenvalues for Triangular Matrices. ENGI 7825: Linear Algebra Review Finding Eigenvalues and Diagonalization

AMS10 HW7 Solutions. All credit is given for effort. (-5 pts for any missing sections) Problem 1 (20 pts) Consider the following matrix 2 A =

Glossary of Linear Algebra Terms. Prepared by Vince Zaccone For Campus Learning Assistance Services at UCSB

Math 314/ Exam 2 Blue Exam Solutions December 4, 2008 Instructor: Dr. S. Cooper. Name:

Linear algebra II Tutorial solutions #1 A = x 1

MATH 304 Linear Algebra Lecture 34: Review for Test 2.

Linear Algebra review Powers of a diagonalizable matrix Spectral decomposition

Homework sheet 4: EIGENVALUES AND EIGENVECTORS. DIAGONALIZATION (with solutions) Year ? Why or why not? 6 9

Linear Algebra. Rekha Santhanam. April 3, Johns Hopkins Univ. Rekha Santhanam (Johns Hopkins Univ.) Linear Algebra April 3, / 7

Finite-Difference Methods (FDM)

Chapters 5 & 6: Theory Review: Solutions Math 308 F Spring 2015

Chapter 3. Determinants and Eigenvalues

进化树构建方法的概率方法 第 4 章 : 进化树构建的概率方法 问题介绍. 部分 lid 修改自 i i f l 的 ih l i

Math 217: Eigenspaces and Characteristic Polynomials Professor Karen Smith

Transcription:

Lecture Note on Linear Algebra 16. Eigenvalues and Eigenvectors Wei-Shi Zheng, wszheng@ieee.org, 2011 November 18, 2011 1 What Do You Learn from This Note In this lecture note, we are considering a very special matrix equation for a given square matrix A: A x = λ x. (1 This equation is very important for developing optimisation algorithm for many engineering problems( 大家将来遇到很多的科学计算优化问题都可以归结为解这个特征向量方程的问题. Basic Concept:Eigenvalue( 特征值, Eigenvector( 特征向量, Characteristic equation( 特征方程 2 Eigenvalue & Eigenvector Definition 1 (eigenvalue( 特征值 and eigenvector( 特征向量. Let A be a n n square matrix. An eigenvalue λ of A is a scalar such that A x = λ x for some non zero vector x, which is called an eigenvector of A corresponding to λ (or a λ eigenvector. Example: For any x R n and λ R, we have (λi n x = λ x. So λ is an eigenvalue of λi n and any non zero vector is a λ eigenvector. Example: Let A R n be non invertible, which is equivalent to Nul(A { 0} =. Then we have A x = 0 = 0 x for any x Nul(A { 0}. So 0 is an 1

eigenvalue of A and any non zero vector in Nul(A { 0} is a 0 eigenvector. Warning: By definition, eigenvalue can take zero scalar whereas eigenvector is restricted to non zero vectors.( 特征值可以是 0, 但特征向量不能是零向量 Examples: Textbook P.303. Theorem 2. If v 1,, v r are eigenvectors that correspond to distinct eigenvalues λ 1,, λ r of an n n matrix A, the set v 1,, v r is linearly independent. Proof. We conduct the proof by contradiction ( 我们用反正法 and induction method. 1. STEP 1: We first prove for two eigenvectors v i1 and v i2 that correspond to different eigenvalues λ i1, λ i2, v i1 and v i2 must be linearly independent. If not, then we have v i1 = c v i2 for some c 0. Also, A v i1 = ca v i2, hence λ i1 v i1 = cλ i2 v i2. Combing with λ i1 v i1 = cλ i1 v i2, so (λ i1 λ i2 v i2 = 0. Since λ i1 λ i2, hence v i2 = 0, which is impossible, as an eigenvector must be non-zero. 2. STEP 2: for any r-1 eigenvector, v i1,, v ir 1 corresponding to r 1 distinct eigenvalues are linearly independent. 3. SETP 3: for given r eigenvectors, If these r vectors are linearly dependent, then there exists a vector v i such that v i is a linear combination of the other eigenvectors, that is there are a series of weights c 1, c 2,, c i 1, c i+1,, c r some of which are not zero having v i = c 1 v 1 + + c i 1 v i 1 + c i+1 v i+1 + + c r v r. (2 Also by multiplying A on both side, we have λ i v i = c 1 λ 1 v 1 + + λ i 1 c i 1 v i 1 + c i+1 λ i+1 v i+1 + + c r λ r v r. (3 In addition, by multiplying λ i on both sides of Eq. (2, then we have λ i v i = c 1 λ i v 1 + + c i 1 λ i v i 1 + c i+1 λ i v i+1 + + c r λ i v r. (4 Combining Eqs. 3 and 5, then we have 0 = c 1 (λ i λ 1 v 1 + +c i 1 (λ i λ i 1 v i 1 +c i+1 (λ i λ i+1 v i+1 + +c r (λ i λ r v r. (5 2

As the r 1 eigenvectors v 1,, v i 1, v i+1,, v r are linearly independent, and not all the weights c 1, c 2,, c i 1, c i+1,, c r are zero, then there must be some weight c k 0 so that λ i λ k. This contradicts to the statement that all the eigenvectors are distinct. Hence all the r eigenvectors that correspond to r different eigenvalues are linearly independent. 3 Computation of Eigenvalues and Eigenvectors Lemma 3. Let A be a n n square matrix. Then for any scalar λ, we have { x A x = λ x} = Nul(λI n A. Proof. { x A x = λ x} = { x λ x A x = 0} = { x (λi n x A x = 0} = { x (λi n A x = 0} = Nul(λI n A. Theorem 4. Let λ be an eigenvalue of a matrix A. Then all the λ eigenvectors of A together with the zero vector form a subspace, which is called the eigenspace of A corresponding to λ (or the λ eigenspace. Proof. Since the set of all λ eigenvectors together with 0 is exactly { x A x = λ x} = Nul(λI n A, which is of course a subspace. Example: Textbook P.304. 3

Question? How to compute the eigenvalues? Theorem 5. Let A be an n n square matrix. Then the following statements are equivalent: 1. A scalar λ is an eigenvalue of A. 2. Nul(λI n A { 0}. 3. dim Nul(λI n A 1. 4. λi n A is not invertible. 5. det(λi n A = 0. Proof. Since λ is an eigenvalue of A iff Nul(λI n A contains a non zero vector iff Nul(λI n A { 0}. So 1. 2., and 2. 3. 4. 5. is straightforward. Example: Find all the eigenvalues of A = eigenspaces. ( 3 1 2 0 and the corresponding Solution. Suppose that λ is ( an eigenvalue of A. Then by Theorem 4, λ 3 1 det(λi n A = 0, that is det = 0. By expanding this determinant, we get a quadratic equation λ 2 3λ + 2 = 0 in λ. By solving this 2 λ equation, we obtain all the eigenvalues of A are 1 and 2. The 1 eigenspace is Nul(I 2 A, that is the solution set of the homogenous equation (I 2 A x = 0. By solving this equation, we obtain the 1 eigenspace is Span{(1 2 T }. Similarly, we can find the 2 eigenspace, which is Span{(1 1 T }. By the Theorem 5, we can easily have: Theorem 6. The eigenvalues of a triangular matrix are the entries on its main diagonal. Motivated by the above example of finding eigenvalues, we have Definition 7 (characteristic equation( 特征方程. Let A be a n n square matrix. The characteristic equation of A is defined to be det(λi n A = 0. 4

Remarks: The eigenvalues of A are precisely the roots of c A (λ in the set of scalars. Hence the number of distinct eigenvalues must be lest than n. Example: Textbook P.313. Example: Find all the eigenvalues of A = eigenspaces over R and C respectively. ( 0 1 1 0 and the corresponding Solution. 1. A R 2 2. ( λ 1 The characteristic polynomial c A (λ = det = λ 1 λ 2 + 1. Since c A (λ has no real root, so no eigenvalue exists in R. 2. A C 2 2. In this case, c A has roots i and i, which are all eigenvalues of A. The i eigenspace is Nul(iI 2 A, which can be computed by solving equation (ii 2 A x = 0 and the result is Span{(i 1 T }. Similarly, we can obtain the ( i eigenspace Span{( i 1 T }. This example illustrates that the existence of eigenvalues depends on the set of scalars (which is in fact a field. Summary of the procedure of computing eigenvalues. For a given square matrix A, its eigenvalues and the corresponding eigenspaces are computed as follows: 1. Expand c A (λ = det(λi n A; 2. Solve c A (λ = 0 in the set of scalars (R or C in this course to obtain all the eigenvalues, say λ 1,..., λ r. 3. For each i = 1,..., r, solve the homogenous equation (λ i I n A x = 0 and the solution set is the λ i eigenspace. 5

4 Diagonalization We are now considering a special matrix factorization for some square matrix using eigenvectors and eigenvalues, which has the following forms A = P DP 1. (6 where P is an invertible matrix and D is a diagonal matrix. This factorization enables us to compute A k by A k = P D k P 1. (7 We are now formally introducing the diagonalization processing. Definition 8 (Similarity. Two square matrix A and B are similar if there is an invertible matrix P such that P 1 AP = B, or say equivalently A = P BP 1. Changing A into P 1 AP is called a similarity transformation. Definition 9 (Diagonalizable Matrix. Let A be a square matrix. A is said to be diagonalisable if A is similar to a diagonal matrix D or equivalently, there exists an invertible matrix P such that P 1 AP is diagonal. Theorem 10. If two n n matrices A and B are similar, they have the same characteristic polynomial and hence the same eigenvalues. Proof. Since A and B are similar, there exists an invertible matrix P such that B = P 1 AP. Therefore, we have B λi = P 1 AP λp 1 IP = P 1 (A λip, So det B = det(p 1 (A λip = det P 1 det(a λi det P = det P 1 det P det(a λi = det(p 1 P det(a λi = det I det(a λi = det(a λi. (8 6

Not all square matrices are diagonalisable, then we have the following question: Question: When can a square matrix A be diagonalized? Theorem 11. Let A R n n. A is diagonalisable iff there exists a basis B = { v 1,..., v n } such that v 1,..., v n are linearly independent eigenvectors of A. Proof. We prove the theorem by the following steps: STEP 1. Suppose that A is diagonalisable. Then there exists an invertible matrix P = ( v 1 v n and a diagonal matrix D = diag(λ 1,..., λ n such that AP = P D. So (A v 1 A v n = AP = P D = (λ 1 v 1 λ n v n. So for all i, v i is a λ i eigenvector of A, also the invertibility of P ensures that { v 1,..., v n } forms a basis. STEP 2. Conversely, let P = ( v 1 v n. For i = 1,..., n, let λ i be the eigenvalue with eigenvector v i. Thus we have AP = P diag(λ 1,..., λ n. Since the eigenvectors are linearly independent, P is invertible, thus we have: A = P diag(λ 1,..., λ n P 1. Remark: The basis v 1 v n in the above theorem is called an eigenvector basis. If A R n n has n distinct eigenvalues then A is diagonal- Corollary 12. isable. Proof. By Theorem 2, { v 1,..., v n } is linearly independent, where for i = 1,..., n, v i is some λ i eigenvector. So { v 1,..., v n } is a basis and A is diagonalisable by by Theorem 9. 7

Question: Can we still perform diagonalization when not all eigenvalues are distinct? Theorem 13. Let λ 1,..., λ r be distinct eigenvalues of A R n n. For i = 1,..., r, let S i = { v i1,..., v ini } be a linearly independent set of λ i eigenspace. Then S = r i=1 S i is linearly independent. Proof. For i = 1,..., r, let c i1,..., c ini be scalars and w i = c i1 v i1 + + c ini v ini. Suppose that w 1 + + w r = 0. We need to show c i1 = = c ini = 0 for all i = 1,..., r. Now for i = 1,..., r, w i is either 0 or a λ i eigenvector. But { w 1,..., w r } is linearly dependent, which forces that w 1 = = w r = 0, otherwise contradicting the result of Theorem 2. So for all i = 1,..., r, c i1 = = c ini = 0 since S i is linearly independent. Corollary 14. Let λ 1,..., λ r be all distinct eigenvalues of A R n n and n i the dimension of λ i eigenspace for i = 1,..., r. Then A is diagonalisable iff r i=1 n i = n. Proof. Suppose A is similar to D, where D is diagonal. Then it is not hard to see that A and D have the same set of eigenvalues and the dimensions of the corresponding eigenspaces are equal (Exercise. But for D, hence for A, we have r i=1 n i = n. Conversely, suppose that B i is a basis of λ i eigenspace for i = 1,..., r. Then B = r i=1 B i is linearly independent by Theorem 13. Also B = r i=1 n i = n. So B is a basis consisting of eigenvectors. So A is diagonalisable by Theorem 11. According to Corollary 14, the following procedure can be used to determined whether or not a given A M n is diagonalisable and in the affirmative case, a matrix P such that P 1 AP is diagonal is computed. Step 1: Compute all eigenvalues λ 1,..., λ r of A. Step 2: For each i = 1,..., r, compute a basis B i of the λ i eigenspace. Step 3: If r i=1 B i = n, then A is not diagonalisable. Otherwise, the matrix P = ( v 1 v n, where { v 1,..., v n } = r i=1 B i, satisfies that P 1 AP is diagonal. Example: Textbook P.324. 8

4.1 A Linear Transformation View of Diagolization Objective: We aim to understand the diagonalization of matrix A from the linear transformation view of point ( 在本节, 我们尝试利用线性变换解释对矩阵 A 的对角话操作. STEP 1: Linear Transformation between Vector Space V and W Let T be the linear transformation between V and W. Let B = { b 1, b 2,, b n } be a basis of vector space V. Let C = { c 1, c 2,, c m } be a basis of vector space W. Then any x V can be represented by a linear combination of b 1, b 2,, b n with weight r 1, r 2,, r n as follows: x = r 1 b1 + + r n bn = B[ x] B. where B = [ b 1, b 2,, b n ] and [ x] B = linear transformation, we have r 1. r n. Then by the definition of T ( x = T (r 1 b1 + + r n bn = r 1 T ( b 1 + + r n T ( b n (9 So, the coordinate of T ( x under basis C in W is [T ( x] C = r 1 [T ( b 1 ] C + + r n [T ( b n ] C = M[x] B, (10 where M = ( [T ( b 1 ] C,, [T ( b n ] C. We call matrix M as the matrix for T relative to the bases B to C. STEP 2: Linear Transformation from V to V (i.e. W = V. In this case, the M matrix is called the matrix for T relative to B, or the B-matrix for T. We re-denote this B-matrix by [T ] B as follows: Then, for all x V, we have [T ] B = ( [T ( b 1 ] B,, [T ( b n ] B. [T ( x] B = [T ] B [ x] B (11 STEP 3: Diagonal Matrix Representation. Now, we reach the main result of this part. 9

Theorem 15. Suppose two square matrices A and C are similar, that is there exists an invertible matrix P such that A = P CP 1, where C is n n matrix. Let B be the basis for R n formed from the columns of P, then C is the B matrix for the transformation T : x A x. Proof. Let P = [ b 1,, b n ], so B = { b 1,, b n }. Note that x = P [ x] B P 1 x = [ x] B. Also, since we are discussing the linear transformation in R n, so we have T ( x = A x. Now we investigate the form of B matrix. That is [T ] B = ( [T ( b 1 ] B,, [T ( b n ] B = ( [A b 1 ] B,, [A b n ] B = ( P 1 A b 1,, P 1 A b n = P 1 A ( b1,, b n = P 1 AP = C. (12 Remark. When C = D in the last theorem, it means diagolizing A amounts to finding a diagonal matrix representation of T : x A x. 5 Complex Eigenvalues Actually all our preceding introduction can be straightforwardly extended to the complex domain, by changing R n to C n and changing R n n to C n n. When the eigenvalue λ is a complex value, we call it the complex eigenvalue and its corresponding vector as complex eigenvector. Rhombicuboctahedron, by Da Vinci 10