Eigenvalues and eigenvectors

Similar documents
AMS526: Numerical Analysis I (Numerical Linear Algebra)

Eigenvalue and Eigenvector Problems

AMS526: Numerical Analysis I (Numerical Linear Algebra for Computational and Data Sciences)

Lecture 10 - Eigenvalues problem

Matrices, Moments and Quadrature, cont d

AMS526: Numerical Analysis I (Numerical Linear Algebra)

Math 102, Winter Final Exam Review. Chapter 1. Matrices and Gaussian Elimination

Linear Algebra, part 2 Eigenvalues, eigenvectors and least squares solutions

Eigenvalue Problems. Eigenvalue problems occur in many areas of science and engineering, such as structural analysis

Scientific Computing: An Introductory Survey

Computational Methods CMSC/AMSC/MAPL 460. Eigenvalues and Eigenvectors. Ramani Duraiswami, Dept. of Computer Science

EIGENVALUE PROBLEMS. EIGENVALUE PROBLEMS p. 1/4

Schur s Triangularization Theorem. Math 422

EECS 275 Matrix Computation

Remark By definition, an eigenvector must be a nonzero vector, but eigenvalue could be zero.

Contents. Preface for the Instructor. Preface for the Student. xvii. Acknowledgments. 1 Vector Spaces 1 1.A R n and C n 2

Linear Algebra: Matrix Eigenvalue Problems

Last Time. Social Network Graphs Betweenness. Graph Laplacian. Girvan-Newman Algorithm. Spectral Bisection

Eigenvalues and Eigenvectors

Numerical Methods for Solving Large Scale Eigenvalue Problems

Matrix Algebra: Summary

Orthogonal iteration to QR

Foundations of Matrix Analysis

Diagonalizing Matrices

Computation of eigenvalues and singular values Recall that your solutions to these questions will not be collected or evaluated.

Review of Linear Algebra

Numerical Methods I Eigenvalue Problems

G1110 & 852G1 Numerical Linear Algebra

Applied Mathematics 205. Unit V: Eigenvalue Problems. Lecturer: Dr. David Knezevic

Remark 1 By definition, an eigenvector must be a nonzero vector, but eigenvalue could be zero.

Math 504 (Fall 2011) 1. (*) Consider the matrices

Preliminary/Qualifying Exam in Numerical Analysis (Math 502a) Spring 2012

LINEAR ALGEBRA 1, 2012-I PARTIAL EXAM 3 SOLUTIONS TO PRACTICE PROBLEMS

DIAGONALIZATION. In order to see the implications of this definition, let us consider the following example Example 1. Consider the matrix

Computational Methods. Eigenvalues and Singular Values

Synopsis of Numerical Linear Algebra

Math 405: Numerical Methods for Differential Equations 2016 W1 Topics 10: Matrix Eigenvalues and the Symmetric QR Algorithm

13-2 Text: 28-30; AB: 1.3.3, 3.2.3, 3.4.2, 3.5, 3.6.2; GvL Eigen2

Review problems for MA 54, Fall 2004.

Ir O D = D = ( ) Section 2.6 Example 1. (Bottom of page 119) dim(v ) = dim(l(v, W )) = dim(v ) dim(f ) = dim(v )

Index. for generalized eigenvalue problem, butterfly form, 211

The German word eigen is cognate with the Old English word āgen, which became owen in Middle English and own in modern English.

Eigenpairs and Similarity Transformations

I. Multiple Choice Questions (Answer any eight)

Chapter 3 Transformations

5 Selected Topics in Numerical Linear Algebra

Computing Eigenvalues and/or Eigenvectors;Part 2, The Power method and QR-algorithm

Numerical Methods - Numerical Linear Algebra

Math 205, Summer I, Week 4b:

MATH 423 Linear Algebra II Lecture 33: Diagonalization of normal operators.

EIGENVALUE PROBLEMS (EVP)

LinGloss. A glossary of linear algebra

MATH 221, Spring Homework 10 Solutions

Quantum Computing Lecture 2. Review of Linear Algebra

University of Colorado at Denver Mathematics Department Applied Linear Algebra Preliminary Exam With Solutions 16 January 2009, 10:00 am 2:00 pm

1. What is the determinant of the following matrix? a 1 a 2 4a 3 2a 2 b 1 b 2 4b 3 2b c 1. = 4, then det

9.1 Eigenvectors and Eigenvalues of a Linear Map

MA 265 FINAL EXAM Fall 2012

The Eigenvalue Problem: Perturbation Theory

Glossary of Linear Algebra Terms. Prepared by Vince Zaccone For Campus Learning Assistance Services at UCSB

Numerical Methods I: Eigenvalues and eigenvectors

Solving large scale eigenvalue problems

Draft. Lecture 14 Eigenvalue Problems. MATH 562 Numerical Analysis II. Songting Luo. Department of Mathematics Iowa State University

1. General Vector Spaces

Linear algebra and applications to graphs Part 1

Linear Algebra 1. M.T.Nair Department of Mathematics, IIT Madras. and in that case x is called an eigenvector of T corresponding to the eigenvalue λ.

Linear Algebra Review. Vectors

ft-uiowa-math2550 Assignment NOTRequiredJustHWformatOfQuizReviewForExam3part2 due 12/31/2014 at 07:10pm CST

Linear Algebra Review

AMS526: Numerical Analysis I (Numerical Linear Algebra)

Lecture 13 Eigenvalue Problems

Online Exercises for Linear Algebra XM511

MATH 240 Spring, Chapter 1: Linear Equations and Matrices

arxiv: v1 [math.na] 5 May 2011

Exercise Set 7.2. Skills

Math 489AB Exercises for Chapter 2 Fall Section 2.3

MATH 31 - ADDITIONAL PRACTICE PROBLEMS FOR FINAL

Math 108b: Notes on the Spectral Theorem

AMS526: Numerical Analysis I (Numerical Linear Algebra)

4.8 Arnoldi Iteration, Krylov Subspaces and GMRES

Chap 3. Linear Algebra

Lecture 2: Numerical linear algebra

The QR Decomposition

Spectral Theorem for Self-adjoint Linear Operators

Lecture # 11 The Power Method for Eigenvalues Part II. The power method find the largest (in magnitude) eigenvalue of. A R n n.

Chapter 5 Eigenvalues and Eigenvectors

AMS526: Numerical Analysis I (Numerical Linear Algebra)

Definitions for Quizzes

Orthogonal iteration to QR

Notes on Eigenvalues, Singular Values and QR

ELE/MCE 503 Linear Algebra Facts Fall 2018

LINEAR ALGEBRA BOOT CAMP WEEK 4: THE SPECTRAL THEOREM

235 Final exam review questions

ALGEBRA QUALIFYING EXAM PROBLEMS LINEAR ALGEBRA

(b) If a multiple of one row of A is added to another row to produce B then det(b) =det(a).

MATH 304 Linear Algebra Lecture 34: Review for Test 2.

IMPORTANT DEFINITIONS AND THEOREMS REFERENCE SHEET

AMS10 HW7 Solutions. All credit is given for effort. (-5 pts for any missing sections) Problem 1 (20 pts) Consider the following matrix 2 A =

IMPORTANT DEFINITIONS AND THEOREMS REFERENCE SHEET

SECTIONS 5.2/5.4 BASIC PROPERTIES OF EIGENVALUES AND EIGENVECTORS / SIMILARITY TRANSFORMATIONS

Transcription:

Chapter 6 Eigenvalues and eigenvectors An eigenvalue of a square matrix represents the linear operator as a scaling of the associated eigenvector, and the action of certain matrices on general vectors can be represented as a scaling of the set of basis vectors used to represent the vector. If iterated some of these basis vectors will be amplified, whereas others will decay, which can charaterize the stability of iterative algorithms, or aphysicalsystemdescribedbythematrix,andalsoimpliesalgorithms for computational approximation of eigenvalues and eigenvectors based on power iteration. We also present the QR algorithm, based on constructing asimilaritytransformofthematrix. 6.1 Eigenvalues and eigenvectors Complex vector spaces In this section we change the focus from real vector spaces to complex vector spaces. We let z 2 C denote a complex scalar, with z thecomplex conjugate of z. The complex vector space C n is defined by the basic operations of componentwise addition and scalar multiplication of complex numbers, and with the transpose of a vector x 2 C n replaced by the adjoint x,corresponding to the transpose with the imaginary parts negated. Similarly, the adjoint of a complex m n matrix A =(a ij )isthen m matrix A =(ā ji ). If A = A the matrix A is Hermitian, andifaa = A A it is normal. The inner product of x, y 2 C n is defined by (x, y) =x y = x i y i, (6.1) i=i 51

52 CHAPTER 6. EIGENVALUES AND EIGENVECTORS with the associated norm for x 2 C n, Matrix spectrum and eigenspaces kxk =(x, x) 1/2. (6.2) We now consider a square matrix A 2 C n n acting on a complex vector space C n. An eigenvector of A is a nonzero vector x 2 C n,suchthat Ax = x, (6.3) with 2 C the corresponding eigenvalue. ThesubspaceofC n spanned by the eigenvectors corresponding to,togetherwiththezerovector,isan eigenspace E,andthesetofalleigenvalues{ j } n is the spectrum of A, denoted by (A). The sum and the product of all eigenvalues are related to the trace and the determinant of A as ny det(a) = j tr(a) = j. (6.4) The eigenspace E is an invariant subspace under A, sothatae E, and dim(e ) is the number of linearly independent eigenvectors corresponding to the eigenvalue,knownasthegeometric multiplicity of. We have that the eigenspace E =null( I A), since ( I A)x =0, and thus for a nonempty eigenspace E, I A is a singular matrix, so that Characteristic polynomial det( I A) =0. (6.5) The characteristic polynomial of the matrix A 2 C n n, is the degree n polynomial p A (z) = det(zi A), (6.6) with z 2 C. For an eigenvalue of A, wethushavethat p A ( )=0, (6.7) and by the fundamental theorem of algebra we can express p A ( )as p A ( )=(z 1)(z 2) (z n), (6.8) where each j is an eigenvalue of A, not necessary unique. The multiplicity of each eigenvalue as a root to the equation p A ( )=0isthealgebraic multiplicity of,whereaneigenvalueissaidtobesimple if its algebraic multiplicity is 1. The algebraic multiplicity of an eigenvalue is at least as great as its geometric multiplicity.

6.1. EIGENVALUES AND EIGENVECTORS 53 Eigenvalue decompositions A defective matrix is a matrix which has one or more defective eigenvalues, whereadefectiveeigenvalueisaneigenvalueforwhichitsalgebraic multiplicity exceeds its geometric multiplicity. Theorem 5 (Eigenvalue decomposition). Each nondefective matrix A 2 C n n has an eigenvalue decomposition A = X X 1, (6.9) where X 2 C n n is a nonsingular matrix with the eigenvectors of A as column vectors, and where 2 C n n is diagonal matrix with the eigenvalues of A on the diagonal. We also say that a nondefective matrix is diagonalizable. Given that the factorization (6.9) exits, we have that which expresses (6.3) as AX = X, (6.10) Ax j = j x j, (6.11) with j the jth diagonal entry of, and x j the jth column of X. For some matrices eigenvectors can be chosen to be pairwise orthogonal, so that a matrix A is unitary diagonalizable, thatis A = Q Q, (6.12) with Q 2 C n n an orthogonal matrix with orthonormal eigenvectors of A as column vectors, and 2 C n n adiagonalmatrixwiththeeigenvaluesof A on the diagonal. Theorem 6. A matrix is unitary diagonalizable if and only if it is normal. Hermitian matrices have real eigenvalues, and thus in the particular case of a real symmetric matrix, no complex vector spaces are needed to characterize the matrix spectrum and eigenspaces. Theorem 7. An Hermitian matrix is unitary diagonalizable with real eigenvalues. Irrespectively if the matrix is nondefective or Hermitian, any square matrix always has a Schur factorization, with the diagonal matrix replaced by an upper triangular matrix.

54 CHAPTER 6. EIGENVALUES AND EIGENVECTORS Theorem 8 (Schur factorization). For every square matrix A there exists a Schur factorization A = QT Q, (6.13) where Q is an orthogonal matrix, and T is an upper triangular matrix with the eigenvalues of A on the diagonal. More generally, if X 2 C n n is nonsingular, the map A 7! X 1 AX is a similarity transformation of A, andwesaythattwomatricesa and B are similar if there exists a similarity transformation such that B = X 1 AX. Theorem 9. Two similar matrices have the same eigenvalues with the same algebraic and geometric multiplicity. 6.2 Eigenvalue algorithms QR algorithm To compute the eigenvalues of a matrix A, onemayseektherootsofthe characteristic polynomial. Although, for a large matrix polynomial root finding is expensive and unstable. Instead the most e cient algorithms are based on computing eigenvalues and eigenvectors by constructing one of the factorizations (6.9), (6.12) or (6.13). We now present the QR algorithm, inwhichaschurfactorization(6.13) of a matrix A is constructed from successive QR factorizations. Algorithm 4: QR algorithm A (0) = A for k =1, 2,... do Q (k) R (k) = A A k = R (k) Q (k) end (k 1) We note that for each iteration A (k) of the algorithm, we have that A (k) = R (k) Q (k) =(Q (k) ) 1 A (k 1) Q (k), (6.14) so that A (k) and A (k 1) are similar, and thus have the same eigenvalues. Under suitable assumptions A k will converge to an upper triangular matrix, or in the case of a Hermitian matrix a diagonal matrix, with the eigenvalues on the diagonal.

6.2. EIGENVALUE ALGORITHMS 55 The basic QR algorithm can be accelerated: (i) by Householder reflectors to reduce the initial matrix A (0) to Hessenberg form, thatisamatrixwith zeros below the first subdiagonal (or in the case of an Hermitian matrix a tridiagonal form), (ii) by introducing a shift to instead of A (k) factorize the matrix A (k) µ (k) I,whichhasidenticaleigenvectors,andwhereµ (k) an eigenvalue estimate, and (iii) if any o -diagonal element is close to zero, both o -diagonal elements are zeroed out to deflate the matrix A (k) into submatrices on which the QR algorithm is then applied. Rayleigh quotient To simplify the presentation, in the rest of this section we restrict attention to matrices that are real and symmetric, for which all eigenvalues j are real and the corresponding eigenvectors q j are orthonormal. We now consider the question: given a vector x 2 R n,whatisthereal number 2 R that best approximate an eigenvalue of A in the sense that kax xk is minimized? If x = q j is an eigenvector of A, then = j is the corresponding eigenvalue. If not, is the solution to the n 1leastsquaresproblem for which the normal equations are given as min kax xk, (6.15) 2R x T Ax = x T x. (6.16) With = r(x), we define the Rayleigh quotient as r(x) = xt Ax x T x, (6.17) where r(x) isanapproximationofaneigenvalue j, ifx is close to the eigenvector q j.infact,r(x) convergesquadraticallytor(q j )= j,thatis as x! q j. Power iteration r(x) r(q j )=O(kx q j k 2 ), (6.18) For a real symmetric n n matrix A, theeigenvectors{q j } n form an orthonormal basis for R n so that we can express any vector v 2 R n in terms of the eigenvectors, v = j q j, (6.19)

56 CHAPTER 6. EIGENVALUES AND EIGENVECTORS with the coordinates j =(v, q j ). Further, we can express the map v 7! Av in terms of the corresponding eigenvalues j, as Av = j Aq j = j j q j, (6.20) and thus the map amounts to a scaling of each eigenvector q j by j. If iterated, this map gives A k v = j A k q j = k j j q j, (6.21) k so that each eigenvector j q j of A k converges to zero if j < 1, or infinity if j > 1. Now assume that 1 =(v, q 1 ) 6= 0, and that the eigenvalues of A are ordered such that 1 > 2 > > n, (6.22) where we say that 1 is the dominant eigenvalue, and q 1 the dominant eigenvector. Thus j/ 1 < 1forallj, whichimpliesthat as k!1.wecanwrite and thus the approximation A k v = k 1( 1 q 1 + ( j / 1 ) k! 0, (6.23) j ( j / 1 ) k q j ), (6.24) j=2 A k v k 1 1 q 1, (6.25) improves as k increases. That is, A k v approaches a multiple of the eigenvector q j,whichcanbeobtainedbynormalizating,sothat v (k) A k v/ka k vk q j, (6.26) from which an approximation of the corresponding eigenvalue can be obtained by the Rayleigh quotient, which leads us to power iteration.

6.2. EIGENVALUE ALGORITHMS 57 Algorithm 5: Power iteration v (0) such that kv (0) k =1 for k =1, 2,... do w = Av (k 1). apply A v (k) = w/kwk. normalize (k) =(v (k) ) T Av (k). Raylegh quotient end Inverse iteration The convergence of the Power iteration to the dominant eigenvector is linear by a constant factor 2/ 1,whereastheconvergencetothedominanteigenvalue is quadratic in the same factor, due to the convergence of Rayleigh quotient. E ciency of the algorithm thus depends on the size of the factor 2/ 1. The idea of inverse iteration, is to apply power iteration to the matrix (A µi) 1,witheigenvalues{( j µ) 1 } and with the same eigenvectors as A, since Av = v, (A µi)v =( µ)v, ( µ) 1 v =(A µi) 1 v. (6.27) With µ an approximation of j,theeigenvalue( j µ) 1 can be expected to be dominant and much larger than the other eigenvalues, which results in an accelerated convergence of power iteration. Rayleigh quotient iteration is inverse iteration where µ is updated to the eigenvalue approximation of the previous step of the iteration, and the convergence to an eigenvalue/eigenvector pair is cubic. Algorithm 6: Rayleigh quotient iteration v (0) such that kv (0) k =1 (0) =(v (0) ) T Av (0) for k =1, 2,... do Solve (A (k 1) )w = v (k 1) for w. apply (A (k 1) I) 1 v (k) = w/kwk. normalize (k) =(v (k) ) T Av (k). Raylegh quotient end

58 CHAPTER 6. EIGENVALUES AND EIGENVECTORS The QR algorithm as a power iteration We now revisit the QR algorithm. Let Q (k) and R (k) be the matrices generated from the (unshifted) QR algorithm, then the matrix products and Q (k) = Q (1) Q (2) Q (k), (6.28) R (k) = R (k) R (k 1) R (1), (6.29) correspond to a QR factorization of the kth power of A, A k = Q (k) R (k), (6.30) which can be proven by induction. That is, the QR algorithm constructs successive orthonormal bases for the powers A k,thusfunctioninglikeapoweriterationthatsimultaneously iterates on the whole set of approximate eigenvectors. Further, the diagonal elements of the kth iterate A (k) are the Rayleigh quotients of A corresponding to the column vectors of Q (k), A (k) =(Q (k) ) T AQ (k), (6.31) and thus the diagonal elements of A (k) converges (quadratically) to the eigenvalues of A. With the accelerations (i)-(iii) the QR algorithm exhibit cubic converge rate in both eigenvalues and eigenvectors.