Numerical Methods I Eigenvalue Problems

Similar documents
Eigenvalue and Eigenvector Problems

Math 504 (Fall 2011) 1. (*) Consider the matrices

Numerical Methods I Singular Value Decomposition

5.3 The Power Method Approximation of the Eigenvalue of Largest Module

EIGENVALUE PROBLEMS. EIGENVALUE PROBLEMS p. 1/4

Numerical Methods I Non-Square and Sparse Linear Systems

Numerical Methods I Solving Square Linear Systems: GEM and LU factorization

Linear algebra & Numerical Analysis

Computation of eigenvalues and singular values Recall that your solutions to these questions will not be collected or evaluated.

ECS231 Handout Subspace projection methods for Solving Large-Scale Eigenvalue Problems. Part I: Review of basic theory of eigenvalue problems

AMS526: Numerical Analysis I (Numerical Linear Algebra)

Applied Mathematics 205. Unit V: Eigenvalue Problems. Lecturer: Dr. David Knezevic

Lecture 2: Numerical linear algebra

Scientific Computing: Dense Linear Systems

Math 108b: Notes on the Spectral Theorem

Eigenvalue Problems. Eigenvalue problems occur in many areas of science and engineering, such as structural analysis

The Eigenvalue Problem: Perturbation Theory

Scientific Computing: An Introductory Survey

Lecture 4 Eigenvalue problems

Math 102, Winter Final Exam Review. Chapter 1. Matrices and Gaussian Elimination

Notes on Eigenvalues, Singular Values and QR

Eigenvalues and eigenvectors

AMS526: Numerical Analysis I (Numerical Linear Algebra)

Scientific Computing: Solving Linear Systems

Lecture 10 - Eigenvalues problem

Numerical methods for eigenvalue problems

Quantum Computing Lecture 2. Review of Linear Algebra

EIGENVALUE PROBLEMS. Background on eigenvalues/ eigenvectors / decompositions. Perturbation analysis, condition numbers..

EIGENVALUE PROBLEMS (EVP)

Numerical Methods I: Eigenvalues and eigenvectors

Numerical Methods - Numerical Linear Algebra

Numerical Methods I Solving Nonlinear Equations

Index. for generalized eigenvalue problem, butterfly form, 211

Applied Mathematics 205. Unit V: Eigenvalue Problems. Lecturer: Dr. David Knezevic

Numerical Methods for Solving Large Scale Eigenvalue Problems

Chap 3. Linear Algebra

EECS 275 Matrix Computation

5 Selected Topics in Numerical Linear Algebra

Eigenvalues and Eigenvectors

Lecture # 11 The Power Method for Eigenvalues Part II. The power method find the largest (in magnitude) eigenvalue of. A R n n.

Draft. Lecture 14 Eigenvalue Problems. MATH 562 Numerical Analysis II. Songting Luo. Department of Mathematics Iowa State University

Course Notes: Week 1

Numerical Solution of Linear Eigenvalue Problems

Eigenproblems II: Computation

Linear algebra and applications to graphs Part 1

Lecture Notes 6: Dynamic Equations Part C: Linear Difference Equation Systems

Computational Methods. Eigenvalues and Singular Values

Orthogonal iteration to QR

Ir O D = D = ( ) Section 2.6 Example 1. (Bottom of page 119) dim(v ) = dim(l(v, W )) = dim(v ) dim(f ) = dim(v )

Foundations of Matrix Analysis

The German word eigen is cognate with the Old English word āgen, which became owen in Middle English and own in modern English.

1. Foundations of Numerics from Advanced Mathematics. Linear Algebra

Schur s Triangularization Theorem. Math 422

MATH 423 Linear Algebra II Lecture 33: Diagonalization of normal operators.

Math 405: Numerical Methods for Differential Equations 2016 W1 Topics 10: Matrix Eigenvalues and the Symmetric QR Algorithm

Linear Algebra Massoud Malek

Elementary linear algebra

Math 408 Advanced Linear Algebra

Lecture 13 Eigenvalue Problems

Linear Algebra: Matrix Eigenvalue Problems

Computing Eigenvalues and/or Eigenvectors;Part 2, The Power method and QR-algorithm

AMS526: Numerical Analysis I (Numerical Linear Algebra)

Math 307 Learning Goals

Review problems for MA 54, Fall 2004.

Exercise Set 7.2. Skills

Last Time. Social Network Graphs Betweenness. Graph Laplacian. Girvan-Newman Algorithm. Spectral Bisection

Matrices, Moments and Quadrature, cont d

Definition (T -invariant subspace) Example. Example

Diagonalization of Matrix

Review of some mathematical tools

Eigenvalues and Eigenvectors

LINEAR ALGEBRA: NUMERICAL METHODS. Version: August 12,

AMS526: Numerical Analysis I (Numerical Linear Algebra for Computational and Data Sciences)

Eigenvalues, Eigenvectors, and Diagonalization

Matrix Algorithms. Volume II: Eigensystems. G. W. Stewart H1HJ1L. University of Maryland College Park, Maryland

Mathematical foundations - linear algebra

QR-decomposition. The QR-decomposition of an n k matrix A, k n, is an n n unitary matrix Q and an n k upper triangular matrix R for which A = QR

arxiv: v1 [math.na] 5 May 2011

Numerical Methods in Matrix Computations

Eigenvalues, Eigenvectors, and Diagonalization

Math 489AB Exercises for Chapter 2 Fall Section 2.3

Krylov Subspace Methods to Calculate PageRank

ECS130 Scientific Computing Handout E February 13, 2017

Notes on basis changes and matrix diagonalization

Linear Algebra 1. M.T.Nair Department of Mathematics, IIT Madras. and in that case x is called an eigenvector of T corresponding to the eigenvalue λ.

Applied Linear Algebra in Geoscience Using MATLAB

Large-scale eigenvalue problems

Linear Algebra Review

Lecture 15, 16: Diagonalization

Preliminary/Qualifying Exam in Numerical Analysis (Math 502a) Spring 2012

1. General Vector Spaces

HOMEWORK PROBLEMS FROM STRANG S LINEAR ALGEBRA AND ITS APPLICATIONS (4TH EDITION)

Mathematical foundations - linear algebra

Math 307 Learning Goals. March 23, 2010

MATH 20F: LINEAR ALGEBRA LECTURE B00 (T. KEMP)

EXAM. Exam 1. Math 5316, Fall December 2, 2012

Section 4.5 Eigenvalues of Symmetric Tridiagonal Matrices

Computational Methods CMSC/AMSC/MAPL 460. Eigenvalues and Eigenvectors. Ramani Duraiswami, Dept. of Computer Science

SUMMARY OF MATH 1600

AMS526: Numerical Analysis I (Numerical Linear Algebra)

Transcription:

Numerical Methods I Eigenvalue Problems Aleksandar Donev Courant Institute, NYU 1 donev@courant.nyu.edu 1 MATH-GA 2011.003 / CSCI-GA 2945.003, Fall 2014 October 2nd, 2014 A. Donev (Courant Institute) Lecture IV 10/2014 1 / 23

Outline 1 Review of Linear Algebra: Eigenvalues 2 Conditioning of Eigenvalue Problems 3 Computing Eigenvalues and Eigenvectors 4 Methods based on QR factorizations 5 Conclusions A. Donev (Courant Institute) Lecture IV 10/2014 2 / 23

Review of Linear Algebra: Eigenvalues Eigenvalue Decomposition For a square matrix A C n n, there exists at least one λ such that Ax = λx (A λi) y = 0 Putting the eigenvectors x j as columns in a matrix X, and the eigenvalues λ j on the diagonal of a diagonal matrix Λ, we get AX = XΛ. A matrix is non-defective or diagonalizable if there exist n linearly independent eigenvectors, i.e., if the matrix X is invertible: X 1 AX = Λ A = XΛX 1. The transformation from A to Λ = X 1 AX is called a similarity transformation and it preserves the eigenspace. A. Donev (Courant Institute) Lecture IV 10/2014 3 / 23

Review of Linear Algebra: Eigenvalues Unitarily Diagonalizable Matrices A matrix is unitarily diagonalizable if there exist n linearly independent orthogonal eigenvectors, i.e., if the matrix X can be chosen to be unitary (orthogonal), X U, where U 1 = U : A = UΛU. Note that unitary matrices generalize orthogonal matrices to the complex domain, so we use adjoints (conjugate transposes) instead of transposes throughout. Theorem: A matrix is unitarily diagonalizable iff it is normal, i.e., it commutes with its adjoint: A A = AA. Theorem: Hermitian (symmetric) matrices, A = A, are unitarily diagonalizable and have real eigenvalues. A. Donev (Courant Institute) Lecture IV 10/2014 4 / 23

Review of Linear Algebra: Eigenvalues Left Eigenvectors The usual eigenvectors are more precisely called right eigenvectors. There is also left eigenvector corresponding to a given eigenvalue λ y A = λy A y = λy. Y A = ΛY For a matrix that is diagonalizable, observe that Y = X 1 and so the left eigenvectors provide no new information. For unitarily diagonalizable matrices, Y = ( X 1) = U, so that the left and right eigenvectors coincide. A. Donev (Courant Institute) Lecture IV 10/2014 5 / 23

Review of Linear Algebra: Eigenvalues Non-diagonalizable Matrices For matrices that are not diagonalizable, one can use Jordan form factorizations, or, more relevant to numerical mathematics, the Schur factorization (decomposition): where T is upper-triangular. A = UTU, The eigenvalues are on the diagonal of T. Note: Observe that A = (UTU ) = UT U so for Hermitian matrices T = T is real diagonal. An important property / use of eigenvalues: A n = (UTU ) (UTU ) (UTU ) = UT (U U) T (U U) TU A n = UT n U A. Donev (Courant Institute) Lecture IV 10/2014 6 / 23

Conditioning of Eigenvalue Problems Sensitivity of Eigenvalues Now consider a perturbation of a diagonalizable matrix δa and see how perturbed the similar matrix becomes: X 1 (A + δa) X = Λ + δλ δλ = X 1 (δa) X δλ X 1 δa X = κ (X) δa Conclusion: The conditioning of the eigenvalue problem is related to the conditioning of the matrix of eigenvectors. If X is unitary then X 2 = 1 (from now on we exclusively work with the 2-norm): Unitarily diagonalizable matrices are always perfectly conditioned! Warning: The absolute error in all eigenvalues is of the same order, meaning that the relative error will be very large for the smallest eigenvalues. A. Donev (Courant Institute) Lecture IV 10/2014 7 / 23

Conditioning of Eigenvalue Problems Sensitivity of Individual Eigenvalues A. Donev (Courant Institute) Lecture IV 10/2014 8 / 23

Conditioning of Eigenvalue Problems Sensitivity of Individual Eigenvalues δλ y (δa) x y x Recalling the Cauchy-Schwartz inequality: y x = x y cos θ xy x y δλ x δa y = δa x y cos θ xy cos θ xy Defining a conditioning number for a given eigenvalue κ (λ, A) = sup δa δλ δa = 1 cos θ xy For unitarily diagonalizable matrices y = x and thus κ (λ, A) = 1: perfectly conditioned! A. Donev (Courant Institute) Lecture IV 10/2014 9 / 23

Conditioning of Eigenvalue Problems Sensitivity of Eigenvectors A priori estimate: The conditioning number for the eigenvector itself depends on the separation between the eigenvalues κ (x, A) = ( ) 1 min λ λ j j This indicates that multiple eigenvalues require care. Even for Hermitian matrices eigenvectors are hard to compute. If there is a defective (non-diagonalizable) matrix with eigenvalue for which the difference between the algebraic and geometric multiplicities is d > 0, then δλ δa 1/(1+d), which means the conditioning number is infinite: Defective eigenvalues are very badly conditioned. A. Donev (Courant Institute) Lecture IV 10/2014 10 / 23

Computing Eigenvalues and Eigenvectors The need for iterative algorithms The eigenvalues are roots of the characteristic polynomial of A, which is generally of order n. According to Abel s theorem, there is no closed-form (rational) solution for n 5. All eigenvalue algorithms must be iterative! This is a fundamental difference from, example, linear solvers. There is an important distinction between iterative methods to: Compute all eigenvalues (similarity transformations). Compute only one or a few eigenvalues, typically the smallest or the largest one (power-like methods). Bounds on eigenvalues are important, e.g., Courant-Fisher theorem for the Rayleigh quotient: min λ r A (x) = x Ax x x max λ A. Donev (Courant Institute) Lecture IV 10/2014 11 / 23

Computing Eigenvalues and Eigenvectors The Power Method Recall that for a diagonalizable matrix A n = XΛ n X 1 and assume λ 1 > λ 2 λ 3 λ n and that the columns of X are normalized, x j = 1. Any initial guess vector q 0 can be represented in the linear basis formed by the eigenvectors q 0 = Xa Recall iterative methods for linear systems: Multiply a vector with the matrix A many times: q k+1 = Aq k q n = A n q 0 = ( XΛ n X 1) Xa = X (Λ n a) A. Donev (Courant Institute) Lecture IV 10/2014 12 / 23

Computing Eigenvalues and Eigenvectors Power Method As n, the eigenvalue of largest modulus λ 0 will dominate, { ( ) n } Λ n = λ n λ2 1Diag 1,,... λ n 1Diag {1, 0,..., 0} λ 1 q n = X (Λ n a) λ n 1X a 1 0. 0 = λn 1x 1 Therefore the normalized iterates converge to the eigenvector: q n = q n q n x 1 The Rayleigh quotient converges to the eigenvalue: r A (q n ) = q naq n q n q n = q na q n λ 1 A. Donev (Courant Institute) Lecture IV 10/2014 13 / 23

Computing Eigenvalues and Eigenvectors An alternative derivation A. Donev (Courant Institute) Lecture IV 10/2014 14 / 23

Computing Eigenvalues and Eigenvectors Power Method Implementation Start with an initial guess q 0, and then iterate: 1 Compute matrix-vector product and normalize it: q k = Aq k 1 Aq k 1 2 Use Raleigh quotient to obtain eigenvalue estimate: ˆλ k = q kaq k 3 Test for convergence: Evaluate the residual r k = Aq k ˆλ k q k and terminate if the error estimate is small enough: λ 1 ˆλ k r k cos θ xy < ε A. Donev (Courant Institute) Lecture IV 10/2014 15 / 23

Computing Eigenvalues and Eigenvectors Convergence Estimates The normalized iterates converge to the eigenvector linearly: ( λ 2 ) k q k (±x 1 ) = O If A is normal the eigenvalue estimate converges a bit faster but still linearly ( λ 2 ) 2k ˆλ k λ 1 O The power method is fast when the dominant eigenvalue is well-separated from the rest (even if it is degenerate). This conclusion is rather general for all iterative methods: Convergence is good for well-separated eigenvalues, bad otherwise. The power method is typically too slow to be used in practice and there are more sophisticated alternatives (Lanczos/Arnoldi iteration). A. Donev (Courant Institute) Lecture IV 10/2014 16 / 23 λ 1 λ 1

Computing Eigenvalues and Eigenvectors Inverse Power Iteration Observe that applying the power method to A 1 will find the largest of λ 1 j, i.e., the smallest eigenvalue (by modulus). If we have an eigenvalue estimate µ λ, then doing the power method for the matrix (A µi) 1 will give the eigenvalue closest to µ. Convergence will be faster if µ is much closer to λ then to other eigenvalues. Recall that in practice (A µi) 1 q is computed by solving a linear system, not matrix inversion (one can reuse an LU factorization)! Finally, if we have an estimate of both the eigenvalue and the eigenvector, we can use Rayleigh Quotient Iteration (see homework). A. Donev (Courant Institute) Lecture IV 10/2014 17 / 23

Methods based on QR factorizations Estimating all eigenvalues / eigenvectors Iterative methods akin the power method are not suitable for estimating all eigenvalues. Basic idea: Build a sequence of matrices A k that all share eigenvalues with A via similarity transformations: A k+1 = P 1 A k P, starting from A 1 = A. A numerically stable and good way to do this is to use the QR factorization: A k = Q k+1 R k+1 A k+1 = Q 1 k+1 A kq k+1 = ( Q 1 k+1 Q k+1) Rk+1 Q k+1 = R k+1 Q k+1. Note that the fact the Q s are orthogonal is crucial to keep the conditioning from getting worse. A. Donev (Courant Institute) Lecture IV 10/2014 18 / 23

Methods based on QR factorizations The basic QR method The behavior of the QR iteration can be understood most transparently as follows [following Trefethen and Bau]: Observation: The range of the matrix A k converges to the space spanned by the eigenvectors of A, with the eigenvectors corresponding to the largest eigenvalues dominating as k (so this is ill-conditioned). Recall: The columns of Q in A = QR form an orthonormal basis for the range of A. Idea: Form a well-conditioned basis for the eigenspace of A by factorizing: A k = Q k R k and then calculate A k = Q 1 k A Q k = Q ka Q k. It is not too hard to show that this produces the same sequence of matrices A k as the QR algorithm. A. Donev (Courant Institute) Lecture IV 10/2014 19 / 23

Methods based on QR factorizations Why the QR algorithm works Summary: The columns of Q k converge to the eigenvectors, and A k = Q ka Q k. We can recognize the above as a matrix of Rayleigh quotients, which for diagonalizable matrices { (A k ) ij = q λ i if i = j i A q j λ i δ ij = 0 if i j showing that (under suitable assumptions): It can also be shown that A k Λ Q k = Q 1 Q 2 Q k X A. Donev (Courant Institute) Lecture IV 10/2014 20 / 23

Methods based on QR factorizations More on QR algorithm The convergence of the basic QR algorithm is closely related to that of the power method: It is only fast if all eigenvalues are well-separated. For more general (non-diagonalizable) matrices in complex arithmetic, the algorithm converges to the Schur decomposition A = UTU, A k T and Q k U. It is possible to implement the algorithm entirely using real arithmetic (no complex numbers). There are several key improvements to the basic method that make this work in practice: Hessenberg matrices for faster QR factorization, shifts and deflation for acceleration. There are other sophisticated algorithms as well, such as the divide-and-conquer algorithm, and the best are implemented in the library LAPACK (MATLAB). A. Donev (Courant Institute) Lecture IV 10/2014 21 / 23

Conclusions Eigenvalues in MATLAB In MATLAB, sophisticated variants of the QR algorithm (LAPACK library) are implemented in the function eig: Λ = eig(a) [X, Λ] = eig(a) For large or sparse matrices, iterative methods based on the Arnoldi iteration (ARPACK library), can be used to obtain a few of the largest/smallest/closest-to-µ eigenvalues: Λ = eigs(a, n eigs ) [X, Λ] = eigs(a, n eigs ) The Schur decomposition is provided by [U, T ] = schur(a). A. Donev (Courant Institute) Lecture IV 10/2014 22 / 23

Conclusions Conclusions/Summary Eigenvalues are well-conditioned for unitarily diagonalizable matrices (includes Hermitian matrices), but ill-conditioned for nearly non-diagonalizable matrices. Eigenvectors are well-conditioned only when eigenvalues are well-separated. Eigenvalue algorithms are always iterative. The power method and its variants can be used to find the largest or smallest eigenvalue, and they converge fast if there is a large separation between the target eigenvalue and nearby ones. Estimating all eigenvalues and/or eigenvectors can be done by combining the power method with QR factorizations. MATLAB has high-quality implementations of sophisticated variants of these algorithms. A. Donev (Courant Institute) Lecture IV 10/2014 23 / 23