Last Time. Social Network Graphs Betweenness. Graph Laplacian. Girvan-Newman Algorithm. Spectral Bisection

Similar documents
Eigenvalue Problems. Eigenvalue problems occur in many areas of science and engineering, such as structural analysis

Scientific Computing: An Introductory Survey

Computational Methods. Eigenvalues and Singular Values

AMS526: Numerical Analysis I (Numerical Linear Algebra)

4.8 Arnoldi Iteration, Krylov Subspaces and GMRES

LARGE SPARSE EIGENVALUE PROBLEMS. General Tools for Solving Large Eigen-Problems

LARGE SPARSE EIGENVALUE PROBLEMS

Krylov Space Methods. Nonstationary sounds good. Radu Trîmbiţaş ( Babeş-Bolyai University) Krylov Space Methods 1 / 17

Math 405: Numerical Methods for Differential Equations 2016 W1 Topics 10: Matrix Eigenvalues and the Symmetric QR Algorithm

Matrices, Moments and Quadrature, cont d

Lecture # 11 The Power Method for Eigenvalues Part II. The power method find the largest (in magnitude) eigenvalue of. A R n n.

EIGENVALUE PROBLEMS. EIGENVALUE PROBLEMS p. 1/4

Computation of eigenvalues and singular values Recall that your solutions to these questions will not be collected or evaluated.

Krylov subspace projection methods

The Lanczos and conjugate gradient algorithms

Matrix Algorithms. Volume II: Eigensystems. G. W. Stewart H1HJ1L. University of Maryland College Park, Maryland

Eigenvalue and Eigenvector Problems

Orthogonal iteration to QR

Eigenvalues and Eigenvectors

AMS526: Numerical Analysis I (Numerical Linear Algebra) Lecture 23: GMRES and Other Krylov Subspace Methods; Preconditioning

13-2 Text: 28-30; AB: 1.3.3, 3.2.3, 3.4.2, 3.5, 3.6.2; GvL Eigen2

Krylov Subspaces. Lab 1. The Arnoldi Iteration

Eigenvalues and eigenvectors

Course Notes: Week 1

MAA507, Power method, QR-method and sparse matrix representation.

Krylov Subspaces. The order-n Krylov subspace of A generated by x is

Linear algebra & Numerical Analysis

ECS231 Handout Subspace projection methods for Solving Large-Scale Eigenvalue Problems. Part I: Review of basic theory of eigenvalue problems

Applied Mathematics 205. Unit V: Eigenvalue Problems. Lecturer: Dr. David Knezevic

Solution of eigenvalue problems. Subspace iteration, The symmetric Lanczos algorithm. Harmonic Ritz values, Jacobi-Davidson s method

Solution of eigenvalue problems. Subspace iteration, The symmetric Lanczos algorithm. Harmonic Ritz values, Jacobi-Davidson s method

AMS526: Numerical Analysis I (Numerical Linear Algebra)

Lecture 4 Eigenvalue problems

Preliminary/Qualifying Exam in Numerical Analysis (Math 502a) Spring 2012

22.4. Numerical Determination of Eigenvalues and Eigenvectors. Introduction. Prerequisites. Learning Outcomes

Eigenvalue Problems CHAPTER 1 : PRELIMINARIES

Linear Algebra: Matrix Eigenvalue Problems

Math 504 (Fall 2011) 1. (*) Consider the matrices

ECS130 Scientific Computing Handout E February 13, 2017

6.4 Krylov Subspaces and Conjugate Gradients

Numerical Methods in Matrix Computations

Iterative methods for Linear System of Equations. Joint Advanced Student School (JASS-2009)

Krylov Subspace Methods that Are Based on the Minimization of the Residual

Iterative methods for Linear System

Index. for generalized eigenvalue problem, butterfly form, 211

The quadratic eigenvalue problem (QEP) is to find scalars λ and nonzero vectors u satisfying

GPU accelerated Arnoldi solver for small batched matrix

1. What is the determinant of the following matrix? a 1 a 2 4a 3 2a 2 b 1 b 2 4b 3 2b c 1. = 4, then det

The German word eigen is cognate with the Old English word āgen, which became owen in Middle English and own in modern English.

Numerical Analysis Lecture Notes

A Cholesky LR algorithm for the positive definite symmetric diagonal-plus-semiseparable eigenproblem

Chapter 7. Iterative methods for large sparse linear systems. 7.1 Sparse matrix algebra. Large sparse matrices

EIGENVALUE PROBLEMS (EVP)

AMS526: Numerical Analysis I (Numerical Linear Algebra)

Krylov Subspace Methods for Large/Sparse Eigenvalue Problems

A HARMONIC RESTARTED ARNOLDI ALGORITHM FOR CALCULATING EIGENVALUES AND DETERMINING MULTIPLICITY

EIGENVALUE PROBLEMS. Background on eigenvalues/ eigenvectors / decompositions. Perturbation analysis, condition numbers..

Solving large scale eigenvalue problems

Direct methods for symmetric eigenvalue problems

ON ORTHOGONAL REDUCTION TO HESSENBERG FORM WITH SMALL BANDWIDTH

Charles University Faculty of Mathematics and Physics DOCTORAL THESIS. Krylov subspace approximations in linear algebraic problems

The Eigenvalue Problem: Perturbation Theory

Notes on Eigenvalues, Singular Values and QR

Lecture 10 - Eigenvalues problem

Numerical Methods - Numerical Linear Algebra

Computing Eigenvalues and/or Eigenvectors;Part 1, Generalities and symmetric matrices

EECS 275 Matrix Computation

Math 489AB Exercises for Chapter 2 Fall Section 2.3

On prescribing Ritz values and GMRES residual norms generated by Arnoldi processes

Scientific Computing with Case Studies SIAM Press, Lecture Notes for Unit VII Sparse Matrix

Iterative Methods for Sparse Linear Systems

Approximating the matrix exponential of an advection-diffusion operator using the incomplete orthogonalization method

Math 108b: Notes on the Spectral Theorem

THE QR METHOD A = Q 1 R 1

Section 4.4 Reduction to Symmetric Tridiagonal Form

Recall : Eigenvalues and Eigenvectors

1. Let m 1 and n 1 be two natural numbers such that m > n. Which of the following is/are true?

Remark By definition, an eigenvector must be a nonzero vector, but eigenvalue could be zero.

Today: eigenvalue sensitivity, eigenvalue algorithms Reminder: midterm starts today

Arnoldi Methods in SLEPc

Numerical Solution of Linear Eigenvalue Problems

Numerical Methods I Eigenvalue Problems

APPLIED NUMERICAL LINEAR ALGEBRA

Section 4.5 Eigenvalues of Symmetric Tridiagonal Matrices

Ma/CS 6b Class 20: Spectral Graph Theory

11.3 Eigenvalues and Eigenvectors of a Tridiagonal Matrix

5.3 The Power Method Approximation of the Eigenvalue of Largest Module

Foundations of Matrix Analysis

Linear Algebra. Brigitte Bidégaray-Fesquet. MSIAM, September Univ. Grenoble Alpes, Laboratoire Jean Kuntzmann, Grenoble.

Orthogonal iteration to QR

Bindel, Fall 2016 Matrix Computations (CS 6210) Notes for

Study Guide for Linear Algebra Exam 2

Sparse matrix methods in quantum chemistry Post-doctorale cursus quantumchemie Han-sur-Lesse, Belgium

FEM and sparse linear system solving

MA 265 FINAL EXAM Fall 2012

MATH 20F: LINEAR ALGEBRA LECTURE B00 (T. KEMP)

Comparing iterative methods to compute the overlap Dirac operator at nonzero chemical potential

Computing Eigenvalues and/or Eigenvectors;Part 1, Generalities and symmetric matrices

SOLVING SPARSE LINEAR SYSTEMS OF EQUATIONS. Chao Yang Computational Research Division Lawrence Berkeley National Laboratory Berkeley, CA, USA

Definition (T -invariant subspace) Example. Example

Transcription:

Eigenvalue Problems

Last Time Social Network Graphs Betweenness Girvan-Newman Algorithm Graph Laplacian Spectral Bisection λ 2, w 2

Today Small deviation into eigenvalue problems

Formulation Standard eigenvalue problem: Given a n n matrix A, find scalar λ and a nonzero vector x such that Ax = λx λ is a eigenvalue, and x is the corresponding eigenvector Spectrum = λ(a) = set of eigenvalues of A Spectral radius = ρ A = max λ λ λ A

Characteristic Polynomial Equation Ax = λx is equivalent to A λi x = 0 Eigenvalues of A are roots of the characteristic polynomial det A λi = 0 The characteristic polynomial is a powerful theoretical tool but usually not useful computationally

Considerations Properties of eigenvalue problem affecting choice of algorithm Are all eigenvalues needed, or only a few? Are only eigenvalues needed, or are corresponding eigenvectors also needed? Is matrix real or complex? Is matrix relatively small and dense, or large and sparse? Does matrix have any special properties, such as symmetry?

Problem Transformations Shift: If Ax = λx and σ is any scalar, then A σi x = λ σ x Inversion: If A is nonsingular and Ax = λx, then λ 0 and A 1 x = 1 λ x Powers: If Ax = λx, then A k x = λ k x Polynomial: If Ax = λx and p(t) is a polynomial, then p A x = p λ x

Similarity Transforms B is similar to A if there is a nonsingular matrix T, such that B = T 1 AT Then, By = λy T 1 ATy = λy A Ty = λ Ty Similarity transformations preserve eigenvalues and eigenvectors are easily recovered

Diagonal form Eigenvalues of a diagonal matrix are the diagonal entries and the eigenvectors are columns of the Identity matrix The diagonal form is highly desirable in simplifying eigenvalue problems for general matrices by similarity transformations But not all matrices are diagonalizable by similarity transformations

Triangular form Any matrix can be transformed into a triangular form by similarity The eigenvalues are simply the diagonal values Eigenvectors are not as obvious, but still easy to compute

Power iteration Simplest method for computing one eigenvalue-eigenvector pair x k = Ax k 1 Converges to multiple of eigenvector corresponding to dominant eigenvalue We have seen this before while computing the Page Rank Proof of convergence?

Convergence of Power iteration Express starting vector x 0 in terms of the eigenvectors of A Then, n x 0 = α i v i i=1 x k = Ax k 1 = A 2 x k 2 = = A k x 0 n i=1 λ i k α i v i = λ n k α n v n + n 1 k λi λ n i=1 α i v i Since λ i λ n < 1, successively higher powers go to zero

Power iteration with shift Convergence rate of power iteration depends on the ratio λ n 1 λ n It is possible to choose a shift, A σi, such that λ n 1 σ λ n σ < λ n 1 λ n so convergence is accelerated Shift must be added back to result to obtain eigenvalue of original matrix

Inverse iteration If the smallest eigenvalues are required rather than the largest, we can make use of the fact that the eigenvalues of A 1 are reciprocals of those of A, so the smallest eigenvalue of A is the reciprocal of the largest eigenvalue of A 1 This leads to the inverse iteration scheme Ay k = x k 1 x k = y k / y k Inverse of A is not computed explicitly, but some factorization of A is used to solve the system at each iteration

Shifted inverse iteration As before, the shifting strategy using a scalar σ can greatly improve convergence It is particularly useful for computing the eigenvector corresponding to the approximate eigenvalue Inverse iteration is also useful for computing the eigenvalue closest to a given value β, since if β is used as the shift, then the desired eigenvalue corresponds to the smallest eigenvalue of the shifted matrix

Deflation Once the dominant eigenvalue and eigenvector λ n, w n have been computed, the remaining eigenvalues can be computed using deflation, which effectively removes the known eigenvalue Let H be any nonsingular matrix such that Hx = αe 1 Then the similarity transform determined by H transforms A into, b T HAH 1 = λ n 0 B Can now work with B to compute the next eigenvalue, Process can be repeated to find additional eigenvalues and eigenvectors

Deflation Alternate approach: let u n be any vector such that u n T w n = λ n Then A w n u n T has eigenvalues λ n 1,, λ 1, 0 Possible choices for u n u n = λ n w n, if A is symmetric and w n is normalized so that w n 2 = 1 u n = λ n y n, where y n is the corresponding left eigenvector (A T y n = λ n y n ) u n = A T e k, if w n is normalized such that w n = 1 and the k th component of w n is 1

QR Iteration Iteratively converges to a triangular or block-triangular form, yielding all eigenvalues of A Starting with A 0 = A, at iteration k compute QR factorization, Q k R k = A k 1 And form the reverse product, A k = R k Q k Product of orthogonal matrices Q k converges to matrix of corresponding eigenvectors If A is symmetric, then symmetry is preserved by the QR iteration, so A k converges to a matrix that is both symmetric and triangular diagonal

Preliminary reductions Efficiency of QR iteration can be enhanced by first transforming the matrix to be as close to triangular form as possible Hessenberg matrix is triangular except for one additional nonzero diagonal immediately adjacent to the main diagonal Symmetric Hessenberg matrix is tridiagonal Any matrix can be reduced to Hessenberg form in finite number of steps using Householder transformations Work per iteration is reduced from O n 3 to O(n 2 ) for general matrices and O(n) for symmetric matrices

Krylov subspace methods Reduces matrix to Hessenberg or tridiagonal form using only matrixvector products For arbitrary starting vector x 0, if K k = x 0 Ax 0 A k 1 x 0 then where C n is upper Hessenberg K n 1 AK n = C n To obtain a better conditioned basis for span(k n ), compute the QR factorization, Q n R n = K n so that with H is upper Hessenberg Q n H AQ n = R n C n R n 1 H

Krylov subspace methods Equating k th columns on each side of the equation AQ n = Q n H yields Aq k = h 1k q 1 + + h kk q k + h k+1,k q k+1 relating q k+1 to preceding vectors q 1,, q k Premultiplying by q H j and using orthonormality h jk = q H j Aq k, j = 1,, k These relationships yield the Arnoldi iteration

Arnoldi iteration x 0 : arbitrary nonzero starting vector q 1 = x 0 / x 0 2 for k = 1,2, u k = Aq k for j = 1 to k h jk = q j H u k u k = u k h jk q j h k+1,k = u k 2 if h k+1,k = 0 then stop q k+1 = u k /h k+1,k

Arnoldi iteration If then Q k = q 1 q k H k = Q k H AQ k is a upper Hessenberg matrix Eigenvalues of H k, called Ritz values, are approximate eigenvalues of A, and Ritz vectors given by Q k y, where y is an eigenvector of H k, are corresponding approximate eigenvectors of A Eigenvalues of H k must be computed by another method, such as QR iteration, but this is an easier problem as k n

Arnoldi iteration Is fairly expensive in both work and storage because each new vector q k must be orthogonalized against all previous columns of Q k, and all must be stored for that purpose Is usually restarted periodically with a carefully chosen starting vector Ritz vectors and values produced are often good approximations to eigenvalues and eigenvectors of A after relatively few iterations

Lanczos iteration Work and storage costs drop dramatically if the matrix is symmetric or Hermitian, since the recurrence has only three terms and H k is tridiagonal q 0, β 0 = 0 and x 0 = arbitrary nonzero starting vector q 1 = x 0 / x 0 2 for k = 1,2, u k = Aq k α k = q k H u k u k = u k β k 1 q k 1 α k q k β k = u k 2 if β k = 0 then stop q k+1 = u k /β k

Lanczos iteration α k and β k are diagonal and subdiagonal entries of symmetric tridiagonal matrix T k As with Arnoldi, Lanczos does not produce eigenvalues and eigenvectors directly, but only the tridiagonal matrix T k, whose eigenvalues and eigenvectors must be computed by another method to obtain Ritz values and vectors If β = 0, then the invariant subspace has already been identified, i.e., the Ritz values and vectors are already exact at that point

Lanczos iteration In principle, if we let Lanczos run until k = n, the resulting tridiagonal matrix would be orthogonally similar to A In practice, rounding errors cause loss of orthogonality Problem can be overcome by reorthogonalizing vectors In practice, this is usually ignored. The resulting approximations are still good

Krylov subspace methods Great advantage of Arnoldi and Lanczos is their ability to produce good approximations to extreme eigenvalues for k n They only require one matrix-vector product per step and little auxiliary storage, so are ideally suited for large sparse matrices If eigenvalues are needed in the middle of the spectrum, say near σ, then the algorithms can be applied to A σi 1, assuming it is practical to solve systems of the form A σi x = y