LARGE SPARSE EIGENVALUE PROBLEMS. General Tools for Solving Large Eigen-Problems

Similar documents
LARGE SPARSE EIGENVALUE PROBLEMS

Solution of eigenvalue problems. Subspace iteration, The symmetric Lanczos algorithm. Harmonic Ritz values, Jacobi-Davidson s method

Solution of eigenvalue problems. Subspace iteration, The symmetric Lanczos algorithm. Harmonic Ritz values, Jacobi-Davidson s method

Last Time. Social Network Graphs Betweenness. Graph Laplacian. Girvan-Newman Algorithm. Spectral Bisection

Eigenvalue Problems CHAPTER 1 : PRELIMINARIES

The Lanczos and conjugate gradient algorithms

AMS526: Numerical Analysis I (Numerical Linear Algebra)

A short course on: Preconditioned Krylov subspace methods. Yousef Saad University of Minnesota Dept. of Computer Science and Engineering

A short course on: Preconditioned Krylov subspace methods. Yousef Saad University of Minnesota Dept. of Computer Science and Engineering

ECS231 Handout Subspace projection methods for Solving Large-Scale Eigenvalue Problems. Part I: Review of basic theory of eigenvalue problems

4.8 Arnoldi Iteration, Krylov Subspaces and GMRES

Eigenvalue Problems. Eigenvalue problems occur in many areas of science and engineering, such as structural analysis

Scientific Computing: An Introductory Survey

13-2 Text: 28-30; AB: 1.3.3, 3.2.3, 3.4.2, 3.5, 3.6.2; GvL Eigen2

Krylov subspace projection methods

Matrix Algorithms. Volume II: Eigensystems. G. W. Stewart H1HJ1L. University of Maryland College Park, Maryland

A short course on: Preconditioned Krylov subspace methods. Yousef Saad University of Minnesota Dept. of Computer Science and Engineering

Numerical Methods in Matrix Computations

Charles University Faculty of Mathematics and Physics DOCTORAL THESIS. Krylov subspace approximations in linear algebraic problems

Index. for generalized eigenvalue problem, butterfly form, 211

Computation of eigenvalues and singular values Recall that your solutions to these questions will not be collected or evaluated.

Large-scale eigenvalue problems

Orthogonal iteration to QR

6.4 Krylov Subspaces and Conjugate Gradients

Iterative methods for Linear System

Arnoldi Methods in SLEPc

MATH 304 Linear Algebra Lecture 20: The Gram-Schmidt process (continued). Eigenvalues and eigenvectors.

Davidson Method CHAPTER 3 : JACOBI DAVIDSON METHOD

ABSTRACT OF DISSERTATION. Ping Zhang

ITERATIVE PROJECTION METHODS FOR SPARSE LINEAR SYSTEMS AND EIGENPROBLEMS CHAPTER 11 : JACOBI DAVIDSON METHOD

A HARMONIC RESTARTED ARNOLDI ALGORITHM FOR CALCULATING EIGENVALUES AND DETERMINING MULTIPLICITY

235 Final exam review questions

AMS526: Numerical Analysis I (Numerical Linear Algebra) Lecture 23: GMRES and Other Krylov Subspace Methods; Preconditioning

Introduction to Arnoldi method

Bindel, Fall 2016 Matrix Computations (CS 6210) Notes for

EIGENVALUE PROBLEMS. Background on eigenvalues/ eigenvectors / decompositions. Perturbation analysis, condition numbers..

The quadratic eigenvalue problem (QEP) is to find scalars λ and nonzero vectors u satisfying

EIGIFP: A MATLAB Program for Solving Large Symmetric Generalized Eigenvalue Problems

Applied Mathematics 205. Unit V: Eigenvalue Problems. Lecturer: Dr. David Knezevic

FEM and sparse linear system solving

Krylov Subspaces. Lab 1. The Arnoldi Iteration

Krylov Subspace Methods that Are Based on the Minimization of the Residual

A Chebyshev-based two-stage iterative method as an alternative to the direct solution of linear systems

Introduction to Iterative Solvers of Linear Systems

Matrices, Moments and Quadrature, cont d

Linear Algebra. Brigitte Bidégaray-Fesquet. MSIAM, September Univ. Grenoble Alpes, Laboratoire Jean Kuntzmann, Grenoble.

Henk van der Vorst. Abstract. We discuss a novel approach for the computation of a number of eigenvalues and eigenvectors

Solutions to Review Problems for Chapter 6 ( ), 7.1

Krylov Subspace Methods for Large/Sparse Eigenvalue Problems

M.A. Botchev. September 5, 2014

The Conjugate Gradient Method

APPLIED NUMERICAL LINEAR ALGEBRA

Singular Value Computation and Subspace Clustering

On Solving Large Algebraic. Riccati Matrix Equations

Approximating the matrix exponential of an advection-diffusion operator using the incomplete orthogonalization method

Krylov Subspaces. The order-n Krylov subspace of A generated by x is

On prescribing Ritz values and GMRES residual norms generated by Arnoldi processes

IDR(s) as a projection method

Numerical Methods I Eigenvalue Problems

I. Multiple Choice Questions (Answer any eight)

Krylov Space Methods. Nonstationary sounds good. Radu Trîmbiţaş ( Babeş-Bolyai University) Krylov Space Methods 1 / 17

ABSTRACT NUMERICAL SOLUTION OF EIGENVALUE PROBLEMS WITH SPECTRAL TRANSFORMATIONS

Sparse matrix methods in quantum chemistry Post-doctorale cursus quantumchemie Han-sur-Lesse, Belgium

Alternative correction equations in the Jacobi-Davidson method

Iterative methods for Linear System of Equations. Joint Advanced Student School (JASS-2009)

Ir O D = D = ( ) Section 2.6 Example 1. (Bottom of page 119) dim(v ) = dim(l(v, W )) = dim(v ) dim(f ) = dim(v )

Numerical Methods for Solving Large Scale Eigenvalue Problems

Eigenvalues and Eigenvectors

Preliminary/Qualifying Exam in Numerical Analysis (Math 502a) Spring 2012

A Jacobi Davidson Method for Nonlinear Eigenproblems

ON RESTARTING THE ARNOLDI METHOD FOR LARGE NONSYMMETRIC EIGENVALUE PROBLEMS

Block Bidiagonal Decomposition and Least Squares Problems

Efficient Methods For Nonlinear Eigenvalue Problems. Diploma Thesis

Numerical Methods - Numerical Linear Algebra

Recent advances in approximation using Krylov subspaces. V. Simoncini. Dipartimento di Matematica, Università di Bologna.

Iterative methods for symmetric eigenvalue problems

Application of Lanczos and Schur vectors in structural dynamics

A DISSERTATION. Extensions of the Conjugate Residual Method. by Tomohiro Sogabe. Presented to

The amount of work to construct each new guess from the previous one should be a small multiple of the number of nonzeros in A.

LINEAR ALGEBRA 1, 2012-I PARTIAL EXAM 3 SOLUTIONS TO PRACTICE PROBLEMS

The Eigenvalue Problem: Perturbation Theory

Least-Squares Systems and The QR factorization

Math 102, Winter Final Exam Review. Chapter 1. Matrices and Gaussian Elimination

1 Last time: least-squares problems

Preface to the Second Edition. Preface to the First Edition

A Tour of the Lanczos Algorithm and its Convergence Guarantees through the Decades

ECS130 Scientific Computing Handout E February 13, 2017

Conceptual Questions for Review

The German word eigen is cognate with the Old English word āgen, which became owen in Middle English and own in modern English.

MATH 20F: LINEAR ALGEBRA LECTURE B00 (T. KEMP)

EECS 275 Matrix Computation

Bindel, Spring 2016 Numerical Analysis (CS 4220) Notes for

DS-GA 1002 Lecture notes 0 Fall Linear Algebra. These notes provide a review of basic concepts in linear algebra.

Key words. conjugate gradients, normwise backward error, incremental norm estimation.

Chapter 6: Orthogonality

Moments, Model Reduction and Nonlinearity in Solving Linear Algebraic Problems

A JACOBI-DAVIDSON ITERATION METHOD FOR LINEAR EIGENVALUE PROBLEMS. GERARD L.G. SLEIJPEN y AND HENK A. VAN DER VORST y

The Pennsylvania State University The Graduate School NEWTON-KRYLOV METHODS FOR THE SOLUTION OF THE. k-eigenvalue PROBLEM IN MULTIGROUP NEUTRONICS

Problem Set (T) If A is an m n matrix, B is an n p matrix and D is a p s matrix, then show

Sensitivity of Gauss-Christoffel quadrature and sensitivity of Jacobi matrices to small changes of spectral data

Transcription:

LARGE SPARSE EIGENVALUE PROBLEMS Projection methods The subspace iteration Krylov subspace methods: Arnoldi and Lanczos Golub-Kahan-Lanczos bidiagonalization General Tools for Solving Large Eigen-Problems Projection techniques Arnoldi, Lanczos, Subspace Iteration; Preconditioninings: shift-and-invert, Polynomials,... Deflation and restarting techniques Computational codes often combine these three ingredients 14-2 Text: 36; AB: 4.6.1, 4.6.7-8, 4.5.4, 4.6.2; Gvl4 10.1,10.5.1 Eigen3 A few popular solution Methods Subspace Iteration [Now less popular sometimes used for validation] Arnoldi s method (or Lanczos) with polynomial acceleration Shift-and-invert and other preconditioners. [Use Arnoldi or Lanczos for (A σi) 1.] Davidson s method and variants, Jacobi-Davidson Specialized method: Automatic Multilevel Substructuring (AMLS). Projection Methods for Eigenvalue Problems Projection method onto K orthogonal to L Given: Two subspaces K and L of same dimension. Approximate eigenpairs λ, ũ, obtained by solving: Find: λ C, ũ K such that( λi A)ũ L Two types of methods: Orthogonal projection methods: Situation when L = K. Oblique projection methods: When L K. First situation leads to Rayleigh-Ritz procedure 14-3 Text: 36; AB: 4.6.1, 4.6.7-8, 4.5.4, 4.6.2; Gvl4 10.1,10.5.1 Eigen3 14-4 Text: 36; AB: 4.6.1, 4.6.7-8, 4.5.4, 4.6.2; Gvl4 10.1,10.5.1 Eigen3

Rayleigh-Ritz projection Given: a subspace X known to contain good approximations to eigenvectors of A. Question: How to extract best approximations to eigenvalues/ eigenvectors from this subspace? Answer: Orthogonal projection method Let Q = [q 1,..., q m ] = orthonormal basis of X Orthogonal projection method onto X yields: Q H (A λi)ũ = 0 Q H AQy = λy where ũ = Qy Known as Rayleigh Ritz process 14-5 Text: 36; AB: 4.6.1, 4.6.7-8, 4.5.4, 4.6.2; Gvl4 10.1,10.5.1 Eigen3 Procedure: 1. Obtain an orthonormal basis of X 2. Compute C = Q H AQ (an m m matrix) 3. Obtain Schur factorization of C, C = Y RY H 4. Compute Ũ = QY Property: if X is (exactly) invariant, then procedure will yield exact eigenvalues and eigenvectors. Proof: Since X is invariant, (A λi)u = Qz for a certain z. Q H Qz = 0 implies z = 0 and therefore (A λi)u = 0. Can use this procedure in conjunction with the subspace obtained from subspace iteration algorithm 14-6 Text: 36; AB: 4.6.1, 4.6.7-8, 4.5.4, 4.6.2; Gvl4 10.1,10.5.1 Eigen3 Subspace Iteration Original idea: Y = A k X projection technique onto a subspace of the form Practically: A k replaced by suitable polynomial Advantages: Easy to implement (in symmetric case); Easy to analyze; Disadvantage: Slow. Often used with polynomial acceleration: A k X replaced by C k (A)X. Typically C k = Chebyshev polynomial. Algorithm: Subspace Iteration with Projection 1. Start: Choose an initial system of vectors X = [x 0,..., x m ] and an initial polynomial C k. 2. Iterate: Until convergence do: (a) Compute Ẑ = C k (A)X. [Simplest case: Ẑ = AX.] (b) Orthonormalize Ẑ: [Z, R Z ] = qr(ẑ, 0) (c) Compute B = Z H AZ (d) Compute the Schur factorization B = Y R B Y H of B (e) Compute X := ZY. (f) Test for convergence. If satisfied stop. Else select a new polynomial C k and continue. 14-7 Text: 36; AB: 4.6.1, 4.6.7-8, 4.5.4, 4.6.2; Gvl4 10.1,10.5.1 Eigen3 14-8 Text: 36; AB: 4.6.1, 4.6.7-8, 4.5.4, 4.6.2; Gvl4 10.1,10.5.1 Eigen3

THEOREM: Let S 0 = span{x 1, x 2,..., x m } and assume that S 0 is such that the vectors {P x i } i=1,...,m are linearly independent where P is the spectral projector associated with λ 1,..., λ m. Let P k the orthogonal projector onto the subspace S k = span{x k }. Then for each eigenvector u i of A, i = 1,..., m, there exists a unique vector s i in the subspace S 0 such that P s i = u i. Moreover, the following inequality is satisfied (I P k )u i 2 u i s i 2 ( λ m+1 λ i where ɛ k tends to zero as k tends to infinity. k k) + ɛ, (1) KRYLOV SUBSPACE METHODS 14-9 Text: 36; AB: 4.6.1, 4.6.7-8, 4.5.4, 4.6.2; Gvl4 10.1,10.5.1 Eigen3 Krylov subspace methods Principle: Projection methods on Krylov subspaces: K m (A, v 1 ) = span{v 1, Av 1,, A m 1 v 1 } The most important class of projection methods [for linear systems and for eigenvalue problems] Variants depend on the subspace L Let µ = deg. of minimal polynom. of v 1. Then: K m = {p(a)v 1 p = polynomial of degree m 1} K m = K µ for all m µ. Moreover, K µ is invariant under A. dim(k m ) = m iff µ m. 14-11 Text: 36; AB: 4.6.1, 4.6.7-8, 4.5.4, 4.6.2; Gvl4 10.1,10.5.1 Eigen3 Arnoldi s algorithm Goal: to compute an orthogonal basis of K m. Input: Initial vector v 1, with v 1 2 = 1 and m. ALGORITHM : 1 For j = 1,..., m do Compute w := Av j For i = 1,..., j, do h j+1,j = w 2 ; v j+1 = w/h j+1,j End Arnoldi s procedure Based on Gram-Schmidt procedure { hi,j := (w, v i ) w := w h i,j v i 14-12 Text: 36; AB: 4.6.1, 4.6.7-8, 4.5.4, 4.6.2; Gvl4 10.1,10.5.1 Eigen3

Result of Arnoldi s algorithm x x x x x x x x x x x x x x x Let: H m = x x x x x x x x x x x x, H m = x x x x x x x x x x x x Results: 1. V m = [v 1, v 2,..., v m ] orthonormal basis of K m. 2. AV m = V m+1 H m = V m H m + h m+1,m v m+1 e T m 3. Vm T AV m = H m H m last row. 14-13 Text: 36; AB: 4.6.1, 4.6.7-8, 4.5.4, 4.6.2; Gvl4 10.1,10.5.1 Eigen3 Application to eigenvalue problems Write approximate eigenvector as ũ = V m y Galerkin condition: (A λi)v m y K m Vm H(A λi)v m y = 0 Approximate eigenvalues are eigenvalues of H m H m y j = λ j y j Associated approximate eigenvectors are ũ j = V m y j Typically a few of the outermost eigenvalues will converge first. 14-14 Text: 36; AB: 4.6.1, 4.6.7-8, 4.5.4, 4.6.2; Gvl4 10.1,10.5.1 Eigen3 Hermitian case: The Lanczos Algorithm The Hessenberg matrix becomes tridiagonal : A = A H and V H m AV m = H m H m = H H m Denote H m by T m and H m by T m. We can write α 1 β 2 β 2 α 2 β 3 T m = β 3 α 3 β 4...... β m α m Relation AV m = V m+1 T m Consequence: three term recurrence ALGORITHM : 2 β j+1 v j+1 = Av j α j v j β j v j 1 Lanczos 1. Choose an initial v 1 with v 1 2 = 1; Set β 1 0, v 0 0 2. For j = 1, 2,..., m Do: 3. w j := Av j β j v j 1 4. α j := (w j, v j ) 5. w j := w j α j v j 6. β j+1 := w j 2. If β j+1 = 0 then Stop 7. v j+1 := w j /β j+1 8. EndDo Hermitian matrix + Arnoldi Hermitian Lanczos 14-15 Text: 36; AB: 4.6.1, 4.6.7-8, 4.5.4, 4.6.2; Gvl4 10.1,10.5.1 Eigen3 14-16 Text: 36; AB: 4.6.1, 4.6.7-8, 4.5.4, 4.6.2; Gvl4 10.1,10.5.1 Eigen3

In theory v i s defined by 3-term recurrence are orthogonal. However: in practice severe loss of orthogonality; Observation [Paige, 1981]: Loss of orthogonality starts suddenly, when the first eigenpair has converged. It is a sign of loss of linear independence of the computed eigenvectors. When orthogonality is lost, then several the copies of the same eigenvalue start appearing. Reorthogonalization Full reorthogonalization reorthogonalize v j+1 against all previous v i s every time. Partial reorthogonalization reorthogonalize v j+1 against all previous v i s only when needed [Parlett & Simon] Selective reorthogonalization reorthogonalize v j+1 against computed eigenvectors [Parlett & Scott] No reorthogonalization Do not reorthogonalize - but take measures to deal with spurious eigenvalues. [Cullum & Willoughby] 14-17 Text: 36; AB: 4.6.1, 4.6.7-8, 4.5.4, 4.6.2; Gvl4 10.1,10.5.1 Eigen3 14-18 Text: 36; AB: 4.6.1, 4.6.7-8, 4.5.4, 4.6.2; Gvl4 10.1,10.5.1 Eigen3 Lanczos Bidiagonalization We now deal with rectangular matrices. Let A R m n. ALGORITHM : 3 Let: Golub-Kahan-Lanczos 1. Choose an initial v 1 with v 1 2 = 1; Set p v 1, β 0 1, u 0 0 2. For k = 1,..., p Do: 3. r := Av k β k 1 u k 1 4. α k = r 2 ; u k = r/α k 5. p = A T u k α k v k 6. β k = p 2 ; v k+1 := p/β k 7. EndDo V p+1 = [v 1, v 2,, v p+1 ] R n (p+1) U p = [u 1, u 2,, u p ] R m p Let: Result: α 1 β 2 α 2 β 3 B p =...... ;...... α p β p+1 ˆB p = B p (:, 1 : p) V p = [v 1, v 2,, v p ] R n p V T p+1 V p+1 = I U T p U p = I AV p = U p ˆB p A T U p = V p+1 B T p 14-19 Text: 36; AB: 4.6.1, 4.6.7-8, 4.5.4, 4.6.2; Gvl4 10.1,10.5.1 Eigen3 14-20 Text: 36; AB: 4.6.1, 4.6.7-8, 4.5.4, 4.6.2; Gvl4 10.1,10.5.1 Eigen3

Observe that : A T (AV p ) = A T (U p ˆB p ) = V p+1 B T p ˆB p B T p ˆB p is a (symmetric) tridiagonal matrix of size (p + 1) p Call this matrix T k. Then: (A T A)V p = V p+1 T p Standard Lanczos relation! Algorithm is equivalent to standard Lanczos applied to A T A. Similar result for the u i s [involves AA T ] Work out the details: What are the entries of T p relative to those of B p? 14-21 Text: 36; AB: 4.6.1, 4.6.7-8, 4.5.4, 4.6.2; Gvl4 10.1,10.5.1 Eigen3