Is there life after the Lanczos method? What is LOBPCG?

Similar documents
Preconditioned Eigenvalue Solvers for electronic structure calculations. Andrew V. Knyazev. Householder Symposium XVI May 26, 2005

Implementation of a preconditioned eigensolver using Hypre

Preconditioned Eigensolver LOBPCG in hypre and PETSc

EIGIFP: A MATLAB Program for Solving Large Symmetric Generalized Eigenvalue Problems

Recent implementations, applications, and extensions of the Locally Optimal Block Preconditioned Conjugate Gradient method LOBPCG

Toward the Optimal Preconditioned Eigensolver: Locally Optimal Block Preconditioned Conjugate Gradient Method

Using the Karush-Kuhn-Tucker Conditions to Analyze the Convergence Rate of Preconditioned Eigenvalue Solvers

Preconditioned Locally Minimal Residual Method for Computing Interior Eigenpairs of Symmetric Operators

Preconditioned inverse iteration and shift-invert Arnoldi method

State-of-the-art numerical solution of large Hermitian eigenvalue problems. Andreas Stathopoulos

A Note on Inverse Iteration

STEEPEST DESCENT AND CONJUGATE GRADIENT METHODS WITH VARIABLE PRECONDITIONING

Domain Decomposition-based contour integration eigenvalue solvers

Solving Symmetric Indefinite Systems with Symmetric Positive Definite Preconditioners

Recent advances in approximation using Krylov subspaces. V. Simoncini. Dipartimento di Matematica, Università di Bologna.

Multigrid absolute value preconditioning

Last Time. Social Network Graphs Betweenness. Graph Laplacian. Girvan-Newman Algorithm. Spectral Bisection

Conjugate-Gradient Eigenvalue Solvers in Computing Electronic Properties of Nanostructure Architectures

Iterative methods for symmetric eigenvalue problems

SOLVING MESH EIGENPROBLEMS WITH MULTIGRID EFFICIENCY

DELFT UNIVERSITY OF TECHNOLOGY

APPLIED NUMERICAL LINEAR ALGEBRA

arxiv: v1 [cs.ms] 18 May 2007

A Chebyshev-based two-stage iterative method as an alternative to the direct solution of linear systems

Iterative Methods for Solving A x = b

The Nullspace free eigenvalue problem and the inexact Shift and invert Lanczos method. V. Simoncini. Dipartimento di Matematica, Università di Bologna

Inexact inverse iteration with preconditioning

Lecture 3: Inexact inverse iteration with preconditioning

Applied Mathematics 205. Unit V: Eigenvalue Problems. Lecturer: Dr. David Knezevic

Contents. Preface... xi. Introduction...

c 2009 Society for Industrial and Applied Mathematics

ETNA Kent State University

The Lanczos and conjugate gradient algorithms

AMS526: Numerical Analysis I (Numerical Linear Algebra) Lecture 23: GMRES and Other Krylov Subspace Methods; Preconditioning

Notes on PCG for Sparse Linear Systems

Rayleigh-Ritz majorization error bounds with applications to FEM and subspace iterations

M.A. Botchev. September 5, 2014

Numerical Methods for Solving Large Scale Eigenvalue Problems

Domain decomposition on different levels of the Jacobi-Davidson method

FEM and sparse linear system solving

1. Introduction. In this paper we consider the large and sparse eigenvalue problem. Ax = λx (1.1) T (λ)x = 0 (1.2)

Iterative methods for Linear System of Equations. Joint Advanced Student School (JASS-2009)

On the Superlinear Convergence of MINRES. Valeria Simoncini and Daniel B. Szyld. Report January 2012

A Tuned Preconditioner for Inexact Inverse Iteration for Generalised Eigenvalue Problems

Lecture 11: CMSC 878R/AMSC698R. Iterative Methods An introduction. Outline. Inverse, LU decomposition, Cholesky, SVD, etc.

Conjugate-Gradient Eigenvalue Solvers in Computing Electronic Properties of Nanostructure Architectures

Space-Variant Computer Vision: A Graph Theoretic Approach

A Tuned Preconditioner for Inexact Inverse Iteration Applied to Hermitian Eigenvalue Problems

LARGE SPARSE EIGENVALUE PROBLEMS. General Tools for Solving Large Eigen-Problems

Conjugate Gradient Method

A CHEBYSHEV-DAVIDSON ALGORITHM FOR LARGE SYMMETRIC EIGENPROBLEMS

ETNA Kent State University

ORIE 6334 Spectral Graph Theory October 13, Lecture 15

LARGE SPARSE EIGENVALUE PROBLEMS

Matrix Algorithms. Volume II: Eigensystems. G. W. Stewart H1HJ1L. University of Maryland College Park, Maryland

Arnoldi Methods in SLEPc

ECS231 Handout Subspace projection methods for Solving Large-Scale Eigenvalue Problems. Part I: Review of basic theory of eigenvalue problems

Numerical Methods - Numerical Linear Algebra

AMS526: Numerical Analysis I (Numerical Linear Algebra)

ETNA Kent State University

PRECONDITIONED ITERATIVE METHODS FOR LINEAR SYSTEMS, EIGENVALUE AND SINGULAR VALUE PROBLEMS. Eugene Vecharynski. M.S., Belarus State University, 2006

Chapter 7 Iterative Techniques in Matrix Algebra

Computational Methods. Eigenvalues and Singular Values

Inexact Inverse Iteration for Symmetric Matrices

IDR(s) Master s thesis Goushani Kisoensingh. Supervisor: Gerard L.G. Sleijpen Department of Mathematics Universiteit Utrecht

6.4 Krylov Subspaces and Conjugate Gradients

WHEN studying distributed simulations of power systems,

A Model-Trust-Region Framework for Symmetric Generalized Eigenvalue Problems

Polynomial Jacobi Davidson Method for Large/Sparse Eigenvalue Problems

c 2006 Society for Industrial and Applied Mathematics

Max Planck Institute Magdeburg Preprints

A geometric theory for preconditioned inverse iteration III: A short and sharp convergence estimate for generalized eigenvalue problems

Algebraic Multigrid as Solvers and as Preconditioner

Numerical Optimization

A Parallel Scalable PETSc-Based Jacobi-Davidson Polynomial Eigensolver with Application in Quantum Dot Simulation

A Domain Decomposition Based Jacobi-Davidson Algorithm for Quantum Dot Simulation

Alternative correction equations in the Jacobi-Davidson method

Key words. conjugate gradients, normwise backward error, incremental norm estimation.

Finite-choice algorithm optimization in Conjugate Gradients

Numerical Methods in Matrix Computations

An Arnoldi Method for Nonlinear Symmetric Eigenvalue Problems

Adaptive Coarse Space Selection in BDDC and FETI-DP Iterative Substructuring Methods: Towards Fast and Robust Solvers

NEW ESTIMATES FOR RITZ VECTORS

A Jacobi-Davidson method for solving complex symmetric eigenvalue problems Arbenz, P.; Hochstenbach, M.E.

Majorization for Changes in Ritz Values and Canonical Angles Between Subspaces (Part I and Part II)

AMS526: Numerical Analysis I (Numerical Linear Algebra)

Model order reduction of large-scale dynamical systems with Jacobi-Davidson style eigensolvers

1. Introduction. In this paper we consider variants of the Jacobi Davidson (JD) algorithm [27] for computing a few eigenpairs of

Motivation: Sparse matrices and numerical PDE's

ABSTRACT OF DISSERTATION. Ping Zhang

DELFT UNIVERSITY OF TECHNOLOGY

DELFT UNIVERSITY OF TECHNOLOGY

A HARMONIC RESTARTED ARNOLDI ALGORITHM FOR CALCULATING EIGENVALUES AND DETERMINING MULTIPLICITY

Iterative Methods for Linear Systems of Equations

Conjugate gradient acceleration of non-linear smoothing filters Iterated edge-preserving smoothing

Linear Solvers. Andrew Hazel

The Deflation Accelerated Schwarz Method for CFD

Incomplete Cholesky preconditioners that exploit the low-rank property

Eigenvalue Problems CHAPTER 1 : PRELIMINARIES

Krylov Subspace Methods to Calculate PageRank

Transcription:

1 Is there life after the Lanczos method? What is LOBPCG? Andrew V. Knyazev Department of Mathematics and Center for Computational Mathematics University of Colorado at Denver SIAM ALA Meeting, July 17, 2003 Supported by the NSF DMS

2 Abstract Let us consider the problem of computing the smallest eigenvalue λ and the corresponding eigenvector x of a real symmetric matrix A. We restrict ourselves to the class of polynomial iterative methods, i. e. methods that generate approximations to x within the Krylov subspaces, and we use the Rayleigh quotient to measure the approximation quality. Under these assumptions, the Lanczos method gives the best possible approximation as it provides the global minimization of the Rayleigh quotient on Krylov subspaces. Is there a possible competitor?

3 Let us consider the local minimization of the Rayleigh quotient on the subspace spanned by the current approximation, the current residual and the previous approximation. This method was suggested in Knyazev, 1986. The preconditioned version is in Knyazev, 1991, the block version in Knyazev, 1998, and a practically stable implementation in Knyazev, 2001. It is called the locally optimal block preconditioned conjugate gradient (LOBPCG) method. The costs per iteration and the memory use in LOBPCG are competitive with those of the Lanczos method. According to preliminary numerical tests, the LOBPCG is capable of capturing the same linear (but not super-linear) convergence speed rate as that of the Lanczos method. It computes an approximation to the eigenvector directly on every iteration and has no numerical stability issues similar to those of the Lanczos method.

4 OUTLINE: 1. What is LOBPCG? 2. A Bit of Theory 3. Block iterations in LOBPCG 4. Comparison of Features with Lanczos and Block Lanczos 5. LOBPCG Software 6. Numerical Examples 7. LOBPCG Applications 8. Conclusions

5 1. What is LOBPCG? The method combines robustness and simplicity of the steepest descent method with a three-term recurrence formula: x (i+1) = w (i) + τ (i) x (i) + γ (i) x (i 1), w (i) = Ax (i) λ (i) x (i), λ (i) = λ(x (i) )=(x (i),ax (i) )/(x (i),x (i) ) with properly chosen scalar iteration parameters τ (i) and γ (i). The easiest and most efficient choice of parameters is based on an idea of local optimality Knyazev 1986, namely, select τ (i) and γ (i) that minimize the Rayleigh quotient λ(x (i+1) )byusingtherayleigh Ritz method. Three-term recurrence + Rayleigh Ritz method = Locally Optimal Conjugate Gradient Method

6 2. A Bit of Theory (Knyazev, Neymeyr 2001) If λ (i) [λ k,λ k+1 [ then it holds for the Rayleigh quotient λ (i+1) that either λ (i+1) <λ k (unless k = 1), or λ (i+1) [λ k,λ (i) [. In the latter case, λ (i+1) λ k q2 λ(i) λ k, (1) λ k+1 λ (i+1) λ k+1 λ (i) where q =1 2 κ(a)+1 ( 1 λ ) k. (2) λ k+1

7 3. Block iterations in LOBPCG A well known idea of using Simultaneous, or Block Iterations provides an important improvement over singe-vector methods, and permits us to compute an m>1 dimensional invariant subspace, rather than one eigenvector at a time. It can also serve as an acceleration technique over a single-vector methods on parallel computers, as convergence for extreme eigenvalues usually increases with the size of the block, and every step can be naturally implemented on wide varieties of multiprocessor computers as well as to take advantage of high level BLAS libraries. The LOBPCG Method, is a straightforward generalization of the singe-vector version enhanced with the Rayleigh Ritz procedure on a larger trial subspace. Compared to other alternatives, e.g., trace minimization methods, LOBPCG demonstrates optimal convergence numerically.

8 4. Comparison of Features with Lanczos and Block Lanczos Property LOBPCG Lanczos B. Lanczos Optimal convergence Almost Yes Yes Superlinear convergence No Yes Yes Missing eigenpairs No May be No Using block vectors BLAS Yes No Yes Algorithm simplicity Yes No No Stability/Extra costs Low High Higher Extra costs for eigenvectors None High Higher Inexact inverse/preconditioning Yes No No Recursive use Yes No Partially

9 5. LOBPCG Software LOBPCG is publicly available as: MATLAB code posted at http://www.mathworks.com/ MPI C code for Hypre posted at http://www.llnl.gov/casc/hypre/ Both are also at the Web page maintained by the speaker, http://math.cudenver.edu/ aknyazev/

10 6. Numerical Examples 7 Point 3 D Laplacian with 1/h = 100, 126, 159, 200, 252 on IBM ASCI BLUE (4 processors per node) at LLNL. The smallest eigenpair with the tolerance of 10 6 using Hypre Schwarz PCG preconditioner is computed. 1500 Total Execution Time Prec. Setup Apply Prec. Lobpcg Lin. Alg. 50 45 Average LOBPCG Execution Time/Iteration Apply Prec. Lobpcg Lin. Alg. 40 Time (Seconds) 1000 500 Time (Seconds) 35 30 25 20 15 10 5 0 2 4 8 16 32 64 Number of Processors 2 4 8 16 32 64 Figure 1: Execution Time to Converge as the Problem Size Doubles 0 Number of Processors

11 7.LOBPCG Applications 1. Electronic structure calculations (joint with John Pask, LLNL, work in progress) 2. Spectral biclustering for microarrays DNA data (joint with Marina Kniazeva, work in progress) 3. Structural dynamics in ANSYS (work in progress) 4. Electromagnetic resonators (Rich Lehoucq and Peter Arbenz) 5. Image segmentation (private communication)

12 8. Conclusion Conditions, favorable for LOBPCG vs. Lanczos: 1. Good preconditioner is available, especially if the preconditioning function can accept block vectors. 2. Computation of eigenvectors is necessary and approximations to eigenvectors of interest are known. 3. No superlinear convergence of Lanczos is expected.