Polynomial filtering for interior eigenvalue problems Yousef Saad Department of Computer Science and Engineering University of Minnesota

Similar documents
Divide and conquer algorithms for large eigenvalue problems Yousef Saad Department of Computer Science and Engineering University of Minnesota

ILAS 2017 July 25, 2017

PASC17 - Lugano June 27, 2017

The EVSL package for symmetric eigenvalue problems Yousef Saad Department of Computer Science and Engineering University of Minnesota

A FILTERED LANCZOS PROCEDURE FOR EXTREME AND INTERIOR EIGENVALUE PROBLEMS

FILTERED CONJUGATE RESIDUAL-TYPE ALGORITHMS WITH APPLICATIONS

Filtered Conjugate Residual-type Algorithms with Applications

A Spectrum Slicing Method for the Kohn-Sham Problem

Polynomial filtered Lanczos iterations with applications in Density Functional Theory

Solution of eigenvalue problems. Subspace iteration, The symmetric Lanczos algorithm. Harmonic Ritz values, Jacobi-Davidson s method

Solution of eigenvalue problems. Subspace iteration, The symmetric Lanczos algorithm. Harmonic Ritz values, Jacobi-Davidson s method

LARGE SPARSE EIGENVALUE PROBLEMS

EIGIFP: A MATLAB Program for Solving Large Symmetric Generalized Eigenvalue Problems

LARGE SPARSE EIGENVALUE PROBLEMS. General Tools for Solving Large Eigen-Problems

DELFT UNIVERSITY OF TECHNOLOGY

Multilevel low-rank approximation preconditioners Yousef Saad Department of Computer Science and Engineering University of Minnesota

Solving Symmetric Indefinite Systems with Symmetric Positive Definite Preconditioners

Domain Decomposition-based contour integration eigenvalue solvers

Block Iterative Eigensolvers for Sequences of Dense Correlated Eigenvalue Problems

Cucheb: A GPU implementation of the filtered Lanczos procedure. Department of Computer Science and Engineering University of Minnesota, Twin Cities

Last Time. Social Network Graphs Betweenness. Graph Laplacian. Girvan-Newman Algorithm. Spectral Bisection

LEAST-SQUARES RATIONAL FILTERS FOR THE SOLUTION OF INTERIOR EIGENVALUE PROBLEMS

Density-Matrix-Based Algorithms for Solving Eingenvalue Problems

Multilevel Low-Rank Preconditioners Yousef Saad Department of Computer Science and Engineering University of Minnesota. Modelling 2014 June 2,2014

Applied Mathematics 205. Unit V: Eigenvalue Problems. Lecturer: Dr. David Knezevic

The amount of work to construct each new guess from the previous one should be a small multiple of the number of nonzeros in A.

Preconditioned Locally Minimal Residual Method for Computing Interior Eigenpairs of Symmetric Operators

Part 3: Trust-region methods for unconstrained optimization. Nick Gould (RAL)

ab initio Electronic Structure Calculations

On the Superlinear Convergence of MINRES. Valeria Simoncini and Daniel B. Szyld. Report January 2012

Improving the performance of applied science numerical simulations: an application to Density Functional Theory

Contents. Preface for the Instructor. Preface for the Student. xvii. Acknowledgments. 1 Vector Spaces 1 1.A R n and C n 2

Iterative methods for symmetric eigenvalue problems

MATH 583A REVIEW SESSION #1

AMS526: Numerical Analysis I (Numerical Linear Algebra) Lecture 23: GMRES and Other Krylov Subspace Methods; Preconditioning

Iterative Methods for Sparse Linear Systems

Fast estimation of approximate matrix ranks using spectral densities

ECS231 Handout Subspace projection methods for Solving Large-Scale Eigenvalue Problems. Part I: Review of basic theory of eigenvalue problems

Inexact and Nonlinear Extensions of the FEAST Eigenvalue Algorithm

Iterative Methods for Linear Systems of Equations

A refined Lanczos method for computing eigenvalues and eigenvectors of unsymmetric matrices

On prescribing Ritz values and GMRES residual norms generated by Arnoldi processes

A knowledge-based approach to high-performance computing in ab initio simulations.

A short course on: Preconditioned Krylov subspace methods. Yousef Saad University of Minnesota Dept. of Computer Science and Engineering

Recent advances in approximation using Krylov subspaces. V. Simoncini. Dipartimento di Matematica, Università di Bologna.

Preconditioned Eigenvalue Solvers for electronic structure calculations. Andrew V. Knyazev. Householder Symposium XVI May 26, 2005

Conjugate-Gradient Eigenvalue Solvers in Computing Electronic Properties of Nanostructure Architectures

A Domain Decomposition Based Jacobi-Davidson Algorithm for Quantum Dot Simulation

Parallel Self-Consistent-Field Calculations via Chebyshev-Filtered Subspace Acceleration

EIGENVALUE PROBLEMS. Background on eigenvalues/ eigenvectors / decompositions. Perturbation analysis, condition numbers..

The Lanczos and conjugate gradient algorithms

Karhunen-Loève Approximation of Random Fields Using Hierarchical Matrix Techniques

Singular Value Decompsition

Dimension reduction methods: Algorithms and Applications Yousef Saad Department of Computer Science and Engineering University of Minnesota

COMPUTING PARTIAL SPECTRA WITH LEAST-SQUARES RATIONAL FILTERS

M.A. Botchev. September 5, 2014

Eigenvalue Problems Computation and Applications

A Tour of the Lanczos Algorithm and its Convergence Guarantees through the Decades

Deflation for inversion with multiple right-hand sides in QCD

Bounding the Spectrum of Large Hermitian Matrices

Notes on PCG for Sparse Linear Systems

The Conjugate Gradient Method

ECS130 Scientific Computing Handout E February 13, 2017

Department of Computer Science and Engineering

1 Conjugate gradients

Eigenvalues and Eigenvectors

1. Introduction. We consider the problem of solving the linear system. Ax = b, (1.1)

Conjugate-Gradient Eigenvalue Solvers in Computing Electronic Properties of Nanostructure Architectures

Computation of Smallest Eigenvalues using Spectral Schur Complements

Is there life after the Lanczos method? What is LOBPCG?

Conjugate Gradient Method

Multigrid absolute value preconditioning

Large-scale eigenvalue problems

A parameter tuning technique of a weighted Jacobi-type preconditioner and its application to supernova simulations

A CHEBYSHEV-DAVIDSON ALGORITHM FOR LARGE SYMMETRIC EIGENPROBLEMS

(a) If A is a 3 by 4 matrix, what does this tell us about its nullspace? Solution: dim N(A) 1, since rank(a) 3. Ax =

Key words. linear equations, polynomial preconditioning, nonsymmetric Lanczos, BiCGStab, IDR

ELSI: A Unified Software Interface for Kohn-Sham Electronic Structure Solvers

The following definition is fundamental.

Efficient Deflation for Communication-Avoiding Krylov Subspace Methods

4.8 Arnoldi Iteration, Krylov Subspaces and GMRES

State-of-the-art numerical solution of large Hermitian eigenvalue problems. Andreas Stathopoulos

Presentation of XLIFE++

Polynomial Filtering in Latent Semantic Indexing for Information Retrieval

Block Krylov Space Solvers: a Survey

Remark By definition, an eigenvector must be a nonzero vector, but eigenvalue could be zero.

Numerical Simulation of Spin Dynamics

GPU ACCELERATED POLYNOMIAL SPECTRAL TRANSFORMATION METHODS

Algorithms that use the Arnoldi Basis

A communication-avoiding thick-restart Lanczos method on a distributed-memory system

Iterative Methods for Solving A x = b

Linear Algebra and its Applications

Key words. conjugate gradients, normwise backward error, incremental norm estimation.

13-2 Text: 28-30; AB: 1.3.3, 3.2.3, 3.4.2, 3.5, 3.6.2; GvL Eigen2

An Estimator for the Diagonal of a Matrix

A Framework for Structured Linearizations of Matrix Polynomials in Various Bases

Linear Solvers. Andrew Hazel

Summary of Iterative Methods for Non-symmetric Linear Equations That Are Related to the Conjugate Gradient (CG) Method

Applied Linear Algebra in Geoscience Using MATLAB

FEM and sparse linear system solving

Transcription:

Polynomial filtering for interior eigenvalue problems Yousef Saad Department of Computer Science and Engineering University of Minnesota SIAM CSE Boston - March 1, 2013

First: Joint work with: Haw-ren Fang Grady Schoefield and Jim Chelikowsky [UT Austin] [windowing into PARSEC] Work supported in part by NSF (to 2012) and now by DOE 2 SIAM-CSE February 27, 2013

Introduction Q: A: How do you compute eigenvalues in the middle of the spectrum of a large Hermitian matrix? Method of choice: Shift and invert + some projection process (Lanczos, subspace iteration..) Main steps: 1) Select a shift (or sequence of shifts) σ; 2) Factor A σi: A σi = LDL T 3) Apply Lanczos algorithm to (A σi) 1 Solves with A σi carried out using factorization Limitation: factorization 3 SIAM-CSE February 27, 2013

Q: What if factoring A is too expensive (e.g., Large 3-D similation)? A: Obvious answer: Use iterative solvers... However: systems are highly indefinite Wont work too well. Digression: Still lots of work to do in iterative solution methods for highly indefinite linear systems 4 SIAM-CSE February 27, 2013

Other common characteristic: Need a very large number of eigenvalues and eigenvectors Applications: Excited states in quantum physics: TDDFT, GW,... Or just plain Density Functional Theory (DFT) Number of wanted eigenvectors is equal to number of occupied states [ == the number of valence electrons in DFT] An example: in real-space code (PARSEC), you can have a Hamiltonian of size a few Millions, and number of ev s in the tens of thousands. 5 SIAM-CSE February 27, 2013

Polynomial filtered Lanczos Possible solution: Use Lanczos with polynomial filtering. Idea not new (and not too popular in the past) 1. Very large problems; What is new? 2. (tens of) Thousands of eigenvalues; 3. Parallelism. Most important factor is 2. Main rationale of polynomial filtering : reduce the cost of orthogonalization Important application: compute the spectrum by pieces [ spectrum slicing a term coined by B. Parlett] 6 SIAM-CSE February 27, 2013

Introduction: What is filtered Lanczos? In short: we just replace (A σi) 1 in S.I. Lanczos by p k (A) where p k (t) = polynomial of degree k We want to compute eigenvalues near σ = 1 of a matrix A with Λ(A) [0, 4]. Use the simple transform: p 2 (t) = 1 α(t σ) 2. For α =.2, σ = 1 you get Use Lanczos with B = p 2 (A). 1 0.8 0.6 0.4 0.2 0 0.2 0.4 0.6 0.8 1 0 0.5 1 1.5 2 2.5 3 3.5 4 Eigenvalues near σ become the dominant ones so Lanczos will work but...... they are now poorly separated slow convergence. 7 SIAM-CSE February 27, 2013

A low-pass filter [a,b]=[0,3], [ξ,η]=[0,1], d=10 1 0.8 base filter ψ(λ) polynomial filter ρ(λ) γ 0.6 f(λ) 0.4 0.2 0-0.2 0 0.5 1 1.5 2 2.5 3 λ 8 SIAM-CSE February 27, 2013

A mid-pass filter [a,b]=[0,3], [ξ,η]=[0.5,1], d=20 1 0.8 base filter ψ(λ) polynomial filter ρ(λ) γ 0.6 f(λ) 0.4 0.2 0-0.2 0 0.5 1 1.5 2 2.5 3 λ 9 SIAM-CSE February 27, 2013

Misconception: High degree polynomials are bad d=1000 Degree 1000 (zoom) f(λ) 1 0.8 0.6 0.4 0.2 0-0.2 base filter ψ(λ) polynomial filter ρ(λ) γ -0.4-0.6 0.8 0.85 0.9 0.95 1 1.05 1.1 1.15 1.2 λ d=1600 Degree 1600 (zoom) f(λ) 1 0.8 0.6 0.4 0.2 0 base filter ψ(λ) polynomial filter ρ(λ) γ -0.2-0.4 0.8 0.85 0.9 0.95 1 1.05 1.1 1.15 1.2 10 SIAM-CSE February 27, 2013 λ

Hypothetical scenario: large A, zillions of wanted e-values Assume A has size 10M (Not huge by todays standard)... and you want to compute 50,000 eigenvalues/vectors (huge for numerical analysits, not for physicists)...... in the lower part of the spectrum - or the middle. By (any) standard methods you will need to orthogonalize at least 50K vectors of size 10M Space is an issue: 4 10 12 bytes = 4TB of mem *just for the basis* Orthogonalization is also an issue: 5 10 16 = 50 PetaOPS. Toward the end, at step k, each orthogonalization step costs about 4kn 200, 000n for k close to 50, 000. 11 SIAM-CSE February 27, 2013

The alternative: Spectrum slicing or windowing Rationale. Eigenvectors on both ends of wanted spectrum need not be orthogonalized against each other : Idea: Get the spectrum by slices or windows Can get a few hundreds or thousands of vectors at a time. 12 SIAM-CSE February 27, 2013

Compute eigenpairs one slice at a time Deceivingly simple looking idea. Issues: Deal with interfaces : duplicate/missing eigenvalues Window size [need estimate of eigenvalues] polynomial degree 13 SIAM-CSE February 27, 2013

Spectrum slicing in PARSEC Being implemented in our code: Pseudopotential Algorithm for Real-Space Electronic Calcultions (PARSEC) See : A Spectrum Slicing Method for the Kohn-Sham Problem, G. Schofield, J. R. Chelikowsky and YS, Computer Physics Comm., vol 183, pp. 487-505. Refer to this paper for details on windowing and initial proof of concept 14 SIAM-CSE February 27, 2013

Computing the polynomials: Jackson-Chebyshev Chebyshev-Jackson approximation of a function f: f(x) k i=0 g k i γ it i (x) γ i = 2 δ i0 π 1 1 1 1 x 2 f(x)dx δ i0 = Kronecker symbol The gi k s attenuate higher order terms in the sum. Attenuation coefficient gi k for 0.2 0.1 k=50,100,150 0 g k i 1 0.9 0.8 0.7 0.6 0.5 0.4 0.3 k=50 k=100 k=150 0 20 40 60 80 100 120 140 160 i (Degree) 15 SIAM-CSE February 27, 2013

Let α k = π k + 2, then : ( ) 1 i sin(α g k i = k+2 k ) cos(iα k ) + 1 k+2 cos(α k) sin(iα k ) sin(α k ) See Electronic structure calculations in plane-wave codes without diagonalization. Laurent O. Jay, Hanchul Kim, YS, and James R. Chelikowsky. Computer Physics Communications, 118:21 30, 1999. 16 SIAM-CSE February 27, 2013

The expansion coefficients γ i When f(x) is a step function on [a, b] [ 1 1]: γ i = 1 (arccos(a) arccos(b)) : i = 0 π ( ) 2 sin(i arccos(a)) sin(i arccos(b)) : i > 0 π i A few examples follow 17 SIAM-CSE February 27, 2013

Computing the polynomials: Jackson-Chebyshev Polynomials of degree 30 for [a, b] = [.3,.6] 1.2 1 Mid pass polynom. filter [ 1.3.6 1]; Degree = 30 Standard Cheb. Jackson Cheb. 0.8 0.6 0.4 0.2 0 0.2 1 0.8 0.6 0.4 0.2 0 0.2 0.4 0.6 0.8 1 18 SIAM-CSE February 27, 2013

1.2 1 Mid pass polynom. filter [ 1.3.6 1]; Degree = 80 Standard Cheb. Jackson Cheb. 0.8 0.6 0.4 0.2 0 0.2 1 0.8 0.6 0.4 0.2 0 0.2 0.4 0.6 0.8 1 19 SIAM-CSE February 27, 2013

1.2 1 Mid pass polynom. filter [ 1.3.6 1]; Degree = 200 Standard Cheb. Jackson Cheb. 0.8 0.6 0.4 0.2 0 0.2 1 0.8 0.6 0.4 0.2 0 0.2 0.4 0.6 0.8 1 20 SIAM-CSE February 27, 2013

How to get the polynomial filter? Second approach Idea: φ First select an ideal filter e.g., a piecewise polynomial function a b For example φ = Hermite interpolating pol. in [0,a], and φ = 1 in [a, b] Referred to as the Base filter 21 SIAM-CSE February 27, 2013

Then approximate base filter by degree k polynomial in a least-squares sense. Can do this without numerical integration a b φ Main advantage: Extremely flexible. Method: Build a sequence of polynomials φ k which approximate the ideal PP filter φ, in the L 2 sense. Again 2 implementations 22 SIAM-CSE February 27, 2013

Define φ k the least-squares polynomial approximation to φ: φ k (t) = k φ, P j P j (t), j=1 where {P j } is a basis of polynomials that is orthonormal for some L 2 inner-product. Method 1: Use Stieljes procedure to computing orthogonal polynomials

ALGORITHM : 1 Stieljes 1. P 0 0, 2. β 1 = S 0, 3. P 1 (t) = 1 β 1 S 0 (t), 4. For j = 2,..., m Do 5. α j = t P j, P j, 6. S j (t) = t P j (t) α j P j (t) β j P j 1 (t), 7. β j+1 = S j, 8. P j+1 (t) = 1 β j+1 S j (t). 9. EndDo 24 SIAM-CSE February 27, 2013

Computation of Stieljes coefficients Problem: To compute the scalars α j and β j+1, of 3-term recurrence (Stieljes) + the expansion coefficients γ j. Need to avoid numerical integration. Solution: define orthogonal polynomials over two (or more) disjoint intervals see similar work YS 83: YS, Iterative solution of indefinite symmetric systems by methods using orthogonal polynomials over two disjoint intervals, SIAM Journal on Numerical Analysis, 20 (1983), pp. 784 811. E. KOKIOPOULOU AND YS, Polynomial Filtering in Latent Semantic Indexing for Information Retrieval, in Proc. ACM-SIGIR Conference on research and development in information retrieval, Sheffield, UK, (2004) 25 SIAM-CSE February 27, 2013

Let large interval be [0, b] should contain Λ(B) Assume 2 subintervals. On subinterval [a l 1, a l ], l = 1, 2 define the inner-product ψ 1, ψ 2 al 1,a l by ψ 1, ψ 2 al 1,a l = al a l 1 ψ 1 (t)ψ 2 (t) (t al 1 )(a l t) dt. Then define the inner product on [0, b] by ψ 1, ψ 2 = a 0 ψ 1 (t)ψ 2 (t) t(a t) dt + ρ b a ψ 1 (t)ψ 2 (t) (t a)(b t) dt. To avoid numerical integration, use a basis of Chebyshev polynomials on interval [YS 83] 26 SIAM-CSE February 27, 2013

Mehod 2 : Filtered CG/CR - like polynomial iterations Want: a CG-like (or CR-like) algorithms for which the inderlying residual polynomial or solution polynomial are Least-squares filter polynomials Seek s to minimize φ(λ) λ s(λ) w with respect to a certain norm. w. Equivalently, minimize (1 φ) (1 λ s(λ)) w over all polynomials s of degree k. Focus on second view-point (residual polynomial) goal is to make r(λ) 1 λs(λ) close to 1 φ. 27 SIAM-CSE February 27, 2013

Recall: Conjugate Residual Algorithm ALGORITHM : 2 Conjugate Residual Algorithm 1. Compute r 0 := b Ax 0, p 0 := r 0 2. For j = 0, 1,..., until convergence Do: 3. α j := (r j, Ar j )/(Ap j, Ap j ) 4. x j+1 := x j + α j p j 5. r j+1 := r j α j Ap j 6. β j := (r j+1, Ar j+1 )/(r j, Ar j ) 7. p j+1 := r j+1 + β j p j 8. Compute Ap j+1 = Ar j+1 + β j Ap j 9. EndDo Think in terms of polynomial iteration 28 SIAM-CSE February 27, 2013

ALGORITHM : 3 Filtered CR polynomial Iteration 1. Compute r 0 := b Ax 0, p 0 := r 0 π 0 = ρ 0 = 1 1.a. Compute λπ 0 2. For j = 0, 1,..., until convergence Do: 3. α j :=< ρ j, λρ j > w / < λπ j, λπ j > w 3.a. α j := α j < 1 φ, λπ j > w / < λπ j, λπ j > w 4. x j+1 := x j + α j p j 5. r j+1 := r j α j Ap j ρ j+1 = ρ j α j λπ j 6. β j :=< ρ j+1, λρ j+1 > w / < ρ j, λρ j > w 7. p j+1 := r j+1 + β j p j π j+1 := ρ j+1 + β j π j 8. Compute λπ j+1 9. EndDo All polynomials expressed in Chebyshev basis - cost of algorithm is negligible [ O(k 2 ) for deg. k. ] 29 SIAM-CSE February 27, 2013

A few mid-pass filters of various degrees Four examples of middle-pass filters ψ(λ) and their polynomial approximations ρ(λ). Degrees 20 and 30 d=20 d=30 1 base filter ψ(λ) polynomial filter ρ(λ) γ 1 base filter ψ(λ) polynomial filter ρ(λ) γ 0.8 0.8 0.6 0.6 f(λ) 0.4 f(λ) 0.4 0.2 0.2 0 0-0.2-0.2-0.4-1.5-1 -0.5 0 0.5 1 1.5 2 2.5 3 λ -0.4-1.5-1 -0.5 0 0.5 1 1.5 2 2.5 3 λ 30 SIAM-CSE February 27, 2013

Degrees 50 and 100 d=50 d=100 1 0.8 base filter ψ(λ) polynomial filter ρ(λ) γ 1 0.8 base filter ψ(λ) polynomial filter ρ(λ) γ 0.6 0.6 f(λ) 0.4 f(λ) 0.4 0.2 0.2 0 0-0.2-1.5-1 -0.5 0 0.5 1 1.5 2 2.5 3 λ -0.2-1.5-1 -0.5 0 0.5 1 1.5 2 2.5 3 λ 31 SIAM-CSE February 27, 2013

Base Filter to build a Mid-Pass filter polynomial We partition [0, b] into five sub-intervals, [0, b] = [0, 1][τ 1, τ 2 ] [τ 2, τ 3 ] [τ 3, τ 4 ] [τ 4, b] 0 τ τ τ τ 1 2 3 4 b Set: ψ(t) = 0 in [0, τ 1 ] [τ 4, b] and ψ(t) = 1 in [τ 2, τ 3 ] Use standard Hermite interpolation to get brigde functions in [τ 1, τ 2 ] and [τ 3, τ 4 ] 32 SIAM-CSE February 27, 2013

References A Filtered Lanczos Procedure for Extreme and Interior Eigenvalue Problems, H. R. Fang and YS, SISC 34(4) A2220-2246 (2012). For details on window-less implementation (one slice) + code Computation of Large Invariant Subspaces Using Polynomial Filtered Lanczos Iterations with Applications in Density Functional Theory, C. Bekas and E. Kokiopoulou and YS, SIMAX 30(1), 397-418 (2008). Filtered Conjugate Residual-type Algorithms with Applications, YS; SIMAX 28 pp. 845-870 (2006) 33 SIAM-CSE February 27, 2013

Tests Test matrices Experiments performed in sequential mode: on two dualcore AMD Opteron(tm) Processors 2214 @ 2.2GHz and 16GB memory. Test matrices: * Five Hamiltonians from electronic structure calculations, * An integer matrix named Andrews, and * A discretized Laplacian (FD) 34 SIAM-CSE February 27, 2013

0 500 1000 1500 2000 2500 3000 3500 4000 4500 5000 0 500 1000 1500 2000 2500 3000 3500 4000 4500 5000 nz = 11584 35 SIAM-CSE February 27, 2013

Matrix characteristics matrix n nnz nnz n full eigen-range Fermi [a, b] n 0 GE87H76 112,985 7,892,195 69.85 [ 1.2140, 32.764] 212 Ge99H100 112,985 8,451,395 74.80 [ 1.2264, 32.703] 248 SI41Ge41H72 185,639 15,011,265 80.86 [ 1.2135, 49.818] 200 Si87H76 240,369 10,661,631 44.36 [ 1.1963, 43.074] 212 Ga41As41H72 268,096 18,488,476 68.96 [ 1.2501, 1300.9] 200 Andrews 60,000 760,154 12.67 [0, 36.485] N/A Laplacian 1,000,000 6,940,000 6.94 [0.00290, 11.997] N/A 36 SIAM-CSE February 27, 2013

Experimental set-up eigen-interval # eig # eig matrix [ξ, η] in [ξ, η] in [a, η] η ξ b a η a b a GE87H76 [ 0.645, 0.0053] 212 318 0.0188 0.0356 Ge99H100 [ 0.65, 0.0096] 250 372 0.0189 0.0359 SI41Ge41H72 [ 0.64, 0.00282] 218 318 0.0125 0.0237 Si87H76 [ 0.66, 0.33] 212 317 0.0075 0.0196 Ga41As41H72 [ 0.64, 0.0] 201 301 0.0005 0.0010 Andrews [4, 5] 1,844 3,751 0.0274 0.1370 Laplacian [1, 1.01] 276 >17,000 0.0008 0.0044 37 SIAM-CSE February 27, 2013

Results for Ge99H100 -set 1 of stats method degree # iter # matvecs memory d = 20 1,020 20,400 1,117 filt. Lan. d = 30 710 21,300 806 (high-pass) d = 50 470 23,500 508 d = 100 340 34,000 440 d = 10 770 7,700 806 filt. Lan. d = 20 600 12,000 688 (low-pass) d = 30 530 15,900 590 d = 50 470 23,500 508 Part. Lanczos 5,140 5,140 4,883 ARPACK 6,233 6,233 1,073 38 SIAM-CSE February 27, 2013

Results for Ge99H100 -CPU times (sec.) method degree ρ(a)v reorth eigvec total d = 20 1,283 77 23 1,417 filt. Lan. d = 30 1,343 55 14 1,440 (high-pass) d = 50 1,411 32 9 1,479 d = 100 1,866 26 7 1,930 d = 10 483 124 21 668 filt. Lan. d = 20 663 57 21 777 (low-pass) d = 30 1,017 49 15 1,123 d = 50 1,254 26 13 1,342 Part. Lanczos 234 1,460 793 2,962 ARPACK 298 17,503 666 18,468 39 SIAM-CSE February 27, 2013

Results for Andrews - set 1 of stats method degree # iter # matvecs memory d = 20 9,440 188,800 4,829 filt. Lan. d = 30 6,040 180,120 2,799 (mid-pass) d = 50 3,800 190,000 1,947 d = 100 2,360 236,000 1,131 d = 10 5,990 59,900 2,799 filt. Lan. d = 20 4,780 95,600 2,334 (high-pass) d = 30 4,360 130,800 2,334 d = 50 4,690 234,500 2,334 Part. Lanczos 22,345 22,345 10,312 ARPACK 30,716 30,716 6,129 40 SIAM-CSE February 27, 2013

Results for Andrews - CPU times (sec.) method degree ρ(a)v reorth eigvec total d = 20 2,797 192 4,834 9,840 filt. Lan. d = 30 2,429 115 2,151 5,279 (mid-pass) d = 50 3,040 65 521 3,810 d = 100 3,757 93 220 4,147 d = 10 1,152 2,911 2,391 7,050 filt. Lan. d = 20 1,335 1,718 1,472 4,874 (high-pass) d = 30 1,806 1,218 1,274 4,576 d = 50 3,187 1,032 1,383 5,918 Part. Lanczos 217 30,455 64,223 112,664 ARPACK 345 423,492 18,094 441,934 41 SIAM-CSE February 27, 2013

Results for Laplacian set 1 of stats method degree # iter # matvecs memory 600 1,400 840,000 10,913 mid-pass filter 1, 000 950 950,000 7,640 1, 600 710 1,136,000 6,358 42 SIAM-CSE February 27, 2013

Results for Laplacian CPU times method degree ρ(a)v reorth eigvec total 600 97,817 927 241 99,279 mid-pass filter 1, 000 119,242 773 162 120,384 1, 600 169,741 722 119 170,856 43 SIAM-CSE February 27, 2013

Conclusion Quite appealing general approach when number of eigenvectors to be computed is large and when Matvec is not too expensise Will not work too well for generalized eigenvalue problem Code available here www.cs.umn.edu/~saad/software/filtlan 44 SIAM-CSE February 27, 2013