A Tuned Preconditioner for Inexact Inverse Iteration Applied to Hermitian Eigenvalue Problems

Similar documents
A Tuned Preconditioner for Inexact Inverse Iteration for Generalised Eigenvalue Problems

Lecture 3: Inexact inverse iteration with preconditioning

Preconditioned inverse iteration and shift-invert Arnoldi method

Inexact inverse iteration with preconditioning

Inexact Inverse Iteration for Symmetric Matrices

Inexact Inverse Iteration for Symmetric Matrices

Inexact inverse iteration for symmetric matrices

RAYLEIGH QUOTIENT ITERATION AND SIMPLIFIED JACOBI-DAVIDSON WITH PRECONDITIONED ITERATIVE SOLVES FOR GENERALISED EIGENVALUE PROBLEMS

c 2006 Society for Industrial and Applied Mathematics

Inner-outer Iterative Methods for Eigenvalue Problems - Convergence and Preconditioning

Recent advances in approximation using Krylov subspaces. V. Simoncini. Dipartimento di Matematica, Università di Bologna.

AN INEXACT INVERSE ITERATION FOR COMPUTING THE SMALLEST EIGENVALUE OF AN IRREDUCIBLE M-MATRIX. 1. Introduction. We consider the eigenvalue problem

Linear algebra issues in Interior Point methods for bound-constrained least-squares problems

ABSTRACT NUMERICAL SOLUTION OF EIGENVALUE PROBLEMS WITH SPECTRAL TRANSFORMATIONS

Conjugate Gradient Method

Alternative correction equations in the Jacobi-Davidson method

ITERATIVE PROJECTION METHODS FOR SPARSE LINEAR SYSTEMS AND EIGENPROBLEMS CHAPTER 11 : JACOBI DAVIDSON METHOD

Is there life after the Lanczos method? What is LOBPCG?

The Conjugate Gradient Method

1. Introduction. In this paper we consider the large and sparse eigenvalue problem. Ax = λx (1.1) T (λ)x = 0 (1.2)

A Model-Trust-Region Framework for Symmetric Generalized Eigenvalue Problems

Model order reduction of large-scale dynamical systems with Jacobi-Davidson style eigensolvers

Davidson Method CHAPTER 3 : JACOBI DAVIDSON METHOD

Solution of eigenvalue problems. Subspace iteration, The symmetric Lanczos algorithm. Harmonic Ritz values, Jacobi-Davidson s method

Controlling inner iterations in the Jacobi Davidson method

Eigenvalue Problems CHAPTER 1 : PRELIMINARIES

The Nullspace free eigenvalue problem and the inexact Shift and invert Lanczos method. V. Simoncini. Dipartimento di Matematica, Università di Bologna

MICHIEL E. HOCHSTENBACH

A robust multilevel approximate inverse preconditioner for symmetric positive definite matrices

CONTROLLING INNER ITERATIONS IN THE JACOBI DAVIDSON METHOD

A Chebyshev-based two-stage iterative method as an alternative to the direct solution of linear systems

Polynomial Jacobi Davidson Method for Large/Sparse Eigenvalue Problems

7.2 Steepest Descent and Preconditioning

CONTROLLING INNER ITERATIONS IN THE JACOBI DAVIDSON METHOD

THEORETICAL AND NUMERICAL COMPARISON OF PROJECTION METHODS DERIVED FROM DEFLATION, DOMAIN DECOMPOSITION AND MULTIGRID METHODS

ETNA Kent State University

AMS526: Numerical Analysis I (Numerical Linear Algebra) Lecture 23: GMRES and Other Krylov Subspace Methods; Preconditioning

Solving Regularized Total Least Squares Problems

Notes on PCG for Sparse Linear Systems

Computational Methods. Eigenvalues and Singular Values

A Note on Inverse Iteration

c 2009 Society for Industrial and Applied Mathematics

Matrix Algorithms. Volume II: Eigensystems. G. W. Stewart H1HJ1L. University of Maryland College Park, Maryland

COMP 558 lecture 18 Nov. 15, 2010

Low Rank Approximation Lecture 7. Daniel Kressner Chair for Numerical Algorithms and HPC Institute of Mathematics, EPFL

A Jacobi Davidson type method for the product eigenvalue problem

6.4 Krylov Subspaces and Conjugate Gradients

Lecture Note 7: Iterative methods for solving linear systems. Xiaoqun Zhang Shanghai Jiao Tong University

HOMOGENEOUS JACOBI DAVIDSON. 1. Introduction. We study a homogeneous Jacobi Davidson variant for the polynomial eigenproblem

Applied Mathematics 205. Unit V: Eigenvalue Problems. Lecturer: Dr. David Knezevic

The Eigenvalue Problem: Perturbation Theory

Finding Rightmost Eigenvalues of Large, Sparse, Nonsymmetric Parameterized Eigenvalue Problems

DELFT UNIVERSITY OF TECHNOLOGY

Linear Systems. Class 27. c 2008 Ron Buckmire. TITLE Projection Matrices and Orthogonal Diagonalization CURRENT READING Poole 5.4

Conjugate gradient method. Descent method. Conjugate search direction. Conjugate Gradient Algorithm (294)

Matrix Algebra: Summary

c 2009 Society for Industrial and Applied Mathematics

SOLVING MESH EIGENPROBLEMS WITH MULTIGRID EFFICIENCY

Linear Algebra and its Applications

Numerical behavior of inexact linear solvers

QR-decomposition. The QR-decomposition of an n k matrix A, k n, is an n n unitary matrix Q and an n k upper triangular matrix R for which A = QR

U AUc = θc. n u = γ i x i,

Lecture Note 13: Eigenvalue Problem for Symmetric Matrices

Universiteit-Utrecht. Department. of Mathematics. The convergence of Jacobi-Davidson for. Hermitian eigenproblems. Jasper van den Eshof.

On the accuracy of saddle point solvers

LINEAR ALGEBRA BOOT CAMP WEEK 4: THE SPECTRAL THEOREM

Lecture 9: Krylov Subspace Methods. 2 Derivation of the Conjugate Gradient Algorithm

A Jacobi Davidson type method for the generalized singular value problem

Using the Karush-Kuhn-Tucker Conditions to Analyze the Convergence Rate of Preconditioned Eigenvalue Solvers

Domain decomposition on different levels of the Jacobi-Davidson method

Computation. For QDA we need to calculate: Lets first consider the case that

Topics. The CG Algorithm Algorithmic Options CG s Two Main Convergence Theorems

Iterative methods for symmetric eigenvalue problems

A Newton-Galerkin-ADI Method for Large-Scale Algebraic Riccati Equations

A Jacobi Davidson Method for Nonlinear Eigenproblems

Symmetric matrices and dot products

Invariant Subspace Computation - A Geometric Approach

Point Distribution Models

An Arnoldi Method for Nonlinear Symmetric Eigenvalue Problems

Numerical Methods I: Eigenvalues and eigenvectors

HARMONIC RAYLEIGH RITZ EXTRACTION FOR THE MULTIPARAMETER EIGENVALUE PROBLEM

Lecture 2: Linear Algebra Review

Incomplete Cholesky preconditioners that exploit the low-rank property

ECS231 Handout Subspace projection methods for Solving Large-Scale Eigenvalue Problems. Part I: Review of basic theory of eigenvalue problems

Structured Preconditioners for Saddle Point Problems

Solution of eigenvalue problems. Subspace iteration, The symmetric Lanczos algorithm. Harmonic Ritz values, Jacobi-Davidson s method

Contents. Preface... xi. Introduction...

STEEPEST DESCENT AND CONJUGATE GRADIENT METHODS WITH VARIABLE PRECONDITIONING

Linear Solvers. Andrew Hazel

DELFT UNIVERSITY OF TECHNOLOGY

Indefinite Preconditioners for PDE-constrained optimization problems. V. Simoncini

Conjugate Gradient (CG) Method

Preconditioning Techniques Analysis for CG Method

of dimension n 1 n 2, one defines the matrix determinants

Scientific Computing with Case Studies SIAM Press, Lecture Notes for Unit VII Sparse Matrix

Numerical Methods for Large Scale Eigenvalue Problems

Numerical Methods - Numerical Linear Algebra

On Lagrange multipliers of trust-region subproblems

Nonlinear Eigenvalue Problems: An Introduction

ABSTRACT OF DISSERTATION. Ping Zhang

Transcription:

A Tuned Preconditioner for Applied to Eigenvalue Problems Department of Mathematical Sciences University of Bath, United Kingdom IWASEP VI May 22-25, 2006 Pennsylvania State University, University Park Joint work with: Alastair Spence

1 2 3 4

1 2 3 4

Problem and Inverse Find an eigenvalue and eigenvector of a positive definite A: Ax = λx, Inverse : A large, sparse. (A σi)y = x Inverse iteration with preconditioned iterative

1 2 3 4

for i = 1 to i max do choose τ (i), σ (i) solve (A σ (i) I)y (i) x (i) τ (i), Rescale x (i+1) = y(i) y (i), Update λ (i+1) = x (i+1)t Ax (i+1), possibly: update the shift σ (i) Test: eigenvalue residual r (i+1) = (A λ (i+1) I)x (i+1). end for

Error indicator Error indicator (Orthogonal decomposition for symmetric A, Parlett) Eigenvalue residual x (i) = cos θ (i) x 1 + sin θ (i) x (i), x(i) x 1. sin θ (i) λ 2 λ (i) r (i) sin θ (i) λ n λ 1

rates of inexact inverse iteration Decreasing tolerance τ (i) = C r (i) = O(sin θ (i) ) 1 For decreasing tolerance τ (i) C r (i) = O(sin θ (i) ) the inexact method recovers the rate of convergence achieved by exact. 2 Fixed shift σ: linear convergence. 3 Rayleigh quotient shift σ (i) = ρ(x (i) ) = x(i)t Ax (i) x (i)t x : cubic (i) convergence for A = A. Fixed tolerance τ (i) = τ 1 Rayleigh quotient shift: quadratic convergence

MINRES (A σi)y = x when A is symmetric Solving a linear system (A σi)y = x standard MINRES for y 0 = 0: x (A σi)y k 2 where κ is the condition number of A σi. Number of inner iterations: then k 1 + κ ( ) k 1 κ 1 x. κ + 1 { log 2 + log x } τ x (A σi)y k τ.

Unpreconditioned with MINRES rates for with MINRES for simple eigenvalue If A is positive definite and has a simple eigenvalue then x (i) (A σi)y (i) k 2 2 λ 1 λ n λ 1 σ ( )k 2 κ1 1 Qx (i) 2. κ 1 + 1 where Q is the orthogonal projection onto span{x 2,..., x n } and κ 1 is the reduced condition number κ 1 = max i=2,...,n λ i σ min i=2,...,n λ i σ. Number of inner for each i for x (i) (A σ (i) I)y (i) τ (i) ( ) k (i) 2 + κ 1 log 2 λ 1 λ n + log Qx(i) 2 λ 1 σ τ (i)

Unpreconditioned with MINRES rates for with MINRES for simple eigenvalue If A is positive definite and has a simple eigenvalue then x (i) (A σi)y (i) k 2 2 λ 1 λ n λ 1 σ ( )k 2 κ1 1 sin θ (i). κ 1 + 1 where Q is the orthogonal projection onto span{x 2,..., x n } and κ 1 is the reduced condition number κ 1 = max i=2,...,n λ i σ min i=2,...,n λ i σ. Number of inner for each i for x (i) (A σ (i) I)y (i) τ (i) ( k (i) 2 + κ 1 log 2 λ 1 λ n + log sin ) θ(i) λ 1 σ τ (i)

1 2 3 4

Incomplete Cholesky A = LL T + E symmetric of (A σi)y (i) = x (i) : Remarks L 1 (A σi)l T ỹ (i) = L 1 x (i), 1 changes number of inner iterations k (i) 2 + κ 1 (log 2 λ 1 λ n + log 2 k (i) increases with i for τ (i) = C r (i). y (i) = L T ỹ (i) L 1 ) λ 1 σ τ (i)

Incomplete Cholesky A = LL T + E symmetric of (A σi)y (i) = x (i) : Remarks L 1 (A σi)l T ỹ (i) = L 1 x (i), 1 changes number of inner iterations k (i) 2 + κ 1 (log 2 λ 1 λ n + log 2 k (i) increases with i for τ (i) = C r (i). y (i) = L T ỹ (i) L 1 ) λ 1 σ τ (i)

Derivation Aims 1 modify L L L 1 (A σi)l T ỹ (i) = L 1 x (i), 2 minor extra computation cost for L y (i) = L T ỹ (i) 3 nice right hand side L 1 x (i) (same behaviour as unpreconditioned, e.g. for fixed shifts k (i) does not increase with i)

Derivation Aims 1 modify L L L 1 (A σi)l T ỹ (i) = L 1 x (i), 2 minor extra computation cost for L y (i) = L T ỹ (i) 3 nice right hand side L 1 x (i) (same behaviour as unpreconditioned, e.g. for fixed shifts k (i) does not increase with i)

Choice of L Condition MINRES indicates that L 1 x (i) should be close to eigenvector of L 1 (A σi)l T Holds if Justification of LL T x (i) = Ax (i) If x (i) = x 1 then LL T x 1 = λ 1 x 1 LL T x (i) = Ax (i) L 1 (A σi)l T L 1 x 1 = λ 1 σ λ 1 L 1 x 1 L 1 (A σi)l T L 1 x (i) = λ 1 σ L 1 x (i) + C r (i) λ 1

Choice of L Condition MINRES indicates that L 1 x (i) should be close to eigenvector of L 1 (A σi)l T Holds if Justification of LL T x (i) = Ax (i) If x (i) = x 1 then LL T x 1 = λ 1 x 1 LL T x (i) = Ax (i) L 1 (A σi)l T L 1 x 1 = λ 1 σ λ 1 L 1 x 1 L 1 (A σi)l T L 1 x (i) = λ 1 σ L 1 x (i) + C r (i) λ 1

How do we achieve LL T x (i) = Ax (i)? Theorem Let x (i) current eigenvector approximation, e (i) = Ax (i) LL T x (i) (known) and L chosen such that L = L + α (i) e (i) (L 1 e (i) ) T with α (i) root of quadratic function we get LL T x (i) = Ax (i). Implementation 1 Note: LL T = LL T 1 + e (i)t x (i) e(i) e (i)t 2 L is a rank-one update of L.

How do we achieve LL T x (i) = Ax (i)? Theorem Let x (i) current eigenvector approximation, e (i) = Ax (i) LL T x (i) (known) and L chosen such that L = L + α (i) e (i) (L 1 e (i) ) T with α (i) root of quadratic function we get LL T x (i) = Ax (i). Implementation 1 Note: LL T = LL T 1 + e (i)t x (i) e(i) e (i)t 2 L is a rank-one update of L.

Implementation General positive definite For MINRES implementation only the evaluation of P 1 is necessary Sherman-Morrison formula where z (i) = P 1 Ax (i). P = P + γ (i) e (i) e (i)t P 1 = P 1 (z(i) x (i) )(z (i) x (i) ) T (z (i) x (i) ) T Ax (i)

rates The 1 outer convergence rate is retained 2 cheap inner are provided ( sin θ k (i) (i) ) C 1 + C 2 log λ 1 σ τ (i) 3 only a single extra back substitution with P = LL T per outer iteration needed

Example SPD matrix from the Matrix Market library (nos5: 3 story building with attached tower) seek eigenvalue near fixed shift σ = 100 A LL T, incomplete Cholesky factorisation (drop tol. = 0.1) compare standard and

Fixed shift with standard incomplete Cholesky total number of inner iterations using standard : 2026 total number of inner iterations using : 779

Comparison of LL T with LL T Spectral properties of preconditioned matrix Let Theorem If σ / Λ(A) then µ, ξ 0 and L 1 (A σi)l T w = µw L 1 (A σi)l T ŵ = ξŵ min µ Λ(L 1 (A σi)l T ) where γ = 1/(e T x) and v = L 1 e. µ ξ ξ γv v,

Comparison of LL T with LL T Spectral properties of preconditioned matrix Let L 1 (A σi)l T w = µw L 1 (A σi)l T ŵ = ξŵ Interlacing property Rewrite second equation Dt = ξ(i + γzz T )t where L 1 (A σi)l T = QDQ T, z = Q T v, (I + αvv T )Qt = ŵ. Interlacing property If γ > 0 eigenvalues are moved towards the origin. If γ < 0 eigenvalues are moved away from the origin.

Comparison of LL T with LL T Spectral properties of preconditioned matrix Let L 1 (A σi)l T w = µw L 1 (A σi)l T ŵ = ξŵ Interlacing property Rewrite second equation Dt = ξ(i + γzz T )t where L 1 (A σi)l T = QDQ T, z = Q T v, (I + αvv T )Qt = ŵ. Interlacing property If γ > 0 eigenvalues are moved towards the origin. If γ < 0 eigenvalues are moved away from the origin.

Comparison of LL T with LL T Interlacing property µ and ξ interlace each other depending on the sign of γ Clustering properties are preserved reduced condition number κ 1 L κ1 L κ1 L (1 + γvt v)

Comparison of LL T with LL T Interlacing property µ and ξ interlace each other depending on the sign of γ Clustering properties are preserved reduced condition number κ 1 L κ1 L κ1 L (1 + γvt v)

Changing the right hand side Approach by Simoncini/Eldén [3] Instead of solving L 1 (A σ (i) I)L T ỹ (i) = L 1 x (i), change the right hand side L 1 (A σ (i) I)L T ỹ (i) = L T x (i), y (i) = L T ỹ (i) y (i) = L T ỹ (i)

Comparison Tuned and Simoncini & Eldén Example nos5.mtx from Matrix Market. Solves to fixed tolerance τ = 0.01. Rayleigh quotient shift. Quadratic convergence for both methods. Simoncini & Eldén Tuned Drop Tolerances Outer 0.25 0.1 0.25 0.1 1 67 62 29 26 2 74 66 56 55 3 85 75 71 67 4 63 18 total 289 203 174 148

1 2 3 4

example Ax = λm x with bcsstk08 (Structural engineering) Figure: Fixed Shift Figure: Rayleigh Quotient Shift

example for the eigenproblem Ax = λm x with bcsstk08 (Structural engineering) Figure: Fixed Shift Figure: Rayleigh Quotient Shift

M. A. Freitag and A. Spence, rates for inexact inverse iteration with application to preconditioned iterative, 2006. Submitted to BIT., A for inexact inverse iteration applied to eigenvalue problems, 2006. Submitted to IMA J. Numer. Anal. V. Simoncini and L. Eldén, Inexact Rayleigh quotient-type methods for eigenvalue computations, BIT, 42 (2002), pp. 159 182.