AMS526: Numerical Analysis I (Numerical Linear Algebra)

Similar documents
AMS526: Numerical Analysis I (Numerical Linear Algebra) Lecture 23: GMRES and Other Krylov Subspace Methods; Preconditioning

Krylov Space Methods. Nonstationary sounds good. Radu Trîmbiţaş ( Babeş-Bolyai University) Krylov Space Methods 1 / 17

AMS526: Numerical Analysis I (Numerical Linear Algebra)

4.8 Arnoldi Iteration, Krylov Subspaces and GMRES

Summary of Iterative Methods for Non-symmetric Linear Equations That Are Related to the Conjugate Gradient (CG) Method

AMS526: Numerical Analysis I (Numerical Linear Algebra)

FEM and sparse linear system solving

Course Notes: Week 1

Iterative Methods for Linear Systems of Equations

Algorithms that use the Arnoldi Basis

Iterative Methods for Sparse Linear Systems

Iterative methods for Linear System of Equations. Joint Advanced Student School (JASS-2009)

Last Time. Social Network Graphs Betweenness. Graph Laplacian. Girvan-Newman Algorithm. Spectral Bisection

AMS526: Numerical Analysis I (Numerical Linear Algebra)

Computation of eigenvalues and singular values Recall that your solutions to these questions will not be collected or evaluated.

Krylov Subspace Methods that Are Based on the Minimization of the Residual

6.4 Krylov Subspaces and Conjugate Gradients

AMS526: Numerical Analysis I (Numerical Linear Algebra)

Iterative methods for Linear System

Computational Linear Algebra

Math 504 (Fall 2011) 1. (*) Consider the matrices

Chapter 7 Iterative Techniques in Matrix Algebra

Computational Linear Algebra

A short course on: Preconditioned Krylov subspace methods. Yousef Saad University of Minnesota Dept. of Computer Science and Engineering

AMS526: Numerical Analysis I (Numerical Linear Algebra for Computational and Data Sciences)

1 Extrapolation: A Hint of Things to Come

Contents. Preface... xi. Introduction...

SOLVING SPARSE LINEAR SYSTEMS OF EQUATIONS. Chao Yang Computational Research Division Lawrence Berkeley National Laboratory Berkeley, CA, USA

AMS526: Numerical Analysis I (Numerical Linear Algebra)

On the influence of eigenvalues on Bi-CG residual norms

Applied Mathematics 205. Unit V: Eigenvalue Problems. Lecturer: Dr. David Knezevic

AMS526: Numerical Analysis I (Numerical Linear Algebra)

CME342 Parallel Methods in Numerical Analysis. Matrix Computation: Iterative Methods II. Sparse Matrix-vector Multiplication.

ITERATIVE METHODS BASED ON KRYLOV SUBSPACES

AMS526: Numerical Analysis I (Numerical Linear Algebra)

Institute for Advanced Computer Studies. Department of Computer Science. Iterative methods for solving Ax = b. GMRES/FOM versus QMR/BiCG

Charles University Faculty of Mathematics and Physics DOCTORAL THESIS. Krylov subspace approximations in linear algebraic problems

Conjugate gradient method. Descent method. Conjugate search direction. Conjugate Gradient Algorithm (294)

AMS526: Numerical Analysis I (Numerical Linear Algebra)

Eigenvalue and Eigenvector Problems

Lecture 17 Methods for System of Linear Equations: Part 2. Songting Luo. Department of Mathematics Iowa State University

Preconditioned inverse iteration and shift-invert Arnoldi method

DELFT UNIVERSITY OF TECHNOLOGY

The Lanczos and conjugate gradient algorithms

AMS526: Numerical Analysis I (Numerical Linear Algebra for Computational and Data Sciences)

Introduction. Chapter One

RESIDUAL SMOOTHING AND PEAK/PLATEAU BEHAVIOR IN KRYLOV SUBSPACE METHODS

ON ORTHOGONAL REDUCTION TO HESSENBERG FORM WITH SMALL BANDWIDTH

Krylov Space Solvers

Bindel, Fall 2016 Matrix Computations (CS 6210) Notes for

A DISSERTATION. Extensions of the Conjugate Residual Method. by Tomohiro Sogabe. Presented to

Topics. The CG Algorithm Algorithmic Options CG s Two Main Convergence Theorems

GMRES: Generalized Minimal Residual Algorithm for Solving Nonsymmetric Linear Systems

M.A. Botchev. September 5, 2014

Course Notes: Week 4

Lecture 8 Fast Iterative Methods for linear systems of equations

The amount of work to construct each new guess from the previous one should be a small multiple of the number of nonzeros in A.

Parallel Numerics, WT 2016/ Iterative Methods for Sparse Linear Systems of Equations. page 1 of 1

Numerical Methods in Matrix Computations

Numerical Methods I: Eigenvalues and eigenvectors

Simple iteration procedure

Preface to the Second Edition. Preface to the First Edition

Solving Symmetric Indefinite Systems with Symmetric Positive Definite Preconditioners

Index. for generalized eigenvalue problem, butterfly form, 211

IDR(s) as a projection method

Lecture 18 Classical Iterative Methods

Multigrid absolute value preconditioning

MATH 304 Linear Algebra Lecture 20: The Gram-Schmidt process (continued). Eigenvalues and eigenvectors.

IDR(s) Master s thesis Goushani Kisoensingh. Supervisor: Gerard L.G. Sleijpen Department of Mathematics Universiteit Utrecht

The quadratic eigenvalue problem (QEP) is to find scalars λ and nonzero vectors u satisfying

On prescribing Ritz values and GMRES residual norms generated by Arnoldi processes

AMS526: Numerical Analysis I (Numerical Linear Algebra)

ITERATIVE METHODS FOR SPARSE LINEAR SYSTEMS

AMS526: Numerical Analysis I (Numerical Linear Algebra)

Steady-State Optimization Lecture 1: A Brief Review on Numerical Linear Algebra Methods

Lecture 11: CMSC 878R/AMSC698R. Iterative Methods An introduction. Outline. Inverse, LU decomposition, Cholesky, SVD, etc.

Chapter 7. Iterative methods for large sparse linear systems. 7.1 Sparse matrix algebra. Large sparse matrices

Arnoldi Methods in SLEPc

Solving Sparse Linear Systems: Iterative methods

Solving Sparse Linear Systems: Iterative methods

Lab 1: Iterative Methods for Solving Linear Systems

5.3 The Power Method Approximation of the Eigenvalue of Largest Module

AMS526: Numerical Analysis I (Numerical Linear Algebra)

AMS526: Numerical Analysis I (Numerical Linear Algebra for Computational and Data Sciences)

ECS130 Scientific Computing Handout E February 13, 2017

Krylov subspace methods for linear systems with tensor product structure

Matrices, Moments and Quadrature, cont d

Research Article Some Generalizations and Modifications of Iterative Methods for Solving Large Sparse Symmetric Indefinite Linear Systems

Conjugate Gradient algorithm. Storage: fixed, independent of number of steps.

KRYLOV SUBSPACE ITERATION

Krylov subspace projection methods

Numerical Methods I Eigenvalue Problems

Research Article MINRES Seed Projection Methods for Solving Symmetric Linear Systems with Multiple Right-Hand Sides

PROJECTED GMRES AND ITS VARIANTS

Eigenvalues and eigenvectors

Introduction to Arnoldi method

Matrix Algorithms. Volume II: Eigensystems. G. W. Stewart H1HJ1L. University of Maryland College Park, Maryland

SOLVING HERMITIAN POSITIVE DEFINITE SYSTEMS USING INDEFINITE INCOMPLETE FACTORIZATIONS

Numerical Methods for Solving Large Scale Eigenvalue Problems

Numerical Methods I Non-Square and Sparse Linear Systems

Transcription:

AMS526: Numerical Analysis I (Numerical Linear Algebra) Lecture 23: GMRES and Other Krylov Subspace Methods Xiangmin Jiao SUNY Stony Brook Xiangmin Jiao Numerical Analysis I 1 / 9

Minimizing Residual CG only works for SPD matrices. It minimizes x x A and φ(x) = 1 2 x T Ax x T b There have been many proposed extensions to nonsymmetric matrices, GMRES, BiCG, etc. GMRES (Generalized Minimal RESiduals) is one of most well known The basic idea of GMRES is to find x n K n that minimizes r n = b Ax n This can be viewed as a least squares problem: Find a vector c s.t. AK n c b is minimized, where K n is the m n Krylov matrix composed of basis vectors of K n, and AK n = Ab A 2 b A n b Orthogonal basis is often used, produced by Arnoldi iteration Xiangmin Jiao Numerical Analysis I 2 / 9

Review: Arnoldi Iteration Let Q n = [q 1 q 2 q n ] be m n matrix with first n columns of Q and H n be (n + 1) n upper-left section of H Start by picking a random q 1 and then determine q 2 and H 1 The nth columns of AQ n = Q n+1 H n can be written as Aq n = h 1n q 1 + + h nn q n + h n+1,n q n+1 Algorithm: Arnoldi Iteration given random nonzero b, let q 1 = b/ b for n = 1 to 1, 2, 3,... v = Aq n for j = 1 to n h jn = q j v v = v h jn q j h n+1,n = v q n+1 = v/h n+1,n Xiangmin Jiao Numerical Analysis I 3 / 9

Minimal Residual with Orthogonal Basis Let Q n be Krylov matrix whose columns q 1, q 2,... span the successive Krylov subspaces Instead of find x n = K n c, find x n = Q n y which minimizes AQ n y b For Arnoldi iteration, we showed that AQ n = Q n+1 H n, so Q n+1 H n y b = minimum Left multiplication by Q n+1 does not change the norm, so H n y Q n+1b = minimum Finally, by construction, Q n+1b = b e 1, so H n y b e 1 = minimum. Xiangmin Jiao Numerical Analysis I 4 / 9

The GMRES Algorithm Algorithm: GMRES q 1 = b/ b for n = 1 to 1, 2, 3,... Step n of Arnoldi iteration Find y to minimize H n y b e 1 = r n x n = Q n y The residual r n does not need to be computed explicitly from x n Xiangmin Jiao Numerical Analysis I 5 / 9

The GMRES Algorithm Least squares problem has Hessenberg structure, solved with QR factorization of H n If QR factorization of H n is constructed from scratch, then it costs O(n 2 ) flops, due to Hessenberg structure However, QR factorization of H n can be updated from that of H n 1, using Givens rotation within O(n) work However, memory and cost grow with n. In practice, restart the algorithm by clearing accumulated data. This might stagnate the method Xiangmin Jiao Numerical Analysis I 6 / 9

GMRES and Polynomial Approximation GMRES can be interpreted as finding polynomial p n P n for n = 1, 2, 3,... where P n = {polynomial p of degree n with p(0) = 1} such that p n (A)b is minimized Note that r = b AK n c, where AK n c = ( c 1 A + c 2 A 2 + + c n A n) b and r = ( 1 c 1 A c 2 A 2 c n A n) b. In other words, p n (z) = 1 z(c 1 + c 2 z + + c n z n ) Invariance of GMRES Scale invariance: If we change A σa and b σb, then r n σr n Invariance under unitary similarity transformations: If change A UAU for some unitary matrix U and b Ub, then r n U r n Xiangmin Jiao Numerical Analysis I 7 / 9

Convergence of GMRES GMRES converges monotonically and it converges after at most m steps Based on a polynomial analysis, diagonalizable A = V ΛV 1 converges as r n κ(v ) inf sup p n (λ i ) b p n P n λ i Λ(A) In other words, if A is not far from normal (i.e., eigenvectors are nearly orthogonal), and if properly normalized degree n polynomials can be found whose size on the spectrum Λ(A) decreases quickly with n, then GMRES converges quickly Xiangmin Jiao Numerical Analysis I 8 / 9

Other Krylov Subspace Methods CG on the Normal Equations (CGN) Solve A Ax = A b using Conjugate Gradients Poor convergence due to squared condition number (i.e., κ(a A) = κ(a) 2 ) One advantage is that it applies least squares problems without modification BiConjugate Gradients (BCG/BiCG) Makes residuals orthogonal to another Krylov subspace, based on A It can be implemented with three-term recurrences, so memory requirements is smaller Convergence sometimes comparable to GMRES, but unpredictable Conjugate Gradients Squared (CGS) Avoids multiplication by A in BCG, sometimes twice as fast convergence as BCG Quasi-Minimal Residuals (QMR) and Stabilized BiCG (Bi-CGSTAB) Variants of BiCG with more regular convergence Xiangmin Jiao Numerical Analysis I 9 / 9