Communication-avoiding Krylov subspace methods

Size: px
Start display at page:

Download "Communication-avoiding Krylov subspace methods"

Transcription

1 Motivation Communication-avoiding Krylov subspace methods Mark University of California Berkeley EECS MS Numerical Libraries Group visit: 28 April 2008

2 Overview Motivation Current Krylov methods: communication-limited Can rearrange them to avoid communication Can do this in a numerically stable way Requires stepping outside the black box

3 Motivation Review Encapsulation Problem Amortize the communication Data dependencies limit reuse Review: Krylov subspace methods Given these kernels: Sparse matrix-vector product (SpMV) operator A (Possibly) a preconditioner operator M 1 Dot product and vector operations Solve Ax = b or Ax = λx iteratively By constructing a basis of span{r, Ar, A 2 r,... } And projecting onto it

4 Motivation Review Encapsulation Problem Amortize the communication Data dependencies limit reuse Encapsulation in standard Krylov methods Krylov methods great for software engineers! All kernels called as black boxes Good challenging to optimize! SpMV needs tuning & domain-specific hints Preconditioners a domain-specific black art Efficient reductions nontrivial on complex architectures

5 Motivation Review Encapsulation Problem Amortize the communication Data dependencies limit reuse Problem: Kernels are communication bound Sparse matrix-vector multiplication Parallel: Latency (communicate with neighbors) Sequential: Bandwidth (read matrix) Much like SpMV Orthogonalization Dot products and norms Θ(1) reductions per vector

6 Motivation Amortize the commmunication Review Encapsulation Problem Amortize the communication Data dependencies limit reuse Our goal: Amortize cost of each kernel over s steps Compute many SpMVs for cost of 1 SpMV Orthogonalize many vectors for cost of 1 reduction which we accomplish by breaking encapsulation, oops!

7 Matrix powers kernel Motivation Review Encapsulation Problem Amortize the communication Data dependencies limit reuse Replace s SpMVs with matrix powers kernel Marghoob s talk (details in Demmel 2007) Compute basis of span{v, Av, A 2 v,..., A s v}... With same communication cost as one SpMV Via overlapping ghost zones + redundant computation Local Dependencies for k=8 Type (1) Remote Dependencies for k=8 Type (2) Remote Dependencies for k=

8 Tall Skinny QR Motivation Review Encapsulation Problem Amortize the communication Data dependencies limit reuse Replace orthogonalization with Tall Skinny QR (TSQR) Demmel et al (coming soon) Block row reduction in one communication step QR factorization is the reduction operator Unconditionally stable Figure: TSQR on a binary tree of 4 processors.

9 What stops us? Motivation Review Encapsulation Problem Amortize the communication Data dependencies limit reuse What stops us from using both kernels in a Krylov method?

10 Data dependencies Motivation Review Encapsulation Problem Amortize the communication Data dependencies limit reuse Krylov methods advance one vector at a time SpMV, then orthogonalize, then SpMV,... Figure: Data dependencies in Krylov subspace methods.

11 Motivation Solution: Solution Example: GMRES Basis condition number Numerical experiments Can break dependency with s-step Krylov methods Mathematically equivalent to original algorithms Compute basis of span{v, Av, A 2 v,..., A s v} Orthogonalize s + 1 vectors Reconstruct upper Hessenberg H (Arnoldi) resp. tridiagonal T (Lanczos) using R factor and basis properties Solve for solution update and continue

12 Motivation Solution Example: GMRES Basis condition number Numerical experiments Example: GMRES

13 Original GMRES Motivation Solution Example: GMRES Basis condition number Numerical experiments 1: for k = 1 to s do 2: w = Av k 1 3: Orthogonalize w against v 0,..., v k 1 (Modified Gram-Schmidt) 4: end for 5: Compute solution using H

14 Motivation Solution Example: GMRES Basis condition number Numerical experiments Version 2: Matrix powers kernel & TSQR 1: W = [v 0, Av 0, A 2 v 0,..., A s v 0 ] 2: [Q, R] = TSQR(W ) 3: Compute H using R 4: Compute solution using H s powers of A for no extra latency cost s steps of QR for one step of latency But...

15 Motivation Basis computation not stable Solution Example: GMRES Basis condition number Numerical experiments v, Av, A 2 v,... looks familiar... It s the power method! Converges to principal eigenvector of A Basis condition number exponential in s

16 Motivation Basis computation not stable Solution Example: GMRES Basis condition number Numerical experiments v, Av, A 2 v,... looks familiar... It s the power method! Converges to principal eigenvector of A Basis condition number exponential in s

17 Motivation Version 3: Different basis Solution Example: GMRES Basis condition number Numerical experiments Just like polynomial interpolation Use a different basis, e.g.: Newton basis W = [v, (A θ 1 I)v, (A θ 2 I)(A θ 1 I)v,... ] Chebyshev basis W = [v, T 1 (v), T 2 (v),... ] Parameters via estimated spectral info... Which comes free with Krylov method!

18 Motivation Basis condition number Solution Example: GMRES Basis condition number Numerical experiments Figure: Condition number of various bases vs. basis length s. Matrix A is D Poisson with Dirac delta right-hand side.

19 Motivation Numerical experiments Solution Example: GMRES Basis condition number Numerical experiments Diagonal matrix, κ 2 (A) = 10 8 s = 24 Newton: basis condition # about Monomial: basis condition # about 10 16

20 Motivation Better basis pays off: restarting Solution Example: GMRES Basis condition number Numerical experiments GMRES(24,1) residuals: cond(a) = 1e8, n=1e4 Standard(24,1) Monomial(24,1) Newton(24,1) Log base 10 of 2 norm relative residual error Iteration count Figure: Restart after every group of s steps

21 Motivation Solution Example: GMRES Basis condition number Numerical experiments Better basis pays off: less restarting 1 0 GMRES(24,8) residuals: cond(a) = 1e8, n=1e4 Standard(24,8) Monomial(24,8) Newton(24,8) Log base 10 of 2 norm relative residual error Iteration count Figure: Restart after 8 groups of s = 24 steps.

22 Motivation Lots of previous work s-step CG Van Rosendale 1983, Chronopoulos et al. 1989, Toledo 1995,... s-step GMRES Walker 1988, De Sturler 1991, Bai et al. 1991, Joubert et al. 1992, Erhel 1995,...

23 Motivation Our advances over previous work Performance: Matrix powers kernel and TSQR Fixed numerical stability problems Need not restart after each group of s Can do preconditioning...

24 Motivation Matrix powers kernel changes Matrix powers kernel Computing this is hard What preconditioners might work? Dense but low rank GMRES with left preconditioning v, M 1 Av, (M 1 A) 2 v,..., (M 1 A) s v CG with split preconditioning v, L 1 AL T v,..., (L 1 AL T ) s v CG with left preconditioning M 1 A not necessarily symmetric! V = [v, M 1 Av,..., (M 1 A) s v], and W = [Av, AM 1 Av,..., (AM 1 ) s Av] We know the necessary adjustments for any basis

25 Motivation Matrix powers kernel changes Matrix powers kernel Computing this is hard What preconditioners might work? Dense but low rank GMRES with left preconditioning v, M 1 Av, (M 1 A) 2 v,..., (M 1 A) s v CG with split preconditioning v, L 1 AL T v,..., (L 1 AL T ) s v CG with left preconditioning M 1 A not necessarily symmetric! V = [v, M 1 Av,..., (M 1 A) s v], and W = [Av, AM 1 Av,..., (AM 1 ) s Av] We know the necessary adjustments for any basis

26 Motivation Computing this is hard Matrix powers kernel Computing this is hard What preconditioners might work? Dense but low rank We couldn t find an author that tried Preconditioner must be part of matrix powers kernel Otherwise, Θ(s) communication steps vs. Θ(1) This changes the abstraction Matrix and preconditioner not separate black boxes anymore! But encapsulation preserved at a different level

27 Motivation Computing this is hard Matrix powers kernel Computing this is hard What preconditioners might work? Dense but low rank We couldn t find an author that tried Preconditioner must be part of matrix powers kernel Otherwise, Θ(s) communication steps vs. Θ(1) This changes the abstraction Matrix and preconditioner not separate black boxes anymore! But encapsulation preserved at a different level

28 Motivation What preconditioners might work? Matrix powers kernel Computing this is hard What preconditioners might work? Dense but low rank Polynomial preconditioning? Natural fit with matrix powers kernel See e.g., Saad 1985 Overlapping Schwarz domain decomposition? Use same overlap regions as matrix powers kernel Sparse approximate inverse? Place nonzeros to avoid communication

29 Beyond sparsity? Motivation Matrix powers kernel Computing this is hard What preconditioners might work? Dense but low rank Sparsity is the wrong abstraction Quickly disappears with matrix powers Doesn t limit communication to nearest neighbors: e.g., x x x x x x A = x x x is O(n) sparse but requires O(n) messages per SpMV

30 Dense but low rank Motivation Matrix powers kernel Computing this is hard What preconditioners might work? Dense but low rank Matrix powers tend to destroy sparsity Tridiagonal + block Jacobi dense after four iterations But ranks of off-diagonal block rows & columns only increase linearly True for general matrices

31 Motivation Rank = communication Matrix powers kernel Computing this is hard What preconditioners might work? Dense but low rank Let r ij = rank of block i, j r ij : # of source vector words proc j sends to proc i in SpMV In matrix powers kernel: proc i redundantly computes s r ij ghost zone values from proc j Bound rank to limit communication

32 Motivation Matrix powers kernel Computing this is hard What preconditioners might work? Dense but low rank Caveat: Communication vs. computation Low off-diagonal block rank only saves communication Need sparsity, else higher computational complexity O( Ω ) term per node of interior domain Ω Asymptotically more complex for 2-, 3-D problems We re still working on this... Figure: Left: low-rank dense off-diagonal blocks. Right: low-rank sparse off-diagonal blocks.

33 Motivation Compatible preconditioners Matrix powers kernel Computing this is hard What preconditioners might work? Dense but low rank Black box in interior, low-rank blocks outside Hierarchical matrices (Hackbusch et al.) Semiseparable matrices Fast multipole method (a) log x y on a straight line Figure: Discretization of log ( x y ) on interval.

34 Motivation Performance tuning (choosing s) Extension to eigensolvers Lanczos biorthogonalization (e.g., Bi-CG) Preconditioner implementations Combine with block Krylov methods Block methods can already use TSQR Does combining block and s-step pay?

35 Motivation Can amortize communication in Krylov methods Break open the global SpMV black box SpMV on interior of local domain can remain closed Rewrite orthogonalization as one reduction Can preserve numerical stability in theory

36 Motivation Mentors and colleagues Advisor: James Demmel (UC Berkeley) Sparse matrix operations: Kathy Yelick (U. Calif. Berkeley and LBNL) Marghoob Mohiyuddin (U. Calif. Berkeley) QR factorization: Julien Langou (U. Colorado Denver) Laura Grigori (INRIA, France)

37 Motivation Acknowledgments Research supported by: Microsoft and Intel (ParLab) ACM/IEEE NSF US Department of Energy

38 Copyright Motivation Copyright 2008 Mark Licensed under Creative Commons Attribution-Share Alike 3.0 United States. See by-sa/3.0/us/ for more information.

39 Appendix Extra slides Block Krylov methods More slides Bibliography Bibliography I Z. BAI, D. HU, AND L. REICHTEL, A Newton basis GMRES implementation, IMA Journal of Numerical Analysis, 14 (1994), pp A. H. BAKER, J. M. DENNIS, AND E. R. JESSUP, On improving linear solver performance: A block variant of GMRES, SIAM J. Sci. Comp., 27 (2006), pp S. BÖRM, L. GRASEDYCK, AND W. HACKBUSCH, Hierarchical matrices. HMatrices.pdf, 2004.

40 Appendix Extra slides Block Krylov methods More slides Bibliography Bibliography II S. CHANDRASEKARAN, M. GU, AND W. LYONS, A fast and stable adaptive solver for hierarchically semi-separable representations, May A. T. CHRONOPOULOS AND C. W. GEAR, s-step iterative methods for symmetric linear systems, J. Comput. Appl. Math., 25 (1989), pp A. T. CHRONOPOULOS AND A. B. KUCHEROV, A parallel Krylov-type method for nonsymmetric linear systems, in High Performance Computing - HiPC 2001: Eighth International Conference, Hyderabad, India, December 17-20, Proceedings, Springer, 2001, pp

41 Appendix Extra slides Block Krylov methods More slides Bibliography Bibliography III E. DE STURLER, A parallel variant of GMRES(m), in Proceedings of the 13th IMACS World Congress on Computation and Applied Mathematics, J. J. H. Miller and R. Vichnevetsky, eds., Dublin, Ireland, 1991, Criterion Press. J. DEMMEL, M. F. HOEMMEN, M. MOHIYUDDIN, AND K. A. YELICK, Avoiding communication in computing Krylov subspaces, Tech. Rep. UCB/EECS , EECS Department, University of California, Berkeley, Oct 2007.

42 Appendix Extra slides Block Krylov methods More slides Bibliography Bibliography IV J. ERHEL, A parallel GMRES version for general sparse matrices, Electronic Transactions on Numerical Analysis, 3 (1995), pp W. GAUTSCHI AND G. INGLESE, Lower bounds for the condition number of Vandermonde matrices, Numer. Math., 52 (1988), pp W. HACKBUSCH, Hierarchische Matrizen Algorithmen und Analysis. hmvorlesung.ps, last accessed 22 May 2006, Jan

43 Appendix Extra slides Block Krylov methods More slides Bibliography Bibliography V W. D. JOUBERT AND G. F. CAREY, Parallelizable restarted iterative methods for nonsymmetric linear systems, Part I: Theory, International Journal of Computer Mathematics, 44 (1992), pp , Parallelizable restarted iterative methods for nonsymmetric linear systems, Part II: Parallel implementation, International Journal of Computer Mathematics, 44 (1992), pp

44 Appendix Extra slides Block Krylov methods More slides Bibliography Bibliography VI C. E. LEISERSON, S. RAO, AND S. TOLEDO, Efficient out-of-core algorithms for linear relaxation using blocking covers, Journal of Computer and System Sciences, 54 (1997), pp G. MEURANT, The block preconditioned conjugate gradient method on vector computers, BIT, 24 (1984), pp D. P. O LEARY, The block conjugate gradient algorithm and related methods, Linear Algebra Appl., 29 (1980), pp

45 Appendix Extra slides Block Krylov methods More slides Bibliography Bibliography VII Y. SAAD, Practical use of polynomial preconditionings for the conjugate gradient method, SIAM J. Sci. Stat. Comput., 6 (1985), pp S. A. TOLEDO, Quantitative performance modeling of scientific computations and creating locality in numerical algorithms, PhD thesis, Massachusetts Institute of Technology, J. VAN ROSENDALE, Minimizing inner product data dependence in conjugate gradient iteration, in Proc. IEEE Internat. Confer. Parallel Processing, 1983.

46 Appendix Extra slides Block Krylov methods More slides Review Problems with block methods Review: Block Krylov methods Build up basis of span{b, A B, A 2 B,... } for block of vectors B Original application: hard eigenproblems Accelerates convergence for multiple / clustered eigenvalues Can also solve Ax = B Best for multiple right-hand sides Can use if only one right-hand side SpMV for multiple vectors No extra latency cost Bandwidth cost scales linearly w/ # vectors

47 Appendix Extra slides Block Krylov methods More slides Review Problems with block methods Problems with block methods for Ax = b (1 of 2) If only one right-hand side: Do one restart cycle with one RHS After each restart cycle, add one error vector to RHS block Higher startup cost than Need s cycles of s until at full block size Whereas, s-step always at full optimization

48 Appendix Extra slides Block Krylov methods More slides Review Problems with block methods Problems with block methods for Ax = b (2 of 2) More complicated convergence & breakdown conditions Convergence benefit isn t as clear as one-vector case Must do deflation Remove linearly dependent vectors from block Need rank-revealing factorization

49 Appendix Extra slides Block Krylov methods More slides Restarting for stability Extra precision for stability Lanczos reorthogonalization What preconditioner structure limits communication? Components Restarting for stability

50 Appendix Extra slides Block Krylov methods More slides Restarting for stability Extra precision for stability Lanczos reorthogonalization What preconditioner structure limits communication? Components Extra precision for stability (1 of 3)

51 Appendix Extra slides Block Krylov methods More slides Restarting for stability Extra precision for stability Lanczos reorthogonalization What preconditioner structure limits communication? Components Extra precision for stability (2 of 3)

52 Appendix Extra slides Block Krylov methods More slides Restarting for stability Extra precision for stability Lanczos reorthogonalization What preconditioner structure limits communication? Components Extra precision for stability (3 of 3)

53 Appendix Extra slides Block Krylov methods More slides Restarting for stability Extra precision for stability Lanczos reorthogonalization What preconditioner structure limits communication? Components Lanczos(s,t) w/ reorthogonalization Get orthogonality estimates from Lanczos recurrence (Paige) Each group of s basis vectors is a TSQR Q factor Best reorthogonalization: Do TSQR of last group to compute Lanczos coefficients Use Lanczos coeffs in Paige s recurrence If last group not orthogonal w.r.t. previous groups Compute it explicitly Orthogonalize against previous t 1 groups Finally take TSQR again of last group Converting all groups of s to explicit storage and redoing TSQR on them all is too expensive & unnecessary

54 Appendix Extra slides Block Krylov methods More slides Restarting for stability Extra precision for stability Lanczos reorthogonalization What preconditioner structure limits communication? Components What preconditioner structure limits communication? Experiment: Partition matrix A & preconditioner M into block rows Compute MA, A(MA), (MA) 2, A(MA) 2,... Look at Fill-in (sparsity) Rank of off-diagonal block rows & columns Matrices Matrix A is tridiagonal Preconditioner M is block diagonal

55 Appendix Extra slides Block Krylov methods More slides Restarting for stability Extra precision for stability Lanczos reorthogonalization What preconditioner structure limits communication? Components Figure: Left is spy plot of A; right is spy plot of M.

56 Appendix Extra slides Block Krylov methods More slides Restarting for stability Extra precision for stability Lanczos reorthogonalization What preconditioner structure limits communication? Components Figure: Left is spy plot of A; right is spy plot of M A. Block row & column ranks of M A are 1, 2, 2, 1.

57 Appendix Extra slides Block Krylov methods More slides Restarting for stability Extra precision for stability Lanczos reorthogonalization What preconditioner structure limits communication? Components Figure: Left is spy plot of A(M A); block row & column ranks are 2, 4, 4, 2. Right is spy plot of (M A) 2 ; block row & column ranks are 2, 4, 4, 2.

58 Appendix Extra slides Block Krylov methods More slides Restarting for stability Extra precision for stability Lanczos reorthogonalization What preconditioner structure limits communication? Components Figure: Left is spy plot of A(M A) 2 ; block row & column ranks are 3, 6, 6, 3. Right is spy plot of (M A) 3 ; block row & column ranks are 3, 6, 6, 3.

59 Appendix Extra slides Block Krylov methods More slides Restarting for stability Extra precision for stability Lanczos reorthogonalization What preconditioner structure limits communication? Components Figure: Left is spy plot of A(M A) 3 ; block row & column ranks are 4, 8, 8, 4. Right is spy plot of (M A) 4 ; block row & column ranks are 4, 8, 8, 4.

60 Dense but low rank Appendix Extra slides Block Krylov methods More slides Restarting for stability Extra precision for stability Lanczos reorthogonalization What preconditioner structure limits communication? Components In matrix powers, sparsity quickly lost But ranks of off-diagonal block rows & columns only increase linearly True for general matrices

61 Appendix Extra slides Block Krylov methods More slides Rank = communication Restarting for stability Extra precision for stability Lanczos reorthogonalization What preconditioner structure limits communication? Components Let r ij = rank of block i, j r ij : # of source vector words proc j sends to proc i in SpMV In matrix powers kernel: proc i redundantly computes s r ij ghost zone values from proc j Bound rank to limit communication in matrix powers kernel

62 Appendix Extra slides Block Krylov methods More slides Restarting for stability Extra precision for stability Lanczos reorthogonalization What preconditioner structure limits communication? Components Components Figure: Components of communication-avoiding Krylov methods.

Minisymposia 9 and 34: Avoiding Communication in Linear Algebra. Jim Demmel UC Berkeley bebop.cs.berkeley.edu

Minisymposia 9 and 34: Avoiding Communication in Linear Algebra. Jim Demmel UC Berkeley bebop.cs.berkeley.edu Minisymposia 9 and 34: Avoiding Communication in Linear Algebra Jim Demmel UC Berkeley bebop.cs.berkeley.edu Motivation (1) Increasing parallelism to exploit From Top500 to multicores in your laptop Exponentially

More information

Exploiting Low-Rank Structure in Computing Matrix Powers with Applications to Preconditioning

Exploiting Low-Rank Structure in Computing Matrix Powers with Applications to Preconditioning Exploiting Low-Rank Structure in Computing Matrix Powers with Applications to Preconditioning Erin C. Carson, Nicholas Knight, James Demmel, Ming Gu U.C. Berkeley SIAM PP 12, Savannah, Georgia, USA, February

More information

Efficient Deflation for Communication-Avoiding Krylov Subspace Methods

Efficient Deflation for Communication-Avoiding Krylov Subspace Methods Efficient Deflation for Communication-Avoiding Krylov Subspace Methods Erin Carson Nicholas Knight, James Demmel Univ. of California, Berkeley Monday, June 24, NASCA 2013, Calais, France Overview We derive

More information

The Lanczos and conjugate gradient algorithms

The Lanczos and conjugate gradient algorithms The Lanczos and conjugate gradient algorithms Gérard MEURANT October, 2008 1 The Lanczos algorithm 2 The Lanczos algorithm in finite precision 3 The nonsymmetric Lanczos algorithm 4 The Golub Kahan bidiagonalization

More information

4.8 Arnoldi Iteration, Krylov Subspaces and GMRES

4.8 Arnoldi Iteration, Krylov Subspaces and GMRES 48 Arnoldi Iteration, Krylov Subspaces and GMRES We start with the problem of using a similarity transformation to convert an n n matrix A to upper Hessenberg form H, ie, A = QHQ, (30) with an appropriate

More information

AMS526: Numerical Analysis I (Numerical Linear Algebra) Lecture 23: GMRES and Other Krylov Subspace Methods; Preconditioning

AMS526: Numerical Analysis I (Numerical Linear Algebra) Lecture 23: GMRES and Other Krylov Subspace Methods; Preconditioning AMS526: Numerical Analysis I (Numerical Linear Algebra) Lecture 23: GMRES and Other Krylov Subspace Methods; Preconditioning Xiangmin Jiao SUNY Stony Brook Xiangmin Jiao Numerical Analysis I 1 / 18 Outline

More information

Arnoldi Methods in SLEPc

Arnoldi Methods in SLEPc Scalable Library for Eigenvalue Problem Computations SLEPc Technical Report STR-4 Available at http://slepc.upv.es Arnoldi Methods in SLEPc V. Hernández J. E. Román A. Tomás V. Vidal Last update: October,

More information

APPLIED NUMERICAL LINEAR ALGEBRA

APPLIED NUMERICAL LINEAR ALGEBRA APPLIED NUMERICAL LINEAR ALGEBRA James W. Demmel University of California Berkeley, California Society for Industrial and Applied Mathematics Philadelphia Contents Preface 1 Introduction 1 1.1 Basic Notation

More information

A communication-avoiding thick-restart Lanczos method on a distributed-memory system

A communication-avoiding thick-restart Lanczos method on a distributed-memory system A communication-avoiding thick-restart Lanczos method on a distributed-memory system Ichitaro Yamazaki and Kesheng Wu Lawrence Berkeley National Laboratory, Berkeley, CA, USA Abstract. The Thick-Restart

More information

M.A. Botchev. September 5, 2014

M.A. Botchev. September 5, 2014 Rome-Moscow school of Matrix Methods and Applied Linear Algebra 2014 A short introduction to Krylov subspaces for linear systems, matrix functions and inexact Newton methods. Plan and exercises. M.A. Botchev

More information

Summary of Iterative Methods for Non-symmetric Linear Equations That Are Related to the Conjugate Gradient (CG) Method

Summary of Iterative Methods for Non-symmetric Linear Equations That Are Related to the Conjugate Gradient (CG) Method Summary of Iterative Methods for Non-symmetric Linear Equations That Are Related to the Conjugate Gradient (CG) Method Leslie Foster 11-5-2012 We will discuss the FOM (full orthogonalization method), CG,

More information

Finite-choice algorithm optimization in Conjugate Gradients

Finite-choice algorithm optimization in Conjugate Gradients Finite-choice algorithm optimization in Conjugate Gradients Jack Dongarra and Victor Eijkhout January 2003 Abstract We present computational aspects of mathematically equivalent implementations of the

More information

SOLVING SPARSE LINEAR SYSTEMS OF EQUATIONS. Chao Yang Computational Research Division Lawrence Berkeley National Laboratory Berkeley, CA, USA

SOLVING SPARSE LINEAR SYSTEMS OF EQUATIONS. Chao Yang Computational Research Division Lawrence Berkeley National Laboratory Berkeley, CA, USA 1 SOLVING SPARSE LINEAR SYSTEMS OF EQUATIONS Chao Yang Computational Research Division Lawrence Berkeley National Laboratory Berkeley, CA, USA 2 OUTLINE Sparse matrix storage format Basic factorization

More information

Numerical Methods in Matrix Computations

Numerical Methods in Matrix Computations Ake Bjorck Numerical Methods in Matrix Computations Springer Contents 1 Direct Methods for Linear Systems 1 1.1 Elements of Matrix Theory 1 1.1.1 Matrix Algebra 2 1.1.2 Vector Spaces 6 1.1.3 Submatrices

More information

Preface to the Second Edition. Preface to the First Edition

Preface to the Second Edition. Preface to the First Edition n page v Preface to the Second Edition Preface to the First Edition xiii xvii 1 Background in Linear Algebra 1 1.1 Matrices................................. 1 1.2 Square Matrices and Eigenvalues....................

More information

Algorithms that use the Arnoldi Basis

Algorithms that use the Arnoldi Basis AMSC 600 /CMSC 760 Advanced Linear Numerical Analysis Fall 2007 Arnoldi Methods Dianne P. O Leary c 2006, 2007 Algorithms that use the Arnoldi Basis Reference: Chapter 6 of Saad The Arnoldi Basis How to

More information

Sparse linear solvers: iterative methods and preconditioning

Sparse linear solvers: iterative methods and preconditioning Sparse linear solvers: iterative methods and preconditioning L Grigori ALPINES INRIA and LJLL, UPMC March 2017 Plan Sparse linear solvers Sparse matrices and graphs Classes of linear solvers Krylov subspace

More information

Course Notes: Week 1

Course Notes: Week 1 Course Notes: Week 1 Math 270C: Applied Numerical Linear Algebra 1 Lecture 1: Introduction (3/28/11) We will focus on iterative methods for solving linear systems of equations (and some discussion of eigenvalues

More information

Linear Solvers. Andrew Hazel

Linear Solvers. Andrew Hazel Linear Solvers Andrew Hazel Introduction Thus far we have talked about the formulation and discretisation of physical problems...... and stopped when we got to a discrete linear system of equations. Introduction

More information

Scientific Computing with Case Studies SIAM Press, Lecture Notes for Unit VII Sparse Matrix

Scientific Computing with Case Studies SIAM Press, Lecture Notes for Unit VII Sparse Matrix Scientific Computing with Case Studies SIAM Press, 2009 http://www.cs.umd.edu/users/oleary/sccswebpage Lecture Notes for Unit VII Sparse Matrix Computations Part 1: Direct Methods Dianne P. O Leary c 2008

More information

The Conjugate Gradient Method

The Conjugate Gradient Method The Conjugate Gradient Method Classical Iterations We have a problem, We assume that the matrix comes from a discretization of a PDE. The best and most popular model problem is, The matrix will be as large

More information

WHEN studying distributed simulations of power systems,

WHEN studying distributed simulations of power systems, 1096 IEEE TRANSACTIONS ON POWER SYSTEMS, VOL 21, NO 3, AUGUST 2006 A Jacobian-Free Newton-GMRES(m) Method with Adaptive Preconditioner and Its Application for Power Flow Calculations Ying Chen and Chen

More information

Contents. Preface... xi. Introduction...

Contents. Preface... xi. Introduction... Contents Preface... xi Introduction... xv Chapter 1. Computer Architectures... 1 1.1. Different types of parallelism... 1 1.1.1. Overlap, concurrency and parallelism... 1 1.1.2. Temporal and spatial parallelism

More information

Conjugate gradient method. Descent method. Conjugate search direction. Conjugate Gradient Algorithm (294)

Conjugate gradient method. Descent method. Conjugate search direction. Conjugate Gradient Algorithm (294) Conjugate gradient method Descent method Hestenes, Stiefel 1952 For A N N SPD In exact arithmetic, solves in N steps In real arithmetic No guaranteed stopping Often converges in many fewer than N steps

More information

Topics. The CG Algorithm Algorithmic Options CG s Two Main Convergence Theorems

Topics. The CG Algorithm Algorithmic Options CG s Two Main Convergence Theorems Topics The CG Algorithm Algorithmic Options CG s Two Main Convergence Theorems What about non-spd systems? Methods requiring small history Methods requiring large history Summary of solvers 1 / 52 Conjugate

More information

Minimizing Communication in Linear Algebra. James Demmel 15 June

Minimizing Communication in Linear Algebra. James Demmel 15 June Minimizing Communication in Linear Algebra James Demmel 15 June 2010 www.cs.berkeley.edu/~demmel 1 Outline What is communication and why is it important to avoid? Direct Linear Algebra Lower bounds on

More information

ITERATIVE METHODS FOR SPARSE LINEAR SYSTEMS

ITERATIVE METHODS FOR SPARSE LINEAR SYSTEMS ITERATIVE METHODS FOR SPARSE LINEAR SYSTEMS YOUSEF SAAD University of Minnesota PWS PUBLISHING COMPANY I(T)P An International Thomson Publishing Company BOSTON ALBANY BONN CINCINNATI DETROIT LONDON MADRID

More information

LARGE SPARSE EIGENVALUE PROBLEMS. General Tools for Solving Large Eigen-Problems

LARGE SPARSE EIGENVALUE PROBLEMS. General Tools for Solving Large Eigen-Problems LARGE SPARSE EIGENVALUE PROBLEMS Projection methods The subspace iteration Krylov subspace methods: Arnoldi and Lanczos Golub-Kahan-Lanczos bidiagonalization General Tools for Solving Large Eigen-Problems

More information

LARGE SPARSE EIGENVALUE PROBLEMS

LARGE SPARSE EIGENVALUE PROBLEMS LARGE SPARSE EIGENVALUE PROBLEMS Projection methods The subspace iteration Krylov subspace methods: Arnoldi and Lanczos Golub-Kahan-Lanczos bidiagonalization 14-1 General Tools for Solving Large Eigen-Problems

More information

6.4 Krylov Subspaces and Conjugate Gradients

6.4 Krylov Subspaces and Conjugate Gradients 6.4 Krylov Subspaces and Conjugate Gradients Our original equation is Ax = b. The preconditioned equation is P Ax = P b. When we write P, we never intend that an inverse will be explicitly computed. P

More information

AMS526: Numerical Analysis I (Numerical Linear Algebra)

AMS526: Numerical Analysis I (Numerical Linear Algebra) AMS526: Numerical Analysis I (Numerical Linear Algebra) Lecture 19: More on Arnoldi Iteration; Lanczos Iteration Xiangmin Jiao Stony Brook University Xiangmin Jiao Numerical Analysis I 1 / 17 Outline 1

More information

Last Time. Social Network Graphs Betweenness. Graph Laplacian. Girvan-Newman Algorithm. Spectral Bisection

Last Time. Social Network Graphs Betweenness. Graph Laplacian. Girvan-Newman Algorithm. Spectral Bisection Eigenvalue Problems Last Time Social Network Graphs Betweenness Girvan-Newman Algorithm Graph Laplacian Spectral Bisection λ 2, w 2 Today Small deviation into eigenvalue problems Formulation Standard eigenvalue

More information

On the influence of eigenvalues on Bi-CG residual norms

On the influence of eigenvalues on Bi-CG residual norms On the influence of eigenvalues on Bi-CG residual norms Jurjen Duintjer Tebbens Institute of Computer Science Academy of Sciences of the Czech Republic duintjertebbens@cs.cas.cz Gérard Meurant 30, rue

More information

Iterative methods for Linear System

Iterative methods for Linear System Iterative methods for Linear System JASS 2009 Student: Rishi Patil Advisor: Prof. Thomas Huckle Outline Basics: Matrices and their properties Eigenvalues, Condition Number Iterative Methods Direct and

More information

IDR(s) as a projection method

IDR(s) as a projection method Delft University of Technology Faculty of Electrical Engineering, Mathematics and Computer Science Delft Institute of Applied Mathematics IDR(s) as a projection method A thesis submitted to the Delft Institute

More information

1 Extrapolation: A Hint of Things to Come

1 Extrapolation: A Hint of Things to Come Notes for 2017-03-24 1 Extrapolation: A Hint of Things to Come Stationary iterations are simple. Methods like Jacobi or Gauss-Seidel are easy to program, and it s (relatively) easy to analyze their convergence.

More information

FEM and sparse linear system solving

FEM and sparse linear system solving FEM & sparse linear system solving, Lecture 9, Nov 19, 2017 1/36 Lecture 9, Nov 17, 2017: Krylov space methods http://people.inf.ethz.ch/arbenz/fem17 Peter Arbenz Computer Science Department, ETH Zürich

More information

Applied Mathematics 205. Unit V: Eigenvalue Problems. Lecturer: Dr. David Knezevic

Applied Mathematics 205. Unit V: Eigenvalue Problems. Lecturer: Dr. David Knezevic Applied Mathematics 205 Unit V: Eigenvalue Problems Lecturer: Dr. David Knezevic Unit V: Eigenvalue Problems Chapter V.4: Krylov Subspace Methods 2 / 51 Krylov Subspace Methods In this chapter we give

More information

arxiv: v1 [hep-lat] 2 May 2012

arxiv: v1 [hep-lat] 2 May 2012 A CG Method for Multiple Right Hand Sides and Multiple Shifts in Lattice QCD Calculations arxiv:1205.0359v1 [hep-lat] 2 May 2012 Fachbereich C, Mathematik und Naturwissenschaften, Bergische Universität

More information

Parallel sparse linear solvers and applications in CFD

Parallel sparse linear solvers and applications in CFD Parallel sparse linear solvers and applications in CFD Jocelyne Erhel Joint work with Désiré Nuentsa Wakam () and Baptiste Poirriez () SAGE team, Inria Rennes, France journée Calcul Intensif Distribué

More information

PROJECTED GMRES AND ITS VARIANTS

PROJECTED GMRES AND ITS VARIANTS PROJECTED GMRES AND ITS VARIANTS Reinaldo Astudillo Brígida Molina rastudillo@kuaimare.ciens.ucv.ve bmolina@kuaimare.ciens.ucv.ve Centro de Cálculo Científico y Tecnológico (CCCT), Facultad de Ciencias,

More information

The amount of work to construct each new guess from the previous one should be a small multiple of the number of nonzeros in A.

The amount of work to construct each new guess from the previous one should be a small multiple of the number of nonzeros in A. AMSC/CMSC 661 Scientific Computing II Spring 2005 Solution of Sparse Linear Systems Part 2: Iterative methods Dianne P. O Leary c 2005 Solving Sparse Linear Systems: Iterative methods The plan: Iterative

More information

MS 28: Scalable Communication-Avoiding and -Hiding Krylov Subspace Methods I

MS 28: Scalable Communication-Avoiding and -Hiding Krylov Subspace Methods I MS 28: Scalable Communication-Avoiding and -Hiding Krylov Subspace Methods I Organizers: Siegfried Cools, University of Antwerp, Belgium Erin C. Carson, New York University, USA 10:50-11:10 High Performance

More information

The quadratic eigenvalue problem (QEP) is to find scalars λ and nonzero vectors u satisfying

The quadratic eigenvalue problem (QEP) is to find scalars λ and nonzero vectors u satisfying I.2 Quadratic Eigenvalue Problems 1 Introduction The quadratic eigenvalue problem QEP is to find scalars λ and nonzero vectors u satisfying where Qλx = 0, 1.1 Qλ = λ 2 M + λd + K, M, D and K are given

More information

CME342 Parallel Methods in Numerical Analysis. Matrix Computation: Iterative Methods II. Sparse Matrix-vector Multiplication.

CME342 Parallel Methods in Numerical Analysis. Matrix Computation: Iterative Methods II. Sparse Matrix-vector Multiplication. CME342 Parallel Methods in Numerical Analysis Matrix Computation: Iterative Methods II Outline: CG & its parallelization. Sparse Matrix-vector Multiplication. 1 Basic iterative methods: Ax = b r = b Ax

More information

Parallel Numerics, WT 2016/ Iterative Methods for Sparse Linear Systems of Equations. page 1 of 1

Parallel Numerics, WT 2016/ Iterative Methods for Sparse Linear Systems of Equations. page 1 of 1 Parallel Numerics, WT 2016/2017 5 Iterative Methods for Sparse Linear Systems of Equations page 1 of 1 Contents 1 Introduction 1.1 Computer Science Aspects 1.2 Numerical Problems 1.3 Graphs 1.4 Loop Manipulations

More information

Iterative methods, preconditioning, and their application to CMB data analysis. Laura Grigori INRIA Saclay

Iterative methods, preconditioning, and their application to CMB data analysis. Laura Grigori INRIA Saclay Iterative methods, preconditioning, and their application to CMB data analysis Laura Grigori INRIA Saclay Plan Motivation Communication avoiding for numerical linear algebra Novel algorithms that minimize

More information

MS4: Minimizing Communication in Numerical Algorithms Part I of II

MS4: Minimizing Communication in Numerical Algorithms Part I of II MS4: Minimizing Communication in Numerical Algorithms Part I of II Organizers: Oded Schwartz (Hebrew University of Jerusalem) and Erin Carson (New York University) Talks: 1. Communication-Avoiding Krylov

More information

Avoiding Communication in Distributed-Memory Tridiagonalization

Avoiding Communication in Distributed-Memory Tridiagonalization Avoiding Communication in Distributed-Memory Tridiagonalization SIAM CSE 15 Nicholas Knight University of California, Berkeley March 14, 2015 Joint work with: Grey Ballard (SNL) James Demmel (UCB) Laura

More information

Iterative Methods for Sparse Linear Systems

Iterative Methods for Sparse Linear Systems Iterative Methods for Sparse Linear Systems Luca Bergamaschi e-mail: berga@dmsa.unipd.it - http://www.dmsa.unipd.it/ berga Department of Mathematical Methods and Models for Scientific Applications University

More information

Chapter 7 Iterative Techniques in Matrix Algebra

Chapter 7 Iterative Techniques in Matrix Algebra Chapter 7 Iterative Techniques in Matrix Algebra Per-Olof Persson persson@berkeley.edu Department of Mathematics University of California, Berkeley Math 128B Numerical Analysis Vector Norms Definition

More information

Contribution of Wo¹niakowski, Strako²,... The conjugate gradient method in nite precision computa

Contribution of Wo¹niakowski, Strako²,... The conjugate gradient method in nite precision computa Contribution of Wo¹niakowski, Strako²,... The conjugate gradient method in nite precision computations ªaw University of Technology Institute of Mathematics and Computer Science Warsaw, October 7, 2006

More information

Key words. linear equations, polynomial preconditioning, nonsymmetric Lanczos, BiCGStab, IDR

Key words. linear equations, polynomial preconditioning, nonsymmetric Lanczos, BiCGStab, IDR POLYNOMIAL PRECONDITIONED BICGSTAB AND IDR JENNIFER A. LOE AND RONALD B. MORGAN Abstract. Polynomial preconditioning is applied to the nonsymmetric Lanczos methods BiCGStab and IDR for solving large nonsymmetric

More information

ON ORTHOGONAL REDUCTION TO HESSENBERG FORM WITH SMALL BANDWIDTH

ON ORTHOGONAL REDUCTION TO HESSENBERG FORM WITH SMALL BANDWIDTH ON ORTHOGONAL REDUCTION TO HESSENBERG FORM WITH SMALL BANDWIDTH V. FABER, J. LIESEN, AND P. TICHÝ Abstract. Numerous algorithms in numerical linear algebra are based on the reduction of a given matrix

More information

S-Step and Communication-Avoiding Iterative Methods

S-Step and Communication-Avoiding Iterative Methods S-Step and Communication-Avoiding Iterative Methods Maxim Naumov NVIDIA, 270 San Tomas Expressway, Santa Clara, CA 95050 Abstract In this paper we make an overview of s-step Conjugate Gradient and develop

More information

Simple iteration procedure

Simple iteration procedure Simple iteration procedure Solve Known approximate solution Preconditionning: Jacobi Gauss-Seidel Lower triangle residue use of pre-conditionner correction residue use of pre-conditionner Convergence Spectral

More information

Communication Avoiding Strategies for the Numerical Kernels in Coupled Physics Simulations

Communication Avoiding Strategies for the Numerical Kernels in Coupled Physics Simulations ExaScience Lab Intel Labs Europe EXASCALE COMPUTING Communication Avoiding Strategies for the Numerical Kernels in Coupled Physics Simulations SIAM Conference on Parallel Processing for Scientific Computing

More information

Deflation Strategies to Improve the Convergence of Communication-Avoiding GMRES

Deflation Strategies to Improve the Convergence of Communication-Avoiding GMRES Deflation Strategies to Improve the Convergence of Communication-Avoiding GMRES Ichitaro Yamazaki, Stanimire Tomov, and Jack Dongarra Department of Electrical Engineering and Computer Science University

More information

Communication avoiding parallel algorithms for dense matrix factorizations

Communication avoiding parallel algorithms for dense matrix factorizations Communication avoiding parallel dense matrix factorizations 1/ 44 Communication avoiding parallel algorithms for dense matrix factorizations Edgar Solomonik Department of EECS, UC Berkeley October 2013

More information

Parallel Iterative Methods for Sparse Linear Systems. H. Martin Bücker Lehrstuhl für Hochleistungsrechnen

Parallel Iterative Methods for Sparse Linear Systems. H. Martin Bücker Lehrstuhl für Hochleistungsrechnen Parallel Iterative Methods for Sparse Linear Systems Lehrstuhl für Hochleistungsrechnen www.sc.rwth-aachen.de RWTH Aachen Large and Sparse Small and Dense Outline Problem with Direct Methods Iterative

More information

Communication-avoiding parallel and sequential QR factorizations

Communication-avoiding parallel and sequential QR factorizations Communication-avoiding parallel and sequential QR factorizations James Demmel, Laura Grigori, Mark Hoemmen, and Julien Langou May 30, 2008 Abstract We present parallel and sequential dense QR factorization

More information

Solution of eigenvalue problems. Subspace iteration, The symmetric Lanczos algorithm. Harmonic Ritz values, Jacobi-Davidson s method

Solution of eigenvalue problems. Subspace iteration, The symmetric Lanczos algorithm. Harmonic Ritz values, Jacobi-Davidson s method Solution of eigenvalue problems Introduction motivation Projection methods for eigenvalue problems Subspace iteration, The symmetric Lanczos algorithm Nonsymmetric Lanczos procedure; Implicit restarts

More information

9.1 Preconditioned Krylov Subspace Methods

9.1 Preconditioned Krylov Subspace Methods Chapter 9 PRECONDITIONING 9.1 Preconditioned Krylov Subspace Methods 9.2 Preconditioned Conjugate Gradient 9.3 Preconditioned Generalized Minimal Residual 9.4 Relaxation Method Preconditioners 9.5 Incomplete

More information

Preconditioned inverse iteration and shift-invert Arnoldi method

Preconditioned inverse iteration and shift-invert Arnoldi method Preconditioned inverse iteration and shift-invert Arnoldi method Melina Freitag Department of Mathematical Sciences University of Bath CSC Seminar Max-Planck-Institute for Dynamics of Complex Technical

More information

Lecture 8: Fast Linear Solvers (Part 7)

Lecture 8: Fast Linear Solvers (Part 7) Lecture 8: Fast Linear Solvers (Part 7) 1 Modified Gram-Schmidt Process with Reorthogonalization Test Reorthogonalization If Av k 2 + δ v k+1 2 = Av k 2 to working precision. δ = 10 3 2 Householder Arnoldi

More information

A Residual Replacement Strategy for Improving the Maximum Attainable Accuracy of Communication- Avoiding Krylov Subspace Methods

A Residual Replacement Strategy for Improving the Maximum Attainable Accuracy of Communication- Avoiding Krylov Subspace Methods A Residual Replacement Strategy for Improving the Maximum Attainable Accuracy of Communication- Avoiding Krylov Subspace Methods Erin Carson James Demmel Electrical Engineering and Computer Sciences University

More information

Key words. conjugate gradients, normwise backward error, incremental norm estimation.

Key words. conjugate gradients, normwise backward error, incremental norm estimation. Proceedings of ALGORITMY 2016 pp. 323 332 ON ERROR ESTIMATION IN THE CONJUGATE GRADIENT METHOD: NORMWISE BACKWARD ERROR PETR TICHÝ Abstract. Using an idea of Duff and Vömel [BIT, 42 (2002), pp. 300 322

More information

Communication-avoiding parallel and sequential QR factorizations

Communication-avoiding parallel and sequential QR factorizations Communication-avoiding parallel and sequential QR factorizations James Demmel Laura Grigori Mark Frederick Hoemmen Julien Langou Electrical Engineering and Computer Sciences University of California at

More information

Charles University Faculty of Mathematics and Physics DOCTORAL THESIS. Krylov subspace approximations in linear algebraic problems

Charles University Faculty of Mathematics and Physics DOCTORAL THESIS. Krylov subspace approximations in linear algebraic problems Charles University Faculty of Mathematics and Physics DOCTORAL THESIS Iveta Hnětynková Krylov subspace approximations in linear algebraic problems Department of Numerical Mathematics Supervisor: Doc. RNDr.

More information

Solution of eigenvalue problems. Subspace iteration, The symmetric Lanczos algorithm. Harmonic Ritz values, Jacobi-Davidson s method

Solution of eigenvalue problems. Subspace iteration, The symmetric Lanczos algorithm. Harmonic Ritz values, Jacobi-Davidson s method Solution of eigenvalue problems Introduction motivation Projection methods for eigenvalue problems Subspace iteration, The symmetric Lanczos algorithm Nonsymmetric Lanczos procedure; Implicit restarts

More information

Iterative Methods for Linear Systems of Equations

Iterative Methods for Linear Systems of Equations Iterative Methods for Linear Systems of Equations Projection methods (3) ITMAN PhD-course DTU 20-10-08 till 24-10-08 Martin van Gijzen 1 Delft University of Technology Overview day 4 Bi-Lanczos method

More information

Algebraic Multigrid as Solvers and as Preconditioner

Algebraic Multigrid as Solvers and as Preconditioner Ò Algebraic Multigrid as Solvers and as Preconditioner Domenico Lahaye domenico.lahaye@cs.kuleuven.ac.be http://www.cs.kuleuven.ac.be/ domenico/ Department of Computer Science Katholieke Universiteit Leuven

More information

AMS526: Numerical Analysis I (Numerical Linear Algebra for Computational and Data Sciences)

AMS526: Numerical Analysis I (Numerical Linear Algebra for Computational and Data Sciences) AMS526: Numerical Analysis I (Numerical Linear Algebra for Computational and Data Sciences) Lecture 19: Computing the SVD; Sparse Linear Systems Xiangmin Jiao Stony Brook University Xiangmin Jiao Numerical

More information

A DISSERTATION. Extensions of the Conjugate Residual Method. by Tomohiro Sogabe. Presented to

A DISSERTATION. Extensions of the Conjugate Residual Method. by Tomohiro Sogabe. Presented to A DISSERTATION Extensions of the Conjugate Residual Method ( ) by Tomohiro Sogabe Presented to Department of Applied Physics, The University of Tokyo Contents 1 Introduction 1 2 Krylov subspace methods

More information

On the loss of orthogonality in the Gram-Schmidt orthogonalization process

On the loss of orthogonality in the Gram-Schmidt orthogonalization process CERFACS Technical Report No. TR/PA/03/25 Luc Giraud Julien Langou Miroslav Rozložník On the loss of orthogonality in the Gram-Schmidt orthogonalization process Abstract. In this paper we study numerical

More information

Communication-Avoiding Krylov Subspace Methods in Theory and Practice. Erin Claire Carson. A dissertation submitted in partial satisfaction of the

Communication-Avoiding Krylov Subspace Methods in Theory and Practice. Erin Claire Carson. A dissertation submitted in partial satisfaction of the Communication-Avoiding Krylov Subspace Methods in Theory and Practice by Erin Claire Carson A dissertation submitted in partial satisfaction of the requirements for the degree of Doctor of Philosophy in

More information

Matrix Algorithms. Volume II: Eigensystems. G. W. Stewart H1HJ1L. University of Maryland College Park, Maryland

Matrix Algorithms. Volume II: Eigensystems. G. W. Stewart H1HJ1L. University of Maryland College Park, Maryland Matrix Algorithms Volume II: Eigensystems G. W. Stewart University of Maryland College Park, Maryland H1HJ1L Society for Industrial and Applied Mathematics Philadelphia CONTENTS Algorithms Preface xv xvii

More information

Solving Sparse Linear Systems: Iterative methods

Solving Sparse Linear Systems: Iterative methods Scientific Computing with Case Studies SIAM Press, 2009 http://www.cs.umd.edu/users/oleary/sccs Lecture Notes for Unit VII Sparse Matrix Computations Part 2: Iterative Methods Dianne P. O Leary c 2008,2010

More information

Solving Sparse Linear Systems: Iterative methods

Solving Sparse Linear Systems: Iterative methods Scientific Computing with Case Studies SIAM Press, 2009 http://www.cs.umd.edu/users/oleary/sccswebpage Lecture Notes for Unit VII Sparse Matrix Computations Part 2: Iterative Methods Dianne P. O Leary

More information

Incomplete Cholesky preconditioners that exploit the low-rank property

Incomplete Cholesky preconditioners that exploit the low-rank property anapov@ulb.ac.be ; http://homepages.ulb.ac.be/ anapov/ 1 / 35 Incomplete Cholesky preconditioners that exploit the low-rank property (theory and practice) Artem Napov Service de Métrologie Nucléaire, Université

More information

Multigrid absolute value preconditioning

Multigrid absolute value preconditioning Multigrid absolute value preconditioning Eugene Vecharynski 1 Andrew Knyazev 2 (speaker) 1 Department of Computer Science and Engineering University of Minnesota 2 Department of Mathematical and Statistical

More information

A Domain Decomposition Based Jacobi-Davidson Algorithm for Quantum Dot Simulation

A Domain Decomposition Based Jacobi-Davidson Algorithm for Quantum Dot Simulation A Domain Decomposition Based Jacobi-Davidson Algorithm for Quantum Dot Simulation Tao Zhao 1, Feng-Nan Hwang 2 and Xiao-Chuan Cai 3 Abstract In this paper, we develop an overlapping domain decomposition

More information

Performance Evaluation of Some Inverse Iteration Algorithms on PowerXCell T M 8i Processor

Performance Evaluation of Some Inverse Iteration Algorithms on PowerXCell T M 8i Processor Performance Evaluation of Some Inverse Iteration Algorithms on PowerXCell T M 8i Processor Masami Takata 1, Hiroyuki Ishigami 2, Kini Kimura 2, and Yoshimasa Nakamura 2 1 Academic Group of Information

More information

Lab 1: Iterative Methods for Solving Linear Systems

Lab 1: Iterative Methods for Solving Linear Systems Lab 1: Iterative Methods for Solving Linear Systems January 22, 2017 Introduction Many real world applications require the solution to very large and sparse linear systems where direct methods such as

More information

ITERATIVE METHODS BASED ON KRYLOV SUBSPACES

ITERATIVE METHODS BASED ON KRYLOV SUBSPACES ITERATIVE METHODS BASED ON KRYLOV SUBSPACES LONG CHEN We shall present iterative methods for solving linear algebraic equation Au = b based on Krylov subspaces We derive conjugate gradient (CG) method

More information

Deflation for inversion with multiple right-hand sides in QCD

Deflation for inversion with multiple right-hand sides in QCD Deflation for inversion with multiple right-hand sides in QCD A Stathopoulos 1, A M Abdel-Rehim 1 and K Orginos 2 1 Department of Computer Science, College of William and Mary, Williamsburg, VA 23187 2

More information

Key words. linear equations, eigenvalues, polynomial preconditioning, GMRES, GMRES-DR, deflation, QCD

Key words. linear equations, eigenvalues, polynomial preconditioning, GMRES, GMRES-DR, deflation, QCD 1 POLYNOMIAL PRECONDITIONED GMRES AND GMRES-DR QUAN LIU, RONALD B. MORGAN, AND WALTER WILCOX Abstract. We look at solving large nonsymmetric systems of linear equations using polynomial preconditioned

More information

From Stationary Methods to Krylov Subspaces

From Stationary Methods to Krylov Subspaces Week 6: Wednesday, Mar 7 From Stationary Methods to Krylov Subspaces Last time, we discussed stationary methods for the iterative solution of linear systems of equations, which can generally be written

More information

AMS526: Numerical Analysis I (Numerical Linear Algebra)

AMS526: Numerical Analysis I (Numerical Linear Algebra) AMS526: Numerical Analysis I (Numerical Linear Algebra) Lecture 23: GMRES and Other Krylov Subspace Methods Xiangmin Jiao SUNY Stony Brook Xiangmin Jiao Numerical Analysis I 1 / 9 Minimizing Residual CG

More information

Krylov Subspace Methods that Are Based on the Minimization of the Residual

Krylov Subspace Methods that Are Based on the Minimization of the Residual Chapter 5 Krylov Subspace Methods that Are Based on the Minimization of the Residual Remark 51 Goal he goal of these methods consists in determining x k x 0 +K k r 0,A such that the corresponding Euclidean

More information

Introduction. Chapter One

Introduction. Chapter One Chapter One Introduction The aim of this book is to describe and explain the beautiful mathematical relationships between matrices, moments, orthogonal polynomials, quadrature rules and the Lanczos and

More information

Iterative Methods for Solving A x = b

Iterative Methods for Solving A x = b Iterative Methods for Solving A x = b A good (free) online source for iterative methods for solving A x = b is given in the description of a set of iterative solvers called templates found at netlib: http

More information

Reduced Synchronization Overhead on. December 3, Abstract. The standard formulation of the conjugate gradient algorithm involves

Reduced Synchronization Overhead on. December 3, Abstract. The standard formulation of the conjugate gradient algorithm involves Lapack Working Note 56 Conjugate Gradient Algorithms with Reduced Synchronization Overhead on Distributed Memory Multiprocessors E. F. D'Azevedo y, V.L. Eijkhout z, C. H. Romine y December 3, 1999 Abstract

More information

Iterative methods for Linear System of Equations. Joint Advanced Student School (JASS-2009)

Iterative methods for Linear System of Equations. Joint Advanced Student School (JASS-2009) Iterative methods for Linear System of Equations Joint Advanced Student School (JASS-2009) Course #2: Numerical Simulation - from Models to Software Introduction In numerical simulation, Partial Differential

More information

Inexactness and flexibility in linear Krylov solvers

Inexactness and flexibility in linear Krylov solvers Inexactness and flexibility in linear Krylov solvers Luc Giraud ENSEEIHT (N7) - IRIT, Toulouse Matrix Analysis and Applications CIRM Luminy - October 15-19, 2007 in honor of Gérard Meurant for his 60 th

More information

Alternative correction equations in the Jacobi-Davidson method

Alternative correction equations in the Jacobi-Davidson method Chapter 2 Alternative correction equations in the Jacobi-Davidson method Menno Genseberger and Gerard Sleijpen Abstract The correction equation in the Jacobi-Davidson method is effective in a subspace

More information

Review: From problem to parallel algorithm

Review: From problem to parallel algorithm Review: From problem to parallel algorithm Mathematical formulations of interesting problems abound Poisson s equation Sources: Electrostatics, gravity, fluid flow, image processing (!) Numerical solution:

More information

Numerical Methods I Non-Square and Sparse Linear Systems

Numerical Methods I Non-Square and Sparse Linear Systems Numerical Methods I Non-Square and Sparse Linear Systems Aleksandar Donev Courant Institute, NYU 1 donev@courant.nyu.edu 1 MATH-GA 2011.003 / CSCI-GA 2945.003, Fall 2014 September 25th, 2014 A. Donev (Courant

More information

A short course on: Preconditioned Krylov subspace methods. Yousef Saad University of Minnesota Dept. of Computer Science and Engineering

A short course on: Preconditioned Krylov subspace methods. Yousef Saad University of Minnesota Dept. of Computer Science and Engineering A short course on: Preconditioned Krylov subspace methods Yousef Saad University of Minnesota Dept. of Computer Science and Engineering Universite du Littoral, Jan 19-3, 25 Outline Part 1 Introd., discretization

More information

Krylov Space Methods. Nonstationary sounds good. Radu Trîmbiţaş ( Babeş-Bolyai University) Krylov Space Methods 1 / 17

Krylov Space Methods. Nonstationary sounds good. Radu Trîmbiţaş ( Babeş-Bolyai University) Krylov Space Methods 1 / 17 Krylov Space Methods Nonstationary sounds good Radu Trîmbiţaş Babeş-Bolyai University Radu Trîmbiţaş ( Babeş-Bolyai University) Krylov Space Methods 1 / 17 Introduction These methods are used both to solve

More information