Incomplete Cholesky preconditioners that exploit the low-rank property

Size: px
Start display at page:

Download "Incomplete Cholesky preconditioners that exploit the low-rank property"

Transcription

1 ; anapov/ 1 / 35 Incomplete Cholesky preconditioners that exploit the low-rank property (theory and practice) Artem Napov Service de Métrologie Nucléaire, Université Libre de Bruxelles PRECONDITIONING (TU EINDHOVEN) JUNE 17, 2015

2 Motivation The solution of a symmetric positive definite (SPD) system Au = b may be obtained with a direct method (usually: Cholesky factorization A = R T R) robust iterative method (usually: preconditioned conjugate gradient) ; efficient if a good SPD preconditioner B A is available, i.e.: 1. Cheap (to construct, store, invert, parallelize) 2. Good approximation of A contradictory goals tradeoff Possible black box design: take an exact factorization (here, Cholesky) ; add approximation (to enforce 1.) incomplete factorization preconditioner B = R T R A. 2 / 35

3 Motivation Incomplete factorizations B = R T R A may perform approximations by dropping individual entries (mainstream) ; using low-rank approximations (emerging). Incomplete factorizations: + are (almost) black box (approximation threshold required) + are (relatively) robust may breakdown (but breakdown-free variants exist) have no guarantee to converge fast; fast convergence if B approximates well A ; that is, if κ(r T AR 1 ) = λ max(r T AR 1 ) λ min (R T AR 1 ) is small. Here we present incomplete factorizations that are breakdown-free and have controllable condition number. 3 / 35

4 4 / 35 Outline Theory... One-level variant basic approach and underlying analysis Multilevel variants extensions of analysis to multi-level setting... and practice Sparse solver motivation and design choices Numerical experiments and comparison with other solvers

5 4 / 35 Outline Theory... One-level variant basic approach and underlying analysis Multilevel variants extensions of analysis to multi-level setting... and practice Sparse solver motivation and design choices Numerical experiments and comparison with other solvers

6 5 / 35 One-level variant : basic ideas Incomplete (block left-looking) Cholesky for each block row k = 1,..., l : 1 factorize 2 solve 3 approximate 4 update (compute new rows) A ( k = 1 )

7 5 / 35 One-level variant : basic ideas Incomplete (block left-looking) Cholesky for each block row k = 1,..., l : 1 factorize 2 solve 3 approximate 4 update (compute new rows) ( k = 1 ) A 11 A 12 A 22 A

8 5 / 35 One-level variant : basic ideas Incomplete (block left-looking) Cholesky for each block row k = 1,..., l : 1 factorize 2 solve 3 approximate 4 update (compute new rows) ( k = 1 ) R 11 A 12 A 22 A 11 = R T 11 R 11 A

9 5 / 35 One-level variant : basic ideas Incomplete (block left-looking) Cholesky for each block row k = 1,..., l : 1 factorize 2 solve 3 approximate 4 update (compute new rows) ( k = 1 ) R 11 R T 11 A 12 A 22 A

10 5 / 35 One-level variant : basic ideas Incomplete (block left-looking) Cholesky for each block row k = 1,..., l : 1 factorize 2 solve 3 approximate 4 update (compute new rows) ( k = 1 ) R 12 A 22 A R 12 R 12 The dropping is orthogonal if R ( ) T R 12 R 12 R R 12 = O. R 12 R 12

11 5 / 35 One-level variant : basic ideas Incomplete (block left-looking) Cholesky for each block row k = 1,..., l : 1 factorize 2 solve 3 approximate 4 update (compute new rows) ( k = 1 ) R 12 A 22 A R 12 R 12 The dropping is orthogonal if R 12 T ( R 12 R 12 ) = O. This implies monotonicity for Schur complement (for any SPD A!) : v T (A 22 R T 12 R 12 )v v T (A 22 R T 12R 12 )v

12 One-level variant : basic ideas Incomplete (block left-looking) Cholesky for each block row k = 1,..., l : 1 factorize 2 solve 3 approximate 4 update (compute new rows) ( k = 1 ) R 12 A 22 A R 12 R 12 Example of orthogonal dropping low-rank approximation via truncated SVD (with absolute threshold tol a ) ) T + U 2 (Σ 2 V2 ) }{{} R 12 = ( U 1 U 2 ) ( Σ 1 Σ 2 ) ( V T 1 V T 2 If Σ 2 < tol a, then = U 1 (Σ 1 V T 1 ) }{{} R 12 R 12 R 12 R 12 R 12 < tol a 5 / 35

13 5 / 35 One-level variant : basic ideas Incomplete (block left-looking) Cholesky for each block row k = 1,..., l : 1 factorize 2 solve 3 approximate 4 update (compute new rows) R 12 A 22 B 1 ( k = 1 ) S 1 = A 22 1 R T 12 1 R 12

14 5 / 35 One-level variant : basic ideas Incomplete (block left-looking) Cholesky for each block row k = 1,..., l : 1 factorize 2 solve 3 approximate 4 update (compute new rows) B 1 ( k = 2 )

15 5 / 35 One-level variant : basic ideas Incomplete (block left-looking) Cholesky for each block row k = 1,..., l : 1 factorize 2 solve 3 approximate 4 update (compute new rows) B 1 ( k = 2 )

16 One-level variant : successive approximations Incomplete (block left-looking) Cholesky for each block row k = 1,..., l : 1 factorize 2 solve 3 approximate 4 update (compute new rows) ( k = 2 ) approximate new rows only One-level (model) variant k = 1 k = 2 k = 3 k = 4 6 / 35

17 7 / 35 Accuracy of individual dropping The orthogonal dropping is assumed, namely R 12 T ( R 12 R 12 ) = O R 12 A 22 B k 1 The accuracy of individual dropping at step k : (R γ k = 12 R ) 12 S 1/2 < 1, B with S B = A 22 R T 12 R 12. R 12 B k

18 Model problem Model Problem: { u u = f in Ω = (0, 1) 2 u n Full system matrix: = 0 on Ω A Ω = A I A II A I,Γ A II,Γ. Ω I Γ Ω II A T I,Γ AT II,Γ A Γ Interface system matrix: A = A Γ A T I,ΓA 1 I A Γ,I A T II,ΓA 1 II A Γ,II Corresponds to: last Schur complement in sparse Cholesky with ND ordering Schur complement system in iterative substructuring 8 / 35

19 9 / 35 Accuracy of individual dropping The orthogonal dropping is assumed, namely R 12 T ( R 12 R 12 ) = O. The accuracy of individual dropping at step k: (R γ k = 12 R ) 12 S 1/2 < 1, with B S B = A 22 R T 12 R 12. Model problem: (k = 1, block size = 10, grid, truncated SVD dropping) r γ B k 1 B k

20 One-level variant: conditioning analysis (for λ max ) Assume A SPD and that dropping is orthogonal Let B k correspond to the preconditioner at the end of step k, with A = B 0 B l = R T R is the final preconditioner. Only newly computed rows are modified; Let λ (k) ( ) max = λ max B 1 (k 1) l B k, λ max = λ max ( B 1 l B k 1 ) B k 1 B k Then λ (k) max λ (k 1) max λ (k) max + g( λ (k) max, γ k ) where g( λ, γ ) = max β>0 2 γ β λ 1 β 2 β 2 + λ 1. B l 10 / 35

21 One-level variant: conditioning analysis (for λ min ) Assume A SPD and that dropping is orthogonal Let B k correspond to the preconditioner at the end of step k, with A = B 0 B l = R T R is the final preconditioner. Only newly computed rows are modified; Let λ (k) min = λ ( ) min B 1 (k) l B k, λ min = λ ( ) min B 1 l B k 1 B k 1 B k Then λ (k) min λ(k 1) min λ (k) min g( λ(k) min, γ k ) where g( λ, γ ) = max β>0 2 γ β λ 1 β 2 β 2 + λ 1. B l 10 / 35

22 One-level variant: conditioning analysis One-level bounds (for λ max ) λ (k) max λ (k 1) max λ (k) max + g( λ (k) max, γ k ) B k 1 where λ (k) max ( ) = λ max B 1 (k 1) l B k, λ max = λ max ( B 1 l B k 1 ) The estimate is sharp: For any γ k a matrix A < 1, k = 1,..., l, there exist a sequence of approximations B k (with B 0 = A) such that the bounds for λ (0) max and λ (0) min are simultaneously reached. B k B l 11 / 35

23 Model problem: numerical experiments Model Problem: { u u = f in Ω = (0, 1) 2 u n Full system matrix: = 0 on Ω A Ω = A I A II A I,Γ A II,Γ. Ω I Γ Ω II N A T I,Γ AT II,Γ A Γ Interface system N N matrix: A = A Γ A T I,ΓA 1 I A Γ,I A T II,ΓA 1 II A Γ,II Algorithmic details: block size = 10 we consider N from 20 to 650 (n from 400 to , l from 1 to 64) maximal ranks remains below 4 12 / 35

24 13 / 35 One-level variant: numerical experiments (tol r = 10 3 ) 10 2 κ(r T AR 1 ) n condition number one-level precond. upper bound one-level

25 14 / 35 Outline Theory... One-level variant basic approach and underlying analysis Multilevel variants extensions of analysis to multi-level setting... and practice Sparse solver motivation and design choices Numerical experiments and comparison with other solvers

26 Multilevel variants : motivation One-level (model) variant k = 1 k = 2 k = 3 k = 4 O(N 3 ) cost, O(N 2 ) memory (although both cost and memory are improved!) N 15 / 35

27 Multilevel variants : motivation General \ Sequentially Semiseparable variant (SSS) [Gu, Li, Vassilevski 10] k = 1 k = 2 k = 3 k = 4 O(rN 2 ) cost, O(rN) memory r - maximal approximation rank N N/r improvement! 16 / 35

28 17 / 35 Multilevel variants : general estimate B 1 Assuming that dropping is orthogonal one has ( ) λ max B 1 k B k 1 = 1 + γk ( ) λ min B 1 k B k 1 = 1 γk B 2 and, hence, κ(r T AR 1 ) l k=1 1 + γ k 1 γ k. B 3 B 4

29 Model problem Model Problem: { u u = f in Ω = (0, 1) 2 u n Full system matrix: = 0 on Ω A Ω = A I A II A I,Γ A II,Γ. Ω I Γ Ω II N A T I,Γ AT II,Γ A Γ Interface system N N matrix: A = A Γ A T I,ΓA 1 I A Γ,I A T II,ΓA 1 II A Γ,II Algorithmic details: block size = 10 we consider N from 20 to 650 (n from 400 to , l from 1 to 64) maximal ranks remains below 4 18 / 35

30 19 / 35 Multilevel variants: numerical experiments κ(r T AR 1 ) n condition number κ(a) SSS precond. (tol r = 10 3 ) upper bound general (10 3 )

31 20 / 35 Multilevel variants: nested subspaces Sequentially Semiseparable variant (SSS) [Gu, Li, Vassilevski 10] k = 1 k = 2 k = 3 k = 4 Nested subspaces assumption R 12 R 12 R 12 R 12 B 1 B 2 B 3 B 4 span( R 12 ) span( ) span( ) span( )

32 Multilevel variants: nested subspaces Sequentially Semiseparable variant (SSS) [Gu, Li, Vassilevski 10] k = 1 k = 2 k = 3 k = 4 Nested subspaces assumption R 12 R 12 R 12 R 12 B 1 B 2 B 3 B 4 span( R 12 ) span( R 12 ) span( ) span( ) 20 / 35

33 Multilevel variants: nested subspaces Sequentially Semiseparable variant (SSS) [Gu, Li, Vassilevski 10] k = 1 k = 2 k = 3 k = 4 Nested subspaces assumption R 12 R 12 R 12 R 12 B 1 B 2 B 3 B 4 span( R 12 ) span( R 12 ) span( R 12 ) span( ) 20 / 35

34 20 / 35 Multilevel variants: nested subspaces Same bounds as for one-level case! One-level bounds (for λ max ) where λ (k) max λ (k) max λ (k 1) max ( ) = λ max B 1 (k 1) l B k, λ max Nested subspaces assumption λ (k) max + g( λ (k) max, γ k ) = λ max ( B 1 l B k 1 ) R 12 R 12 R 12 R 12 B 1 B 2 B 3 B 4 span( R 12 ) span( R 12 ) span( R 12 ) span( )

35 20 / 35 Multilevel variants: nested subspaces Same bounds as for one-level case! One-level bounds (for λ min ) where λ (k) min λ (k) min = λ min λ(k 1) min ( ) B 1 (k 1) l B k, λ min Nested subspaces assumption λ (k) min g( λ(k) min, γ k ) = λ min ( B 1 l B k 1 ) R 12 R 12 R 12 R 12 B 1 B 2 B 3 B 4 span( R 12 ) span( R 12 ) span( R 12 ) span( )

36 21 / 35 Multilevel variants: numerical experiments (tol r = 10 3 ) κ(r T AR 1 ) n condition number SSS precond. one-level precond. upper bound one-level one-level

37 22 / 35 Outline Theory... One-level variant basic approach and underlying analysis Multilevel variants extensions of analysis to multi-level setting... and practice Sparse solver motivation and design choices Numerical experiments and comparison with other solvers

38 23 / 35 Sparse solver: factorization MAIN FEATURES: symmetric left-looking block factorization accumulate updates before block row computation nested dissection (ND) ordering induces block structure, enforces sparsity, and reduces operation count orthogonal low-rank approximations by truncated SVD or rank-revealing QR

39 23 / 35 Sparse solver: factorization MAIN FEATURES: symmetric left-looking block factorization accumulate updates before block row computation nested dissection (ND) ordering induces block structure, enforces sparsity, and reduces operation count orthogonal low-rank approximations by truncated SVD or rank-revealing QR Γ 3 Γ 2 Ω 2 Ω 3 Ω 4 Ω 1 Γ 1 } } } } } } Ω 1 Ω 2 Γ 1 Ω 3 Ω 4 Γ 2 Γ 3

40 23 / 35 Sparse solver: factorization MAIN FEATURES: symmetric left-looking block factorization accumulate updates before block row computation nested dissection (ND) ordering induces block structure, enforces sparsity, and reduces operation count orthogonal low-rank approximations by truncated SVD or rank-revealing QR

41 24 / 35 Sparse solver: factorization SPECIAL FEATURES: rank-revealing reordering inside ND blocks ensures the algebraic character of the solver symbolic compression ensures that memory usage decreases adaptive block size faster than fixed block size

42 25 / 35 Sparse solver: revealing low-rank Model problem analysis [Chandrasekaran et al., 10]: rank is bounded proportionally to the number of connections between the corresp. subset and the remaining nodes of separator Implemented: recursive edge bisection (via METIS) 2D model separators:

43 Sparse solver: revealing low-rank Model problem analysis [Chandrasekaran et al., 10]: rank is bounded proportionally to the number of connections between the corresp. subset and the remaining nodes of separator Implemented: recursive edge bisection (via METIS) 2D model separators: 25 / 35

44 Sparse solver: revealing low-rank Model problem analysis [Chandrasekaran et al., 10]: rank is bounded proportionally to the number of connections between the corresp. subset and the remaining nodes of separator Implemented: recursive edge bisection (via METIS) 2D model separators: } } 25 / 35

45 Sparse solver: revealing low-rank Model problem analysis [Chandrasekaran et al., 10]: rank is bounded proportionally to the number of connections between the corresp. subset and the remaining nodes of separator Implemented: recursive edge bisection (via METIS) 2D model separators: } } } } } } 25 / 35

46 Sparse solver: revealing low-rank Model problem analysis [Chandrasekaran et al., 10]: rank is bounded proportionally to the number of connections between the corresp. subset and the remaining nodes of separator Implemented: recursive edge bisection (via METIS) 2D separators from Scotch: connections of length 2 25 / 35

47 Sparse solver: revealing low-rank Model problem analysis [Chandrasekaran et al., 10]: rank is bounded proportionally to the number of connections between the corresp. subset and the remaining nodes of separator Implemented: recursive edge bisection (via METIS) 2D separators from Scotch: connections of length 2 25 / 35

48 Sparse solver: revealing low-rank Model problem analysis [Chandrasekaran et al., 10]: rank is bounded proportionally to the number of connections between the corresp. subset and the remaining nodes of separator Implemented: recursive edge bisection (via METIS) 3D model separators: 25 / 35

49 Sparse solver: revealing low-rank Model problem analysis [Chandrasekaran et al., 10]: rank is bounded proportionally to the number of connections between the corresp. subset and the remaining nodes of separator Implemented: recursive edge bisection (via METIS) 3D model separators: } } } } } } 25 / 35

50 26 / 35 Sparse solver: symbolic compression If for a low-rank approximation R 12 R 12 Q one has ( nnz R 12 ) nnz( R 12 ) + nnz( Q ), then the low-rank approximation is replaced by R 12 = I R 12. This low-rank approximation is (trivially) orthogonal. No need to store I does not increase memory use.

51 27 / 35 Sparse solver: adaptive block size Via symbolic compression (s.c.), block subdivision can evolve: } } } } }s.c. }s.c. } } } } We use this evolution: by automatically applying s.c. after few occurrences at the same level; (more successive s.c. occur, (exponentially) more s.c. are skipped to save computations); to adjust the minimal block size for the following separators (this also saves computations).

52 28 / 35 Outline Theory... One-level variant basic approach and underlying analysis Multilevel variants extensions of analysis to multi-level setting... and practice Sparse solver motivation and design choices Numerical experiments and comparison with other solvers

53 Numerical experiments: setting Solver parameters: the preconditioner, denoted by SIC, is used with approximation based on rank-revealing QR with column pivoting and absolute approximation threshold ; adaptive block size (initially block size is set to 16); PCG as outer iteration. Test problems: all SPD matrices from University of Florida Sparse Matrix Collection 1 with n > 10 5 and random rhs (excluded: thermomech_tk and bmw7st_1, both have κ(a) ). Experimental setting: time reported for best threshold value (chosen among {10 10, 10 9,..., }) stopping criterion : 10 6 relative residual decrease; hardware : Intel Xeon L5420, 2.5 GHz, 16 GB RAM / 35

54 30 / 35 Comparison with (exact) Cholesky 100 chol SIC time (sec.) per million nnz problems by nnz

55 31 / 35 Comparison with (unpreconditioned) CG 100 CG SIC time (sec.) per million nnz problems by nnz

56 32 / 35 Comparison with ILUPACK ILUPACK SIC time (sec.) per million nnz problems by nnz

57 33 / 35 Conclusions In theory... We have presented a conditioning analysis for incomplete Cholesky factorizations preconditioner based on orthogonal dropping Only requirement on A: should be SPD The analysis relates the condition number of the preconditioned system to the individual dropping accuracies (namely, to γ k s) The analysis is sharp in the one-level case One-level bound can be extended to the multilevel setting if an additional nested subspace assumption holds; it naturally holds for the presented preconditioner

58 34 / 35 Conclusions... and in practice We have presented a preliminary implementation of incomplete Cholesky factorizations preconditioner based on orthogonal dropping The solver targets at sparse matrices; it uses sparsity structure to reduce the block rank during the factorization Preliminary numerical experiments demonstrate that the solver is competitive

59 Further details Theory: A. Napov, Conditioning analysis of incomplete Cholesky factorizations with orthogonal dropping, SIAM J. Matrix Anal. Appl. A. Napov, Conditioning analysis of incomplete Cholesky factorizations with orthogonal dropping II: nested subspaces (in preparation) Related approaches (for dense matrices): M. Gu, X. S. Li, P. Vassilevski, Direction-preserving and Schur-monotonic semiseparable approximations of symmetric positive definite matrices, SIAM J. Matrix Anal. Appl. J. Xia, M. Gu, Robust approximate Cholesky factorization of rank-structured symmetric positive definite matrices, SIAM J. Matrix Anal. Appl. Rank revealing based on sparsity: A. Napov, X. S. Li, An algebraic multifrontal preconditioner that exploits the low-rank property, Numer. Lin. Alg. Appl. (to appear). 35 / 35

J.I. Aliaga 1 M. Bollhöfer 2 A.F. Martín 1 E.S. Quintana-Ortí 1. March, 2009

J.I. Aliaga 1 M. Bollhöfer 2 A.F. Martín 1 E.S. Quintana-Ortí 1. March, 2009 Parallel Preconditioning of Linear Systems based on ILUPACK for Multithreaded Architectures J.I. Aliaga M. Bollhöfer 2 A.F. Martín E.S. Quintana-Ortí Deparment of Computer Science and Engineering, Univ.

More information

A robust multilevel approximate inverse preconditioner for symmetric positive definite matrices

A robust multilevel approximate inverse preconditioner for symmetric positive definite matrices DICEA DEPARTMENT OF CIVIL, ENVIRONMENTAL AND ARCHITECTURAL ENGINEERING PhD SCHOOL CIVIL AND ENVIRONMENTAL ENGINEERING SCIENCES XXX CYCLE A robust multilevel approximate inverse preconditioner for symmetric

More information

Numerical Methods I Non-Square and Sparse Linear Systems

Numerical Methods I Non-Square and Sparse Linear Systems Numerical Methods I Non-Square and Sparse Linear Systems Aleksandar Donev Courant Institute, NYU 1 donev@courant.nyu.edu 1 MATH-GA 2011.003 / CSCI-GA 2945.003, Fall 2014 September 25th, 2014 A. Donev (Courant

More information

AMS526: Numerical Analysis I (Numerical Linear Algebra for Computational and Data Sciences)

AMS526: Numerical Analysis I (Numerical Linear Algebra for Computational and Data Sciences) AMS526: Numerical Analysis I (Numerical Linear Algebra for Computational and Data Sciences) Lecture 19: Computing the SVD; Sparse Linear Systems Xiangmin Jiao Stony Brook University Xiangmin Jiao Numerical

More information

Multilevel low-rank approximation preconditioners Yousef Saad Department of Computer Science and Engineering University of Minnesota

Multilevel low-rank approximation preconditioners Yousef Saad Department of Computer Science and Engineering University of Minnesota Multilevel low-rank approximation preconditioners Yousef Saad Department of Computer Science and Engineering University of Minnesota SIAM CSE Boston - March 1, 2013 First: Joint work with Ruipeng Li Work

More information

A sparse multifrontal solver using hierarchically semi-separable frontal matrices

A sparse multifrontal solver using hierarchically semi-separable frontal matrices A sparse multifrontal solver using hierarchically semi-separable frontal matrices Pieter Ghysels Lawrence Berkeley National Laboratory Joint work with: Xiaoye S. Li (LBNL), Artem Napov (ULB), François-Henry

More information

Direct and Incomplete Cholesky Factorizations with Static Supernodes

Direct and Incomplete Cholesky Factorizations with Static Supernodes Direct and Incomplete Cholesky Factorizations with Static Supernodes AMSC 661 Term Project Report Yuancheng Luo 2010-05-14 Introduction Incomplete factorizations of sparse symmetric positive definite (SSPD)

More information

Parallelization of Multilevel Preconditioners Constructed from Inverse-Based ILUs on Shared-Memory Multiprocessors

Parallelization of Multilevel Preconditioners Constructed from Inverse-Based ILUs on Shared-Memory Multiprocessors Parallelization of Multilevel Preconditioners Constructed from Inverse-Based ILUs on Shared-Memory Multiprocessors J.I. Aliaga 1 M. Bollhöfer 2 A.F. Martín 1 E.S. Quintana-Ortí 1 1 Deparment of Computer

More information

Multilevel Preconditioning of Graph-Laplacians: Polynomial Approximation of the Pivot Blocks Inverses

Multilevel Preconditioning of Graph-Laplacians: Polynomial Approximation of the Pivot Blocks Inverses Multilevel Preconditioning of Graph-Laplacians: Polynomial Approximation of the Pivot Blocks Inverses P. Boyanova 1, I. Georgiev 34, S. Margenov, L. Zikatanov 5 1 Uppsala University, Box 337, 751 05 Uppsala,

More information

The Conjugate Gradient Method

The Conjugate Gradient Method The Conjugate Gradient Method Classical Iterations We have a problem, We assume that the matrix comes from a discretization of a PDE. The best and most popular model problem is, The matrix will be as large

More information

AMS526: Numerical Analysis I (Numerical Linear Algebra) Lecture 23: GMRES and Other Krylov Subspace Methods; Preconditioning

AMS526: Numerical Analysis I (Numerical Linear Algebra) Lecture 23: GMRES and Other Krylov Subspace Methods; Preconditioning AMS526: Numerical Analysis I (Numerical Linear Algebra) Lecture 23: GMRES and Other Krylov Subspace Methods; Preconditioning Xiangmin Jiao SUNY Stony Brook Xiangmin Jiao Numerical Analysis I 1 / 18 Outline

More information

Fast algorithms for hierarchically semiseparable matrices

Fast algorithms for hierarchically semiseparable matrices NUMERICAL LINEAR ALGEBRA WITH APPLICATIONS Numer. Linear Algebra Appl. 2010; 17:953 976 Published online 22 December 2009 in Wiley Online Library (wileyonlinelibrary.com)..691 Fast algorithms for hierarchically

More information

Contents. Preface... xi. Introduction...

Contents. Preface... xi. Introduction... Contents Preface... xi Introduction... xv Chapter 1. Computer Architectures... 1 1.1. Different types of parallelism... 1 1.1.1. Overlap, concurrency and parallelism... 1 1.1.2. Temporal and spatial parallelism

More information

Robust solution of Poisson-like problems with aggregation-based AMG

Robust solution of Poisson-like problems with aggregation-based AMG Robust solution of Poisson-like problems with aggregation-based AMG Yvan Notay Université Libre de Bruxelles Service de Métrologie Nucléaire Paris, January 26, 215 Supported by the Belgian FNRS http://homepages.ulb.ac.be/

More information

Solving Large Nonlinear Sparse Systems

Solving Large Nonlinear Sparse Systems Solving Large Nonlinear Sparse Systems Fred W. Wubs and Jonas Thies Computational Mechanics & Numerical Mathematics University of Groningen, the Netherlands f.w.wubs@rug.nl Centre for Interdisciplinary

More information

EFFECTIVE AND ROBUST PRECONDITIONING OF GENERAL SPD MATRICES VIA STRUCTURED INCOMPLETE FACTORIZATION

EFFECTIVE AND ROBUST PRECONDITIONING OF GENERAL SPD MATRICES VIA STRUCTURED INCOMPLETE FACTORIZATION EFFECVE AND ROBUS PRECONDONNG OF GENERAL SPD MARCES VA SRUCURED NCOMPLEE FACORZAON JANLN XA AND ZXNG XN Abstract For general symmetric positive definite SPD matrices, we present a framework for designing

More information

Partial Left-Looking Structured Multifrontal Factorization & Algorithms for Compressed Sensing. Cinna Julie Wu

Partial Left-Looking Structured Multifrontal Factorization & Algorithms for Compressed Sensing. Cinna Julie Wu Partial Left-Looking Structured Multifrontal Factorization & Algorithms for Compressed Sensing by Cinna Julie Wu A dissertation submitted in partial satisfaction of the requirements for the degree of Doctor

More information

SOLVING SPARSE LINEAR SYSTEMS OF EQUATIONS. Chao Yang Computational Research Division Lawrence Berkeley National Laboratory Berkeley, CA, USA

SOLVING SPARSE LINEAR SYSTEMS OF EQUATIONS. Chao Yang Computational Research Division Lawrence Berkeley National Laboratory Berkeley, CA, USA 1 SOLVING SPARSE LINEAR SYSTEMS OF EQUATIONS Chao Yang Computational Research Division Lawrence Berkeley National Laboratory Berkeley, CA, USA 2 OUTLINE Sparse matrix storage format Basic factorization

More information

Enhancing Scalability of Sparse Direct Methods

Enhancing Scalability of Sparse Direct Methods Journal of Physics: Conference Series 78 (007) 0 doi:0.088/7-6596/78//0 Enhancing Scalability of Sparse Direct Methods X.S. Li, J. Demmel, L. Grigori, M. Gu, J. Xia 5, S. Jardin 6, C. Sovinec 7, L.-Q.

More information

Preconditioning Techniques Analysis for CG Method

Preconditioning Techniques Analysis for CG Method Preconditioning Techniques Analysis for CG Method Huaguang Song Department of Computer Science University of California, Davis hso@ucdavis.edu Abstract Matrix computation issue for solve linear system

More information

Linear Solvers. Andrew Hazel

Linear Solvers. Andrew Hazel Linear Solvers Andrew Hazel Introduction Thus far we have talked about the formulation and discretisation of physical problems...... and stopped when we got to a discrete linear system of equations. Introduction

More information

FINE-GRAINED PARALLEL INCOMPLETE LU FACTORIZATION

FINE-GRAINED PARALLEL INCOMPLETE LU FACTORIZATION FINE-GRAINED PARALLEL INCOMPLETE LU FACTORIZATION EDMOND CHOW AND AFTAB PATEL Abstract. This paper presents a new fine-grained parallel algorithm for computing an incomplete LU factorization. All nonzeros

More information

FINE-GRAINED PARALLEL INCOMPLETE LU FACTORIZATION

FINE-GRAINED PARALLEL INCOMPLETE LU FACTORIZATION FINE-GRAINED PARALLEL INCOMPLETE LU FACTORIZATION EDMOND CHOW AND AFTAB PATEL Abstract. This paper presents a new fine-grained parallel algorithm for computing an incomplete LU factorization. All nonzeros

More information

Multipréconditionnement adaptatif pour les méthodes de décomposition de domaine. Nicole Spillane (CNRS, CMAP, École Polytechnique)

Multipréconditionnement adaptatif pour les méthodes de décomposition de domaine. Nicole Spillane (CNRS, CMAP, École Polytechnique) Multipréconditionnement adaptatif pour les méthodes de décomposition de domaine Nicole Spillane (CNRS, CMAP, École Polytechnique) C. Bovet (ONERA), P. Gosselet (ENS Cachan), A. Parret Fréaud (SafranTech),

More information

An Efficient Solver for Sparse Linear Systems based on Rank-Structured Cholesky Factorization

An Efficient Solver for Sparse Linear Systems based on Rank-Structured Cholesky Factorization An Efficient Solver for Sparse Linear Systems based on Rank-Structured Cholesky Factorization David Bindel Department of Computer Science Cornell University 15 March 2016 (TSIMF) Rank-Structured Cholesky

More information

Low-Rank Approximations, Random Sampling and Subspace Iteration

Low-Rank Approximations, Random Sampling and Subspace Iteration Modified from Talk in 2012 For RNLA study group, March 2015 Ming Gu, UC Berkeley mgu@math.berkeley.edu Low-Rank Approximations, Random Sampling and Subspace Iteration Content! Approaches for low-rank matrix

More information

Numerical Methods in Matrix Computations

Numerical Methods in Matrix Computations Ake Bjorck Numerical Methods in Matrix Computations Springer Contents 1 Direct Methods for Linear Systems 1 1.1 Elements of Matrix Theory 1 1.1.1 Matrix Algebra 2 1.1.2 Vector Spaces 6 1.1.3 Submatrices

More information

c 2015 Society for Industrial and Applied Mathematics

c 2015 Society for Industrial and Applied Mathematics SIAM J. SCI. COMPUT. Vol. 37, No. 2, pp. C169 C193 c 2015 Society for Industrial and Applied Mathematics FINE-GRAINED PARALLEL INCOMPLETE LU FACTORIZATION EDMOND CHOW AND AFTAB PATEL Abstract. This paper

More information

Aggregation-based algebraic multigrid

Aggregation-based algebraic multigrid Aggregation-based algebraic multigrid from theory to fast solvers Yvan Notay Université Libre de Bruxelles Service de Métrologie Nucléaire CEMRACS, Marseille, July 18, 2012 Supported by the Belgian FNRS

More information

AMS Mathematics Subject Classification : 65F10,65F50. Key words and phrases: ILUS factorization, preconditioning, Schur complement, 1.

AMS Mathematics Subject Classification : 65F10,65F50. Key words and phrases: ILUS factorization, preconditioning, Schur complement, 1. J. Appl. Math. & Computing Vol. 15(2004), No. 1, pp. 299-312 BILUS: A BLOCK VERSION OF ILUS FACTORIZATION DAVOD KHOJASTEH SALKUYEH AND FAEZEH TOUTOUNIAN Abstract. ILUS factorization has many desirable

More information

FAST STRUCTURED EIGENSOLVER FOR DISCRETIZED PARTIAL DIFFERENTIAL OPERATORS ON GENERAL MESHES

FAST STRUCTURED EIGENSOLVER FOR DISCRETIZED PARTIAL DIFFERENTIAL OPERATORS ON GENERAL MESHES Proceedings of the Project Review, Geo-Mathematical Imaging Group Purdue University, West Lafayette IN, Vol. 1 2012 pp. 123-132. FAST STRUCTURED EIGENSOLVER FOR DISCRETIZED PARTIAL DIFFERENTIAL OPERATORS

More information

Course Notes: Week 1

Course Notes: Week 1 Course Notes: Week 1 Math 270C: Applied Numerical Linear Algebra 1 Lecture 1: Introduction (3/28/11) We will focus on iterative methods for solving linear systems of equations (and some discussion of eigenvalues

More information

Preconditioning Techniques for Large Linear Systems Part III: General-Purpose Algebraic Preconditioners

Preconditioning Techniques for Large Linear Systems Part III: General-Purpose Algebraic Preconditioners Preconditioning Techniques for Large Linear Systems Part III: General-Purpose Algebraic Preconditioners Michele Benzi Department of Mathematics and Computer Science Emory University Atlanta, Georgia, USA

More information

Notes on PCG for Sparse Linear Systems

Notes on PCG for Sparse Linear Systems Notes on PCG for Sparse Linear Systems Luca Bergamaschi Department of Civil Environmental and Architectural Engineering University of Padova e-mail luca.bergamaschi@unipd.it webpage www.dmsa.unipd.it/

More information

Robust Preconditioned Conjugate Gradient for the GPU and Parallel Implementations

Robust Preconditioned Conjugate Gradient for the GPU and Parallel Implementations Robust Preconditioned Conjugate Gradient for the GPU and Parallel Implementations Rohit Gupta, Martin van Gijzen, Kees Vuik GPU Technology Conference 2012, San Jose CA. GPU Technology Conference 2012,

More information

A robust inner-outer HSS preconditioner

A robust inner-outer HSS preconditioner NUMERICAL LINEAR ALGEBRA WIH APPLICAIONS Numer. Linear Algebra Appl. 2011; 00:1 0 [Version: 2002/09/18 v1.02] A robust inner-outer HSS preconditioner Jianlin Xia 1 1 Department of Mathematics, Purdue University,

More information

arxiv: v1 [cs.na] 20 Jul 2015

arxiv: v1 [cs.na] 20 Jul 2015 AN EFFICIENT SOLVER FOR SPARSE LINEAR SYSTEMS BASED ON RANK-STRUCTURED CHOLESKY FACTORIZATION JEFFREY N. CHADWICK AND DAVID S. BINDEL arxiv:1507.05593v1 [cs.na] 20 Jul 2015 Abstract. Direct factorization

More information

Master Thesis Literature Study Presentation

Master Thesis Literature Study Presentation Master Thesis Literature Study Presentation Delft University of Technology The Faculty of Electrical Engineering, Mathematics and Computer Science January 29, 2010 Plaxis Introduction Plaxis Finite Element

More information

Multilevel Low-Rank Preconditioners Yousef Saad Department of Computer Science and Engineering University of Minnesota. Modelling 2014 June 2,2014

Multilevel Low-Rank Preconditioners Yousef Saad Department of Computer Science and Engineering University of Minnesota. Modelling 2014 June 2,2014 Multilevel Low-Rank Preconditioners Yousef Saad Department of Computer Science and Engineering University of Minnesota Modelling 24 June 2,24 Dedicated to Owe Axelsson at the occasion of his 8th birthday

More information

SOLVING MESH EIGENPROBLEMS WITH MULTIGRID EFFICIENCY

SOLVING MESH EIGENPROBLEMS WITH MULTIGRID EFFICIENCY SOLVING MESH EIGENPROBLEMS WITH MULTIGRID EFFICIENCY KLAUS NEYMEYR ABSTRACT. Multigrid techniques can successfully be applied to mesh eigenvalue problems for elliptic differential operators. They allow

More information

Improvements for Implicit Linear Equation Solvers

Improvements for Implicit Linear Equation Solvers Improvements for Implicit Linear Equation Solvers Roger Grimes, Bob Lucas, Clement Weisbecker Livermore Software Technology Corporation Abstract Solving large sparse linear systems of equations is often

More information

ITERATIVE METHODS BASED ON KRYLOV SUBSPACES

ITERATIVE METHODS BASED ON KRYLOV SUBSPACES ITERATIVE METHODS BASED ON KRYLOV SUBSPACES LONG CHEN We shall present iterative methods for solving linear algebraic equation Au = b based on Krylov subspaces We derive conjugate gradient (CG) method

More information

Incomplete LU Preconditioning and Error Compensation Strategies for Sparse Matrices

Incomplete LU Preconditioning and Error Compensation Strategies for Sparse Matrices Incomplete LU Preconditioning and Error Compensation Strategies for Sparse Matrices Eun-Joo Lee Department of Computer Science, East Stroudsburg University of Pennsylvania, 327 Science and Technology Center,

More information

Lecture 18 Classical Iterative Methods

Lecture 18 Classical Iterative Methods Lecture 18 Classical Iterative Methods MIT 18.335J / 6.337J Introduction to Numerical Methods Per-Olof Persson November 14, 2006 1 Iterative Methods for Linear Systems Direct methods for solving Ax = b,

More information

Scientific Computing with Case Studies SIAM Press, Lecture Notes for Unit VII Sparse Matrix

Scientific Computing with Case Studies SIAM Press, Lecture Notes for Unit VII Sparse Matrix Scientific Computing with Case Studies SIAM Press, 2009 http://www.cs.umd.edu/users/oleary/sccswebpage Lecture Notes for Unit VII Sparse Matrix Computations Part 1: Direct Methods Dianne P. O Leary c 2008

More information

From Direct to Iterative Substructuring: some Parallel Experiences in 2 and 3D

From Direct to Iterative Substructuring: some Parallel Experiences in 2 and 3D From Direct to Iterative Substructuring: some Parallel Experiences in 2 and 3D Luc Giraud N7-IRIT, Toulouse MUMPS Day October 24, 2006, ENS-INRIA, Lyon, France Outline 1 General Framework 2 The direct

More information

The new challenges to Krylov subspace methods Yousef Saad Department of Computer Science and Engineering University of Minnesota

The new challenges to Krylov subspace methods Yousef Saad Department of Computer Science and Engineering University of Minnesota The new challenges to Krylov subspace methods Yousef Saad Department of Computer Science and Engineering University of Minnesota SIAM Applied Linear Algebra Valencia, June 18-22, 2012 Introduction Krylov

More information

Parallel Algorithms for Solution of Large Sparse Linear Systems with Applications

Parallel Algorithms for Solution of Large Sparse Linear Systems with Applications Parallel Algorithms for Solution of Large Sparse Linear Systems with Applications Murat Manguoğlu Department of Computer Engineering Middle East Technical University, Ankara, Turkey Prace workshop: HPC

More information

Iterative solution methods and their rate of convergence

Iterative solution methods and their rate of convergence Uppsala University Graduate School in Mathematics and Computing Institute for Information Technology Numerical Linear Algebra FMB and MN Fall 2007 Mandatory Assignment 3a: Iterative solution methods and

More information

FINDING PARALLELISM IN GENERAL-PURPOSE LINEAR PROGRAMMING

FINDING PARALLELISM IN GENERAL-PURPOSE LINEAR PROGRAMMING FINDING PARALLELISM IN GENERAL-PURPOSE LINEAR PROGRAMMING Daniel Thuerck 1,2 (advisors Michael Goesele 1,2 and Marc Pfetsch 1 ) Maxim Naumov 3 1 Graduate School of Computational Engineering, TU Darmstadt

More information

Lecture 9: Numerical Linear Algebra Primer (February 11st)

Lecture 9: Numerical Linear Algebra Primer (February 11st) 10-725/36-725: Convex Optimization Spring 2015 Lecture 9: Numerical Linear Algebra Primer (February 11st) Lecturer: Ryan Tibshirani Scribes: Avinash Siravuru, Guofan Wu, Maosheng Liu Note: LaTeX template

More information

On the design of parallel linear solvers for large scale problems

On the design of parallel linear solvers for large scale problems On the design of parallel linear solvers for large scale problems ICIAM - August 2015 - Mini-Symposium on Recent advances in matrix computations for extreme-scale computers M. Faverge, X. Lacoste, G. Pichon,

More information

MULTI-LAYER HIERARCHICAL STRUCTURES AND FACTORIZATIONS

MULTI-LAYER HIERARCHICAL STRUCTURES AND FACTORIZATIONS MULTI-LAYER HIERARCHICAL STRUCTURES AND FACTORIZATIONS JIANLIN XIA Abstract. We propose multi-layer hierarchically semiseparable MHS structures for the fast factorizations of dense matrices arising from

More information

Leveraging Task-Parallelism in Energy-Efficient ILU Preconditioners

Leveraging Task-Parallelism in Energy-Efficient ILU Preconditioners Leveraging Task-Parallelism in Energy-Efficient ILU Preconditioners José I. Aliaga Leveraging task-parallelism in energy-efficient ILU preconditioners Universidad Jaime I (Castellón, Spain) José I. Aliaga

More information

arxiv: v1 [hep-lat] 2 May 2012

arxiv: v1 [hep-lat] 2 May 2012 A CG Method for Multiple Right Hand Sides and Multiple Shifts in Lattice QCD Calculations arxiv:1205.0359v1 [hep-lat] 2 May 2012 Fachbereich C, Mathematik und Naturwissenschaften, Bergische Universität

More information

Jae Heon Yun and Yu Du Han

Jae Heon Yun and Yu Du Han Bull. Korean Math. Soc. 39 (2002), No. 3, pp. 495 509 MODIFIED INCOMPLETE CHOLESKY FACTORIZATION PRECONDITIONERS FOR A SYMMETRIC POSITIVE DEFINITE MATRIX Jae Heon Yun and Yu Du Han Abstract. We propose

More information

An efficient multigrid solver based on aggregation

An efficient multigrid solver based on aggregation An efficient multigrid solver based on aggregation Yvan Notay Université Libre de Bruxelles Service de Métrologie Nucléaire Graz, July 4, 2012 Co-worker: Artem Napov Supported by the Belgian FNRS http://homepages.ulb.ac.be/

More information

Chapter 7 Iterative Techniques in Matrix Algebra

Chapter 7 Iterative Techniques in Matrix Algebra Chapter 7 Iterative Techniques in Matrix Algebra Per-Olof Persson persson@berkeley.edu Department of Mathematics University of California, Berkeley Math 128B Numerical Analysis Vector Norms Definition

More information

Performance Evaluation of GPBiCGSafe Method without Reverse-Ordered Recurrence for Realistic Problems

Performance Evaluation of GPBiCGSafe Method without Reverse-Ordered Recurrence for Realistic Problems Performance Evaluation of GPBiCGSafe Method without Reverse-Ordered Recurrence for Realistic Problems Seiji Fujino, Takashi Sekimoto Abstract GPBiCG method is an attractive iterative method for the solution

More information

Jacobi-Davidson Eigensolver in Cusolver Library. Lung-Sheng Chien, NVIDIA

Jacobi-Davidson Eigensolver in Cusolver Library. Lung-Sheng Chien, NVIDIA Jacobi-Davidson Eigensolver in Cusolver Library Lung-Sheng Chien, NVIDIA lchien@nvidia.com Outline CuSolver library - cusolverdn: dense LAPACK - cusolversp: sparse LAPACK - cusolverrf: refactorization

More information

An Efficient Solver for Sparse Linear Systems based on Rank-Structured Cholesky Factorization

An Efficient Solver for Sparse Linear Systems based on Rank-Structured Cholesky Factorization An Efficient Solver for Sparse Linear Systems based on Rank-Structured Cholesky Factorization David Bindel and Jeffrey Chadwick Department of Computer Science Cornell University 30 October 2015 (Department

More information

Key words. conjugate gradients, normwise backward error, incremental norm estimation.

Key words. conjugate gradients, normwise backward error, incremental norm estimation. Proceedings of ALGORITMY 2016 pp. 323 332 ON ERROR ESTIMATION IN THE CONJUGATE GRADIENT METHOD: NORMWISE BACKWARD ERROR PETR TICHÝ Abstract. Using an idea of Duff and Vömel [BIT, 42 (2002), pp. 300 322

More information

Fine-Grained Parallel Algorithms for Incomplete Factorization Preconditioning

Fine-Grained Parallel Algorithms for Incomplete Factorization Preconditioning Fine-Grained Parallel Algorithms for Incomplete Factorization Preconditioning Edmond Chow School of Computational Science and Engineering Georgia Institute of Technology, USA SPPEXA Symposium TU München,

More information

Algorithm for Sparse Approximate Inverse Preconditioners in the Conjugate Gradient Method

Algorithm for Sparse Approximate Inverse Preconditioners in the Conjugate Gradient Method Algorithm for Sparse Approximate Inverse Preconditioners in the Conjugate Gradient Method Ilya B. Labutin A.A. Trofimuk Institute of Petroleum Geology and Geophysics SB RAS, 3, acad. Koptyug Ave., Novosibirsk

More information

Topics. The CG Algorithm Algorithmic Options CG s Two Main Convergence Theorems

Topics. The CG Algorithm Algorithmic Options CG s Two Main Convergence Theorems Topics The CG Algorithm Algorithmic Options CG s Two Main Convergence Theorems What about non-spd systems? Methods requiring small history Methods requiring large history Summary of solvers 1 / 52 Conjugate

More information

Lecture 17: Iterative Methods and Sparse Linear Algebra

Lecture 17: Iterative Methods and Sparse Linear Algebra Lecture 17: Iterative Methods and Sparse Linear Algebra David Bindel 25 Mar 2014 Logistics HW 3 extended to Wednesday after break HW 4 should come out Monday after break Still need project description

More information

OUTLINE ffl CFD: elliptic pde's! Ax = b ffl Basic iterative methods ffl Krylov subspace methods ffl Preconditioning techniques: Iterative methods ILU

OUTLINE ffl CFD: elliptic pde's! Ax = b ffl Basic iterative methods ffl Krylov subspace methods ffl Preconditioning techniques: Iterative methods ILU Preconditioning Techniques for Solving Large Sparse Linear Systems Arnold Reusken Institut für Geometrie und Praktische Mathematik RWTH-Aachen OUTLINE ffl CFD: elliptic pde's! Ax = b ffl Basic iterative

More information

Algebraic Multigrid as Solvers and as Preconditioner

Algebraic Multigrid as Solvers and as Preconditioner Ò Algebraic Multigrid as Solvers and as Preconditioner Domenico Lahaye domenico.lahaye@cs.kuleuven.ac.be http://www.cs.kuleuven.ac.be/ domenico/ Department of Computer Science Katholieke Universiteit Leuven

More information

A Robust Preconditioned Iterative Method for the Navier-Stokes Equations with High Reynolds Numbers

A Robust Preconditioned Iterative Method for the Navier-Stokes Equations with High Reynolds Numbers Applied and Computational Mathematics 2017; 6(4): 202-207 http://www.sciencepublishinggroup.com/j/acm doi: 10.11648/j.acm.20170604.18 ISSN: 2328-5605 (Print); ISSN: 2328-5613 (Online) A Robust Preconditioned

More information

A dissection solver with kernel detection for unsymmetric matrices in FreeFem++

A dissection solver with kernel detection for unsymmetric matrices in FreeFem++ . p.1/21 11 Dec. 2014, LJLL, Paris FreeFem++ workshop A dissection solver with kernel detection for unsymmetric matrices in FreeFem++ Atsushi Suzuki Atsushi.Suzuki@ann.jussieu.fr Joint work with François-Xavier

More information

A Newton-Galerkin-ADI Method for Large-Scale Algebraic Riccati Equations

A Newton-Galerkin-ADI Method for Large-Scale Algebraic Riccati Equations A Newton-Galerkin-ADI Method for Large-Scale Algebraic Riccati Equations Peter Benner Max-Planck-Institute for Dynamics of Complex Technical Systems Computational Methods in Systems and Control Theory

More information

Estimating the Largest Elements of a Matrix

Estimating the Largest Elements of a Matrix Estimating the Largest Elements of a Matrix Samuel Relton samuel.relton@manchester.ac.uk @sdrelton samrelton.com blog.samrelton.com Joint work with Nick Higham nick.higham@manchester.ac.uk May 12th, 2016

More information

The flexible incomplete LU preconditioner for large nonsymmetric linear systems. Takatoshi Nakamura Takashi Nodera

The flexible incomplete LU preconditioner for large nonsymmetric linear systems. Takatoshi Nakamura Takashi Nodera Research Report KSTS/RR-15/006 The flexible incomplete LU preconditioner for large nonsymmetric linear systems by Takatoshi Nakamura Takashi Nodera Takatoshi Nakamura School of Fundamental Science and

More information

Sparsity-Preserving Difference of Positive Semidefinite Matrix Representation of Indefinite Matrices

Sparsity-Preserving Difference of Positive Semidefinite Matrix Representation of Indefinite Matrices Sparsity-Preserving Difference of Positive Semidefinite Matrix Representation of Indefinite Matrices Jaehyun Park June 1 2016 Abstract We consider the problem of writing an arbitrary symmetric matrix as

More information

9.1 Preconditioned Krylov Subspace Methods

9.1 Preconditioned Krylov Subspace Methods Chapter 9 PRECONDITIONING 9.1 Preconditioned Krylov Subspace Methods 9.2 Preconditioned Conjugate Gradient 9.3 Preconditioned Generalized Minimal Residual 9.4 Relaxation Method Preconditioners 9.5 Incomplete

More information

Conjugate gradient method. Descent method. Conjugate search direction. Conjugate Gradient Algorithm (294)

Conjugate gradient method. Descent method. Conjugate search direction. Conjugate Gradient Algorithm (294) Conjugate gradient method Descent method Hestenes, Stiefel 1952 For A N N SPD In exact arithmetic, solves in N steps In real arithmetic No guaranteed stopping Often converges in many fewer than N steps

More information

A Cholesky LR algorithm for the positive definite symmetric diagonal-plus-semiseparable eigenproblem

A Cholesky LR algorithm for the positive definite symmetric diagonal-plus-semiseparable eigenproblem A Cholesky LR algorithm for the positive definite symmetric diagonal-plus-semiseparable eigenproblem Bor Plestenjak Department of Mathematics University of Ljubljana Slovenia Ellen Van Camp and Marc Van

More information

Fine-grained Parallel Incomplete LU Factorization

Fine-grained Parallel Incomplete LU Factorization Fine-grained Parallel Incomplete LU Factorization Edmond Chow School of Computational Science and Engineering Georgia Institute of Technology Sparse Days Meeting at CERFACS June 5-6, 2014 Contribution

More information

Preface to the Second Edition. Preface to the First Edition

Preface to the Second Edition. Preface to the First Edition n page v Preface to the Second Edition Preface to the First Edition xiii xvii 1 Background in Linear Algebra 1 1.1 Matrices................................. 1 1.2 Square Matrices and Eigenvalues....................

More information

Linear algebra issues in Interior Point methods for bound-constrained least-squares problems

Linear algebra issues in Interior Point methods for bound-constrained least-squares problems Linear algebra issues in Interior Point methods for bound-constrained least-squares problems Stefania Bellavia Dipartimento di Energetica S. Stecco Università degli Studi di Firenze Joint work with Jacek

More information

In order to solve the linear system KL M N when K is nonsymmetric, we can solve the equivalent system

In order to solve the linear system KL M N when K is nonsymmetric, we can solve the equivalent system !"#$% "&!#' (%)!#" *# %)%(! #! %)!#" +, %"!"#$ %*&%! $#&*! *# %)%! -. -/ 0 -. 12 "**3! * $!#%+,!2!#% 44" #% &#33 # 4"!#" "%! "5"#!!#6 -. - #% " 7% "3#!#3! - + 87&2! * $!#% 44" ) 3( $! # % %#!!#%+ 9332!

More information

AMS526: Numerical Analysis I (Numerical Linear Algebra)

AMS526: Numerical Analysis I (Numerical Linear Algebra) AMS526: Numerical Analysis I (Numerical Linear Algebra) Lecture 24: Preconditioning and Multigrid Solver Xiangmin Jiao SUNY Stony Brook Xiangmin Jiao Numerical Analysis I 1 / 5 Preconditioning Motivation:

More information

Iterative Methods for Solving A x = b

Iterative Methods for Solving A x = b Iterative Methods for Solving A x = b A good (free) online source for iterative methods for solving A x = b is given in the description of a set of iterative solvers called templates found at netlib: http

More information

FEM and sparse linear system solving

FEM and sparse linear system solving FEM & sparse linear system solving, Lecture 9, Nov 19, 2017 1/36 Lecture 9, Nov 17, 2017: Krylov space methods http://people.inf.ethz.ch/arbenz/fem17 Peter Arbenz Computer Science Department, ETH Zürich

More information

A SPARSE APPROXIMATE INVERSE PRECONDITIONER FOR NONSYMMETRIC LINEAR SYSTEMS

A SPARSE APPROXIMATE INVERSE PRECONDITIONER FOR NONSYMMETRIC LINEAR SYSTEMS INTERNATIONAL JOURNAL OF NUMERICAL ANALYSIS AND MODELING, SERIES B Volume 5, Number 1-2, Pages 21 30 c 2014 Institute for Scientific Computing and Information A SPARSE APPROXIMATE INVERSE PRECONDITIONER

More information

Fast matrix algebra for dense matrices with rank-deficient off-diagonal blocks

Fast matrix algebra for dense matrices with rank-deficient off-diagonal blocks CHAPTER 2 Fast matrix algebra for dense matrices with rank-deficient off-diagonal blocks Chapter summary: The chapter describes techniques for rapidly performing algebraic operations on dense matrices

More information

An efficient multigrid method for graph Laplacian systems II: robust aggregation

An efficient multigrid method for graph Laplacian systems II: robust aggregation An efficient multigrid method for graph Laplacian systems II: robust aggregation Artem Napov and Yvan Notay Service de Métrologie Nucléaire Université Libre de Bruxelles (C.P. 165/84) 50, Av. F.D. Roosevelt,

More information

FPGA Implementation of a Predictive Controller

FPGA Implementation of a Predictive Controller FPGA Implementation of a Predictive Controller SIAM Conference on Optimization 2011, Darmstadt, Germany Minisymposium on embedded optimization Juan L. Jerez, George A. Constantinides and Eric C. Kerrigan

More information

Conjugate Gradient Method

Conjugate Gradient Method Conjugate Gradient Method direct and indirect methods positive definite linear systems Krylov sequence spectral analysis of Krylov sequence preconditioning Prof. S. Boyd, EE364b, Stanford University Three

More information

Network Analysis at IIT Bombay

Network Analysis at IIT Bombay H. Narayanan Department of Electrical Engineering Indian Institute of Technology, Bombay October, 2007 Plan Speed-up of Network Analysis by exploiting Topological Methods. Adaptation of standard linear

More information

Various ways to use a second level preconditioner

Various ways to use a second level preconditioner Various ways to use a second level preconditioner C. Vuik 1, J.M. Tang 1, R. Nabben 2, and Y. Erlangga 3 1 Delft University of Technology Delft Institute of Applied Mathematics 2 Technische Universität

More information

Block Bidiagonal Decomposition and Least Squares Problems

Block Bidiagonal Decomposition and Least Squares Problems Block Bidiagonal Decomposition and Least Squares Problems Åke Björck Department of Mathematics Linköping University Perspectives in Numerical Analysis, Helsinki, May 27 29, 2008 Outline Bidiagonal Decomposition

More information

Utilisation de la compression low-rank pour réduire la complexité du solveur PaStiX

Utilisation de la compression low-rank pour réduire la complexité du solveur PaStiX Utilisation de la compression low-rank pour réduire la complexité du solveur PaStiX 26 Septembre 2018 - JCAD 2018 - Lyon Grégoire Pichon, Mathieu Faverge, Pierre Ramet, Jean Roman Outline 1. Context 2.

More information

On the Preconditioning of the Block Tridiagonal Linear System of Equations

On the Preconditioning of the Block Tridiagonal Linear System of Equations On the Preconditioning of the Block Tridiagonal Linear System of Equations Davod Khojasteh Salkuyeh Department of Mathematics, University of Mohaghegh Ardabili, PO Box 179, Ardabil, Iran E-mail: khojaste@umaacir

More information

Implementation of a preconditioned eigensolver using Hypre

Implementation of a preconditioned eigensolver using Hypre Implementation of a preconditioned eigensolver using Hypre Andrew V. Knyazev 1, and Merico E. Argentati 1 1 Department of Mathematics, University of Colorado at Denver, USA SUMMARY This paper describes

More information

An Empirical Comparison of Graph Laplacian Solvers

An Empirical Comparison of Graph Laplacian Solvers An Empirical Comparison of Graph Laplacian Solvers Kevin Deweese 1 Erik Boman 2 John Gilbert 1 1 Department of Computer Science University of California, Santa Barbara 2 Scalable Algorithms Department

More information

Structure preserving preconditioner for the incompressible Navier-Stokes equations

Structure preserving preconditioner for the incompressible Navier-Stokes equations Structure preserving preconditioner for the incompressible Navier-Stokes equations Fred W. Wubs and Jonas Thies Computational Mechanics & Numerical Mathematics University of Groningen, the Netherlands

More information

Exploiting off-diagonal rank structures in the solution of linear matrix equations

Exploiting off-diagonal rank structures in the solution of linear matrix equations Stefano Massei Exploiting off-diagonal rank structures in the solution of linear matrix equations Based on joint works with D. Kressner (EPFL), M. Mazza (IPP of Munich), D. Palitta (IDCTS of Magdeburg)

More information

IMPROVING THE PERFORMANCE OF SPARSE LU MATRIX FACTORIZATION USING A SUPERNODAL ALGORITHM

IMPROVING THE PERFORMANCE OF SPARSE LU MATRIX FACTORIZATION USING A SUPERNODAL ALGORITHM IMPROVING THE PERFORMANCE OF SPARSE LU MATRIX FACTORIZATION USING A SUPERNODAL ALGORITHM Bogdan OANCEA PhD, Associate Professor, Artife University, Bucharest, Romania E-mail: oanceab@ie.ase.ro Abstract:

More information

Preconditioners for the incompressible Navier Stokes equations

Preconditioners for the incompressible Navier Stokes equations Preconditioners for the incompressible Navier Stokes equations C. Vuik M. ur Rehman A. Segal Delft Institute of Applied Mathematics, TU Delft, The Netherlands SIAM Conference on Computational Science and

More information