Incomplete Cholesky preconditioners that exploit the low-rank property
|
|
- Rolf Doyle
- 5 years ago
- Views:
Transcription
1 ; anapov/ 1 / 35 Incomplete Cholesky preconditioners that exploit the low-rank property (theory and practice) Artem Napov Service de Métrologie Nucléaire, Université Libre de Bruxelles PRECONDITIONING (TU EINDHOVEN) JUNE 17, 2015
2 Motivation The solution of a symmetric positive definite (SPD) system Au = b may be obtained with a direct method (usually: Cholesky factorization A = R T R) robust iterative method (usually: preconditioned conjugate gradient) ; efficient if a good SPD preconditioner B A is available, i.e.: 1. Cheap (to construct, store, invert, parallelize) 2. Good approximation of A contradictory goals tradeoff Possible black box design: take an exact factorization (here, Cholesky) ; add approximation (to enforce 1.) incomplete factorization preconditioner B = R T R A. 2 / 35
3 Motivation Incomplete factorizations B = R T R A may perform approximations by dropping individual entries (mainstream) ; using low-rank approximations (emerging). Incomplete factorizations: + are (almost) black box (approximation threshold required) + are (relatively) robust may breakdown (but breakdown-free variants exist) have no guarantee to converge fast; fast convergence if B approximates well A ; that is, if κ(r T AR 1 ) = λ max(r T AR 1 ) λ min (R T AR 1 ) is small. Here we present incomplete factorizations that are breakdown-free and have controllable condition number. 3 / 35
4 4 / 35 Outline Theory... One-level variant basic approach and underlying analysis Multilevel variants extensions of analysis to multi-level setting... and practice Sparse solver motivation and design choices Numerical experiments and comparison with other solvers
5 4 / 35 Outline Theory... One-level variant basic approach and underlying analysis Multilevel variants extensions of analysis to multi-level setting... and practice Sparse solver motivation and design choices Numerical experiments and comparison with other solvers
6 5 / 35 One-level variant : basic ideas Incomplete (block left-looking) Cholesky for each block row k = 1,..., l : 1 factorize 2 solve 3 approximate 4 update (compute new rows) A ( k = 1 )
7 5 / 35 One-level variant : basic ideas Incomplete (block left-looking) Cholesky for each block row k = 1,..., l : 1 factorize 2 solve 3 approximate 4 update (compute new rows) ( k = 1 ) A 11 A 12 A 22 A
8 5 / 35 One-level variant : basic ideas Incomplete (block left-looking) Cholesky for each block row k = 1,..., l : 1 factorize 2 solve 3 approximate 4 update (compute new rows) ( k = 1 ) R 11 A 12 A 22 A 11 = R T 11 R 11 A
9 5 / 35 One-level variant : basic ideas Incomplete (block left-looking) Cholesky for each block row k = 1,..., l : 1 factorize 2 solve 3 approximate 4 update (compute new rows) ( k = 1 ) R 11 R T 11 A 12 A 22 A
10 5 / 35 One-level variant : basic ideas Incomplete (block left-looking) Cholesky for each block row k = 1,..., l : 1 factorize 2 solve 3 approximate 4 update (compute new rows) ( k = 1 ) R 12 A 22 A R 12 R 12 The dropping is orthogonal if R ( ) T R 12 R 12 R R 12 = O. R 12 R 12
11 5 / 35 One-level variant : basic ideas Incomplete (block left-looking) Cholesky for each block row k = 1,..., l : 1 factorize 2 solve 3 approximate 4 update (compute new rows) ( k = 1 ) R 12 A 22 A R 12 R 12 The dropping is orthogonal if R 12 T ( R 12 R 12 ) = O. This implies monotonicity for Schur complement (for any SPD A!) : v T (A 22 R T 12 R 12 )v v T (A 22 R T 12R 12 )v
12 One-level variant : basic ideas Incomplete (block left-looking) Cholesky for each block row k = 1,..., l : 1 factorize 2 solve 3 approximate 4 update (compute new rows) ( k = 1 ) R 12 A 22 A R 12 R 12 Example of orthogonal dropping low-rank approximation via truncated SVD (with absolute threshold tol a ) ) T + U 2 (Σ 2 V2 ) }{{} R 12 = ( U 1 U 2 ) ( Σ 1 Σ 2 ) ( V T 1 V T 2 If Σ 2 < tol a, then = U 1 (Σ 1 V T 1 ) }{{} R 12 R 12 R 12 R 12 R 12 < tol a 5 / 35
13 5 / 35 One-level variant : basic ideas Incomplete (block left-looking) Cholesky for each block row k = 1,..., l : 1 factorize 2 solve 3 approximate 4 update (compute new rows) R 12 A 22 B 1 ( k = 1 ) S 1 = A 22 1 R T 12 1 R 12
14 5 / 35 One-level variant : basic ideas Incomplete (block left-looking) Cholesky for each block row k = 1,..., l : 1 factorize 2 solve 3 approximate 4 update (compute new rows) B 1 ( k = 2 )
15 5 / 35 One-level variant : basic ideas Incomplete (block left-looking) Cholesky for each block row k = 1,..., l : 1 factorize 2 solve 3 approximate 4 update (compute new rows) B 1 ( k = 2 )
16 One-level variant : successive approximations Incomplete (block left-looking) Cholesky for each block row k = 1,..., l : 1 factorize 2 solve 3 approximate 4 update (compute new rows) ( k = 2 ) approximate new rows only One-level (model) variant k = 1 k = 2 k = 3 k = 4 6 / 35
17 7 / 35 Accuracy of individual dropping The orthogonal dropping is assumed, namely R 12 T ( R 12 R 12 ) = O R 12 A 22 B k 1 The accuracy of individual dropping at step k : (R γ k = 12 R ) 12 S 1/2 < 1, B with S B = A 22 R T 12 R 12. R 12 B k
18 Model problem Model Problem: { u u = f in Ω = (0, 1) 2 u n Full system matrix: = 0 on Ω A Ω = A I A II A I,Γ A II,Γ. Ω I Γ Ω II A T I,Γ AT II,Γ A Γ Interface system matrix: A = A Γ A T I,ΓA 1 I A Γ,I A T II,ΓA 1 II A Γ,II Corresponds to: last Schur complement in sparse Cholesky with ND ordering Schur complement system in iterative substructuring 8 / 35
19 9 / 35 Accuracy of individual dropping The orthogonal dropping is assumed, namely R 12 T ( R 12 R 12 ) = O. The accuracy of individual dropping at step k: (R γ k = 12 R ) 12 S 1/2 < 1, with B S B = A 22 R T 12 R 12. Model problem: (k = 1, block size = 10, grid, truncated SVD dropping) r γ B k 1 B k
20 One-level variant: conditioning analysis (for λ max ) Assume A SPD and that dropping is orthogonal Let B k correspond to the preconditioner at the end of step k, with A = B 0 B l = R T R is the final preconditioner. Only newly computed rows are modified; Let λ (k) ( ) max = λ max B 1 (k 1) l B k, λ max = λ max ( B 1 l B k 1 ) B k 1 B k Then λ (k) max λ (k 1) max λ (k) max + g( λ (k) max, γ k ) where g( λ, γ ) = max β>0 2 γ β λ 1 β 2 β 2 + λ 1. B l 10 / 35
21 One-level variant: conditioning analysis (for λ min ) Assume A SPD and that dropping is orthogonal Let B k correspond to the preconditioner at the end of step k, with A = B 0 B l = R T R is the final preconditioner. Only newly computed rows are modified; Let λ (k) min = λ ( ) min B 1 (k) l B k, λ min = λ ( ) min B 1 l B k 1 B k 1 B k Then λ (k) min λ(k 1) min λ (k) min g( λ(k) min, γ k ) where g( λ, γ ) = max β>0 2 γ β λ 1 β 2 β 2 + λ 1. B l 10 / 35
22 One-level variant: conditioning analysis One-level bounds (for λ max ) λ (k) max λ (k 1) max λ (k) max + g( λ (k) max, γ k ) B k 1 where λ (k) max ( ) = λ max B 1 (k 1) l B k, λ max = λ max ( B 1 l B k 1 ) The estimate is sharp: For any γ k a matrix A < 1, k = 1,..., l, there exist a sequence of approximations B k (with B 0 = A) such that the bounds for λ (0) max and λ (0) min are simultaneously reached. B k B l 11 / 35
23 Model problem: numerical experiments Model Problem: { u u = f in Ω = (0, 1) 2 u n Full system matrix: = 0 on Ω A Ω = A I A II A I,Γ A II,Γ. Ω I Γ Ω II N A T I,Γ AT II,Γ A Γ Interface system N N matrix: A = A Γ A T I,ΓA 1 I A Γ,I A T II,ΓA 1 II A Γ,II Algorithmic details: block size = 10 we consider N from 20 to 650 (n from 400 to , l from 1 to 64) maximal ranks remains below 4 12 / 35
24 13 / 35 One-level variant: numerical experiments (tol r = 10 3 ) 10 2 κ(r T AR 1 ) n condition number one-level precond. upper bound one-level
25 14 / 35 Outline Theory... One-level variant basic approach and underlying analysis Multilevel variants extensions of analysis to multi-level setting... and practice Sparse solver motivation and design choices Numerical experiments and comparison with other solvers
26 Multilevel variants : motivation One-level (model) variant k = 1 k = 2 k = 3 k = 4 O(N 3 ) cost, O(N 2 ) memory (although both cost and memory are improved!) N 15 / 35
27 Multilevel variants : motivation General \ Sequentially Semiseparable variant (SSS) [Gu, Li, Vassilevski 10] k = 1 k = 2 k = 3 k = 4 O(rN 2 ) cost, O(rN) memory r - maximal approximation rank N N/r improvement! 16 / 35
28 17 / 35 Multilevel variants : general estimate B 1 Assuming that dropping is orthogonal one has ( ) λ max B 1 k B k 1 = 1 + γk ( ) λ min B 1 k B k 1 = 1 γk B 2 and, hence, κ(r T AR 1 ) l k=1 1 + γ k 1 γ k. B 3 B 4
29 Model problem Model Problem: { u u = f in Ω = (0, 1) 2 u n Full system matrix: = 0 on Ω A Ω = A I A II A I,Γ A II,Γ. Ω I Γ Ω II N A T I,Γ AT II,Γ A Γ Interface system N N matrix: A = A Γ A T I,ΓA 1 I A Γ,I A T II,ΓA 1 II A Γ,II Algorithmic details: block size = 10 we consider N from 20 to 650 (n from 400 to , l from 1 to 64) maximal ranks remains below 4 18 / 35
30 19 / 35 Multilevel variants: numerical experiments κ(r T AR 1 ) n condition number κ(a) SSS precond. (tol r = 10 3 ) upper bound general (10 3 )
31 20 / 35 Multilevel variants: nested subspaces Sequentially Semiseparable variant (SSS) [Gu, Li, Vassilevski 10] k = 1 k = 2 k = 3 k = 4 Nested subspaces assumption R 12 R 12 R 12 R 12 B 1 B 2 B 3 B 4 span( R 12 ) span( ) span( ) span( )
32 Multilevel variants: nested subspaces Sequentially Semiseparable variant (SSS) [Gu, Li, Vassilevski 10] k = 1 k = 2 k = 3 k = 4 Nested subspaces assumption R 12 R 12 R 12 R 12 B 1 B 2 B 3 B 4 span( R 12 ) span( R 12 ) span( ) span( ) 20 / 35
33 Multilevel variants: nested subspaces Sequentially Semiseparable variant (SSS) [Gu, Li, Vassilevski 10] k = 1 k = 2 k = 3 k = 4 Nested subspaces assumption R 12 R 12 R 12 R 12 B 1 B 2 B 3 B 4 span( R 12 ) span( R 12 ) span( R 12 ) span( ) 20 / 35
34 20 / 35 Multilevel variants: nested subspaces Same bounds as for one-level case! One-level bounds (for λ max ) where λ (k) max λ (k) max λ (k 1) max ( ) = λ max B 1 (k 1) l B k, λ max Nested subspaces assumption λ (k) max + g( λ (k) max, γ k ) = λ max ( B 1 l B k 1 ) R 12 R 12 R 12 R 12 B 1 B 2 B 3 B 4 span( R 12 ) span( R 12 ) span( R 12 ) span( )
35 20 / 35 Multilevel variants: nested subspaces Same bounds as for one-level case! One-level bounds (for λ min ) where λ (k) min λ (k) min = λ min λ(k 1) min ( ) B 1 (k 1) l B k, λ min Nested subspaces assumption λ (k) min g( λ(k) min, γ k ) = λ min ( B 1 l B k 1 ) R 12 R 12 R 12 R 12 B 1 B 2 B 3 B 4 span( R 12 ) span( R 12 ) span( R 12 ) span( )
36 21 / 35 Multilevel variants: numerical experiments (tol r = 10 3 ) κ(r T AR 1 ) n condition number SSS precond. one-level precond. upper bound one-level one-level
37 22 / 35 Outline Theory... One-level variant basic approach and underlying analysis Multilevel variants extensions of analysis to multi-level setting... and practice Sparse solver motivation and design choices Numerical experiments and comparison with other solvers
38 23 / 35 Sparse solver: factorization MAIN FEATURES: symmetric left-looking block factorization accumulate updates before block row computation nested dissection (ND) ordering induces block structure, enforces sparsity, and reduces operation count orthogonal low-rank approximations by truncated SVD or rank-revealing QR
39 23 / 35 Sparse solver: factorization MAIN FEATURES: symmetric left-looking block factorization accumulate updates before block row computation nested dissection (ND) ordering induces block structure, enforces sparsity, and reduces operation count orthogonal low-rank approximations by truncated SVD or rank-revealing QR Γ 3 Γ 2 Ω 2 Ω 3 Ω 4 Ω 1 Γ 1 } } } } } } Ω 1 Ω 2 Γ 1 Ω 3 Ω 4 Γ 2 Γ 3
40 23 / 35 Sparse solver: factorization MAIN FEATURES: symmetric left-looking block factorization accumulate updates before block row computation nested dissection (ND) ordering induces block structure, enforces sparsity, and reduces operation count orthogonal low-rank approximations by truncated SVD or rank-revealing QR
41 24 / 35 Sparse solver: factorization SPECIAL FEATURES: rank-revealing reordering inside ND blocks ensures the algebraic character of the solver symbolic compression ensures that memory usage decreases adaptive block size faster than fixed block size
42 25 / 35 Sparse solver: revealing low-rank Model problem analysis [Chandrasekaran et al., 10]: rank is bounded proportionally to the number of connections between the corresp. subset and the remaining nodes of separator Implemented: recursive edge bisection (via METIS) 2D model separators:
43 Sparse solver: revealing low-rank Model problem analysis [Chandrasekaran et al., 10]: rank is bounded proportionally to the number of connections between the corresp. subset and the remaining nodes of separator Implemented: recursive edge bisection (via METIS) 2D model separators: 25 / 35
44 Sparse solver: revealing low-rank Model problem analysis [Chandrasekaran et al., 10]: rank is bounded proportionally to the number of connections between the corresp. subset and the remaining nodes of separator Implemented: recursive edge bisection (via METIS) 2D model separators: } } 25 / 35
45 Sparse solver: revealing low-rank Model problem analysis [Chandrasekaran et al., 10]: rank is bounded proportionally to the number of connections between the corresp. subset and the remaining nodes of separator Implemented: recursive edge bisection (via METIS) 2D model separators: } } } } } } 25 / 35
46 Sparse solver: revealing low-rank Model problem analysis [Chandrasekaran et al., 10]: rank is bounded proportionally to the number of connections between the corresp. subset and the remaining nodes of separator Implemented: recursive edge bisection (via METIS) 2D separators from Scotch: connections of length 2 25 / 35
47 Sparse solver: revealing low-rank Model problem analysis [Chandrasekaran et al., 10]: rank is bounded proportionally to the number of connections between the corresp. subset and the remaining nodes of separator Implemented: recursive edge bisection (via METIS) 2D separators from Scotch: connections of length 2 25 / 35
48 Sparse solver: revealing low-rank Model problem analysis [Chandrasekaran et al., 10]: rank is bounded proportionally to the number of connections between the corresp. subset and the remaining nodes of separator Implemented: recursive edge bisection (via METIS) 3D model separators: 25 / 35
49 Sparse solver: revealing low-rank Model problem analysis [Chandrasekaran et al., 10]: rank is bounded proportionally to the number of connections between the corresp. subset and the remaining nodes of separator Implemented: recursive edge bisection (via METIS) 3D model separators: } } } } } } 25 / 35
50 26 / 35 Sparse solver: symbolic compression If for a low-rank approximation R 12 R 12 Q one has ( nnz R 12 ) nnz( R 12 ) + nnz( Q ), then the low-rank approximation is replaced by R 12 = I R 12. This low-rank approximation is (trivially) orthogonal. No need to store I does not increase memory use.
51 27 / 35 Sparse solver: adaptive block size Via symbolic compression (s.c.), block subdivision can evolve: } } } } }s.c. }s.c. } } } } We use this evolution: by automatically applying s.c. after few occurrences at the same level; (more successive s.c. occur, (exponentially) more s.c. are skipped to save computations); to adjust the minimal block size for the following separators (this also saves computations).
52 28 / 35 Outline Theory... One-level variant basic approach and underlying analysis Multilevel variants extensions of analysis to multi-level setting... and practice Sparse solver motivation and design choices Numerical experiments and comparison with other solvers
53 Numerical experiments: setting Solver parameters: the preconditioner, denoted by SIC, is used with approximation based on rank-revealing QR with column pivoting and absolute approximation threshold ; adaptive block size (initially block size is set to 16); PCG as outer iteration. Test problems: all SPD matrices from University of Florida Sparse Matrix Collection 1 with n > 10 5 and random rhs (excluded: thermomech_tk and bmw7st_1, both have κ(a) ). Experimental setting: time reported for best threshold value (chosen among {10 10, 10 9,..., }) stopping criterion : 10 6 relative residual decrease; hardware : Intel Xeon L5420, 2.5 GHz, 16 GB RAM / 35
54 30 / 35 Comparison with (exact) Cholesky 100 chol SIC time (sec.) per million nnz problems by nnz
55 31 / 35 Comparison with (unpreconditioned) CG 100 CG SIC time (sec.) per million nnz problems by nnz
56 32 / 35 Comparison with ILUPACK ILUPACK SIC time (sec.) per million nnz problems by nnz
57 33 / 35 Conclusions In theory... We have presented a conditioning analysis for incomplete Cholesky factorizations preconditioner based on orthogonal dropping Only requirement on A: should be SPD The analysis relates the condition number of the preconditioned system to the individual dropping accuracies (namely, to γ k s) The analysis is sharp in the one-level case One-level bound can be extended to the multilevel setting if an additional nested subspace assumption holds; it naturally holds for the presented preconditioner
58 34 / 35 Conclusions... and in practice We have presented a preliminary implementation of incomplete Cholesky factorizations preconditioner based on orthogonal dropping The solver targets at sparse matrices; it uses sparsity structure to reduce the block rank during the factorization Preliminary numerical experiments demonstrate that the solver is competitive
59 Further details Theory: A. Napov, Conditioning analysis of incomplete Cholesky factorizations with orthogonal dropping, SIAM J. Matrix Anal. Appl. A. Napov, Conditioning analysis of incomplete Cholesky factorizations with orthogonal dropping II: nested subspaces (in preparation) Related approaches (for dense matrices): M. Gu, X. S. Li, P. Vassilevski, Direction-preserving and Schur-monotonic semiseparable approximations of symmetric positive definite matrices, SIAM J. Matrix Anal. Appl. J. Xia, M. Gu, Robust approximate Cholesky factorization of rank-structured symmetric positive definite matrices, SIAM J. Matrix Anal. Appl. Rank revealing based on sparsity: A. Napov, X. S. Li, An algebraic multifrontal preconditioner that exploits the low-rank property, Numer. Lin. Alg. Appl. (to appear). 35 / 35
J.I. Aliaga 1 M. Bollhöfer 2 A.F. Martín 1 E.S. Quintana-Ortí 1. March, 2009
Parallel Preconditioning of Linear Systems based on ILUPACK for Multithreaded Architectures J.I. Aliaga M. Bollhöfer 2 A.F. Martín E.S. Quintana-Ortí Deparment of Computer Science and Engineering, Univ.
More informationA robust multilevel approximate inverse preconditioner for symmetric positive definite matrices
DICEA DEPARTMENT OF CIVIL, ENVIRONMENTAL AND ARCHITECTURAL ENGINEERING PhD SCHOOL CIVIL AND ENVIRONMENTAL ENGINEERING SCIENCES XXX CYCLE A robust multilevel approximate inverse preconditioner for symmetric
More informationNumerical Methods I Non-Square and Sparse Linear Systems
Numerical Methods I Non-Square and Sparse Linear Systems Aleksandar Donev Courant Institute, NYU 1 donev@courant.nyu.edu 1 MATH-GA 2011.003 / CSCI-GA 2945.003, Fall 2014 September 25th, 2014 A. Donev (Courant
More informationAMS526: Numerical Analysis I (Numerical Linear Algebra for Computational and Data Sciences)
AMS526: Numerical Analysis I (Numerical Linear Algebra for Computational and Data Sciences) Lecture 19: Computing the SVD; Sparse Linear Systems Xiangmin Jiao Stony Brook University Xiangmin Jiao Numerical
More informationMultilevel low-rank approximation preconditioners Yousef Saad Department of Computer Science and Engineering University of Minnesota
Multilevel low-rank approximation preconditioners Yousef Saad Department of Computer Science and Engineering University of Minnesota SIAM CSE Boston - March 1, 2013 First: Joint work with Ruipeng Li Work
More informationA sparse multifrontal solver using hierarchically semi-separable frontal matrices
A sparse multifrontal solver using hierarchically semi-separable frontal matrices Pieter Ghysels Lawrence Berkeley National Laboratory Joint work with: Xiaoye S. Li (LBNL), Artem Napov (ULB), François-Henry
More informationDirect and Incomplete Cholesky Factorizations with Static Supernodes
Direct and Incomplete Cholesky Factorizations with Static Supernodes AMSC 661 Term Project Report Yuancheng Luo 2010-05-14 Introduction Incomplete factorizations of sparse symmetric positive definite (SSPD)
More informationParallelization of Multilevel Preconditioners Constructed from Inverse-Based ILUs on Shared-Memory Multiprocessors
Parallelization of Multilevel Preconditioners Constructed from Inverse-Based ILUs on Shared-Memory Multiprocessors J.I. Aliaga 1 M. Bollhöfer 2 A.F. Martín 1 E.S. Quintana-Ortí 1 1 Deparment of Computer
More informationMultilevel Preconditioning of Graph-Laplacians: Polynomial Approximation of the Pivot Blocks Inverses
Multilevel Preconditioning of Graph-Laplacians: Polynomial Approximation of the Pivot Blocks Inverses P. Boyanova 1, I. Georgiev 34, S. Margenov, L. Zikatanov 5 1 Uppsala University, Box 337, 751 05 Uppsala,
More informationThe Conjugate Gradient Method
The Conjugate Gradient Method Classical Iterations We have a problem, We assume that the matrix comes from a discretization of a PDE. The best and most popular model problem is, The matrix will be as large
More informationAMS526: Numerical Analysis I (Numerical Linear Algebra) Lecture 23: GMRES and Other Krylov Subspace Methods; Preconditioning
AMS526: Numerical Analysis I (Numerical Linear Algebra) Lecture 23: GMRES and Other Krylov Subspace Methods; Preconditioning Xiangmin Jiao SUNY Stony Brook Xiangmin Jiao Numerical Analysis I 1 / 18 Outline
More informationFast algorithms for hierarchically semiseparable matrices
NUMERICAL LINEAR ALGEBRA WITH APPLICATIONS Numer. Linear Algebra Appl. 2010; 17:953 976 Published online 22 December 2009 in Wiley Online Library (wileyonlinelibrary.com)..691 Fast algorithms for hierarchically
More informationContents. Preface... xi. Introduction...
Contents Preface... xi Introduction... xv Chapter 1. Computer Architectures... 1 1.1. Different types of parallelism... 1 1.1.1. Overlap, concurrency and parallelism... 1 1.1.2. Temporal and spatial parallelism
More informationRobust solution of Poisson-like problems with aggregation-based AMG
Robust solution of Poisson-like problems with aggregation-based AMG Yvan Notay Université Libre de Bruxelles Service de Métrologie Nucléaire Paris, January 26, 215 Supported by the Belgian FNRS http://homepages.ulb.ac.be/
More informationSolving Large Nonlinear Sparse Systems
Solving Large Nonlinear Sparse Systems Fred W. Wubs and Jonas Thies Computational Mechanics & Numerical Mathematics University of Groningen, the Netherlands f.w.wubs@rug.nl Centre for Interdisciplinary
More informationEFFECTIVE AND ROBUST PRECONDITIONING OF GENERAL SPD MATRICES VIA STRUCTURED INCOMPLETE FACTORIZATION
EFFECVE AND ROBUS PRECONDONNG OF GENERAL SPD MARCES VA SRUCURED NCOMPLEE FACORZAON JANLN XA AND ZXNG XN Abstract For general symmetric positive definite SPD matrices, we present a framework for designing
More informationPartial Left-Looking Structured Multifrontal Factorization & Algorithms for Compressed Sensing. Cinna Julie Wu
Partial Left-Looking Structured Multifrontal Factorization & Algorithms for Compressed Sensing by Cinna Julie Wu A dissertation submitted in partial satisfaction of the requirements for the degree of Doctor
More informationSOLVING SPARSE LINEAR SYSTEMS OF EQUATIONS. Chao Yang Computational Research Division Lawrence Berkeley National Laboratory Berkeley, CA, USA
1 SOLVING SPARSE LINEAR SYSTEMS OF EQUATIONS Chao Yang Computational Research Division Lawrence Berkeley National Laboratory Berkeley, CA, USA 2 OUTLINE Sparse matrix storage format Basic factorization
More informationEnhancing Scalability of Sparse Direct Methods
Journal of Physics: Conference Series 78 (007) 0 doi:0.088/7-6596/78//0 Enhancing Scalability of Sparse Direct Methods X.S. Li, J. Demmel, L. Grigori, M. Gu, J. Xia 5, S. Jardin 6, C. Sovinec 7, L.-Q.
More informationPreconditioning Techniques Analysis for CG Method
Preconditioning Techniques Analysis for CG Method Huaguang Song Department of Computer Science University of California, Davis hso@ucdavis.edu Abstract Matrix computation issue for solve linear system
More informationLinear Solvers. Andrew Hazel
Linear Solvers Andrew Hazel Introduction Thus far we have talked about the formulation and discretisation of physical problems...... and stopped when we got to a discrete linear system of equations. Introduction
More informationFINE-GRAINED PARALLEL INCOMPLETE LU FACTORIZATION
FINE-GRAINED PARALLEL INCOMPLETE LU FACTORIZATION EDMOND CHOW AND AFTAB PATEL Abstract. This paper presents a new fine-grained parallel algorithm for computing an incomplete LU factorization. All nonzeros
More informationFINE-GRAINED PARALLEL INCOMPLETE LU FACTORIZATION
FINE-GRAINED PARALLEL INCOMPLETE LU FACTORIZATION EDMOND CHOW AND AFTAB PATEL Abstract. This paper presents a new fine-grained parallel algorithm for computing an incomplete LU factorization. All nonzeros
More informationMultipréconditionnement adaptatif pour les méthodes de décomposition de domaine. Nicole Spillane (CNRS, CMAP, École Polytechnique)
Multipréconditionnement adaptatif pour les méthodes de décomposition de domaine Nicole Spillane (CNRS, CMAP, École Polytechnique) C. Bovet (ONERA), P. Gosselet (ENS Cachan), A. Parret Fréaud (SafranTech),
More informationAn Efficient Solver for Sparse Linear Systems based on Rank-Structured Cholesky Factorization
An Efficient Solver for Sparse Linear Systems based on Rank-Structured Cholesky Factorization David Bindel Department of Computer Science Cornell University 15 March 2016 (TSIMF) Rank-Structured Cholesky
More informationLow-Rank Approximations, Random Sampling and Subspace Iteration
Modified from Talk in 2012 For RNLA study group, March 2015 Ming Gu, UC Berkeley mgu@math.berkeley.edu Low-Rank Approximations, Random Sampling and Subspace Iteration Content! Approaches for low-rank matrix
More informationNumerical Methods in Matrix Computations
Ake Bjorck Numerical Methods in Matrix Computations Springer Contents 1 Direct Methods for Linear Systems 1 1.1 Elements of Matrix Theory 1 1.1.1 Matrix Algebra 2 1.1.2 Vector Spaces 6 1.1.3 Submatrices
More informationc 2015 Society for Industrial and Applied Mathematics
SIAM J. SCI. COMPUT. Vol. 37, No. 2, pp. C169 C193 c 2015 Society for Industrial and Applied Mathematics FINE-GRAINED PARALLEL INCOMPLETE LU FACTORIZATION EDMOND CHOW AND AFTAB PATEL Abstract. This paper
More informationAggregation-based algebraic multigrid
Aggregation-based algebraic multigrid from theory to fast solvers Yvan Notay Université Libre de Bruxelles Service de Métrologie Nucléaire CEMRACS, Marseille, July 18, 2012 Supported by the Belgian FNRS
More informationAMS Mathematics Subject Classification : 65F10,65F50. Key words and phrases: ILUS factorization, preconditioning, Schur complement, 1.
J. Appl. Math. & Computing Vol. 15(2004), No. 1, pp. 299-312 BILUS: A BLOCK VERSION OF ILUS FACTORIZATION DAVOD KHOJASTEH SALKUYEH AND FAEZEH TOUTOUNIAN Abstract. ILUS factorization has many desirable
More informationFAST STRUCTURED EIGENSOLVER FOR DISCRETIZED PARTIAL DIFFERENTIAL OPERATORS ON GENERAL MESHES
Proceedings of the Project Review, Geo-Mathematical Imaging Group Purdue University, West Lafayette IN, Vol. 1 2012 pp. 123-132. FAST STRUCTURED EIGENSOLVER FOR DISCRETIZED PARTIAL DIFFERENTIAL OPERATORS
More informationCourse Notes: Week 1
Course Notes: Week 1 Math 270C: Applied Numerical Linear Algebra 1 Lecture 1: Introduction (3/28/11) We will focus on iterative methods for solving linear systems of equations (and some discussion of eigenvalues
More informationPreconditioning Techniques for Large Linear Systems Part III: General-Purpose Algebraic Preconditioners
Preconditioning Techniques for Large Linear Systems Part III: General-Purpose Algebraic Preconditioners Michele Benzi Department of Mathematics and Computer Science Emory University Atlanta, Georgia, USA
More informationNotes on PCG for Sparse Linear Systems
Notes on PCG for Sparse Linear Systems Luca Bergamaschi Department of Civil Environmental and Architectural Engineering University of Padova e-mail luca.bergamaschi@unipd.it webpage www.dmsa.unipd.it/
More informationRobust Preconditioned Conjugate Gradient for the GPU and Parallel Implementations
Robust Preconditioned Conjugate Gradient for the GPU and Parallel Implementations Rohit Gupta, Martin van Gijzen, Kees Vuik GPU Technology Conference 2012, San Jose CA. GPU Technology Conference 2012,
More informationA robust inner-outer HSS preconditioner
NUMERICAL LINEAR ALGEBRA WIH APPLICAIONS Numer. Linear Algebra Appl. 2011; 00:1 0 [Version: 2002/09/18 v1.02] A robust inner-outer HSS preconditioner Jianlin Xia 1 1 Department of Mathematics, Purdue University,
More informationarxiv: v1 [cs.na] 20 Jul 2015
AN EFFICIENT SOLVER FOR SPARSE LINEAR SYSTEMS BASED ON RANK-STRUCTURED CHOLESKY FACTORIZATION JEFFREY N. CHADWICK AND DAVID S. BINDEL arxiv:1507.05593v1 [cs.na] 20 Jul 2015 Abstract. Direct factorization
More informationMaster Thesis Literature Study Presentation
Master Thesis Literature Study Presentation Delft University of Technology The Faculty of Electrical Engineering, Mathematics and Computer Science January 29, 2010 Plaxis Introduction Plaxis Finite Element
More informationMultilevel Low-Rank Preconditioners Yousef Saad Department of Computer Science and Engineering University of Minnesota. Modelling 2014 June 2,2014
Multilevel Low-Rank Preconditioners Yousef Saad Department of Computer Science and Engineering University of Minnesota Modelling 24 June 2,24 Dedicated to Owe Axelsson at the occasion of his 8th birthday
More informationSOLVING MESH EIGENPROBLEMS WITH MULTIGRID EFFICIENCY
SOLVING MESH EIGENPROBLEMS WITH MULTIGRID EFFICIENCY KLAUS NEYMEYR ABSTRACT. Multigrid techniques can successfully be applied to mesh eigenvalue problems for elliptic differential operators. They allow
More informationImprovements for Implicit Linear Equation Solvers
Improvements for Implicit Linear Equation Solvers Roger Grimes, Bob Lucas, Clement Weisbecker Livermore Software Technology Corporation Abstract Solving large sparse linear systems of equations is often
More informationITERATIVE METHODS BASED ON KRYLOV SUBSPACES
ITERATIVE METHODS BASED ON KRYLOV SUBSPACES LONG CHEN We shall present iterative methods for solving linear algebraic equation Au = b based on Krylov subspaces We derive conjugate gradient (CG) method
More informationIncomplete LU Preconditioning and Error Compensation Strategies for Sparse Matrices
Incomplete LU Preconditioning and Error Compensation Strategies for Sparse Matrices Eun-Joo Lee Department of Computer Science, East Stroudsburg University of Pennsylvania, 327 Science and Technology Center,
More informationLecture 18 Classical Iterative Methods
Lecture 18 Classical Iterative Methods MIT 18.335J / 6.337J Introduction to Numerical Methods Per-Olof Persson November 14, 2006 1 Iterative Methods for Linear Systems Direct methods for solving Ax = b,
More informationScientific Computing with Case Studies SIAM Press, Lecture Notes for Unit VII Sparse Matrix
Scientific Computing with Case Studies SIAM Press, 2009 http://www.cs.umd.edu/users/oleary/sccswebpage Lecture Notes for Unit VII Sparse Matrix Computations Part 1: Direct Methods Dianne P. O Leary c 2008
More informationFrom Direct to Iterative Substructuring: some Parallel Experiences in 2 and 3D
From Direct to Iterative Substructuring: some Parallel Experiences in 2 and 3D Luc Giraud N7-IRIT, Toulouse MUMPS Day October 24, 2006, ENS-INRIA, Lyon, France Outline 1 General Framework 2 The direct
More informationThe new challenges to Krylov subspace methods Yousef Saad Department of Computer Science and Engineering University of Minnesota
The new challenges to Krylov subspace methods Yousef Saad Department of Computer Science and Engineering University of Minnesota SIAM Applied Linear Algebra Valencia, June 18-22, 2012 Introduction Krylov
More informationParallel Algorithms for Solution of Large Sparse Linear Systems with Applications
Parallel Algorithms for Solution of Large Sparse Linear Systems with Applications Murat Manguoğlu Department of Computer Engineering Middle East Technical University, Ankara, Turkey Prace workshop: HPC
More informationIterative solution methods and their rate of convergence
Uppsala University Graduate School in Mathematics and Computing Institute for Information Technology Numerical Linear Algebra FMB and MN Fall 2007 Mandatory Assignment 3a: Iterative solution methods and
More informationFINDING PARALLELISM IN GENERAL-PURPOSE LINEAR PROGRAMMING
FINDING PARALLELISM IN GENERAL-PURPOSE LINEAR PROGRAMMING Daniel Thuerck 1,2 (advisors Michael Goesele 1,2 and Marc Pfetsch 1 ) Maxim Naumov 3 1 Graduate School of Computational Engineering, TU Darmstadt
More informationLecture 9: Numerical Linear Algebra Primer (February 11st)
10-725/36-725: Convex Optimization Spring 2015 Lecture 9: Numerical Linear Algebra Primer (February 11st) Lecturer: Ryan Tibshirani Scribes: Avinash Siravuru, Guofan Wu, Maosheng Liu Note: LaTeX template
More informationOn the design of parallel linear solvers for large scale problems
On the design of parallel linear solvers for large scale problems ICIAM - August 2015 - Mini-Symposium on Recent advances in matrix computations for extreme-scale computers M. Faverge, X. Lacoste, G. Pichon,
More informationMULTI-LAYER HIERARCHICAL STRUCTURES AND FACTORIZATIONS
MULTI-LAYER HIERARCHICAL STRUCTURES AND FACTORIZATIONS JIANLIN XIA Abstract. We propose multi-layer hierarchically semiseparable MHS structures for the fast factorizations of dense matrices arising from
More informationLeveraging Task-Parallelism in Energy-Efficient ILU Preconditioners
Leveraging Task-Parallelism in Energy-Efficient ILU Preconditioners José I. Aliaga Leveraging task-parallelism in energy-efficient ILU preconditioners Universidad Jaime I (Castellón, Spain) José I. Aliaga
More informationarxiv: v1 [hep-lat] 2 May 2012
A CG Method for Multiple Right Hand Sides and Multiple Shifts in Lattice QCD Calculations arxiv:1205.0359v1 [hep-lat] 2 May 2012 Fachbereich C, Mathematik und Naturwissenschaften, Bergische Universität
More informationJae Heon Yun and Yu Du Han
Bull. Korean Math. Soc. 39 (2002), No. 3, pp. 495 509 MODIFIED INCOMPLETE CHOLESKY FACTORIZATION PRECONDITIONERS FOR A SYMMETRIC POSITIVE DEFINITE MATRIX Jae Heon Yun and Yu Du Han Abstract. We propose
More informationAn efficient multigrid solver based on aggregation
An efficient multigrid solver based on aggregation Yvan Notay Université Libre de Bruxelles Service de Métrologie Nucléaire Graz, July 4, 2012 Co-worker: Artem Napov Supported by the Belgian FNRS http://homepages.ulb.ac.be/
More informationChapter 7 Iterative Techniques in Matrix Algebra
Chapter 7 Iterative Techniques in Matrix Algebra Per-Olof Persson persson@berkeley.edu Department of Mathematics University of California, Berkeley Math 128B Numerical Analysis Vector Norms Definition
More informationPerformance Evaluation of GPBiCGSafe Method without Reverse-Ordered Recurrence for Realistic Problems
Performance Evaluation of GPBiCGSafe Method without Reverse-Ordered Recurrence for Realistic Problems Seiji Fujino, Takashi Sekimoto Abstract GPBiCG method is an attractive iterative method for the solution
More informationJacobi-Davidson Eigensolver in Cusolver Library. Lung-Sheng Chien, NVIDIA
Jacobi-Davidson Eigensolver in Cusolver Library Lung-Sheng Chien, NVIDIA lchien@nvidia.com Outline CuSolver library - cusolverdn: dense LAPACK - cusolversp: sparse LAPACK - cusolverrf: refactorization
More informationAn Efficient Solver for Sparse Linear Systems based on Rank-Structured Cholesky Factorization
An Efficient Solver for Sparse Linear Systems based on Rank-Structured Cholesky Factorization David Bindel and Jeffrey Chadwick Department of Computer Science Cornell University 30 October 2015 (Department
More informationKey words. conjugate gradients, normwise backward error, incremental norm estimation.
Proceedings of ALGORITMY 2016 pp. 323 332 ON ERROR ESTIMATION IN THE CONJUGATE GRADIENT METHOD: NORMWISE BACKWARD ERROR PETR TICHÝ Abstract. Using an idea of Duff and Vömel [BIT, 42 (2002), pp. 300 322
More informationFine-Grained Parallel Algorithms for Incomplete Factorization Preconditioning
Fine-Grained Parallel Algorithms for Incomplete Factorization Preconditioning Edmond Chow School of Computational Science and Engineering Georgia Institute of Technology, USA SPPEXA Symposium TU München,
More informationAlgorithm for Sparse Approximate Inverse Preconditioners in the Conjugate Gradient Method
Algorithm for Sparse Approximate Inverse Preconditioners in the Conjugate Gradient Method Ilya B. Labutin A.A. Trofimuk Institute of Petroleum Geology and Geophysics SB RAS, 3, acad. Koptyug Ave., Novosibirsk
More informationTopics. The CG Algorithm Algorithmic Options CG s Two Main Convergence Theorems
Topics The CG Algorithm Algorithmic Options CG s Two Main Convergence Theorems What about non-spd systems? Methods requiring small history Methods requiring large history Summary of solvers 1 / 52 Conjugate
More informationLecture 17: Iterative Methods and Sparse Linear Algebra
Lecture 17: Iterative Methods and Sparse Linear Algebra David Bindel 25 Mar 2014 Logistics HW 3 extended to Wednesday after break HW 4 should come out Monday after break Still need project description
More informationOUTLINE ffl CFD: elliptic pde's! Ax = b ffl Basic iterative methods ffl Krylov subspace methods ffl Preconditioning techniques: Iterative methods ILU
Preconditioning Techniques for Solving Large Sparse Linear Systems Arnold Reusken Institut für Geometrie und Praktische Mathematik RWTH-Aachen OUTLINE ffl CFD: elliptic pde's! Ax = b ffl Basic iterative
More informationAlgebraic Multigrid as Solvers and as Preconditioner
Ò Algebraic Multigrid as Solvers and as Preconditioner Domenico Lahaye domenico.lahaye@cs.kuleuven.ac.be http://www.cs.kuleuven.ac.be/ domenico/ Department of Computer Science Katholieke Universiteit Leuven
More informationA Robust Preconditioned Iterative Method for the Navier-Stokes Equations with High Reynolds Numbers
Applied and Computational Mathematics 2017; 6(4): 202-207 http://www.sciencepublishinggroup.com/j/acm doi: 10.11648/j.acm.20170604.18 ISSN: 2328-5605 (Print); ISSN: 2328-5613 (Online) A Robust Preconditioned
More informationA dissection solver with kernel detection for unsymmetric matrices in FreeFem++
. p.1/21 11 Dec. 2014, LJLL, Paris FreeFem++ workshop A dissection solver with kernel detection for unsymmetric matrices in FreeFem++ Atsushi Suzuki Atsushi.Suzuki@ann.jussieu.fr Joint work with François-Xavier
More informationA Newton-Galerkin-ADI Method for Large-Scale Algebraic Riccati Equations
A Newton-Galerkin-ADI Method for Large-Scale Algebraic Riccati Equations Peter Benner Max-Planck-Institute for Dynamics of Complex Technical Systems Computational Methods in Systems and Control Theory
More informationEstimating the Largest Elements of a Matrix
Estimating the Largest Elements of a Matrix Samuel Relton samuel.relton@manchester.ac.uk @sdrelton samrelton.com blog.samrelton.com Joint work with Nick Higham nick.higham@manchester.ac.uk May 12th, 2016
More informationThe flexible incomplete LU preconditioner for large nonsymmetric linear systems. Takatoshi Nakamura Takashi Nodera
Research Report KSTS/RR-15/006 The flexible incomplete LU preconditioner for large nonsymmetric linear systems by Takatoshi Nakamura Takashi Nodera Takatoshi Nakamura School of Fundamental Science and
More informationSparsity-Preserving Difference of Positive Semidefinite Matrix Representation of Indefinite Matrices
Sparsity-Preserving Difference of Positive Semidefinite Matrix Representation of Indefinite Matrices Jaehyun Park June 1 2016 Abstract We consider the problem of writing an arbitrary symmetric matrix as
More information9.1 Preconditioned Krylov Subspace Methods
Chapter 9 PRECONDITIONING 9.1 Preconditioned Krylov Subspace Methods 9.2 Preconditioned Conjugate Gradient 9.3 Preconditioned Generalized Minimal Residual 9.4 Relaxation Method Preconditioners 9.5 Incomplete
More informationConjugate gradient method. Descent method. Conjugate search direction. Conjugate Gradient Algorithm (294)
Conjugate gradient method Descent method Hestenes, Stiefel 1952 For A N N SPD In exact arithmetic, solves in N steps In real arithmetic No guaranteed stopping Often converges in many fewer than N steps
More informationA Cholesky LR algorithm for the positive definite symmetric diagonal-plus-semiseparable eigenproblem
A Cholesky LR algorithm for the positive definite symmetric diagonal-plus-semiseparable eigenproblem Bor Plestenjak Department of Mathematics University of Ljubljana Slovenia Ellen Van Camp and Marc Van
More informationFine-grained Parallel Incomplete LU Factorization
Fine-grained Parallel Incomplete LU Factorization Edmond Chow School of Computational Science and Engineering Georgia Institute of Technology Sparse Days Meeting at CERFACS June 5-6, 2014 Contribution
More informationPreface to the Second Edition. Preface to the First Edition
n page v Preface to the Second Edition Preface to the First Edition xiii xvii 1 Background in Linear Algebra 1 1.1 Matrices................................. 1 1.2 Square Matrices and Eigenvalues....................
More informationLinear algebra issues in Interior Point methods for bound-constrained least-squares problems
Linear algebra issues in Interior Point methods for bound-constrained least-squares problems Stefania Bellavia Dipartimento di Energetica S. Stecco Università degli Studi di Firenze Joint work with Jacek
More informationIn order to solve the linear system KL M N when K is nonsymmetric, we can solve the equivalent system
!"#$% "&!#' (%)!#" *# %)%(! #! %)!#" +, %"!"#$ %*&%! $#&*! *# %)%! -. -/ 0 -. 12 "**3! * $!#%+,!2!#% 44" #% ! # 4"!#" "%! "5"#!!#6 -. - #% " 7% "3#!#3! - + 87&2! * $!#% 44" ) 3( $! # % %#!!#%+ 9332!
More informationAMS526: Numerical Analysis I (Numerical Linear Algebra)
AMS526: Numerical Analysis I (Numerical Linear Algebra) Lecture 24: Preconditioning and Multigrid Solver Xiangmin Jiao SUNY Stony Brook Xiangmin Jiao Numerical Analysis I 1 / 5 Preconditioning Motivation:
More informationIterative Methods for Solving A x = b
Iterative Methods for Solving A x = b A good (free) online source for iterative methods for solving A x = b is given in the description of a set of iterative solvers called templates found at netlib: http
More informationFEM and sparse linear system solving
FEM & sparse linear system solving, Lecture 9, Nov 19, 2017 1/36 Lecture 9, Nov 17, 2017: Krylov space methods http://people.inf.ethz.ch/arbenz/fem17 Peter Arbenz Computer Science Department, ETH Zürich
More informationA SPARSE APPROXIMATE INVERSE PRECONDITIONER FOR NONSYMMETRIC LINEAR SYSTEMS
INTERNATIONAL JOURNAL OF NUMERICAL ANALYSIS AND MODELING, SERIES B Volume 5, Number 1-2, Pages 21 30 c 2014 Institute for Scientific Computing and Information A SPARSE APPROXIMATE INVERSE PRECONDITIONER
More informationFast matrix algebra for dense matrices with rank-deficient off-diagonal blocks
CHAPTER 2 Fast matrix algebra for dense matrices with rank-deficient off-diagonal blocks Chapter summary: The chapter describes techniques for rapidly performing algebraic operations on dense matrices
More informationAn efficient multigrid method for graph Laplacian systems II: robust aggregation
An efficient multigrid method for graph Laplacian systems II: robust aggregation Artem Napov and Yvan Notay Service de Métrologie Nucléaire Université Libre de Bruxelles (C.P. 165/84) 50, Av. F.D. Roosevelt,
More informationFPGA Implementation of a Predictive Controller
FPGA Implementation of a Predictive Controller SIAM Conference on Optimization 2011, Darmstadt, Germany Minisymposium on embedded optimization Juan L. Jerez, George A. Constantinides and Eric C. Kerrigan
More informationConjugate Gradient Method
Conjugate Gradient Method direct and indirect methods positive definite linear systems Krylov sequence spectral analysis of Krylov sequence preconditioning Prof. S. Boyd, EE364b, Stanford University Three
More informationNetwork Analysis at IIT Bombay
H. Narayanan Department of Electrical Engineering Indian Institute of Technology, Bombay October, 2007 Plan Speed-up of Network Analysis by exploiting Topological Methods. Adaptation of standard linear
More informationVarious ways to use a second level preconditioner
Various ways to use a second level preconditioner C. Vuik 1, J.M. Tang 1, R. Nabben 2, and Y. Erlangga 3 1 Delft University of Technology Delft Institute of Applied Mathematics 2 Technische Universität
More informationBlock Bidiagonal Decomposition and Least Squares Problems
Block Bidiagonal Decomposition and Least Squares Problems Åke Björck Department of Mathematics Linköping University Perspectives in Numerical Analysis, Helsinki, May 27 29, 2008 Outline Bidiagonal Decomposition
More informationUtilisation de la compression low-rank pour réduire la complexité du solveur PaStiX
Utilisation de la compression low-rank pour réduire la complexité du solveur PaStiX 26 Septembre 2018 - JCAD 2018 - Lyon Grégoire Pichon, Mathieu Faverge, Pierre Ramet, Jean Roman Outline 1. Context 2.
More informationOn the Preconditioning of the Block Tridiagonal Linear System of Equations
On the Preconditioning of the Block Tridiagonal Linear System of Equations Davod Khojasteh Salkuyeh Department of Mathematics, University of Mohaghegh Ardabili, PO Box 179, Ardabil, Iran E-mail: khojaste@umaacir
More informationImplementation of a preconditioned eigensolver using Hypre
Implementation of a preconditioned eigensolver using Hypre Andrew V. Knyazev 1, and Merico E. Argentati 1 1 Department of Mathematics, University of Colorado at Denver, USA SUMMARY This paper describes
More informationAn Empirical Comparison of Graph Laplacian Solvers
An Empirical Comparison of Graph Laplacian Solvers Kevin Deweese 1 Erik Boman 2 John Gilbert 1 1 Department of Computer Science University of California, Santa Barbara 2 Scalable Algorithms Department
More informationStructure preserving preconditioner for the incompressible Navier-Stokes equations
Structure preserving preconditioner for the incompressible Navier-Stokes equations Fred W. Wubs and Jonas Thies Computational Mechanics & Numerical Mathematics University of Groningen, the Netherlands
More informationExploiting off-diagonal rank structures in the solution of linear matrix equations
Stefano Massei Exploiting off-diagonal rank structures in the solution of linear matrix equations Based on joint works with D. Kressner (EPFL), M. Mazza (IPP of Munich), D. Palitta (IDCTS of Magdeburg)
More informationIMPROVING THE PERFORMANCE OF SPARSE LU MATRIX FACTORIZATION USING A SUPERNODAL ALGORITHM
IMPROVING THE PERFORMANCE OF SPARSE LU MATRIX FACTORIZATION USING A SUPERNODAL ALGORITHM Bogdan OANCEA PhD, Associate Professor, Artife University, Bucharest, Romania E-mail: oanceab@ie.ase.ro Abstract:
More informationPreconditioners for the incompressible Navier Stokes equations
Preconditioners for the incompressible Navier Stokes equations C. Vuik M. ur Rehman A. Segal Delft Institute of Applied Mathematics, TU Delft, The Netherlands SIAM Conference on Computational Science and
More information