Accelerating the spike family of algorithms by solving linear systems with multiple right-hand sides

Size: px
Start display at page:

Download "Accelerating the spike family of algorithms by solving linear systems with multiple right-hand sides"

Transcription

1 Accelerating the spike family of algorithms by solving linear systems with multiple right-hand sides Dept. of Computer Engineering & Informatics University of Patras Graduate Program in Computer Science & Technology July 30, 2014

2 Outline 1 Introduction 2 Parallel solution of banded linear systems - the Spike method 3 Solvers for linear systems with multiple right-hand sides 4 Numerical experiments 5 Bibliography

3 Goal and motivation I Important problem in NLA Efficient solution of AX = F where A R n n is a sparse matrix and F R n s is the matrix of right-hand sides. Applications Numerous... Modeling of heat propagation Weather forecasting Fluid dynamics Quantum chromodynamics (QCD) Simulation of electronic circuits Analysis of civil structures

4 Goal and motivation II The problem is very important... Numerous methods based on different assumptions... In this thesis Goal: To extend the Spike framework by using techniques for the solution of linear systems with multiple right-hand sides.

5 Characteristics of the problem The coefficient matrix A can be: A general sparse matrix nnz(a) = O(n) Banded with semi-bandwidth m (if m n nnz(a) = O(n)). The banded coefficient matrix can be: nz = nz = 167 Figure : Dense within the band Figure : Sparse within the band We focus in symmetric and positive definite (SPD) matrices.

6 Outline 1 Introduction 2 Parallel solution of banded linear systems - the Spike method 3 Solvers for linear systems with multiple right-hand sides 4 Numerical experiments 5 Bibliography

7

8 The Spike algorithm I The Spike [SamPol11,PolSam06] algorithm is a parallel algorithm for the solution of linear systems of the form AX = F with banded A of semi-bandwidth m, m n. Banded matrix Block tridiagonal matrix It can be divided in two phases Pre-processing Post-processing

9 An example for p = 4 partitions A 1 B 1 C 2 A 2 B 2 C 3 A 3 B 3 C 4 A 4 X 1 X 2 X 3 X 4 = F 1 F 2 F 3 F 4 A X F A j R nj nj, F j R nj s, B j, C j R m m

10 The Spike algorithm II A = DS Matrix A is factorized as A = DS (1) where D R n n is block diagonal and S R n n block tridiagonal. A1 B1 C2 A2 B2 C3 A3 B3 C4 A4 = A1 A2 A3 A4 I1 V1 W2 I2 V2 W3 I3 V3 W4 I4 A D S It holds that V j, W j R n/p m.

11 The Spike algorithm III Pre-processing stage Solution of the local linear systems A j[v j, W j] = [ˆBj, Ĉ j], j = 1,..., p. (2) Notation: ˆBj = [0 (n m) m ; B j], Ĉ j = [C j; 0 (n m) m ] Post-processing stage Solution of A jg j = F j, j = 1,..., p. (3) and SX = G. (4) Solution of (3) is performed locally in each processor. Solution of (4) demands communication between neighboring processors.

12 Solution of SX = G I Partitioning of V j, W j, X j, G j V j = V (1) j V ( ) j V (q) j, Wj = W (1) j W ( ) j W (q) j where V (1) j, V ( ) j, V (q) j and W (1), W ( ) j, W (q) j represent the first m, middle n j 2m and last m rows of V j, W j respectively. We assume the same partition for X j, G j as well. S rx r = G r, example for p = 3 I V (q) I 0 V (1) I V (q) W (1) 3 I W (1) W (q) X (q) 1 X (1) 2 X (q) 2 X (1) 3 = G (q) 1 G (1) 2 G (q) 2 G (1) 3. (5)

13 Solution of SX = G II Intermediate blocks of rows are computed locally as X 1 = G 1 V 1 X (q) 2 j = 1 X j = G j V j X (q) j+1 W j X (1) j 1 2 j p 1 X p = G p W px (1) p 1 j = p

14 Outline 1 Introduction 2 Parallel solution of banded linear systems - the Spike method 3 Solvers for linear systems with multiple right-hand sides 4 Numerical experiments 5 Bibliography

15 The Spike algorithm and the solvers for mrhs I Local linear systems in Spike algorithm The Spike algorithm leads to the solution of a linear system of the form A j[v j, W j, G j] = [ˆBj, Ĉ j, F j]. (6) Each local linear system contains 2m + s right-hand sides Obvious approach: Each right-hand side is solved on its own Solution of each right-hand side by direct methods (by factorizing A j) Solution of each right-hand side by using iterative methods

16 Solution methods Direct methods By LU decomposition CC: O(n 3 j ) Sparse structure? Pardiso, MUMPS, SuperLu... Banded structure? CC: O(n jm 2 ) By Cholesky decomposition P ja j = L ju j. (7) A j = L jl j. (8) Iterative methods Stationary point methods Krylov subspace methods

17 Krylov subspace iterative methods Solution of A jx = b Solution is sought in the Krylov subspace, If q(a j) = α 0I n + α 1A j α nj A nj j In the i-th iteration K µ(a j, b) = span{b, A jb,..., A µ 1 j b}. = 0, then A 1 j b = 1 nj 1 i=0 α i+1a i j. (9) α 0 x i x 0 + K i, r i = (b A jx i) C i. Different choices for subspace C i: C i K i or C i A jk i.

18 The Spike algorithm and the mrhs solvers II But matrix A j is the same for each right-hand side Direct methods Factorization is performed only once, P ja j = L ju j More right-hand sides better overhead amortization Iterative methods Can we employ something similar?

19 The Spike algorithm and the mrhs solvers III Goal: Reduce the solution cost of J linear systems How: Multiple Right Hand Sides Solvers Cost MRH SOLVER K J # rhs Ideally (J K) J

20 The Spike algorithm and the mrhs solvers IV A case example Suppose solving the following linear system, A[x (1), x (2) ] = [f (1), f (2) ]. First approach: solution of each Ax (i) = f (i), i = 1, 2 independently (one by one) A very simple concept but... if f (2) = γf (1) then x (2) = γx (1)!

21 The Spike algorithm and the mrhs solvers IV A case example Suppose solving the following linear system, A[x (1), x (2) ] = [f (1), f (2) ]. First approach: solution of each Ax (i) = f (i), i = 1, 2 independently (one by one) A very simple concept but... if f (2) = γf (1) then x (2) = γx (1)! MRHs Solvers: Categories The above example, although trivial, shows the path Right-hand sides can be co-linear, orthogonal... Two main classes, a) seed methods and b) block methods

22 Seed methods for the solution of mrhs problems I Seed methods - brief description Based on an "artificial" augmentation of the search subspace Solution of the first right-hand side returns an orthonormal basis V (1) i of K i(a, r (1) 0 ). Then with ( x (1,j) 0 = x (1,j) 0 + V (1) i ( T (1) i = V (1) i Procedure continues till all mrhs are solved T (1) i ) 1 ( V (1) i The j-th right-hand side is projected on the subspaces Krylov K i1 (A, r (1,1) 0 ),..., K ij 1 (A, r (j 2,j 1) 0 ) ) (1,j) r 0 (10) ) (1) AV i. (11)

23 Seed methods for the solution of mrhs problems II Algorithm 1 Generic algorithm of seed methods (seed) [SPM89,Saad87,SimG95] Input: A R n n, F R n s, X 0 R n s, tol Output: X R n s for j = 1,..., s do for i = 1,... do the j-th linear system performs the i-th iteration for g = j + 1,..., s do update solution of x (j,g) i end for end for end for

24 Seed methods for the solution of mrhs problems III Algorithm 2 The Single-Seed Conjugate Gradient algorithm (sscg) [ChanWan97] Input: A R n n, F R n s, X 0 R n s, tol R Output: X R n s R 0 = F AX 0 P 0 = R 0 for j = 1,..., s do while r (j) i / f (j) tol do α = ((r (j) i ) r (j) i )/((p (j) i ) Ap (j) x (j) = i+1 x(j) i + p (j) i α; r (j) = i+1 r(j) i β = ((r (j) i+1 ) r (j) i+1 )/((r(j) i ) r (j) i ) p (j) = i+1 r(j) + i+1 p(j) i β i ) Ap (j) i for k = j + 1,..., s do η (j,k) i = (p (j) i ) r (j,k) i /(p (j) i ) Ap (j) i x (j,k) i+1 = x (j,k) i + η (j,k) i p (j) i ; r (j,k) i+1 = r (j,k) i end for end while end for α η (j,k) i Ap (j) i

25 Seed methods for the solution of mrhs problems IV Theorem (ChanWan97) Let F = [f (1),..., f (s) ] and suppose that rank(f) = k, k < s. Then, there exists α, independent of the number of steps m p, such that for the residual of the non-solved rhs we have r (j,k) 0 α j h=1 βmp+1, k = j + 1,..., s, where value β mp+1 stems from Lanczos if we consider that the h-th rhs is solved by the Lanczos method. Proof. See [ChanWan97].

26 Block methods for the solution of the mrhs problem Block methods [Olea80,Gut09,SimG96a,SimG96b] A generalization of the classic approach. In step i, solution of the j-th right-hand side is sought in the following Krylov subspace, K B µ(a, R 0) = span{r 0, AR 0,..., A µ 1 R 0}, and x (j) i x (j) 0 + K B i (A, R 0). Dimension of the search subspace is at most s times bigger than previously, K B µ(a, R 0) K µ(a, r (1) 0 ) K µ(a, r (s) 0 ). Guaranteed to converge in at most n/s steps.

27 Numerical solution of a Laplacian PDE - discretized by finite differences n x = n y = n z = 15 AX = F, s = 50 rel. residual q=1 q=10 q=25 q= index of iterations Figure : Norm of relative residual for different blocksizes

28 An approach based on matrix inversion I For an SPD matrix H we have (H 1 ) ij Cλ i j, i, j = 1,..., n with λ = ( ) 2/m κ(h) 1. κ(h) + 1 Spikes V j, W j mult/ion of A 1 j with B j, C j We can approximate submatrices of A 1 j Can we take advantage of it? by lower rank matrices

29 An approach based on matrix inversion II Express A j as a 2 2 block matrix A j = [ (1) A j A (2) ] j A (3) j A (4) j (12) where A (1) j R (nj z) (nj z), A (2) j R (nj z) z, A (3) j R z (nj z), A (4) j R z z and z m, then A 1 j = [ (A (1) j A (2) j (A (4) j ) 1 A (3) j ) 1 A (2) j (A (4) j ) 1 (A (4) j ) 1 + (A (4) j ) 1 A (3) j (A (1) j A (2) j (A (4) j ) 1 A (3) j ) 1 A (2) j (A (4) j ) 1 ]

30 An approach based on matrix inversion III Spike V j can be computed as A (4) j Y (1) j = Bj (13) S (4) A Y (2) j = A (2) j Y (1) j (14) j A (4) j Y (3) j = A (3) j Y (2) j (15) where S (4) A = (A (1) j A (2) j (A (4) j ) 1 A (3) j ) is the Schur complement of A (4) j and by Bj we j denote the (upwards) extension of B j by z m null rows.

31 An approach based on matrix inversion IV In the same manner, we can express A j as a 2 2 block matrix [ (1) A j A (2) ] j A j = A (3) j A (4) j (16) where A (1) j R z z, A (2) j R z (nj z), A (3) j R (nj z) z, A (4) j R (nj z) (nj z). Then, ] A 1 j = [ (A (1) j ) 1 + (A (1) j ) 1 A (2) j (A (4) j A (3) j (A (1) j ) 1 A (2) j ) 1 A (3) j (A (1) j ) 1 (A (4) j A (3) j (A (1) j ) 1 A (2) j ) 1 A (3) j (A (1) j ) 1

32 An approach based on matrix inversion V Spike W j can be computed as A (1) j Y (1) j = Cj (17) S (1) A Y (2) j = A (3) j Y (1) j (18) j A (1) j Y (3) j = A (2) j Y (2) j (19) where S (1) A = (A (4) j A (3) j (A (1) j ) 1 A (2) j ) is the Schur complement of A (1) j and by Cj we j denote the (downwards) extension of C j by z m null rows.

33 Algorithm 3 The sc-spike algorithm Input : A j R n n, B j R m m, C j R m m, z Z Output : V j R n m, W j R n m {Each processor j:} {compute V j (if j < p):} A (4) j Y (1) j = ˆBj Y (2) j = proj(a (1) j A (2) j (A (4) j ) 1 A (3) j, A (2) j Y (1) j ) A (4) j Y (3) j = A (3) j Y (2) j {compute W j (if j > 1):} A (1) j Y (1) j = Ĉ j Y (2) j = proj(a (4) j A (3) j (A (1) j ) 1 A (2) j, A (3) j Y (1) j ) A (1) j Y (3) j = A (2) j Y (2) j Solve A jg j = F j by an iterative method Solve SX = G

34 Algorithm 4 proj Input: K R n n, L R n s Output: Y R n s 1. [Q, R] = QR(L) 2. l = {i : l (i) belongs in basis of L} 3. Solve KY l = L l 4. D = L l (L l L l ) 1 (L l L / l ) 5. Y / l = Y l D

35 Outline 1 Introduction 2 Parallel solution of banded linear systems - the Spike method 3 Solvers for linear systems with multiple right-hand sides 4 Numerical experiments 5 Bibliography

36 Computational system and software Computational system Experiments were performed in Purdue University 2 Inter(R) Xeon(R) processors with a total of 12 cores in 2.8 GHz 2 GBs RAM per core Software Fortran90 Intel(R) MPI Library for Linux, 64-bit, Version 4.0 mpiifort Fortran MPI compiler

37 Numerical experiments with banded matrices I About In this set of experiments we study the sc-spike algorithm Matrices were selected from the University of Florida Sparse Matrix Collection (UFSC) [DavHu11] We also present experiments with Toeplitz matrices

38 Numerical experiments with banded matrices II Table : Characteristics of the matrices used (UFSC) A n nnz(a) semi-bandwidth (m) Application 1. BenElechi1 245,874 13M 821 2D/3D 2. kim2 456, M D/3D 3. af_5_k , M 859 Structural problem 4. mc2depi 525, M 513 2D/3D

39 Matrices from UFSC I p=2 p= p=2 p=4 Numerical rank Numerical rank z x 10 4 (a) BenElechi z x 10 4 (b) mc2depi Figure : Rank of submatrix (V 1) 1:n 1:m j z (p = 2, 4) for the BenElechi1 and mc2depi matrices.

40 Matrices from UFSC II p=3 p= p=3 p=4 Numerical rank Numerical rank z x 10 4 (a) af_5_k z x 10 4 (b) kim2 Figure : Rank of submatrix (V 1) 1:m 1:n j z (p = 2, 4) for the shipsec5 and kim2 matrices.

41 Toeplitz I a ij = randn(1), a ii = δ n i=1,i j a ij numerical rank m=100 m=200 m=300 m=400 numerical rank m=100 m=200 m=300 m= z x 10 4 (a) δ = 1/ z x 10 4 (b) δ = 1 + Figure : Rank of submatrix (V 1) 1:m 1:n j z for a Toeplitz matrix with varying m and degree of diagonal dominance.

42 Toeplitz II a ij = 1/ i j, a ii = δ n i=1,i j a ij numerical rank m=100 m=200 m=300 m=400 numerical rank m=100 m=200 m=300 m= z x 10 4 (a) δ = 1/ z x 10 4 (b) δ = 1 + Figure : Rank of submatrix (V 1) 1:m 1:n j z for a Toeplitz matrix with varying m and degree of diagonal dominance.

43 Numerical experiments with banded matrices III About Matrices were selected from UFSC F = rand(n, 10) List of algorithms Alg1: All multiple right-hand sides are solved by BCG Alg2: All multiple right-hand sides are solved by BCG, preconditioned by IC(0) Cholesky: All multiple right-hand sides are solved by Cholesky

44 Numerical experiments with banded matrices IV Table : Running times of spike algorithm for varying number of partitions (s = 10). p = 2 p = 4 p = 8 Cholesky Alg1 Alg2 Cholesky Alg1 Alg2 Cholesky Alg1 Alg2 bcsstk fv t2dah_e crystm wathen wathen jnbrng apache F F F 105.2

45 The Spike algorithm as a preconditioner The Spike algorithm can be used as a banded preconditioner M We need M F A F We need pre-processing of A... Reordering? RCM, Fiedler? Weighted Fiedler (WSO) [Mang10]?

46 Introduction Parallel solution of banded linear systems - the Spike method Solvers for linear systems with multiple right-hand sides Numerical experiments Bibliography Reordering of sparse matrices I 4 0 x (a) No reordering (b) RCM nz = x 10 (c) WSO Figure : Non-zero pattern before and after RCM and WSO permutations for matrix opf10000

47 Reordering of sparse matrices II (a) No reordering (b) RCM (c) WSO Figure : Non-zero pattern before and after RCM and WSO permutations for matrix nd3k

48 The PSpike algorithm Algorithmic sketch 1 Reordering of the coefficient matrix A 2 Extraction of a banded preconditioner M with semi-bandwidth ˆm 3 Solve M 1 (PAP ) X = M 1 F by a Krylov subspacec method. Preconditioner is applied through Spike. 4 Form final solution, X = P X.

49 Numerical experiments with general sparse matrices About Matrices were selected from UFSC Different reordering methods Extraction of a banded preconditioner M with semi-bandwidth ˆm 1 (ˆm 2) such that M F 0.95 A F ( M F 0.99 A F) Reordering methods RCM Fiedler WSO

50 Table : Characteristics of the matrices selected from UFSC n Application 1. bcsstk Structural problem 2. eurqsa 7245 Structural problem 3. OPF_ Power electricity 4. dubcova D/3D 5. raefsky Structural problem

51 Table : Original semi-bandwidth m and semi-bandwidths ˆm 1, ˆm 2 so that M F 0.95 A F and M F 0.99 A F after symmetric reorderings RCM Fiedler WSO m ˆm 1 ˆm 2 m ˆm 1 ˆm 2 m ˆm 1 ˆm 2 bcsstk eurqsa OPF_ dubcova raefsky

52 numerical value the singular values of spike V 1 symrcm Fiedler WSO numerical value 3.5 x the singular values of spike V 1 symrcm Fiedler WSO index of singular values (a) raefsky4 (ˆm 1) index of singular values (b) raefsky4 (ˆm 2)

53 numerical value the singular values of spike V 1 symrcm Fiedler WSO numerical value the singular values of spike V 1 symrcm Fiedler WSO index of singular values (c) dubcova1 (ˆm 1) index of singular values (d) dubcova1 (ˆm 2)

54 5 x the singular values of spike V 1 symrcm Fiedler WSO 6 x the singular values of spike V 1 symrcm Fiedler WSO numerical value 3 2 numerical value index of singular values (e) bcsstk28 (ˆm 1) index of singular values (f) bcsstk28 (ˆm 2)

55 2.5 x the singular values of spike V 1 symrcm Fiedler WSO 2.5 x the singular values of spike V 1 symrcm Fiedler WSO numerical value numerical value index of singular values index of singular values (g) eurqsa (ˆm 1) (h) eurqsa (ˆm 2)

56 2.5 3 x 104 the singular values of spike V 1 symrcm Fiedler WSO x 104 the singular values of spike V 1 symrcm Fiedler WSO numerical value numerical value index of singular values index of singular values (i) OPF_3574 (ˆm 1) (j) OPF_3574 (ˆm 2)

57 Talks and manuscripts Talks NUMAN 2012, September 2012, Ioannina, Greece Manuscripts VK, F. Saied, E. Galllopoulos and A. Sameh. Accelerating Spike family of algorithms by solving linear systems with multiple right-hand sides, Technical Report

58 Acknowledgements PurdueU A. Klinvex, A. Sameh, F. Saied, Y. Zhu, G. Kollias UPatras E. Gallopoulos, E. Kontopoulou, C. Zaroliagis, A. Boudouvis

59 Conclusion In this thesis We studied the combination of Spike algorithms with techniques for solving linear systems with multiple right-hand sides We saw that under certain conditions, replacing direct solvers by these techniques can lead to more efficient schemes In future we seek to develop schemes for accelerating the Spike algorithm when not all multiple right-hand sides are available at once

60 Outline 1 Introduction 2 Parallel solution of banded linear systems - the Spike method 3 Solvers for linear systems with multiple right-hand sides 4 Numerical experiments 5 Bibliography

61 Bibliography I T. Chan and L. Wan, Analysis of projection methods for solving linear systems with multiple right-hand sides. SIAM J. Sci. Comput., 18(6), pp , T. A. Davis and Y. Hu, The university of Florida sparse matrix collection. ACM Trans. Math. Soft., 38(1), pp. 1-25, D. P. O Leary, The block conjugate gradient algorithms and related methods. Linear Algebra and its Applications, 29, pp , F. Zhang, Schur complement and its applications. Springer, E. Polizzi and A. Sameh, A parallel hybrid banded system solver: the SPIKE algorithm. Parallel Comput., 32(2), pp , 2006, Elsevier Science Publishers B. V. M. Manguoglu, M. Koyutürk, A. Sameh and A. Grama, Weighted Matrix Ordering and Parallel Banded Preconditioners for Iterative Linear System Solvers. SIAM J. Scientific Computing, 32(3), pp , M. Gutknecht and T. Schmelzer, The block grade of a Krylov subspace. Linear Alg. and its Appl., 430(11), pp , 2009.

62 Bibliography II C. F. Smith, A. F. Peterson and R. Mittra, A conjugate gradient algorithm for the treatment of multiple incident electromagnetic fields. IEEE Trans. Antennas and Propagation, 37, pp , V. Simoncini and E. Gallopoulos, An iterative method for nonsymmetric systems with multiple right-hand sides SIAM J. Sci. Comput, 16, pp , Y. Saad, On the Lanczos method for solving symmetric linear systems with multiple right-hand sides. Math. Comp., 48, pp , E. Gallopoulos, S. Hatzimihail and V. Kalantzis, Towards Lightweight Projection Methods for Linear Systems with Multiple Right-Hand Sides. Midwest Numerical Analysis Days 2011, presentation. V. Simoncini and E. Gallopoulos, Convergence properties of block GMRES and matrix polynomials. Linear Alg. and its Appl., 247, pp , V. Simoncini and E. Gallopoulos, A hybrid block GMRES method for nonsymmetric systems with multiple right-hand sides. J. of Comput. and Applied Mathematics, 66(1), pp , 1996.

AMS526: Numerical Analysis I (Numerical Linear Algebra for Computational and Data Sciences)

AMS526: Numerical Analysis I (Numerical Linear Algebra for Computational and Data Sciences) AMS526: Numerical Analysis I (Numerical Linear Algebra for Computational and Data Sciences) Lecture 19: Computing the SVD; Sparse Linear Systems Xiangmin Jiao Stony Brook University Xiangmin Jiao Numerical

More information

Parallel Algorithms for Solution of Large Sparse Linear Systems with Applications

Parallel Algorithms for Solution of Large Sparse Linear Systems with Applications Parallel Algorithms for Solution of Large Sparse Linear Systems with Applications Murat Manguoğlu Department of Computer Engineering Middle East Technical University, Ankara, Turkey Prace workshop: HPC

More information

arxiv: v1 [hep-lat] 2 May 2012

arxiv: v1 [hep-lat] 2 May 2012 A CG Method for Multiple Right Hand Sides and Multiple Shifts in Lattice QCD Calculations arxiv:1205.0359v1 [hep-lat] 2 May 2012 Fachbereich C, Mathematik und Naturwissenschaften, Bergische Universität

More information

AMS526: Numerical Analysis I (Numerical Linear Algebra) Lecture 23: GMRES and Other Krylov Subspace Methods; Preconditioning

AMS526: Numerical Analysis I (Numerical Linear Algebra) Lecture 23: GMRES and Other Krylov Subspace Methods; Preconditioning AMS526: Numerical Analysis I (Numerical Linear Algebra) Lecture 23: GMRES and Other Krylov Subspace Methods; Preconditioning Xiangmin Jiao SUNY Stony Brook Xiangmin Jiao Numerical Analysis I 1 / 18 Outline

More information

Sparse Matrix Computations in Arterial Fluid Mechanics

Sparse Matrix Computations in Arterial Fluid Mechanics Sparse Matrix Computations in Arterial Fluid Mechanics Murat Manguoğlu Middle East Technical University, Turkey Kenji Takizawa Ahmed Sameh Tayfun Tezduyar Waseda University, Japan Purdue University, USA

More information

Algebra C Numerical Linear Algebra Sample Exam Problems

Algebra C Numerical Linear Algebra Sample Exam Problems Algebra C Numerical Linear Algebra Sample Exam Problems Notation. Denote by V a finite-dimensional Hilbert space with inner product (, ) and corresponding norm. The abbreviation SPD is used for symmetric

More information

Incomplete Cholesky preconditioners that exploit the low-rank property

Incomplete Cholesky preconditioners that exploit the low-rank property anapov@ulb.ac.be ; http://homepages.ulb.ac.be/ anapov/ 1 / 35 Incomplete Cholesky preconditioners that exploit the low-rank property (theory and practice) Artem Napov Service de Métrologie Nucléaire, Université

More information

SOLVING SPARSE LINEAR SYSTEMS OF EQUATIONS. Chao Yang Computational Research Division Lawrence Berkeley National Laboratory Berkeley, CA, USA

SOLVING SPARSE LINEAR SYSTEMS OF EQUATIONS. Chao Yang Computational Research Division Lawrence Berkeley National Laboratory Berkeley, CA, USA 1 SOLVING SPARSE LINEAR SYSTEMS OF EQUATIONS Chao Yang Computational Research Division Lawrence Berkeley National Laboratory Berkeley, CA, USA 2 OUTLINE Sparse matrix storage format Basic factorization

More information

The Lanczos and conjugate gradient algorithms

The Lanczos and conjugate gradient algorithms The Lanczos and conjugate gradient algorithms Gérard MEURANT October, 2008 1 The Lanczos algorithm 2 The Lanczos algorithm in finite precision 3 The nonsymmetric Lanczos algorithm 4 The Golub Kahan bidiagonalization

More information

Contents. Preface... xi. Introduction...

Contents. Preface... xi. Introduction... Contents Preface... xi Introduction... xv Chapter 1. Computer Architectures... 1 1.1. Different types of parallelism... 1 1.1.1. Overlap, concurrency and parallelism... 1 1.1.2. Temporal and spatial parallelism

More information

Block Bidiagonal Decomposition and Least Squares Problems

Block Bidiagonal Decomposition and Least Squares Problems Block Bidiagonal Decomposition and Least Squares Problems Åke Björck Department of Mathematics Linköping University Perspectives in Numerical Analysis, Helsinki, May 27 29, 2008 Outline Bidiagonal Decomposition

More information

Research Article MINRES Seed Projection Methods for Solving Symmetric Linear Systems with Multiple Right-Hand Sides

Research Article MINRES Seed Projection Methods for Solving Symmetric Linear Systems with Multiple Right-Hand Sides Mathematical Problems in Engineering, Article ID 357874, 6 pages http://dx.doi.org/10.1155/2014/357874 Research Article MINRES Seed Projection Methods for Solving Symmetric Linear Systems with Multiple

More information

Preface to the Second Edition. Preface to the First Edition

Preface to the Second Edition. Preface to the First Edition n page v Preface to the Second Edition Preface to the First Edition xiii xvii 1 Background in Linear Algebra 1 1.1 Matrices................................. 1 1.2 Square Matrices and Eigenvalues....................

More information

Key words. conjugate gradients, normwise backward error, incremental norm estimation.

Key words. conjugate gradients, normwise backward error, incremental norm estimation. Proceedings of ALGORITMY 2016 pp. 323 332 ON ERROR ESTIMATION IN THE CONJUGATE GRADIENT METHOD: NORMWISE BACKWARD ERROR PETR TICHÝ Abstract. Using an idea of Duff and Vömel [BIT, 42 (2002), pp. 300 322

More information

Linear Solvers. Andrew Hazel

Linear Solvers. Andrew Hazel Linear Solvers Andrew Hazel Introduction Thus far we have talked about the formulation and discretisation of physical problems...... and stopped when we got to a discrete linear system of equations. Introduction

More information

On the Preconditioning of the Block Tridiagonal Linear System of Equations

On the Preconditioning of the Block Tridiagonal Linear System of Equations On the Preconditioning of the Block Tridiagonal Linear System of Equations Davod Khojasteh Salkuyeh Department of Mathematics, University of Mohaghegh Ardabili, PO Box 179, Ardabil, Iran E-mail: khojaste@umaacir

More information

Lecture 8: Fast Linear Solvers (Part 7)

Lecture 8: Fast Linear Solvers (Part 7) Lecture 8: Fast Linear Solvers (Part 7) 1 Modified Gram-Schmidt Process with Reorthogonalization Test Reorthogonalization If Av k 2 + δ v k+1 2 = Av k 2 to working precision. δ = 10 3 2 Householder Arnoldi

More information

S N. hochdimensionaler Lyapunov- und Sylvestergleichungen. Peter Benner. Mathematik in Industrie und Technik Fakultät für Mathematik TU Chemnitz

S N. hochdimensionaler Lyapunov- und Sylvestergleichungen. Peter Benner. Mathematik in Industrie und Technik Fakultät für Mathematik TU Chemnitz Ansätze zur numerischen Lösung hochdimensionaler Lyapunov- und Sylvestergleichungen Peter Benner Mathematik in Industrie und Technik Fakultät für Mathematik TU Chemnitz S N SIMULATION www.tu-chemnitz.de/~benner

More information

A robust multilevel approximate inverse preconditioner for symmetric positive definite matrices

A robust multilevel approximate inverse preconditioner for symmetric positive definite matrices DICEA DEPARTMENT OF CIVIL, ENVIRONMENTAL AND ARCHITECTURAL ENGINEERING PhD SCHOOL CIVIL AND ENVIRONMENTAL ENGINEERING SCIENCES XXX CYCLE A robust multilevel approximate inverse preconditioner for symmetric

More information

Scientific Computing with Case Studies SIAM Press, Lecture Notes for Unit VII Sparse Matrix

Scientific Computing with Case Studies SIAM Press, Lecture Notes for Unit VII Sparse Matrix Scientific Computing with Case Studies SIAM Press, 2009 http://www.cs.umd.edu/users/oleary/sccswebpage Lecture Notes for Unit VII Sparse Matrix Computations Part 1: Direct Methods Dianne P. O Leary c 2008

More information

Domain Decomposition-based contour integration eigenvalue solvers

Domain Decomposition-based contour integration eigenvalue solvers Domain Decomposition-based contour integration eigenvalue solvers Vassilis Kalantzis joint work with Yousef Saad Computer Science and Engineering Department University of Minnesota - Twin Cities, USA SIAM

More information

Direct solution methods for sparse matrices. p. 1/49

Direct solution methods for sparse matrices. p. 1/49 Direct solution methods for sparse matrices p. 1/49 p. 2/49 Direct solution methods for sparse matrices Solve Ax = b, where A(n n). (1) Factorize A = LU, L lower-triangular, U upper-triangular. (2) Solve

More information

AMS526: Numerical Analysis I (Numerical Linear Algebra)

AMS526: Numerical Analysis I (Numerical Linear Algebra) AMS526: Numerical Analysis I (Numerical Linear Algebra) Lecture 3: Positive-Definite Systems; Cholesky Factorization Xiangmin Jiao Stony Brook University Xiangmin Jiao Numerical Analysis I 1 / 11 Symmetric

More information

M.A. Botchev. September 5, 2014

M.A. Botchev. September 5, 2014 Rome-Moscow school of Matrix Methods and Applied Linear Algebra 2014 A short introduction to Krylov subspaces for linear systems, matrix functions and inexact Newton methods. Plan and exercises. M.A. Botchev

More information

Numerical Methods I Non-Square and Sparse Linear Systems

Numerical Methods I Non-Square and Sparse Linear Systems Numerical Methods I Non-Square and Sparse Linear Systems Aleksandar Donev Courant Institute, NYU 1 donev@courant.nyu.edu 1 MATH-GA 2011.003 / CSCI-GA 2945.003, Fall 2014 September 25th, 2014 A. Donev (Courant

More information

J.I. Aliaga 1 M. Bollhöfer 2 A.F. Martín 1 E.S. Quintana-Ortí 1. March, 2009

J.I. Aliaga 1 M. Bollhöfer 2 A.F. Martín 1 E.S. Quintana-Ortí 1. March, 2009 Parallel Preconditioning of Linear Systems based on ILUPACK for Multithreaded Architectures J.I. Aliaga M. Bollhöfer 2 A.F. Martín E.S. Quintana-Ortí Deparment of Computer Science and Engineering, Univ.

More information

Numerical Methods in Matrix Computations

Numerical Methods in Matrix Computations Ake Bjorck Numerical Methods in Matrix Computations Springer Contents 1 Direct Methods for Linear Systems 1 1.1 Elements of Matrix Theory 1 1.1.1 Matrix Algebra 2 1.1.2 Vector Spaces 6 1.1.3 Submatrices

More information

ANONSINGULAR tridiagonal linear system of the form

ANONSINGULAR tridiagonal linear system of the form Generalized Diagonal Pivoting Methods for Tridiagonal Systems without Interchanges Jennifer B. Erway, Roummel F. Marcia, and Joseph A. Tyson Abstract It has been shown that a nonsingular symmetric tridiagonal

More information

Applied Mathematics 205. Unit V: Eigenvalue Problems. Lecturer: Dr. David Knezevic

Applied Mathematics 205. Unit V: Eigenvalue Problems. Lecturer: Dr. David Knezevic Applied Mathematics 205 Unit V: Eigenvalue Problems Lecturer: Dr. David Knezevic Unit V: Eigenvalue Problems Chapter V.4: Krylov Subspace Methods 2 / 51 Krylov Subspace Methods In this chapter we give

More information

Comparison of Fixed Point Methods and Krylov Subspace Methods Solving Convection-Diffusion Equations

Comparison of Fixed Point Methods and Krylov Subspace Methods Solving Convection-Diffusion Equations American Journal of Computational Mathematics, 5, 5, 3-6 Published Online June 5 in SciRes. http://www.scirp.org/journal/ajcm http://dx.doi.org/.436/ajcm.5.5 Comparison of Fixed Point Methods and Krylov

More information

Accelerating data uncertainty quantification by solving linear systems with multiple right-hand sides

Accelerating data uncertainty quantification by solving linear systems with multiple right-hand sides Accelerating data uncertainty quantification by solving linear systems with multiple right-hand sides Vassilis Kalantzis, Costas Bekas, Alessandro Curioni, and Efstratios Gallopoulos January 2013 EPrint

More information

ITERATIVE METHODS FOR SPARSE LINEAR SYSTEMS

ITERATIVE METHODS FOR SPARSE LINEAR SYSTEMS ITERATIVE METHODS FOR SPARSE LINEAR SYSTEMS YOUSEF SAAD University of Minnesota PWS PUBLISHING COMPANY I(T)P An International Thomson Publishing Company BOSTON ALBANY BONN CINCINNATI DETROIT LONDON MADRID

More information

4.8 Arnoldi Iteration, Krylov Subspaces and GMRES

4.8 Arnoldi Iteration, Krylov Subspaces and GMRES 48 Arnoldi Iteration, Krylov Subspaces and GMRES We start with the problem of using a similarity transformation to convert an n n matrix A to upper Hessenberg form H, ie, A = QHQ, (30) with an appropriate

More information

AMS Mathematics Subject Classification : 65F10,65F50. Key words and phrases: ILUS factorization, preconditioning, Schur complement, 1.

AMS Mathematics Subject Classification : 65F10,65F50. Key words and phrases: ILUS factorization, preconditioning, Schur complement, 1. J. Appl. Math. & Computing Vol. 15(2004), No. 1, pp. 299-312 BILUS: A BLOCK VERSION OF ILUS FACTORIZATION DAVOD KHOJASTEH SALKUYEH AND FAEZEH TOUTOUNIAN Abstract. ILUS factorization has many desirable

More information

Preconditioning Techniques Analysis for CG Method

Preconditioning Techniques Analysis for CG Method Preconditioning Techniques Analysis for CG Method Huaguang Song Department of Computer Science University of California, Davis hso@ucdavis.edu Abstract Matrix computation issue for solve linear system

More information

ON ORTHOGONAL REDUCTION TO HESSENBERG FORM WITH SMALL BANDWIDTH

ON ORTHOGONAL REDUCTION TO HESSENBERG FORM WITH SMALL BANDWIDTH ON ORTHOGONAL REDUCTION TO HESSENBERG FORM WITH SMALL BANDWIDTH V. FABER, J. LIESEN, AND P. TICHÝ Abstract. Numerous algorithms in numerical linear algebra are based on the reduction of a given matrix

More information

Summary of Iterative Methods for Non-symmetric Linear Equations That Are Related to the Conjugate Gradient (CG) Method

Summary of Iterative Methods for Non-symmetric Linear Equations That Are Related to the Conjugate Gradient (CG) Method Summary of Iterative Methods for Non-symmetric Linear Equations That Are Related to the Conjugate Gradient (CG) Method Leslie Foster 11-5-2012 We will discuss the FOM (full orthogonalization method), CG,

More information

A Robust Preconditioned Iterative Method for the Navier-Stokes Equations with High Reynolds Numbers

A Robust Preconditioned Iterative Method for the Navier-Stokes Equations with High Reynolds Numbers Applied and Computational Mathematics 2017; 6(4): 202-207 http://www.sciencepublishinggroup.com/j/acm doi: 10.11648/j.acm.20170604.18 ISSN: 2328-5605 (Print); ISSN: 2328-5613 (Online) A Robust Preconditioned

More information

9.1 Preconditioned Krylov Subspace Methods

9.1 Preconditioned Krylov Subspace Methods Chapter 9 PRECONDITIONING 9.1 Preconditioned Krylov Subspace Methods 9.2 Preconditioned Conjugate Gradient 9.3 Preconditioned Generalized Minimal Residual 9.4 Relaxation Method Preconditioners 9.5 Incomplete

More information

Parallel Numerics, WT 2016/ Iterative Methods for Sparse Linear Systems of Equations. page 1 of 1

Parallel Numerics, WT 2016/ Iterative Methods for Sparse Linear Systems of Equations. page 1 of 1 Parallel Numerics, WT 2016/2017 5 Iterative Methods for Sparse Linear Systems of Equations page 1 of 1 Contents 1 Introduction 1.1 Computer Science Aspects 1.2 Numerical Problems 1.3 Graphs 1.4 Loop Manipulations

More information

Chebyshev semi-iteration in Preconditioning

Chebyshev semi-iteration in Preconditioning Report no. 08/14 Chebyshev semi-iteration in Preconditioning Andrew J. Wathen Oxford University Computing Laboratory Tyrone Rees Oxford University Computing Laboratory Dedicated to Victor Pereyra on his

More information

A short course on: Preconditioned Krylov subspace methods. Yousef Saad University of Minnesota Dept. of Computer Science and Engineering

A short course on: Preconditioned Krylov subspace methods. Yousef Saad University of Minnesota Dept. of Computer Science and Engineering A short course on: Preconditioned Krylov subspace methods Yousef Saad University of Minnesota Dept. of Computer Science and Engineering Universite du Littoral, Jan 19-3, 25 Outline Part 1 Introd., discretization

More information

RESIDUAL SMOOTHING AND PEAK/PLATEAU BEHAVIOR IN KRYLOV SUBSPACE METHODS

RESIDUAL SMOOTHING AND PEAK/PLATEAU BEHAVIOR IN KRYLOV SUBSPACE METHODS RESIDUAL SMOOTHING AND PEAK/PLATEAU BEHAVIOR IN KRYLOV SUBSPACE METHODS HOMER F. WALKER Abstract. Recent results on residual smoothing are reviewed, and it is observed that certain of these are equivalent

More information

Parallel sparse linear solvers and applications in CFD

Parallel sparse linear solvers and applications in CFD Parallel sparse linear solvers and applications in CFD Jocelyne Erhel Joint work with Désiré Nuentsa Wakam () and Baptiste Poirriez () SAGE team, Inria Rennes, France journée Calcul Intensif Distribué

More information

Numerical behavior of inexact linear solvers

Numerical behavior of inexact linear solvers Numerical behavior of inexact linear solvers Miro Rozložník joint results with Zhong-zhi Bai and Pavel Jiránek Institute of Computer Science, Czech Academy of Sciences, Prague, Czech Republic The fourth

More information

EVALUATING SPARSE LINEAR SYSTEM SOLVERS ON SCALABLE PARALLEL ARCHITECTURES

EVALUATING SPARSE LINEAR SYSTEM SOLVERS ON SCALABLE PARALLEL ARCHITECTURES AFRL-RI-RS-TR-2008-273 Final Technical Report October 2008 EVALUATING SPARSE LINEAR SYSTEM SOLVERS ON SCALABLE PARALLEL ARCHITECTURES Purdue University Sponsored by Defense Advanced Research Projects Agency

More information

The flexible incomplete LU preconditioner for large nonsymmetric linear systems. Takatoshi Nakamura Takashi Nodera

The flexible incomplete LU preconditioner for large nonsymmetric linear systems. Takatoshi Nakamura Takashi Nodera Research Report KSTS/RR-15/006 The flexible incomplete LU preconditioner for large nonsymmetric linear systems by Takatoshi Nakamura Takashi Nodera Takatoshi Nakamura School of Fundamental Science and

More information

Iterative Methods for Sparse Linear Systems

Iterative Methods for Sparse Linear Systems Iterative Methods for Sparse Linear Systems Luca Bergamaschi e-mail: berga@dmsa.unipd.it - http://www.dmsa.unipd.it/ berga Department of Mathematical Methods and Models for Scientific Applications University

More information

Lecture 17: Iterative Methods and Sparse Linear Algebra

Lecture 17: Iterative Methods and Sparse Linear Algebra Lecture 17: Iterative Methods and Sparse Linear Algebra David Bindel 25 Mar 2014 Logistics HW 3 extended to Wednesday after break HW 4 should come out Monday after break Still need project description

More information

THE solution of the absolute value equation (AVE) of

THE solution of the absolute value equation (AVE) of The nonlinear HSS-like iterative method for absolute value equations Mu-Zheng Zhu Member, IAENG, and Ya-E Qi arxiv:1403.7013v4 [math.na] 2 Jan 2018 Abstract Salkuyeh proposed the Picard-HSS iteration method

More information

FEM and sparse linear system solving

FEM and sparse linear system solving FEM & sparse linear system solving, Lecture 9, Nov 19, 2017 1/36 Lecture 9, Nov 17, 2017: Krylov space methods http://people.inf.ethz.ch/arbenz/fem17 Peter Arbenz Computer Science Department, ETH Zürich

More information

TR A Comparison of the Performance of SaP::GPU and Intel s Math Kernel Library (MKL) for Solving Dense Banded Linear Systems

TR A Comparison of the Performance of SaP::GPU and Intel s Math Kernel Library (MKL) for Solving Dense Banded Linear Systems TR-0-07 A Comparison of the Performance of ::GPU and Intel s Math Kernel Library (MKL) for Solving Dense Banded Linear Systems Ang Li, Omkar Deshmukh, Radu Serban, Dan Negrut May, 0 Abstract ::GPU is a

More information

The amount of work to construct each new guess from the previous one should be a small multiple of the number of nonzeros in A.

The amount of work to construct each new guess from the previous one should be a small multiple of the number of nonzeros in A. AMSC/CMSC 661 Scientific Computing II Spring 2005 Solution of Sparse Linear Systems Part 2: Iterative methods Dianne P. O Leary c 2005 Solving Sparse Linear Systems: Iterative methods The plan: Iterative

More information

APPLIED NUMERICAL LINEAR ALGEBRA

APPLIED NUMERICAL LINEAR ALGEBRA APPLIED NUMERICAL LINEAR ALGEBRA James W. Demmel University of California Berkeley, California Society for Industrial and Applied Mathematics Philadelphia Contents Preface 1 Introduction 1 1.1 Basic Notation

More information

Preconditioned GMRES Revisited

Preconditioned GMRES Revisited Preconditioned GMRES Revisited Roland Herzog Kirk Soodhalter UBC (visiting) RICAM Linz Preconditioning Conference 2017 Vancouver August 01, 2017 Preconditioned GMRES Revisited Vancouver 1 / 32 Table of

More information

Communication-avoiding Krylov subspace methods

Communication-avoiding Krylov subspace methods Motivation Communication-avoiding Krylov subspace methods Mark mhoemmen@cs.berkeley.edu University of California Berkeley EECS MS Numerical Libraries Group visit: 28 April 2008 Overview Motivation Current

More information

Fast Iterative Solution of Saddle Point Problems

Fast Iterative Solution of Saddle Point Problems Michele Benzi Department of Mathematics and Computer Science Emory University Atlanta, GA Acknowledgments NSF (Computational Mathematics) Maxim Olshanskii (Mech-Math, Moscow State U.) Zhen Wang (PhD student,

More information

Block Krylov Space Solvers: a Survey

Block Krylov Space Solvers: a Survey Seminar for Applied Mathematics ETH Zurich Nagoya University 8 Dec. 2005 Partly joint work with Thomas Schmelzer, Oxford University Systems with multiple RHSs Given is a nonsingular linear system with

More information

Arnoldi Methods in SLEPc

Arnoldi Methods in SLEPc Scalable Library for Eigenvalue Problem Computations SLEPc Technical Report STR-4 Available at http://slepc.upv.es Arnoldi Methods in SLEPc V. Hernández J. E. Román A. Tomás V. Vidal Last update: October,

More information

Performance Evaluation of GPBiCGSafe Method without Reverse-Ordered Recurrence for Realistic Problems

Performance Evaluation of GPBiCGSafe Method without Reverse-Ordered Recurrence for Realistic Problems Performance Evaluation of GPBiCGSafe Method without Reverse-Ordered Recurrence for Realistic Problems Seiji Fujino, Takashi Sekimoto Abstract GPBiCG method is an attractive iterative method for the solution

More information

On the influence of eigenvalues on Bi-CG residual norms

On the influence of eigenvalues on Bi-CG residual norms On the influence of eigenvalues on Bi-CG residual norms Jurjen Duintjer Tebbens Institute of Computer Science Academy of Sciences of the Czech Republic duintjertebbens@cs.cas.cz Gérard Meurant 30, rue

More information

Parallelization of Multilevel Preconditioners Constructed from Inverse-Based ILUs on Shared-Memory Multiprocessors

Parallelization of Multilevel Preconditioners Constructed from Inverse-Based ILUs on Shared-Memory Multiprocessors Parallelization of Multilevel Preconditioners Constructed from Inverse-Based ILUs on Shared-Memory Multiprocessors J.I. Aliaga 1 M. Bollhöfer 2 A.F. Martín 1 E.S. Quintana-Ortí 1 1 Deparment of Computer

More information

WEIGHTED MATRIX ORDERING AND PARALLEL BANDED PRECONDITIONERS FOR ITERATIVE LINEAR SYSTEM SOLVERS

WEIGHTED MATRIX ORDERING AND PARALLEL BANDED PRECONDITIONERS FOR ITERATIVE LINEAR SYSTEM SOLVERS SIAM J. SCI. COMPUT. Vol. 32, No. 3, pp. 1201 1216 c 2010 Society for Industrial and Applied Mathematics WEIGHTED MATRIX ORDERING AND PARALLEL BANDED PRECONDITIONERS FOR ITERATIVE LINEAR SYSTEM SOLVERS

More information

Numerical Methods - Numerical Linear Algebra

Numerical Methods - Numerical Linear Algebra Numerical Methods - Numerical Linear Algebra Y. K. Goh Universiti Tunku Abdul Rahman 2013 Y. K. Goh (UTAR) Numerical Methods - Numerical Linear Algebra I 2013 1 / 62 Outline 1 Motivation 2 Solving Linear

More information

SPARSE SOLVERS POISSON EQUATION. Margreet Nool. November 9, 2015 FOR THE. CWI, Multiscale Dynamics

SPARSE SOLVERS POISSON EQUATION. Margreet Nool. November 9, 2015 FOR THE. CWI, Multiscale Dynamics SPARSE SOLVERS FOR THE POISSON EQUATION Margreet Nool CWI, Multiscale Dynamics November 9, 2015 OUTLINE OF THIS TALK 1 FISHPACK, LAPACK, PARDISO 2 SYSTEM OVERVIEW OF CARTESIUS 3 POISSON EQUATION 4 SOLVERS

More information

Lecture 11: CMSC 878R/AMSC698R. Iterative Methods An introduction. Outline. Inverse, LU decomposition, Cholesky, SVD, etc.

Lecture 11: CMSC 878R/AMSC698R. Iterative Methods An introduction. Outline. Inverse, LU decomposition, Cholesky, SVD, etc. Lecture 11: CMSC 878R/AMSC698R Iterative Methods An introduction Outline Direct Solution of Linear Systems Inverse, LU decomposition, Cholesky, SVD, etc. Iterative methods for linear systems Why? Matrix

More information

Indefinite and physics-based preconditioning

Indefinite and physics-based preconditioning Indefinite and physics-based preconditioning Jed Brown VAW, ETH Zürich 2009-01-29 Newton iteration Standard form of a nonlinear system F (u) 0 Iteration Solve: Update: J(ũ)u F (ũ) ũ + ũ + u Example (p-bratu)

More information

OUTLINE ffl CFD: elliptic pde's! Ax = b ffl Basic iterative methods ffl Krylov subspace methods ffl Preconditioning techniques: Iterative methods ILU

OUTLINE ffl CFD: elliptic pde's! Ax = b ffl Basic iterative methods ffl Krylov subspace methods ffl Preconditioning techniques: Iterative methods ILU Preconditioning Techniques for Solving Large Sparse Linear Systems Arnold Reusken Institut für Geometrie und Praktische Mathematik RWTH-Aachen OUTLINE ffl CFD: elliptic pde's! Ax = b ffl Basic iterative

More information

Some minimization problems

Some minimization problems Week 13: Wednesday, Nov 14 Some minimization problems Last time, we sketched the following two-step strategy for approximating the solution to linear systems via Krylov subspaces: 1. Build a sequence of

More information

PROJECTED GMRES AND ITS VARIANTS

PROJECTED GMRES AND ITS VARIANTS PROJECTED GMRES AND ITS VARIANTS Reinaldo Astudillo Brígida Molina rastudillo@kuaimare.ciens.ucv.ve bmolina@kuaimare.ciens.ucv.ve Centro de Cálculo Científico y Tecnológico (CCCT), Facultad de Ciencias,

More information

Solving Large Nonlinear Sparse Systems

Solving Large Nonlinear Sparse Systems Solving Large Nonlinear Sparse Systems Fred W. Wubs and Jonas Thies Computational Mechanics & Numerical Mathematics University of Groningen, the Netherlands f.w.wubs@rug.nl Centre for Interdisciplinary

More information

Incomplete LU Preconditioning and Error Compensation Strategies for Sparse Matrices

Incomplete LU Preconditioning and Error Compensation Strategies for Sparse Matrices Incomplete LU Preconditioning and Error Compensation Strategies for Sparse Matrices Eun-Joo Lee Department of Computer Science, East Stroudsburg University of Pennsylvania, 327 Science and Technology Center,

More information

Variants of BiCGSafe method using shadow three-term recurrence

Variants of BiCGSafe method using shadow three-term recurrence Variants of BiCGSafe method using shadow three-term recurrence Seiji Fujino, Takashi Sekimoto Research Institute for Information Technology, Kyushu University Ryoyu Systems Co. Ltd. (Kyushu University)

More information

Krylov Subspace Methods that Are Based on the Minimization of the Residual

Krylov Subspace Methods that Are Based on the Minimization of the Residual Chapter 5 Krylov Subspace Methods that Are Based on the Minimization of the Residual Remark 51 Goal he goal of these methods consists in determining x k x 0 +K k r 0,A such that the corresponding Euclidean

More information

ETNA Kent State University

ETNA Kent State University Z Electronic Transactions on Numerical Analysis. olume 16, pp. 191, 003. Copyright 003,. ISSN 10689613. etnamcs.kent.edu A BLOCK ERSION O BICGSTAB OR LINEAR SYSTEMS WITH MULTIPLE RIGHTHAND SIDES A. EL

More information

Fine-Grained Parallel Algorithms for Incomplete Factorization Preconditioning

Fine-Grained Parallel Algorithms for Incomplete Factorization Preconditioning Fine-Grained Parallel Algorithms for Incomplete Factorization Preconditioning Edmond Chow School of Computational Science and Engineering Georgia Institute of Technology, USA SPPEXA Symposium TU München,

More information

Preconditioners for the incompressible Navier Stokes equations

Preconditioners for the incompressible Navier Stokes equations Preconditioners for the incompressible Navier Stokes equations C. Vuik M. ur Rehman A. Segal Delft Institute of Applied Mathematics, TU Delft, The Netherlands SIAM Conference on Computational Science and

More information

Preconditioning for Nonsymmetry and Time-dependence

Preconditioning for Nonsymmetry and Time-dependence Preconditioning for Nonsymmetry and Time-dependence Andy Wathen Oxford University, UK joint work with Jen Pestana and Elle McDonald Jeju, Korea, 2015 p.1/24 Iterative methods For self-adjoint problems/symmetric

More information

On solving linear systems arising from Shishkin mesh discretizations

On solving linear systems arising from Shishkin mesh discretizations On solving linear systems arising from Shishkin mesh discretizations Petr Tichý Faculty of Mathematics and Physics, Charles University joint work with Carlos Echeverría, Jörg Liesen, and Daniel Szyld October

More information

Iterative Methods for Linear Systems of Equations

Iterative Methods for Linear Systems of Equations Iterative Methods for Linear Systems of Equations Projection methods (3) ITMAN PhD-course DTU 20-10-08 till 24-10-08 Martin van Gijzen 1 Delft University of Technology Overview day 4 Bi-Lanczos method

More information

Multilevel low-rank approximation preconditioners Yousef Saad Department of Computer Science and Engineering University of Minnesota

Multilevel low-rank approximation preconditioners Yousef Saad Department of Computer Science and Engineering University of Minnesota Multilevel low-rank approximation preconditioners Yousef Saad Department of Computer Science and Engineering University of Minnesota SIAM CSE Boston - March 1, 2013 First: Joint work with Ruipeng Li Work

More information

1 Extrapolation: A Hint of Things to Come

1 Extrapolation: A Hint of Things to Come Notes for 2017-03-24 1 Extrapolation: A Hint of Things to Come Stationary iterations are simple. Methods like Jacobi or Gauss-Seidel are easy to program, and it s (relatively) easy to analyze their convergence.

More information

Topics. The CG Algorithm Algorithmic Options CG s Two Main Convergence Theorems

Topics. The CG Algorithm Algorithmic Options CG s Two Main Convergence Theorems Topics The CG Algorithm Algorithmic Options CG s Two Main Convergence Theorems What about non-spd systems? Methods requiring small history Methods requiring large history Summary of solvers 1 / 52 Conjugate

More information

Lecture 1: Numerical Issues from Inverse Problems (Parameter Estimation, Regularization Theory, and Parallel Algorithms)

Lecture 1: Numerical Issues from Inverse Problems (Parameter Estimation, Regularization Theory, and Parallel Algorithms) Lecture 1: Numerical Issues from Inverse Problems (Parameter Estimation, Regularization Theory, and Parallel Algorithms) Youzuo Lin 1 Joint work with: Rosemary A. Renaut 2 Brendt Wohlberg 1 Hongbin Guo

More information

6.4 Krylov Subspaces and Conjugate Gradients

6.4 Krylov Subspaces and Conjugate Gradients 6.4 Krylov Subspaces and Conjugate Gradients Our original equation is Ax = b. The preconditioned equation is P Ax = P b. When we write P, we never intend that an inverse will be explicitly computed. P

More information

Linear algebra issues in Interior Point methods for bound-constrained least-squares problems

Linear algebra issues in Interior Point methods for bound-constrained least-squares problems Linear algebra issues in Interior Point methods for bound-constrained least-squares problems Stefania Bellavia Dipartimento di Energetica S. Stecco Università degli Studi di Firenze Joint work with Jacek

More information

Preliminary/Qualifying Exam in Numerical Analysis (Math 502a) Spring 2012

Preliminary/Qualifying Exam in Numerical Analysis (Math 502a) Spring 2012 Instructions Preliminary/Qualifying Exam in Numerical Analysis (Math 502a) Spring 2012 The exam consists of four problems, each having multiple parts. You should attempt to solve all four problems. 1.

More information

Solving PDEs with CUDA Jonathan Cohen

Solving PDEs with CUDA Jonathan Cohen Solving PDEs with CUDA Jonathan Cohen jocohen@nvidia.com NVIDIA Research PDEs (Partial Differential Equations) Big topic Some common strategies Focus on one type of PDE in this talk Poisson Equation Linear

More information

A Domain Decomposition Based Jacobi-Davidson Algorithm for Quantum Dot Simulation

A Domain Decomposition Based Jacobi-Davidson Algorithm for Quantum Dot Simulation A Domain Decomposition Based Jacobi-Davidson Algorithm for Quantum Dot Simulation Tao Zhao 1, Feng-Nan Hwang 2 and Xiao-Chuan Cai 3 Abstract In this paper, we develop an overlapping domain decomposition

More information

Notes on PCG for Sparse Linear Systems

Notes on PCG for Sparse Linear Systems Notes on PCG for Sparse Linear Systems Luca Bergamaschi Department of Civil Environmental and Architectural Engineering University of Padova e-mail luca.bergamaschi@unipd.it webpage www.dmsa.unipd.it/

More information

On the Modification of an Eigenvalue Problem that Preserves an Eigenspace

On the Modification of an Eigenvalue Problem that Preserves an Eigenspace Purdue University Purdue e-pubs Department of Computer Science Technical Reports Department of Computer Science 2009 On the Modification of an Eigenvalue Problem that Preserves an Eigenspace Maxim Maumov

More information

A stable variant of Simpler GMRES and GCR

A stable variant of Simpler GMRES and GCR A stable variant of Simpler GMRES and GCR Miroslav Rozložník joint work with Pavel Jiránek and Martin H. Gutknecht Institute of Computer Science, Czech Academy of Sciences, Prague, Czech Republic miro@cs.cas.cz,

More information

DELFT UNIVERSITY OF TECHNOLOGY

DELFT UNIVERSITY OF TECHNOLOGY DELFT UNIVERSITY OF TECHNOLOGY REPORT 16-02 The Induced Dimension Reduction method applied to convection-diffusion-reaction problems R. Astudillo and M. B. van Gijzen ISSN 1389-6520 Reports of the Delft

More information

Enhancing Scalability of Sparse Direct Methods

Enhancing Scalability of Sparse Direct Methods Journal of Physics: Conference Series 78 (007) 0 doi:0.088/7-6596/78//0 Enhancing Scalability of Sparse Direct Methods X.S. Li, J. Demmel, L. Grigori, M. Gu, J. Xia 5, S. Jardin 6, C. Sovinec 7, L.-Q.

More information

Recycling Bi-Lanczos Algorithms: BiCG, CGS, and BiCGSTAB

Recycling Bi-Lanczos Algorithms: BiCG, CGS, and BiCGSTAB Recycling Bi-Lanczos Algorithms: BiCG, CGS, and BiCGSTAB Kapil Ahuja Thesis submitted to the Faculty of the Virginia Polytechnic Institute and State University in partial fulfillment of the requirements

More information

S-Step and Communication-Avoiding Iterative Methods

S-Step and Communication-Avoiding Iterative Methods S-Step and Communication-Avoiding Iterative Methods Maxim Naumov NVIDIA, 270 San Tomas Expressway, Santa Clara, CA 95050 Abstract In this paper we make an overview of s-step Conjugate Gradient and develop

More information

Conjugate gradient method. Descent method. Conjugate search direction. Conjugate Gradient Algorithm (294)

Conjugate gradient method. Descent method. Conjugate search direction. Conjugate Gradient Algorithm (294) Conjugate gradient method Descent method Hestenes, Stiefel 1952 For A N N SPD In exact arithmetic, solves in N steps In real arithmetic No guaranteed stopping Often converges in many fewer than N steps

More information

Contribution of Wo¹niakowski, Strako²,... The conjugate gradient method in nite precision computa

Contribution of Wo¹niakowski, Strako²,... The conjugate gradient method in nite precision computa Contribution of Wo¹niakowski, Strako²,... The conjugate gradient method in nite precision computations ªaw University of Technology Institute of Mathematics and Computer Science Warsaw, October 7, 2006

More information

Master Thesis Literature Study Presentation

Master Thesis Literature Study Presentation Master Thesis Literature Study Presentation Delft University of Technology The Faculty of Electrical Engineering, Mathematics and Computer Science January 29, 2010 Plaxis Introduction Plaxis Finite Element

More information

The antitriangular factorisation of saddle point matrices

The antitriangular factorisation of saddle point matrices The antitriangular factorisation of saddle point matrices J. Pestana and A. J. Wathen August 29, 2013 Abstract Mastronardi and Van Dooren [this journal, 34 (2013) pp. 173 196] recently introduced the block

More information