On the accuracy of saddle point solvers

Size: px
Start display at page:

Download "On the accuracy of saddle point solvers"

Transcription

1 On the accuracy of saddle point solvers Miro Rozložník joint results with Valeria Simoncini and Pavel Jiránek Institute of Computer Science, Czech Academy of Sciences, Prague, Czech Republic Seminar at the Chinese Academy of Sciences, Beijing, July 11-16, 2010

2 Saddle point problems We consider a saddle point problem with the symmetric 2 2 block form ( ) ( ) ( A B x f B T =. 0 y 0) A is a square n n nonsingular (symmetric positive definite) matrix, B is a rectangular n m matrix of (full column) rank m. Applications: mixed finite element approximations, weighted least squares, constrained optimization etc. [Benzi,Golub, Liesen, 2005]. Numerous schemes: block diagonal preconditioners, block triangular preconditioners, constraint preconditioning, Hermitian/skew-Hermitian preconditioning and other splittings, combination preconditioning References: [Bramble and Pasciak, 1988], [Silvester and Wathen, 1993, 1994], [Elman, Silvester and Wathen, 2002, 2005], [Kay, Loghin and Wathen, 2002], [Keller, Gould and Wathen 2000], [Perugia, Simoncini, Arioli, 1999], [Gould, Hribar and Nocedal, 2001], [Stoll, Wathen, 2008],...

3 Symmetric indefinite system, symmetric positive definite preconditioner A = ( ) A B B T P = R 0 T R A symmetric indefinite, P positive definite (R nonsingular) ( R T AR 1) ( ) ( ) x f R = R T y 0 R T AR 1 is symmetric indefinite!

4 Iterative solution of preconditioned (symmetric indefinite) system Preconditioned MINRES is the MINRES on R T AR 1, minimizes the P 1 = R 1 R T -norm of the residual on K n(p 1 A, P 1 r 0 ) H-MINRES on P 1 A with H = P 1 CG applied to indefinite system with R T AR 1 : CG iterate exists at least at every second step (tridiagonal form T n is nonsingular at least at every second step) [Paige, Saunders, 1975] peak/plateau behavior: CG converges fast MINRES is not much better than CG CG norm increases (peak) MINRES stagnates (plateau) [Greenbaum, Cullum, 1996]

5 Symmetric indefinite system, indefinite or nonsymmetric preconditioner P symmetric indefinite or nonsymmetric P 1 A ( ) ( ) x f = P 1 y 0 ( ) ( ) AP 1 x P = y ( ) f 0 P 1 A and AP 1 are nonsymmetric!

6 Iterative solution of preconditioned nonsymmetric system, positive definite inner product The existence of a short-term recurrence solution methods to solve the system with P 1 A or AP 1 for arbitrary right-hand side vector [Faber, Manteuffel 1984, Liesen, Strakoš, 2006] Matrices P 1 A or AP 1 can be symmetric (self-adjoint) in a given inner product induced by the symmetric positive definite H. Then three term-recurrence method can be applied H(P 1 A) = (P 1 A) T H (P T H) T A = A(P T H) H(AP 1 ) = (AP 1 ) T H HAP 1 = P T AH H(P 1 A) symmetric indefinite: MINRES applied to H(P 1 A) and preconditioned with H H-MINRES on P 1 A H(P 1 A) positive definite: CG applied to H(P 1 A) and preconditioned with H; works on K n(p 1 A, P 1 r 0 ) and can be seen as the CG scheme applied to P 1 A with a nonstandard inner product H H-CG on P 1 A

7 Iterative solution of preconditioned nonsymmetric system, symmetric bilinear form if there exists a symmetric indefinite H such that H(P 1 A) = (P 1 A) T H = [H(P 1 A)] T [(AP 1 ) T H] T = H(AP 1 ) = (AP 1 ) T H is symmetric indefinite MINRES method applied to H(P 1 A) or H(AP 1 ) symmetric indefinite preconditioner H = P 1 = (P 1 ) T so that (P 1 ) T (P 1 )A = A(P 1 ) T (P 1 ) (P 1 ) T AP 1 = P 1 AP 1 right vs left preconditioning for symmetric P P 1 K n (AP 1, r 0 ) = K n (P 1 A, P 1 r 0 ) (AP 1 ) T = (P 1 ) T A = P 1 A

8 Iterative solution of preconditioned nonsymmetric system, symmetric bilinear form H-symmetric variant of the nonsymmetric Lanczos process: AP 1 V n = V n+1 T n+1,n, (AP 1 ) T W n = W n+1 Tn+1,n W T n V n = I = W n = HV n [Freund, Nachtigal, 1995] H-symmetric variant of Bi-CG H-symmetric variant of QMR ITFQMR [Freund, Nachtigal, 1995] QMR-from-BiCG: H-symmetric Bi-CG + QMR-smoothing = H-symmetric QMR [Freund, Nachtigal, 1995, Walker, Zhou 1994] peak/plateau behavior: QMR does not improve the convergence of Bi-CG (Bi-CG converges fast QMR is not much better, Bi-CG norm increases quasi-residual of QMR stagnates) [Greenbaum, Cullum, 1996]

9 Simplified Bi-CG algorithm is a preconditioned CG algorithm H = P 1 -symmetric variant of two-term Bi-CG on AP 1 is the Hestenes-Stiefel CG algorithm on A preconditioned with P P ( 1 ) -symmetric Bi-CG(AP ( ) 1 ) x0 x0, r y 0 = b A 0 y 0 PCG(A) with P 1 P 1 p 0 = P 1 r 0, p 0 = r 0 = P 1 p 0 z 0 = P 1 r 0 k = 0, 1,... ( α k = (r ) k, r k ( )/(AP ) 1 p k, p k ) α k = (r k, z k )/(AP 1 p k, P 1 p k ) xk+1 xk = + α y k+1 y k P 1 p k k r k+1 = r k α k AP 1 p k r k+1 = P 1 r k+1 z k+1 = P 1 r k+1 β k = (r k+1, r k+1 )/(r k, r k ) β k = (r k+1, z k+1 )/(r k, z k ) P 1 p k+1 = P 1 r k+1 + β k P 1 p k P 1 p k+1 = z k+1 + β k P 1 p k p k+1 = P 1 p k+1

10 Saddle point problem and indefinite constraint preconditioner ( ) ( ) A B x B T = 0 y ( I B B T 0 P = ( ) f g ), H = P 1 PCG applied to indefinite system with indefinite preconditioner; will not work for arbitrary ( ) right-hand ( ) side, particular right-hand side or initial guess: x0 s0, r y 0 =, here g = 0 and x = y 0 = 0 [Lukšan, Vlček, 1998], [Gould, Keller, Wathen 2000] [Perugia, Simoncini, Arioli, 1999], [R, Simoncini, 2002]

11 Saddle point problem and indefinite constraint preconditioner - preconditioned system ( A B B T 0 ) ( ) x = y ( ) f, P = 0 ( I ) B B T 0 ( ) A(I Π) + Π (A I)B(B AP 1 = T B) 1 0 I Π = B(B T B) 1 B T - orth. projector onto span(b)

12 Indefinite constraint preconditioner: spectral properties of preconditioned system AP 1 nonsymmetric and non-diagonalizable! but it has a nice spectrum: σ(ap 1 ) {1} σ(a(i Π) + Π) {1} σ((i Π)A(I Π)) {0} and only 2 by 2 Jordan blocks! [Lukšan, Vlček 1998], [Gould, Wathen, Keller, 1999], [Perugia, Simoncini 1999]

13 Basic properties of any Krylov method with the constraint preconditioner e k+1 = ( ) x xk+1 y y k+1 ( f r k+1 = 0 ) ( A B B T 0 ) ( ) xk+1 y k+1 r 0 = ( s0 ) r 0 k+1 = 0 B T (x x k+1 ) = 0 x k+1 Null(B T )! ( sk+1 )

14 The energy-norm of the error in the preconditioned CG method r T k+1 P 1 r j = 0, j = 0,..., k x k+1 is an iterate from CG applied to (I Π)A(I Π)x = (I Π)f! satisfying x x k+1 A = min u x0 +span{(i Π)s x j } u A [Lukšan, Vlček 1998], [Gould, Wathen, Keller, 1999]

15 The residual norm in the preconditioned CG method r k+1 = x k+1 x 0 but in general y k+1 y which is reflected in ( sk+1 0 ) 0! but under appropriate scaling yes!

16 The residual norm in the preconditioned CG method x k+1 x x x k+1 = φ k+1 ((I Π)A(I Π))(x x 0 ) r k+1 = φ k+1 (A(I Π) + Π)s 0 σ((i Π)A(I Π)) σ(a(i Π) + Π) {1} σ((i Π)αA(I ( ) Π)) {0} r k+1 = sk+1 0! 0

17 How to avoid the misconvergence of the scheme Scaling by a constant α > 0 such that {1} conv(σ((i Π)αA(I Π)) {0}) ( ) ( ) ( A B x f B T = 0 y 0) ( ) ( ) αa B x B T = 0 αy ( ) αf 0 v : (I Π)v = 0, α = 1 ((I Π)v, A(I Π)v)! Scaling by a diagonal A (diag(a)) 1/2 A(diag(A)) 1/2 often gives what we want! Different direction vector so that r k+1 = s k+1 is locally minimized! y k+1 = y k + (B T B) 1 B T s k [Braess, Deuflhard,Lipikov 1999], [Hribar, Gould, Nocedal, 1999] [Jiránek, R, 2008]

18 Numerical example A = tridiag(1, 4, 1) R 25,25, B = rand(25, 5) R 25,5 f = rand(25, 1) R 25 σ(a) [2.0146, ] ( ) ( ) 1 αa B I B α = 1/τ σ( B T 0 B T ) 0 1/100 [0.0207, ] {1} 1/10 [0.2067, ] {1} 1/4 [0.5170, ] 1 {1} [2.0678, ] 4 {1} [8.2712, ]

19 τ =100 residual norms r k τ =1 τ = iteration number

20 A norm of errors x x k A τ =100 τ = τ = iteration number

21 Inexact saddle point solvers 1. exact method: exact constraint preconditioning, exact arithmetic : outer iteration for solving the preconditioned system; 2. inexact method with approximate or incomplete factorization scheme to solve inner problems with (B T B) 1 : structure-based or with appropriate dropping criterion; inner iteration method 3. the rounding errors: finite precision arithmetic. References: [Gould, Hribar and Nocedal, 2001], [R, Simoncini, 2002] with the use of [Greenbaum 1994,1997], [Sleijpen, et al. 1994]

22 Delay of convergence and limit on the final accuracy

23 Preconditioned CG in finite precision arithmetic ( xk+1 ȳ k+1 ), r k+1 = ( s (1) k+1 s (2) k+1 x x k+1 A γ 1 Π(x x k+1 ) + γ 2 (I Π)A(I Π)(x x k+1 ) ) Exact arithmetic: Π(x x k+1 ) = 0 (I Π)A(I Π)(x x k+1 ) 0

24 Forward error of computed approximate solution: departure from the null-space of B T + projection of the residual onto it x x k+1 A γ 3 B T (x x k+1 ) + γ 2 (I Π)(f A x k+1 Bȳ k+1 ) can be monitored by easily computable quantities: B T (x x k+1 ) s (2) k+1 (I Π)(f A x k+1 Bȳ k+1 ) (I Π) s (1) k+1

25 Maximum attainable accuracy of the scheme ( f 0) (f A x k+1 Bȳ k+1 ) s (1) k+1, B T (x x k+1 ) s (2) k+1 ( A B B T 0 ) ) ( xk+1 ȳ k+1 ) (1) ( s k+1 s (2) k+1 c 1 εκ(a) max j=0,...,k+1 r j [Greenbaum 1994,1997], [Sleijpen, et al. 1994] good scaling: r j 0 nearly monotonically r 0 max j=0,...,k+1 r j

26 r k 10 4 convergence characteristics Π x k (I Π) ( f A x k B y k ) x x k A iteration number

27 r k x x k A convergence characteristics (I Π) ( f A x k B y k ) Π x k iteration number

28 convergence characteristics r k x x k A (I Π) ( f A x k B y k ) Π x k iteration number

29 Conclusions Short-term recurrence methods are applicable for saddle point problems with indefinite preconditioning at a cost comparable to that of symmetric solvers. There is a tight connection between the simplified Bi-CG algorithm and the classical CG. The convergence of CG applied to saddle point problem with indefinite preconditioner for all right-hand side vectors is not guaranteed. For a particular set of right-hand sides the convergence can be achieved by the appropriate scaling of the saddle point problem or by a different back-substitution formula for dual unknowns. Since the numerical behavior of CG in finite precision arithmetic depends heavily on the size of computed residuals, a good scaling of the problems leads to approximate solutions satisfying both two block equations to the working accuracy.

30 Thank you for your attention. miro M. Rozložník and V. Simoncini, Krylov subspace methods for saddle point problems with indefinite preconditioning,siam J. Matrix Anal. Appl., 24 (2002), pp P. Jiránek and M. Rozložník. Maximum attainable accuracy of inexact saddle point solvers. SIAM J. Matrix Anal. Appl., 29(4): , P. Jiránek and M. Rozložník. Limiting accuracy of segregated solution methods for nonsymmetric saddle point problems. J. Comput. Appl. Math. 215 (2008), pp

31 References I M. Arioli. The use of QR factorization in sparse quadratic programming and backward error issues. SIAM J. Matrix Anal. Appl., 21(3): , Z. Bai, G. H. Golub, and M. Ng. Hermitian and skew-hermitian splitting methods for non-hermitian positive definite linear systems. SIAM J. Matrix Anal. Appl., 24: , A. Greenbaum. Estimating the attainable accuracy of recursively computed residual methods. SIAM J. Matrix Anal. Appl., 18(3): , C. Vuik and A. Saghir. The krylov accelerated simple(r) method for incompressible flow. Technical Report 02-01, Delft University of Technology, 2002.

32 Null-space projection method compute x N(B T ) as a solution of the projected system (I Π)A(I Π)x = (I Π)f, compute y as a solution of the least squares problem By f Ax, Π = B(B T B) 1 B T is the orthogonal projector onto R(B). Results for schemes, where the least squares with B are solved inexactly. Every computed approximate solution v of a least squares problem Bv c is interpreted as an exact solution of a perturbed least squares (B + B) v c + c, B τ B, c τ c, τκ(b) 1.

33 Null-space projection method choose x 0, solve By 0 f Ax 0 compute α k and p (x) k N(B T ) x k+1 = x k + α k p (x) k solve Bp (y) k r (x) k α k Ap (x) k back-substitution: A: y k+1 = y k + p (y) k, inner iteration B: solve By k+1 f Ax k+1, C: solve Bv k f Ax k+1 By k, y k+1 = y k + v k. r (x) k+1 = r(x) k α k Ap (x) k Bp (y) k outer iteration

34 Accuracy in the saddle point system f Ax k By k r (x) k O(α 3)κ(B) 1 τκ(b) ( f + A X k), B T x k O(τ)κ(B) 1 τκ(b) B X k, X k max{ x i i = 0, 1,..., k}. Back-substitution scheme α 3 A: Generic update y k+1 = y k + p (y) u k B: Direct substitution y k+1 = B τ (f Ax k+1 ) C: Corrected dir. subst. y k+1 = y k + B u (f Ax k+1 By k ) } additional least square with B

35 Maximum attainable accuracy of inexact null-space projection schemes The limiting (maximum attainable) accuracy is measured by the ultimate (asymptotic) values of: 1. the true projected residual: (I Π)f (I Π)A(I Π)x k ; 2. the residuals in the saddle point system: f Ax k By k and B T x k ; 3. the forward errors: x x k and y y k. Numerical experiments: a small model example A = tridiag(1, 4, 1) R , B = rand(100, 20), f = rand(100, 1), κ(a) = A A 1 = , κ(b) = B B =

36 Generic update: y k+1 = y k + p (y) k τ = 10 2 relative residual norms f Ax k By k / f Ax 0 By 0, r (x) k / r(x) τ = O(u), τ = 10 10, τ=10 6, τ=10 2 residual norm B T x k τ = 10 6 τ = τ = O(u) iteration number iteration number

37 Direct substitution: y k+1 = B (f Ax k+1 ) 10 0 τ = τ = 10 2 relative residual norms f Ax k By k / f Ax 0 By 0, r (x) k / r(x) τ = 10 6 τ = τ = O(u) residual norm B T x k τ = 10 6 τ = τ = O(u) iteration number iteration number

38 Corrected direct substitution: y k+1 = y k + B (f Ax k+1 By k ) τ = 10 2 relative residual norms f Ax k By k / f Ax 0 By 0, r (x) k / r(x) τ = O(u), τ = 10 10, τ=10 6, τ=10 2 residual norm B T x k τ = 10 6 τ = τ = O(u) iteration number iteration number

39 Schur complement reduction method Compute y as a solution of the Schur complement system compute x as a solution of B T A 1 By = B T A 1 f, Ax = f By. inexact solution of systems with A: every computed solution û of Au = b is interpreted an exact solution of a perturbed system (A + A)û = b + b, A τ A, b τ b, τκ(a) 1.

40 Iterative solution of the Schur complement system choose y 0, solve Ax 0 = f By 0 compute α k and p (y) k y k+1 = y k + α k p (y) k solve Ap (x) k = Bp (y) k back-substitution: A: x k+1 = x k + α k p (x) k, inner iteration B: solve Ax k+1 = f By k+1, C: solve Au k = f Ax k By k+1, x k+1 = x k + u k. r (y) k+1 = r(y) k α k B T p (x) k outer iteration

41 Maximum attainable accuracy of inexact Schur complement schemes The limiting (maximum attainable) accuracy is measured by the ultimate (asymptotic) values of: 1. the Schur complement residual: B T A 1 f B T A 1 By k ; 2. the residuals in the saddle point system: f Ax k By k and B T x k ; 3. the forward errors: x x k and y y k. Numerical experiments: a small model example A = tridiag(1, 4, 1) R , B = rand(100, 20), f = rand(100, 1), κ(a) = A A 1 = , κ(b) = B B =

42 Accuracy in the outer iteration process B T A 1 f + B T A 1 By k r (y) k O(τ)κ(A) 1 τκ(a) A 1 B ( f + B Y k ). Y k max{ y i i = 0, 1,..., k}. relative residual norms B T A 1 f B T A 1 By k / B T A 1 f B T A 1 By 0, r (y) k / r(y) τ = 10 2 τ = 10 6 τ = τ = O(u) iteration number k B T (A + A) 1 Bŷ = B T (A + A) 1 f, B T A 1 f B T A 1 Bŷ τκ(a) 1 τκ(a) A 1 B 2 ŷ.

43 Accuracy in the saddle point system f Ax k By k O(α 1)κ(A) 1 τκ(a) ( f + B Y k), B T x k r (y) k O(α 2)κ(A) 1 τκ(a) A 1 B ( f + B Y k ), Y k max{ y i i = 0, 1,..., k}. Back-substitution scheme α 1 α 2 A: Generic update x k+1 = x k + α k p (x) τ u k B: Direct substitution x k+1 = A 1 τ τ (f By k+1 ) C: Corrected dir. subst. x k+1 = x k + A 1 u τ (f Ax k By k+1 ) } additional system with A B T A 1 f + B T A 1 By k = B T x k B T A 1 (f Ax k By k )

44 Generic update: x k+1 = x k + α k p (x) k 10 0 τ = residual norm f Ax k By k τ = 10 6 τ = τ = O(u) relative residual norms B T x k / B T x 0, r k (y) / r0 (y) τ = O(u), τ = 10 10, τ=10 6, τ= iteration number k iteration number k

45 Direct substitution: x k+1 = A 1 (f By k+1 ) 10 0 τ = τ = 10 2 residual norm f Ax k By k τ = 10 6 τ = τ = O(u) relative residual norms B T x k / B T x 0, r k (y) / r0 (y) τ = 10 6 τ = τ = O(u) iteration number k iteration number k

46 Corrected direct substitution: x k+1 = x k + A 1 (f Ax k By k+1 ) τ = 10 2 residual norm f Ax k By k τ = O(u), τ = 10 10, τ=10 6, τ=10 2 relative residual norms B T x k / B T x 0, r k (y) / r0 (y) τ = 10 6 τ = τ = O(u) iteration number k iteration number k

47 Related results in the context of saddle-point problems and Krylov subspace methods General framework of inexact Krylov subspace methods: in exact arithmetic the effects of relaxation in matrix-vector multiplication on the ultimate accuracy of several solvers [?], [?]. The effects of rounding errors in the Schur complement reduction (block LU decomposition) method and the null-space method [?], [Arioli, 2000], the maximum attainable accuracy studied in terms of the user tolerance specified in the outer iteration [?], [?]. Error analysis in computing the projections into the null-space and constraint preconditioning, limiting accuracy of the preconditioned CG, residual update strategy when solving constrained quadratic programming problems [?], or in cascadic multigrid method for elliptic problems [?]. Theory for a general class of iterative methods based on coupled two-term recursions, all bounds of the limiting accuracy depend on the maximum norm of computed iterates, fixed matrix-vector multiplication, cf. [Greenbaum, 1997].

48 General comments and considerations, future work new value = old value + small correction Fixed-precision iterative refinement for improving the computed solution x old to a system Ax = b: solving update equations Az corr = r that have residual r = b Ay old as a right-hand side to obtain x new = x old + z corr, see [?], [?]. Stationary iterative methods for Ax = b and their maximum attainable accuracy [?]: assuming splitting A = M N and inexact solution of systems with M, use x new = x old + M 1 (b Ax old ) rather than x new = M 1 (Nx old + b), [?]. Two-step splitting iteration framework: A = M 1 N 1 = M 2 N 2 assuming inexact solution of systems with M 1 and M 2, reformulation of M 1 x 1/2 = N 1 x old + b, M 2 x new = N 2 x 1/2 + b, Hermitian/skew-Hermitian splitting (HSS) iteration [Bai, Golub, and Ng, 2003]. Inexact preconditioners for saddle point problems: SIMPLE and SIMPLE(R) type algorithms [Vuik and Saghir, 2002] and constraint preconditioners [?].

Numerical behavior of inexact linear solvers

Numerical behavior of inexact linear solvers Numerical behavior of inexact linear solvers Miro Rozložník joint results with Zhong-zhi Bai and Pavel Jiránek Institute of Computer Science, Czech Academy of Sciences, Prague, Czech Republic The fourth

More information

AMS526: Numerical Analysis I (Numerical Linear Algebra) Lecture 23: GMRES and Other Krylov Subspace Methods; Preconditioning

AMS526: Numerical Analysis I (Numerical Linear Algebra) Lecture 23: GMRES and Other Krylov Subspace Methods; Preconditioning AMS526: Numerical Analysis I (Numerical Linear Algebra) Lecture 23: GMRES and Other Krylov Subspace Methods; Preconditioning Xiangmin Jiao SUNY Stony Brook Xiangmin Jiao Numerical Analysis I 1 / 18 Outline

More information

Structured Preconditioners for Saddle Point Problems

Structured Preconditioners for Saddle Point Problems Structured Preconditioners for Saddle Point Problems V. Simoncini Dipartimento di Matematica Università di Bologna valeria@dm.unibo.it p. 1 Collaborators on this project Mario Arioli, RAL, UK Michele Benzi,

More information

A stable variant of Simpler GMRES and GCR

A stable variant of Simpler GMRES and GCR A stable variant of Simpler GMRES and GCR Miroslav Rozložník joint work with Pavel Jiránek and Martin H. Gutknecht Institute of Computer Science, Czech Academy of Sciences, Prague, Czech Republic miro@cs.cas.cz,

More information

Structured Preconditioners for Saddle Point Problems

Structured Preconditioners for Saddle Point Problems Structured Preconditioners for Saddle Point Problems V. Simoncini Dipartimento di Matematica Università di Bologna valeria@dm.unibo.it. p.1/16 Collaborators on this project Mario Arioli, RAL, UK Michele

More information

Iterative solution of saddle point problems

Iterative solution of saddle point problems Iterative solution of saddle point problems Miroslav Rozložník, Pavel Jiránek Institute of Computer Science, Czech Academy of Sciences, Prague and Faculty of Mechatronics and Interdisciplinary Engineering

More information

Journal of Computational and Applied Mathematics. Optimization of the parameterized Uzawa preconditioners for saddle point matrices

Journal of Computational and Applied Mathematics. Optimization of the parameterized Uzawa preconditioners for saddle point matrices Journal of Computational Applied Mathematics 6 (009) 136 154 Contents lists available at ScienceDirect Journal of Computational Applied Mathematics journal homepage: wwwelseviercom/locate/cam Optimization

More information

The antitriangular factorisation of saddle point matrices

The antitriangular factorisation of saddle point matrices The antitriangular factorisation of saddle point matrices J. Pestana and A. J. Wathen August 29, 2013 Abstract Mastronardi and Van Dooren [this journal, 34 (2013) pp. 173 196] recently introduced the block

More information

Combination Preconditioning of saddle-point systems for positive definiteness

Combination Preconditioning of saddle-point systems for positive definiteness Combination Preconditioning of saddle-point systems for positive definiteness Andy Wathen Oxford University, UK joint work with Jen Pestana Eindhoven, 2012 p.1/30 compute iterates with residuals Krylov

More information

Efficient Solvers for the Navier Stokes Equations in Rotation Form

Efficient Solvers for the Navier Stokes Equations in Rotation Form Efficient Solvers for the Navier Stokes Equations in Rotation Form Computer Research Institute Seminar Purdue University March 4, 2005 Michele Benzi Emory University Atlanta, GA Thanks to: NSF (MPS/Computational

More information

Some Preconditioning Techniques for Saddle Point Problems

Some Preconditioning Techniques for Saddle Point Problems Some Preconditioning Techniques for Saddle Point Problems Michele Benzi 1 and Andrew J. Wathen 2 1 Department of Mathematics and Computer Science, Emory University, Atlanta, Georgia 30322, USA. benzi@mathcs.emory.edu

More information

HOW TO MAKE SIMPLER GMRES AND GCR MORE STABLE

HOW TO MAKE SIMPLER GMRES AND GCR MORE STABLE HOW TO MAKE SIMPLER GMRES AND GCR MORE STABLE PAVEL JIRÁNEK, MIROSLAV ROZLOŽNÍK, AND MARTIN H. GUTKNECHT Abstract. In this paper we analyze the numerical behavior of several minimum residual methods, which

More information

Numerical Methods in Matrix Computations

Numerical Methods in Matrix Computations Ake Bjorck Numerical Methods in Matrix Computations Springer Contents 1 Direct Methods for Linear Systems 1 1.1 Elements of Matrix Theory 1 1.1.1 Matrix Algebra 2 1.1.2 Vector Spaces 6 1.1.3 Submatrices

More information

Fast Iterative Solution of Saddle Point Problems

Fast Iterative Solution of Saddle Point Problems Michele Benzi Department of Mathematics and Computer Science Emory University Atlanta, GA Acknowledgments NSF (Computational Mathematics) Maxim Olshanskii (Mech-Math, Moscow State U.) Zhen Wang (PhD student,

More information

RESIDUAL SMOOTHING AND PEAK/PLATEAU BEHAVIOR IN KRYLOV SUBSPACE METHODS

RESIDUAL SMOOTHING AND PEAK/PLATEAU BEHAVIOR IN KRYLOV SUBSPACE METHODS RESIDUAL SMOOTHING AND PEAK/PLATEAU BEHAVIOR IN KRYLOV SUBSPACE METHODS HOMER F. WALKER Abstract. Recent results on residual smoothing are reviewed, and it is observed that certain of these are equivalent

More information

Chebyshev semi-iteration in Preconditioning

Chebyshev semi-iteration in Preconditioning Report no. 08/14 Chebyshev semi-iteration in Preconditioning Andrew J. Wathen Oxford University Computing Laboratory Tyrone Rees Oxford University Computing Laboratory Dedicated to Victor Pereyra on his

More information

Department of Computer Science, University of Illinois at Urbana-Champaign

Department of Computer Science, University of Illinois at Urbana-Champaign Department of Computer Science, University of Illinois at Urbana-Champaign Probing for Schur Complements and Preconditioning Generalized Saddle-Point Problems Eric de Sturler, sturler@cs.uiuc.edu, http://www-faculty.cs.uiuc.edu/~sturler

More information

Preface to the Second Edition. Preface to the First Edition

Preface to the Second Edition. Preface to the First Edition n page v Preface to the Second Edition Preface to the First Edition xiii xvii 1 Background in Linear Algebra 1 1.1 Matrices................................. 1 1.2 Square Matrices and Eigenvalues....................

More information

Recent computational developments in Krylov Subspace Methods for linear systems. Valeria Simoncini and Daniel B. Szyld

Recent computational developments in Krylov Subspace Methods for linear systems. Valeria Simoncini and Daniel B. Szyld Recent computational developments in Krylov Subspace Methods for linear systems Valeria Simoncini and Daniel B. Szyld A later version appeared in : Numerical Linear Algebra w/appl., 2007, v. 14(1), pp.1-59.

More information

Preconditioning for Nonsymmetry and Time-dependence

Preconditioning for Nonsymmetry and Time-dependence Preconditioning for Nonsymmetry and Time-dependence Andy Wathen Oxford University, UK joint work with Jen Pestana and Elle McDonald Jeju, Korea, 2015 p.1/24 Iterative methods For self-adjoint problems/symmetric

More information

On the influence of eigenvalues on Bi-CG residual norms

On the influence of eigenvalues on Bi-CG residual norms On the influence of eigenvalues on Bi-CG residual norms Jurjen Duintjer Tebbens Institute of Computer Science Academy of Sciences of the Czech Republic duintjertebbens@cs.cas.cz Gérard Meurant 30, rue

More information

c 2008 Society for Industrial and Applied Mathematics

c 2008 Society for Industrial and Applied Mathematics SIAM J. MATRIX ANAL. APPL. Vol. 30, No. 4, pp. 1483 1499 c 2008 Society for Industrial and Applied Mathematics HOW TO MAKE SIMPLER GMRES AND GCR MORE STABLE PAVEL JIRÁNEK, MIROSLAV ROZLOŽNÍK, AND MARTIN

More information

Chapter 7 Iterative Techniques in Matrix Algebra

Chapter 7 Iterative Techniques in Matrix Algebra Chapter 7 Iterative Techniques in Matrix Algebra Per-Olof Persson persson@berkeley.edu Department of Mathematics University of California, Berkeley Math 128B Numerical Analysis Vector Norms Definition

More information

On the interplay between discretization and preconditioning of Krylov subspace methods

On the interplay between discretization and preconditioning of Krylov subspace methods On the interplay between discretization and preconditioning of Krylov subspace methods Josef Málek and Zdeněk Strakoš Nečas Center for Mathematical Modeling Charles University in Prague and Czech Academy

More information

ON ORTHOGONAL REDUCTION TO HESSENBERG FORM WITH SMALL BANDWIDTH

ON ORTHOGONAL REDUCTION TO HESSENBERG FORM WITH SMALL BANDWIDTH ON ORTHOGONAL REDUCTION TO HESSENBERG FORM WITH SMALL BANDWIDTH V. FABER, J. LIESEN, AND P. TICHÝ Abstract. Numerous algorithms in numerical linear algebra are based on the reduction of a given matrix

More information

Indefinite Preconditioners for PDE-constrained optimization problems. V. Simoncini

Indefinite Preconditioners for PDE-constrained optimization problems. V. Simoncini Indefinite Preconditioners for PDE-constrained optimization problems V. Simoncini Dipartimento di Matematica, Università di Bologna, Italy valeria.simoncini@unibo.it Partly joint work with Debora Sesana,

More information

An advanced ILU preconditioner for the incompressible Navier-Stokes equations

An advanced ILU preconditioner for the incompressible Navier-Stokes equations An advanced ILU preconditioner for the incompressible Navier-Stokes equations M. ur Rehman C. Vuik A. Segal Delft Institute of Applied Mathematics, TU delft The Netherlands Computational Methods with Applications,

More information

Lanczos tridigonalization and Golub - Kahan bidiagonalization: Ideas, connections and impact

Lanczos tridigonalization and Golub - Kahan bidiagonalization: Ideas, connections and impact Lanczos tridigonalization and Golub - Kahan bidiagonalization: Ideas, connections and impact Zdeněk Strakoš Academy of Sciences and Charles University, Prague http://www.cs.cas.cz/ strakos Hong Kong, February

More information

Spectral Properties of Saddle Point Linear Systems and Relations to Iterative Solvers Part I: Spectral Properties. V. Simoncini

Spectral Properties of Saddle Point Linear Systems and Relations to Iterative Solvers Part I: Spectral Properties. V. Simoncini Spectral Properties of Saddle Point Linear Systems and Relations to Iterative Solvers Part I: Spectral Properties V. Simoncini Dipartimento di Matematica, Università di ologna valeria@dm.unibo.it 1 Outline

More information

Iterative Methods and Multigrid

Iterative Methods and Multigrid Iterative Methods and Multigrid Part 3: Preconditioning 2 Eric de Sturler Preconditioning The general idea behind preconditioning is that convergence of some method for the linear system Ax = b can be

More information

Key words. conjugate gradients, normwise backward error, incremental norm estimation.

Key words. conjugate gradients, normwise backward error, incremental norm estimation. Proceedings of ALGORITMY 2016 pp. 323 332 ON ERROR ESTIMATION IN THE CONJUGATE GRADIENT METHOD: NORMWISE BACKWARD ERROR PETR TICHÝ Abstract. Using an idea of Duff and Vömel [BIT, 42 (2002), pp. 300 322

More information

Preconditioned inverse iteration and shift-invert Arnoldi method

Preconditioned inverse iteration and shift-invert Arnoldi method Preconditioned inverse iteration and shift-invert Arnoldi method Melina Freitag Department of Mathematical Sciences University of Bath CSC Seminar Max-Planck-Institute for Dynamics of Complex Technical

More information

Contents. Preface... xi. Introduction...

Contents. Preface... xi. Introduction... Contents Preface... xi Introduction... xv Chapter 1. Computer Architectures... 1 1.1. Different types of parallelism... 1 1.1.1. Overlap, concurrency and parallelism... 1 1.1.2. Temporal and spatial parallelism

More information

A Robust Preconditioned Iterative Method for the Navier-Stokes Equations with High Reynolds Numbers

A Robust Preconditioned Iterative Method for the Navier-Stokes Equations with High Reynolds Numbers Applied and Computational Mathematics 2017; 6(4): 202-207 http://www.sciencepublishinggroup.com/j/acm doi: 10.11648/j.acm.20170604.18 ISSN: 2328-5605 (Print); ISSN: 2328-5613 (Online) A Robust Preconditioned

More information

Iterative methods for Linear System of Equations. Joint Advanced Student School (JASS-2009)

Iterative methods for Linear System of Equations. Joint Advanced Student School (JASS-2009) Iterative methods for Linear System of Equations Joint Advanced Student School (JASS-2009) Course #2: Numerical Simulation - from Models to Software Introduction In numerical simulation, Partial Differential

More information

M.A. Botchev. September 5, 2014

M.A. Botchev. September 5, 2014 Rome-Moscow school of Matrix Methods and Applied Linear Algebra 2014 A short introduction to Krylov subspaces for linear systems, matrix functions and inexact Newton methods. Plan and exercises. M.A. Botchev

More information

On the Superlinear Convergence of MINRES. Valeria Simoncini and Daniel B. Szyld. Report January 2012

On the Superlinear Convergence of MINRES. Valeria Simoncini and Daniel B. Szyld. Report January 2012 On the Superlinear Convergence of MINRES Valeria Simoncini and Daniel B. Szyld Report 12-01-11 January 2012 This report is available in the World Wide Web at http://www.math.temple.edu/~szyld 0 Chapter

More information

A Review of Preconditioning Techniques for Steady Incompressible Flow

A Review of Preconditioning Techniques for Steady Incompressible Flow Zeist 2009 p. 1/43 A Review of Preconditioning Techniques for Steady Incompressible Flow David Silvester School of Mathematics University of Manchester Zeist 2009 p. 2/43 PDEs Review : 1984 2005 Update

More information

The amount of work to construct each new guess from the previous one should be a small multiple of the number of nonzeros in A.

The amount of work to construct each new guess from the previous one should be a small multiple of the number of nonzeros in A. AMSC/CMSC 661 Scientific Computing II Spring 2005 Solution of Sparse Linear Systems Part 2: Iterative methods Dianne P. O Leary c 2005 Solving Sparse Linear Systems: Iterative methods The plan: Iterative

More information

FEM and sparse linear system solving

FEM and sparse linear system solving FEM & sparse linear system solving, Lecture 9, Nov 19, 2017 1/36 Lecture 9, Nov 17, 2017: Krylov space methods http://people.inf.ethz.ch/arbenz/fem17 Peter Arbenz Computer Science Department, ETH Zürich

More information

Principles and Analysis of Krylov Subspace Methods

Principles and Analysis of Krylov Subspace Methods Principles and Analysis of Krylov Subspace Methods Zdeněk Strakoš Institute of Computer Science, Academy of Sciences, Prague www.cs.cas.cz/~strakos Ostrava, February 2005 1 With special thanks to C.C.

More information

Inexactness and flexibility in linear Krylov solvers

Inexactness and flexibility in linear Krylov solvers Inexactness and flexibility in linear Krylov solvers Luc Giraud ENSEEIHT (N7) - IRIT, Toulouse Matrix Analysis and Applications CIRM Luminy - October 15-19, 2007 in honor of Gérard Meurant for his 60 th

More information

Iterative methods for Linear System

Iterative methods for Linear System Iterative methods for Linear System JASS 2009 Student: Rishi Patil Advisor: Prof. Thomas Huckle Outline Basics: Matrices and their properties Eigenvalues, Condition Number Iterative Methods Direct and

More information

4.6 Iterative Solvers for Linear Systems

4.6 Iterative Solvers for Linear Systems 4.6 Iterative Solvers for Linear Systems Why use iterative methods? Virtually all direct methods for solving Ax = b require O(n 3 ) floating point operations. In practical applications the matrix A often

More information

Regularized HSS iteration methods for saddle-point linear systems

Regularized HSS iteration methods for saddle-point linear systems BIT Numer Math DOI 10.1007/s10543-016-0636-7 Regularized HSS iteration methods for saddle-point linear systems Zhong-Zhi Bai 1 Michele Benzi 2 Received: 29 January 2016 / Accepted: 20 October 2016 Springer

More information

Topics. The CG Algorithm Algorithmic Options CG s Two Main Convergence Theorems

Topics. The CG Algorithm Algorithmic Options CG s Two Main Convergence Theorems Topics The CG Algorithm Algorithmic Options CG s Two Main Convergence Theorems What about non-spd systems? Methods requiring small history Methods requiring large history Summary of solvers 1 / 52 Conjugate

More information

Conjugate gradient method. Descent method. Conjugate search direction. Conjugate Gradient Algorithm (294)

Conjugate gradient method. Descent method. Conjugate search direction. Conjugate Gradient Algorithm (294) Conjugate gradient method Descent method Hestenes, Stiefel 1952 For A N N SPD In exact arithmetic, solves in N steps In real arithmetic No guaranteed stopping Often converges in many fewer than N steps

More information

THE solution of the absolute value equation (AVE) of

THE solution of the absolute value equation (AVE) of The nonlinear HSS-like iterative method for absolute value equations Mu-Zheng Zhu Member, IAENG, and Ya-E Qi arxiv:1403.7013v4 [math.na] 2 Jan 2018 Abstract Salkuyeh proposed the Picard-HSS iteration method

More information

Recent advances in approximation using Krylov subspaces. V. Simoncini. Dipartimento di Matematica, Università di Bologna.

Recent advances in approximation using Krylov subspaces. V. Simoncini. Dipartimento di Matematica, Università di Bologna. Recent advances in approximation using Krylov subspaces V. Simoncini Dipartimento di Matematica, Università di Bologna and CIRSA, Ravenna, Italy valeria@dm.unibo.it 1 The framework It is given an operator

More information

9.1 Preconditioned Krylov Subspace Methods

9.1 Preconditioned Krylov Subspace Methods Chapter 9 PRECONDITIONING 9.1 Preconditioned Krylov Subspace Methods 9.2 Preconditioned Conjugate Gradient 9.3 Preconditioned Generalized Minimal Residual 9.4 Relaxation Method Preconditioners 9.5 Incomplete

More information

Preconditioners for the incompressible Navier Stokes equations

Preconditioners for the incompressible Navier Stokes equations Preconditioners for the incompressible Navier Stokes equations C. Vuik M. ur Rehman A. Segal Delft Institute of Applied Mathematics, TU Delft, The Netherlands SIAM Conference on Computational Science and

More information

Preconditioned GMRES Revisited

Preconditioned GMRES Revisited Preconditioned GMRES Revisited Roland Herzog Kirk Soodhalter UBC (visiting) RICAM Linz Preconditioning Conference 2017 Vancouver August 01, 2017 Preconditioned GMRES Revisited Vancouver 1 / 32 Table of

More information

INCOMPLETE FACTORIZATION CONSTRAINT PRECONDITIONERS FOR SADDLE-POINT MATRICES

INCOMPLETE FACTORIZATION CONSTRAINT PRECONDITIONERS FOR SADDLE-POINT MATRICES INCOMPLEE FACORIZAION CONSRAIN PRECONDIIONERS FOR SADDLE-POIN MARICES H. S. DOLLAR AND A. J. WAHEN Abstract. We consider the application of the conjugate gradient method to the solution of large symmetric,

More information

Performance Comparison of Relaxation Methods with Singular and Nonsingular Preconditioners for Singular Saddle Point Problems

Performance Comparison of Relaxation Methods with Singular and Nonsingular Preconditioners for Singular Saddle Point Problems Applied Mathematical Sciences, Vol. 10, 2016, no. 30, 1477-1488 HIKARI Ltd, www.m-hikari.com http://dx.doi.org/10.12988/ams.2016.6269 Performance Comparison of Relaxation Methods with Singular and Nonsingular

More information

Contribution of Wo¹niakowski, Strako²,... The conjugate gradient method in nite precision computa

Contribution of Wo¹niakowski, Strako²,... The conjugate gradient method in nite precision computa Contribution of Wo¹niakowski, Strako²,... The conjugate gradient method in nite precision computations ªaw University of Technology Institute of Mathematics and Computer Science Warsaw, October 7, 2006

More information

AMS526: Numerical Analysis I (Numerical Linear Algebra)

AMS526: Numerical Analysis I (Numerical Linear Algebra) AMS526: Numerical Analysis I (Numerical Linear Algebra) Lecture 23: GMRES and Other Krylov Subspace Methods Xiangmin Jiao SUNY Stony Brook Xiangmin Jiao Numerical Analysis I 1 / 9 Minimizing Residual CG

More information

arxiv: v2 [math.na] 1 Feb 2013

arxiv: v2 [math.na] 1 Feb 2013 A FRAMEWORK FOR DEFLATED AND AUGMENTED KRYLOV SUBSPACE METHODS ANDRÉ GAUL, MARTIN H. GUTKNECHT, JÖRG LIESEN AND REINHARD NABBEN arxiv:1206.1506v2 [math.na] 1 Feb 2013 Abstract. We consider deflation and

More information

Summary of Iterative Methods for Non-symmetric Linear Equations That Are Related to the Conjugate Gradient (CG) Method

Summary of Iterative Methods for Non-symmetric Linear Equations That Are Related to the Conjugate Gradient (CG) Method Summary of Iterative Methods for Non-symmetric Linear Equations That Are Related to the Conjugate Gradient (CG) Method Leslie Foster 11-5-2012 We will discuss the FOM (full orthogonalization method), CG,

More information

ITERATIVE METHODS FOR SPARSE LINEAR SYSTEMS

ITERATIVE METHODS FOR SPARSE LINEAR SYSTEMS ITERATIVE METHODS FOR SPARSE LINEAR SYSTEMS YOUSEF SAAD University of Minnesota PWS PUBLISHING COMPANY I(T)P An International Thomson Publishing Company BOSTON ALBANY BONN CINCINNATI DETROIT LONDON MADRID

More information

SOLVING HERMITIAN POSITIVE DEFINITE SYSTEMS USING INDEFINITE INCOMPLETE FACTORIZATIONS

SOLVING HERMITIAN POSITIVE DEFINITE SYSTEMS USING INDEFINITE INCOMPLETE FACTORIZATIONS SOLVING HERMITIAN POSITIVE DEFINITE SYSTEMS USING INDEFINITE INCOMPLETE FACTORIZATIONS HAIM AVRON, ANSHUL GUPTA, AND SIVAN TOLEDO Abstract. Incomplete LDL factorizations sometimes produce an indenite preconditioner

More information

Key words. inf-sup constant, iterative solvers, preconditioning, saddle point problems

Key words. inf-sup constant, iterative solvers, preconditioning, saddle point problems NATURAL PRECONDITIONING AND ITERATIVE METHODS FOR SADDLE POINT SYSTEMS JENNIFER PESTANA AND ANDREW J. WATHEN Abstract. The solution of quadratic or locally quadratic extremum problems subject to linear(ized)

More information

Splitting Iteration Methods for Positive Definite Linear Systems

Splitting Iteration Methods for Positive Definite Linear Systems Splitting Iteration Methods for Positive Definite Linear Systems Zhong-Zhi Bai a State Key Lab. of Sci./Engrg. Computing Inst. of Comput. Math. & Sci./Engrg. Computing Academy of Mathematics and System

More information

OUTLINE ffl CFD: elliptic pde's! Ax = b ffl Basic iterative methods ffl Krylov subspace methods ffl Preconditioning techniques: Iterative methods ILU

OUTLINE ffl CFD: elliptic pde's! Ax = b ffl Basic iterative methods ffl Krylov subspace methods ffl Preconditioning techniques: Iterative methods ILU Preconditioning Techniques for Solving Large Sparse Linear Systems Arnold Reusken Institut für Geometrie und Praktische Mathematik RWTH-Aachen OUTLINE ffl CFD: elliptic pde's! Ax = b ffl Basic iterative

More information

Linear algebra issues in Interior Point methods for bound-constrained least-squares problems

Linear algebra issues in Interior Point methods for bound-constrained least-squares problems Linear algebra issues in Interior Point methods for bound-constrained least-squares problems Stefania Bellavia Dipartimento di Energetica S. Stecco Università degli Studi di Firenze Joint work with Jacek

More information

Block Bidiagonal Decomposition and Least Squares Problems

Block Bidiagonal Decomposition and Least Squares Problems Block Bidiagonal Decomposition and Least Squares Problems Åke Björck Department of Mathematics Linköping University Perspectives in Numerical Analysis, Helsinki, May 27 29, 2008 Outline Bidiagonal Decomposition

More information

Parallel Numerics, WT 2016/ Iterative Methods for Sparse Linear Systems of Equations. page 1 of 1

Parallel Numerics, WT 2016/ Iterative Methods for Sparse Linear Systems of Equations. page 1 of 1 Parallel Numerics, WT 2016/2017 5 Iterative Methods for Sparse Linear Systems of Equations page 1 of 1 Contents 1 Introduction 1.1 Computer Science Aspects 1.2 Numerical Problems 1.3 Graphs 1.4 Loop Manipulations

More information

The Lanczos and conjugate gradient algorithms

The Lanczos and conjugate gradient algorithms The Lanczos and conjugate gradient algorithms Gérard MEURANT October, 2008 1 The Lanczos algorithm 2 The Lanczos algorithm in finite precision 3 The nonsymmetric Lanczos algorithm 4 The Golub Kahan bidiagonalization

More information

Lecture 3: Inexact inverse iteration with preconditioning

Lecture 3: Inexact inverse iteration with preconditioning Lecture 3: Department of Mathematical Sciences CLAPDE, Durham, July 2008 Joint work with M. Freitag (Bath), and M. Robbé & M. Sadkane (Brest) 1 Introduction 2 Preconditioned GMRES for Inverse Power Method

More information

Iterative Methods for Sparse Linear Systems

Iterative Methods for Sparse Linear Systems Iterative Methods for Sparse Linear Systems Luca Bergamaschi e-mail: berga@dmsa.unipd.it - http://www.dmsa.unipd.it/ berga Department of Mathematical Methods and Models for Scientific Applications University

More information

Comparison of Fixed Point Methods and Krylov Subspace Methods Solving Convection-Diffusion Equations

Comparison of Fixed Point Methods and Krylov Subspace Methods Solving Convection-Diffusion Equations American Journal of Computational Mathematics, 5, 5, 3-6 Published Online June 5 in SciRes. http://www.scirp.org/journal/ajcm http://dx.doi.org/.436/ajcm.5.5 Comparison of Fixed Point Methods and Krylov

More information

Journal of Computational and Applied Mathematics. Multigrid method for solving convection-diffusion problems with dominant convection

Journal of Computational and Applied Mathematics. Multigrid method for solving convection-diffusion problems with dominant convection Journal of Computational and Applied Mathematics 226 (2009) 77 83 Contents lists available at ScienceDirect Journal of Computational and Applied Mathematics journal homepage: www.elsevier.com/locate/cam

More information

The Nullspace free eigenvalue problem and the inexact Shift and invert Lanczos method. V. Simoncini. Dipartimento di Matematica, Università di Bologna

The Nullspace free eigenvalue problem and the inexact Shift and invert Lanczos method. V. Simoncini. Dipartimento di Matematica, Università di Bologna The Nullspace free eigenvalue problem and the inexact Shift and invert Lanczos method V. Simoncini Dipartimento di Matematica, Università di Bologna and CIRSA, Ravenna, Italy valeria@dm.unibo.it 1 The

More information

Convergence Properties of Preconditioned Hermitian and Skew-Hermitian Splitting Methods for Non-Hermitian Positive Semidefinite Matrices

Convergence Properties of Preconditioned Hermitian and Skew-Hermitian Splitting Methods for Non-Hermitian Positive Semidefinite Matrices Convergence Properties of Preconditioned Hermitian and Skew-Hermitian Splitting Methods for Non-Hermitian Positive Semidefinite Matrices Zhong-Zhi Bai 1 Department of Mathematics, Fudan University Shanghai

More information

ON AUGMENTED LAGRANGIAN METHODS FOR SADDLE-POINT LINEAR SYSTEMS WITH SINGULAR OR SEMIDEFINITE (1,1) BLOCKS * 1. Introduction

ON AUGMENTED LAGRANGIAN METHODS FOR SADDLE-POINT LINEAR SYSTEMS WITH SINGULAR OR SEMIDEFINITE (1,1) BLOCKS * 1. Introduction Journal of Computational Mathematics Vol.xx, No.x, 200x, 1 9. http://www.global-sci.org/jcm doi:10.4208/jcm.1401-cr7 ON AUGMENED LAGRANGIAN MEHODS FOR SADDLE-POIN LINEAR SYSEMS WIH SINGULAR OR SEMIDEFINIE

More information

Key words. GMRES method, convergence bounds, worst-case GMRES, ideal GMRES, field of values

Key words. GMRES method, convergence bounds, worst-case GMRES, ideal GMRES, field of values THE FIELD OF VALUES BOUNDS ON IDEAL GMRES JÖRG LIESEN AND PETR TICHÝ 27.03.2018) Abstract. A widely known result of Elman, and its improvements due to Starke, Eiermann and Ernst, gives a bound on the worst-case

More information

Stabilization and Acceleration of Algebraic Multigrid Method

Stabilization and Acceleration of Algebraic Multigrid Method Stabilization and Acceleration of Algebraic Multigrid Method Recursive Projection Algorithm A. Jemcov J.P. Maruszewski Fluent Inc. October 24, 2006 Outline 1 Need for Algorithm Stabilization and Acceleration

More information

Mathematics and Computer Science

Mathematics and Computer Science Technical Report TR-2010-026 On Preconditioned MHSS Iteration Methods for Complex Symmetric Linear Systems by Zhong-Zhi Bai, Michele Benzi, Fang Chen Mathematics and Computer Science EMORY UNIVERSITY On

More information

Linear and Non-Linear Preconditioning

Linear and Non-Linear Preconditioning and Non- martin.gander@unige.ch University of Geneva July 2015 Joint work with Victorita Dolean, Walid Kheriji, Felix Kwok and Roland Masson (1845): First Idea of After preconditioning, it takes only three

More information

6.4 Krylov Subspaces and Conjugate Gradients

6.4 Krylov Subspaces and Conjugate Gradients 6.4 Krylov Subspaces and Conjugate Gradients Our original equation is Ax = b. The preconditioned equation is P Ax = P b. When we write P, we never intend that an inverse will be explicitly computed. P

More information

CONVERGENCE BOUNDS FOR PRECONDITIONED GMRES USING ELEMENT-BY-ELEMENT ESTIMATES OF THE FIELD OF VALUES

CONVERGENCE BOUNDS FOR PRECONDITIONED GMRES USING ELEMENT-BY-ELEMENT ESTIMATES OF THE FIELD OF VALUES European Conference on Computational Fluid Dynamics ECCOMAS CFD 2006 P. Wesseling, E. Oñate and J. Périaux (Eds) c TU Delft, The Netherlands, 2006 CONVERGENCE BOUNDS FOR PRECONDITIONED GMRES USING ELEMENT-BY-ELEMENT

More information

A DISSERTATION. Extensions of the Conjugate Residual Method. by Tomohiro Sogabe. Presented to

A DISSERTATION. Extensions of the Conjugate Residual Method. by Tomohiro Sogabe. Presented to A DISSERTATION Extensions of the Conjugate Residual Method ( ) by Tomohiro Sogabe Presented to Department of Applied Physics, The University of Tokyo Contents 1 Introduction 1 2 Krylov subspace methods

More information

DELFT UNIVERSITY OF TECHNOLOGY

DELFT UNIVERSITY OF TECHNOLOGY DELFT UNIVERSITY OF TECHNOLOGY REPORT 13-10 Comparison of some preconditioners for the incompressible Navier-Stokes equations X. He and C. Vuik ISSN 1389-6520 Reports of the Delft Institute of Applied

More information

NUMERICS OF THE GRAM-SCHMIDT ORTHOGONALIZATION PROCESS

NUMERICS OF THE GRAM-SCHMIDT ORTHOGONALIZATION PROCESS NUMERICS OF THE GRAM-SCHMIDT ORTHOGONALIZATION PROCESS Miro Rozložník Institute of Computer Science, Czech Academy of Sciences, Prague, Czech Republic email: miro@cs.cas.cz joint results with Luc Giraud,

More information

ETNA Kent State University

ETNA Kent State University Electronic ransactions on Numerical Analysis. Volume 2, pp. 57-75, March 1994. Copyright 1994,. ISSN 1068-9613. ENA MINIMIZAION PROPERIES AND SHOR RECURRENCES FOR KRYLOV SUBSPACE MEHODS RÜDIGER WEISS Dedicated

More information

On the Vorobyev method of moments

On the Vorobyev method of moments On the Vorobyev method of moments Zdeněk Strakoš Charles University in Prague and Czech Academy of Sciences http://www.karlin.mff.cuni.cz/ strakos Conference in honor of Volker Mehrmann Berlin, May 2015

More information

JURJEN DUINTJER TEBBENS

JURJEN DUINTJER TEBBENS Proceedings of ALGORITMY 005 pp. 40 49 GMRES ACCELERATION ANALYSIS FOR A CONVECTION DIFFUSION MODEL PROBLEM JURJEN DUINTJER TEBBENS Abstract. When we apply the GMRES method [4] to linear systems arising

More information

Preliminary/Qualifying Exam in Numerical Analysis (Math 502a) Spring 2012

Preliminary/Qualifying Exam in Numerical Analysis (Math 502a) Spring 2012 Instructions Preliminary/Qualifying Exam in Numerical Analysis (Math 502a) Spring 2012 The exam consists of four problems, each having multiple parts. You should attempt to solve all four problems. 1.

More information

The Mixed Finite Element Multigrid Preconditioned Minimum Residual Method for Stokes Equations

The Mixed Finite Element Multigrid Preconditioned Minimum Residual Method for Stokes Equations The Mixed Finite Element Multigrid Preconditioned Minimum Residual Method for Stokes Equations K. Muzhinji, S. Shateyi, and S, S. Motsa 2 University of Venda, Department of Mathematics, P Bag X5050, Thohoyandou

More information

Solving Sparse Linear Systems: Iterative methods

Solving Sparse Linear Systems: Iterative methods Scientific Computing with Case Studies SIAM Press, 2009 http://www.cs.umd.edu/users/oleary/sccs Lecture Notes for Unit VII Sparse Matrix Computations Part 2: Iterative Methods Dianne P. O Leary c 2008,2010

More information

Solving Sparse Linear Systems: Iterative methods

Solving Sparse Linear Systems: Iterative methods Scientific Computing with Case Studies SIAM Press, 2009 http://www.cs.umd.edu/users/oleary/sccswebpage Lecture Notes for Unit VII Sparse Matrix Computations Part 2: Iterative Methods Dianne P. O Leary

More information

Master Thesis Literature Study Presentation

Master Thesis Literature Study Presentation Master Thesis Literature Study Presentation Delft University of Technology The Faculty of Electrical Engineering, Mathematics and Computer Science January 29, 2010 Plaxis Introduction Plaxis Finite Element

More information

Block-triangular preconditioners for PDE-constrained optimization

Block-triangular preconditioners for PDE-constrained optimization NUMERICAL LINEAR ALGEBRA WITH APPLICATIONS Numer. Linear Algebra Appl. () Published online in Wiley InterScience (www.interscience.wiley.com). DOI:./nla.693 Block-triangular preconditioners for PDE-constrained

More information

Generalized MINRES or Generalized LSQR?

Generalized MINRES or Generalized LSQR? Generalized MINRES or Generalized LSQR? Michael Saunders Systems Optimization Laboratory (SOL) Institute for Computational Mathematics and Engineering (ICME) Stanford University New Frontiers in Numerical

More information

Stopping criteria for iterations in finite element methods

Stopping criteria for iterations in finite element methods Stopping criteria for iterations in finite element methods M. Arioli, D. Loghin and A. J. Wathen CERFACS Technical Report TR/PA/03/21 Abstract This work extends the results of Arioli [1], [2] on stopping

More information

SEMI-CONVERGENCE ANALYSIS OF THE INEXACT UZAWA METHOD FOR SINGULAR SADDLE POINT PROBLEMS

SEMI-CONVERGENCE ANALYSIS OF THE INEXACT UZAWA METHOD FOR SINGULAR SADDLE POINT PROBLEMS REVISTA DE LA UNIÓN MATEMÁTICA ARGENTINA Vol. 53, No. 1, 2012, 61 70 SEMI-CONVERGENCE ANALYSIS OF THE INEXACT UZAWA METHOD FOR SINGULAR SADDLE POINT PROBLEMS JIAN-LEI LI AND TING-ZHU HUANG Abstract. Recently,

More information

Iterative solvers for saddle point algebraic linear systems: tools of the trade. V. Simoncini

Iterative solvers for saddle point algebraic linear systems: tools of the trade. V. Simoncini Iterative solvers for saddle point algebraic linear systems: tools of the trade V. Simoncini Dipartimento di Matematica, Università di Bologna and CIRSA, Ravenna, Italy valeria@dm.unibo.it 1 The problem

More information

Iterative methods for positive definite linear systems with a complex shift

Iterative methods for positive definite linear systems with a complex shift Iterative methods for positive definite linear systems with a complex shift William McLean, University of New South Wales Vidar Thomée, Chalmers University November 4, 2011 Outline 1. Numerical solution

More information

Iterative Methods for Linear Systems of Equations

Iterative Methods for Linear Systems of Equations Iterative Methods for Linear Systems of Equations Projection methods (3) ITMAN PhD-course DTU 20-10-08 till 24-10-08 Martin van Gijzen 1 Delft University of Technology Overview day 4 Bi-Lanczos method

More information

On the loss of orthogonality in the Gram-Schmidt orthogonalization process

On the loss of orthogonality in the Gram-Schmidt orthogonalization process CERFACS Technical Report No. TR/PA/03/25 Luc Giraud Julien Langou Miroslav Rozložník On the loss of orthogonality in the Gram-Schmidt orthogonalization process Abstract. In this paper we study numerical

More information

Preconditioning Techniques Analysis for CG Method

Preconditioning Techniques Analysis for CG Method Preconditioning Techniques Analysis for CG Method Huaguang Song Department of Computer Science University of California, Davis hso@ucdavis.edu Abstract Matrix computation issue for solve linear system

More information