On solving linear systems arising from Shishkin mesh discretizations
|
|
- Ashlie Jenkins
- 5 years ago
- Views:
Transcription
1 On solving linear systems arising from Shishkin mesh discretizations Petr Tichý Faculty of Mathematics and Physics, Charles University joint work with Carlos Echeverría, Jörg Liesen, and Daniel Szyld October 20, 2016, Prague Seminar of Numerical Mathematics, KNM MFF UK 1
2 Problem formulation Convection-diffusion boundary value problem ɛ u + α u + β u = f in Ω = (0, 1), u(0) = u 0, u(1) = u 1, α > 0, β 0 constants, the problem is convection dominated 0 < ɛ α. [Stynes, 2005] (Acta Numerica) [Roos, Stynes, and Tobiska, 1996, 2008] (book) [Miller, O Riordan, and Shishkin, 1996] (book) 2
3 Solution and boundary layers ɛ = 0.01, α = 1, β = 0, u(0) = u(1) = 0. There are small subregions where the solution has a large gradient. 3
4 Numerical solution, equidistant mesh Standard techniques: u (ih) u i+1 u i 1 2h, u (ih) u i u i 1. h Unnatural oscillations or cannot resolve the layers. Remedy: stabilization or non-equidistant mesh. We study discretizations for a Shishkin mesh. 4
5 Outline 1 Shishkin mesh and discretization 2 How to solve the linear system? 3 Multiplicative Schwarz method 4 Convergence analysis 5 Schwarz method as a preconditioner 6 Numerical examples 5
6 Shishkin mesh on [0, 1] Piecewise equidistant N even, define the transition point 1 τ and n by { 1 τ min 2, ɛ } α 2 ln N, n N 2. If ɛ α, then 1 τ is close to 1. Next define H and h by H 1 τ n, h τ n and consider the Shishkin mesh, x 0 = 0, x i ih, x n+i x n + ih, i = 1,..., n. 6
7 Discretization on the Shishkin mesh - details For simplicity u(0) = u(1) = 0 The upwind difference scheme is given by ɛ δ 2 xu i + α D x u i + βu i = f i, u 0 = u N = 0, and the central difference scheme by where ɛ δ 2 xu i + αd 0 x + βu i = f i, u 0 = u N = 0, δ 2 xu i = 2u i 1 (H + h)h 2u i Hh + 2u i+1 (H + h)h, i = n, and Dx u i = u i u i 1, 1 i n, D 0 H xu i = u i+1 u i 1 h + H, i = n. surveys [Linss, Stynes, 2001], [Stynes, 2005], [Kopteva, O Riordan, 2010] 7
8 Shishkin mesh and ɛ-uniform convergence The upwind difference scheme ɛ δ 2 xu i + α D x u i + βu i = f i, u 0 = u N = 0. There exists a constant C such that ( ) ln N u(x i ) u i C, i = 0,..., N, N see, e.g., [Stynes, 2005]. 8
9 Shishkin mesh and ɛ-uniform convergence The upwind difference scheme ɛ δ 2 xu i + α D x u i + βu i = f i, u 0 = u N = 0. There exists a constant C such that ( ) ln N u(x i ) u i C, i = 0,..., N, N see, e.g., [Stynes, 2005]. The central difference scheme ɛ δ 2 xu i + α D 0 xu i + βu i = f i, u 0 = u N = 0. There exists a constant C such that ( ) ln N 2 u(x i ) u i C, i = 0,..., N, N [Andreev and Kopteva, 1996], a difficult proof, the scheme does not satisfy a discrete maximum principle. [Kopteva and Linss, 2001]. 8
10 Outline 1 Shishkin mesh and discretization 2 How to solve the linear system? 3 Multiplicative Schwarz method 4 Convergence analysis 5 Schwarz method as a preconditioner 6 Numerical examples 9
11 Structure of the matrix A = a H b H c H bh c H a H b H c a b c h a h b h c h bh c h a h 10
12 Entries The upwind scheme c H = ɛ H 2 α H, 2ɛ c = H(H + h) α H, c h = ɛ h 2 α h, a H = 2ɛ H 2 + α H + β, b H = ɛ H 2, a = 2ɛ hh + α H + β, b = 2ɛ h(h + h), a h = 2ɛ h 2 + α h + β, b h = ɛ h 2. The central difference scheme c H = ɛ H 2 α 2H, 2ɛ c = H(H + h) α h + H, c h = ɛ h 2 α 2h, b H = ɛ H 2 + α 2H, a = 2ɛ hh + β, b = 2ɛ h(h + h) + a H = 2ɛ H 2 + β, a h = 2ɛ h 2 + β, b h = ɛ h 2 + α 2h. α h + H, 11
13 Matrix properties nonsymmetric A is M-matrix for the upwind scheme A is not an M-matrix for the central difference scheme A is highly nonnormal. Consider ɛu + u = 1, u(0) = 0, u(1) = 0, ɛ = 10 8 and N = 46, and the spectral decomposition A = Y DY 1. 12
14 Matrix properties nonsymmetric A is M-matrix for the upwind scheme A is not an M-matrix for the central difference scheme A is highly nonnormal. Consider ɛu + u = 1, u(0) = 0, u(1) = 0, ɛ = 10 8 and N = 46, and the spectral decomposition A = Y DY 1. upwind upwind sc. central central sc. κ(a) κ(y )
15 Solving linear system using GMRES 10 0 Convergence of GMRES for the upwind scheme relative residuial norm GMRES, A original GMRES, A scaled number of iterations 13
16 Outline 1 Shishkin mesh and discretization 2 How to solve the linear system? 3 Multiplicative Schwarz method 4 Convergence analysis 5 Schwarz method as a preconditioner 6 Numerical examples 14
17 Multiplicative Schwarz method Idea of solving Ax = b Given an approximation x (k), then x = x (k) + y and y satisfies Ay = b Ax (k). 15
18 Multiplicative Schwarz method Idea of solving Ax = b Given an approximation x (k), then x = x (k) + y and y satisfies Ay = b Ax (k). [ Restriction operators R 1 = I n 0 ] [, R 2 = 0 I n ]. 15
19 Multiplicative Schwarz method Idea of solving Ax = b Given an approximation x (k), then x = x (k) + y and y satisfies Ay = b Ax (k). [ Restriction operators R 1 = Solve on the first domain I n 0 ] [, R 2 = 0 I n ]. (R 1 AR T 1 ) ỹ = R 1 (b Ax (k) ) and approximate y by prolongation of ỹ, i.e., by R1 T ỹ. Define x (k+ 1 2 ) = x (k) + R T 1 (R 1 AR T 1 ) 1 R 1 (b Ax (k) ). 15
20 Multiplicative Schwarz method Idea of solving Ax = b Given an approximation x (k), then x = x (k) + y and y satisfies Ay = b Ax (k). [ Restriction operators R 1 = Solve on the first domain I n 0 ] [, R 2 = 0 I n ]. (R 1 AR T 1 ) ỹ = R 1 (b Ax (k) ) and approximate y by prolongation of ỹ, i.e., by R1 T ỹ. Define x (k+ 1 2 ) = x (k) + R T 1 (R 1 AR T 1 ) 1 R 1 (b Ax (k) ). Similarly, use x (k+ 1 2 ) on the second domain and prolong, x (k+1) = x (k+ 1 2 ) + R T 2 (R 2 AR T 2 ) 1 R 2 (b Ax (k+ 1 2 )). 15
21 Multiplicative Schwarz method Formalism Define P i = R T i A 1 i R i A, A i R i AR T i, i = 1, 2. The multiplicative Schwarz is the iterative scheme x (k+1) = T x (k) + v, T = (I P 2 )(I P 1 ), where v is defined such that the scheme is consistent. 16
22 Multiplicative Schwarz method Formalism Define P i = R T i A 1 i R i A, A i R i AR T i, i = 1, 2. The multiplicative Schwarz is the iterative scheme x (k+1) = T x (k) + v, T = (I P 2 )(I P 1 ), where v is defined such that the scheme is consistent. Hence, and Is it convergent in our case? x x (k+1) = T k+1 (x x 0 ), x x (k+1) T k+1 x x 0. 16
23 Multiplicative Schwarz method Experiment 10 0 upwind difference scheme 10 2 Relative Residual GMRES Multiplicative Schwarz Iterations 17
24 Outline 1 Shishkin mesh and discretization 2 How to solve the linear system? 3 Multiplicative Schwarz method 4 Convergence analysis 5 Schwarz method as a preconditioner 6 Numerical examples 18
25 Convergence analysis Exploiting the structure x x (k+1) x x 0 T k+1. Using T = (I P 2 )(I P 1 ) we are able to show that t 1. T = t n = t e T n+1.. t N 1 19
26 Convergence analysis Exploiting the structure x x (k+1) x x 0 T k+1. Using T = (I P 2 )(I P 1 ) we are able to show that t 1. T = t n = t e T n+1.. t N 1 Therefore, T 2 = t (e T n+1 t) et n+1 = t n+1 T, and T k+1 = t n+1 k T. How to bound t n+1, and T in a convenient norm ( )? 19
27 Convergence analysis Details A = A H b H c a b c h A h. Let m n 1, ρ t n+1... the convergence factor. Then, ( ) bb H A 1 H ρ = ( ) a cb H A 1 H m,m m,m ( ) cc h A 1 h a bc h ( 1,1 ) A 1 h 1,1. 20
28 Convergence analysis Bounding ( A 1 H )m,m and ( ) A 1 h 1,1 A matrix B = [b i,j ] is called a nonsingular M-matrix when B is nonsingular, b i,i > 0 for all i, b i,j 0 for all i j, and B 1 0 (elementwise). 21
29 Convergence analysis Bounding ( A 1 H )m,m and ( ) A 1 h 1,1 A matrix B = [b i,j ] is called a nonsingular M-matrix when B is nonsingular, b i,i > 0 for all i, b i,j 0 for all i j, and B 1 0 (elementwise). If A H and A h are nonsingular M-matrices, then using [Nabben 1999], ( { } 1 A 1 H )m,m min b H, 1, c H ( A 1 h )1,1 min { 1 } b h, 1 c h. A sufficient condition: The sign conditions & irreducibly diagonal dominant nonsingular M-matrix. [Meurant, 1996], [Hackbusch, 2010] 21
30 Convergence analysis The upwind scheme The matrices A H and A h are M-matrices, and we know that e (k+1) e (0) ρ k T. 22
31 Convergence analysis The upwind scheme The matrices A H and A h are M-matrices, and we know that e (k+1) e (0) ρ k T. Theorem (the upwind scheme) [Echeverría, Liesen, T., Szyld, 2016] For the upwind scheme we have ρ ɛ ɛ + αh ɛ ɛ + α, N and ɛ T ɛ + αh. 22
32 Convergence analysis The central difference scheme A h is still an M-matrix. If αh > 2ɛ, i.e. b H > 0, then A H is not an M-matrix... the most common situation from a practical point of view. 23
33 Convergence analysis The central difference scheme A h is still an M-matrix. If αh > 2ɛ, i.e. b H > 0, then A H is not an M-matrix... the most common situation from a practical point of view. Recall ( ) bb H A 1 H ρ = ( ) a cb H A 1 H m,m m,m ( ) cc h A 1 h a bc h ( 1,1 ) A 1 h 1,1 How to bound ( A 1 ) H m,m?... results by [Usmani 1994]. 23
34 Convergence analysis The central difference scheme A h is still an M-matrix. If αh > 2ɛ, i.e. b H > 0, then A H is not an M-matrix... the most common situation from a practical point of view. Recall ( ) bb H A 1 H ρ = ( ) a cb H A 1 H m,m m,m ( ) cc h A 1 h a bc h ( 1,1 ) A 1 h 1,1 How to bound ( A 1 ) H m,m?... results by [Usmani 1994] We proved: If m = N/2 1 is even, then b H (A 1 H ) m,m 1 b H m ch c H bh + b H m < 2mɛ ɛ + αh. ch 2. 23
35 Convergence analysis The central difference scheme A h is M-matrix, if αh > 2ɛ, A H is not an M-matrix. Theorem (the central diff. scheme) [Echeverría, Liesen, T., Szyld, 2016] Let m = N/2 1 be even, and let αh > 2ɛ. For the central differences we have ρ < 2mɛ ɛ + αh 2 < N ɛ ɛ + α, N T < 2. Thus, the error of the multiplicative Schwarz method satisfies e (k+1) ( 2mɛ e (0) < 2 ɛ + αh 2 ) k. 24
36 Remarks on diagonally scaled linear algebraic systems DAx = Db The ill-conditioning can be avoided by diagonal scaling [Roos 1996]: Ax = b is multiplied from the left by d H I m D = d. d h I m 25
37 Remarks on diagonally scaled linear algebraic systems DAx = Db The ill-conditioning can be avoided by diagonal scaling [Roos 1996]: Ax = b is multiplied from the left by d H I m D = d. d h I m Such a scaling preserves the Toeplitz structure and the M-matrix property of the submatrices. Analysis depends on these properties and on ratios between elements in the same row such as b/a and b H /c H. These ratios are invariant under diagonal scaling. 25
38 Remarks on diagonally scaled linear algebraic systems DAx = Db The ill-conditioning can be avoided by diagonal scaling [Roos 1996]: Ax = b is multiplied from the left by d H I m D = d. d h I m Such a scaling preserves the Toeplitz structure and the M-matrix property of the submatrices. Analysis depends on these properties and on ratios between elements in the same row such as b/a and b H /c H. These ratios are invariant under diagonal scaling. Consequently, all convergence bounds hold for the multiplicative Schwarz method applied to DAx = Db. 25
39 Outline 1 Shishkin mesh and discretization 2 How to solve the linear system? 3 Multiplicative Schwarz method 4 Convergence analysis 5 Schwarz method as a preconditioner 6 Numerical examples 26
40 Schwarz method as a preconditioner We have consistent scheme x (k+1) = T x (k) + v. Hence, x solves Ax = b and also the preconditioned system (I T )x = v. 27
41 Schwarz method as a preconditioner We have consistent scheme x (k+1) = T x (k) + v. Hence, x solves Ax = b and also the preconditioned system (I T )x = v. We can formally define a preconditioner M such Ax = b M 1 Ax = M 1 b (I T )x = v. Clearly M = A(I T ) 1. 27
42 Schwarz method as a preconditioner We have consistent scheme x (k+1) = T x (k) + v. Hence, x solves Ax = b and also the preconditioned system (I T )x = v. We can formally define a preconditioner M such Ax = b M 1 Ax = M 1 b (I T )x = v. Clearly M = A(I T ) 1. Then x (k+1) = x (k) + (I T )(x x (k) ) = x (k) + M 1 r (k). 27
43 Schwarz method as a preconditioner for GMRES The multiplicative Schwarz method as well as GMRES applied to the preconditioned system obtain their approximations from the same Krylov subspace. In terms of the residual norm, the preconditioned GMRES will always converge faster than the multiplicative Schwarz. 28
44 Schwarz method as a preconditioner for GMRES The multiplicative Schwarz method as well as GMRES applied to the preconditioned system obtain their approximations from the same Krylov subspace. In terms of the residual norm, the preconditioned GMRES will always converge faster than the multiplicative Schwarz. Moreover, in this case, the iteration matrix T has rank-one structure, and dim (K k (I T, r 0 )) 2. Therefore, GMRES converges in at most 2 steps,... a motivation for more dimensional cases. 28
45 How to multiply by T Schwarz or preconditioned GMRES only multiply by T, T = (I P 2 )(I P 1 ), P i = R T i (R i AR T i ) 1 R i A, i.e., to solve systems of the form (m = n 1 = N 2 1) A H c b H a [ y1:m y m+1 ] = [ z1:m z m+1 ]. 29
46 How to multiply by T Schwarz or preconditioned GMRES only multiply by T, T = (I P 2 )(I P 1 ), P i = R T i (R i AR T i ) 1 R i A, i.e., to solve systems of the form (m = n 1 = N 2 1) A H c b H a Using the Schur complement, [ y1:m y m+1 ] = [ z1:m z m+1 ( A H b ) Hc a e me T b H m y 1:m = z 1:m z m+1 e m. a H Then apply Sherman-Morrison formula. We need only to solve systems with A H (Toeplitz)! ]. 29
47 Outline 1 Shishkin mesh and discretization 2 How to solve the linear system? 3 Multiplicative Schwarz method 4 Convergence analysis 5 Schwarz method as a preconditioner 6 Numerical examples 30
48 Numerical examples Consider i.e. ɛu + u = 1, u(0) = 0, u(1) = 0, α = 1, β = 0, f(x) 1. Choose N = 198, various values of ɛ. upwind central differences ɛ ρ up our bound ρ cd our bound ρ up < ɛ ɛ + αh, ρ cd < 2mɛ ɛ + αh 2. 31
49 Numerical examples Upwind, ɛ = Upwind T=(I P 2 )(I P 1 ) error bound T=(I P 1 )(I P 2 ) error bound Error norms and bounds k 32
50 Numerical examples Upwind, ɛ = Upwind T=(I P 2 )(I P 1 ) error bound T=(I P 1 )(I P 2 ) error bound Error norms and bounds k 33
51 Numerical examples Central differences, ɛ = Central differences T=(I P 2 )(I P 1 ) T=(I P 1 )(I P 2 ) error bound Error norms and bounds k 34
52 Numerical examples Central differences, ɛ = Central differences T=(I P 2 )(I P 1 ) T=(I P 1 )(I P 2 ) error bound Error norms and bounds k 35
53 Conclusions and further work We considered finite difference discretizations of the 1D singularly-perturbed convection-diffusion equation posed on a Shishkin mesh. 36
54 Conclusions and further work We considered finite difference discretizations of the 1D singularly-perturbed convection-diffusion equation posed on a Shishkin mesh. The matrices that arise from the upwind and the central difference schemes are nonsymmetric and highly nonnormal. 36
55 Conclusions and further work We considered finite difference discretizations of the 1D singularly-perturbed convection-diffusion equation posed on a Shishkin mesh. The matrices that arise from the upwind and the central difference schemes are nonsymmetric and highly nonnormal. For the upwind scheme, we proved rapid convergence of the multiplicative Schwarz method in the most relevant case Nɛ < α. 36
56 Conclusions and further work We considered finite difference discretizations of the 1D singularly-perturbed convection-diffusion equation posed on a Shishkin mesh. The matrices that arise from the upwind and the central difference schemes are nonsymmetric and highly nonnormal. For the upwind scheme, we proved rapid convergence of the multiplicative Schwarz method in the most relevant case Nɛ < α. The convergence for the central difference scheme is slower, but still rapid, when N 2 ɛ < α and if N/2 1 is even. 36
57 Conclusions and further work We considered finite difference discretizations of the 1D singularly-perturbed convection-diffusion equation posed on a Shishkin mesh. The matrices that arise from the upwind and the central difference schemes are nonsymmetric and highly nonnormal. For the upwind scheme, we proved rapid convergence of the multiplicative Schwarz method in the most relevant case Nɛ < α. The convergence for the central difference scheme is slower, but still rapid, when N 2 ɛ < α and if N/2 1 is even. Thanks to the rank-one structure of T, the preconditioned GMRES converges in two steps. 36
58 Conclusions and further work We considered finite difference discretizations of the 1D singularly-perturbed convection-diffusion equation posed on a Shishkin mesh. The matrices that arise from the upwind and the central difference schemes are nonsymmetric and highly nonnormal. For the upwind scheme, we proved rapid convergence of the multiplicative Schwarz method in the most relevant case Nɛ < α. The convergence for the central difference scheme is slower, but still rapid, when N 2 ɛ < α and if N/2 1 is even. Thanks to the rank-one structure of T, the preconditioned GMRES converges in two steps. Inspired by 1D case (preconditioner, low-rank structure), we can study 2D case. 36
59 Related papers C. Echeverría, J. Liesen, P. Tichý, and D. Szyld, [Convergence of the multiplicative Schwarz method for singularly perturbed convection-diffusion problems discretized on a Shishkin mesh, (2016), in preparation] J. Miller, E. O Riordan, and G. Shishkin, [Fitted Numerical Methods for Singular Perturbation Problems: Error Estimates in the Maximum Norm for Linear Problems in One and Two Dimensions, World Scientific, 1996.] H-G. Roos, M. Stynes, L. Tobiska, [Robust Numerical Methods for Singularly Perturbed Differential Equations, second edition, Springer Series in Computational Mathematics, Springer-Verlag, Berlin, 2008, 604 pp.] M. Stynes, [Steady-state convection-diffusion problems, Acta Numerica, 14 (2005), pp ] Thank you for your attention! 37
Abstract. 1. Introduction
Journal of Computational Mathematics Vol.28, No.2, 2010, 273 288. http://www.global-sci.org/jcm doi:10.4208/jcm.2009.10-m2870 UNIFORM SUPERCONVERGENCE OF GALERKIN METHODS FOR SINGULARLY PERTURBED PROBLEMS
More informationOptimal Left and Right Additive Schwarz Preconditioning for Minimal Residual Methods with Euclidean and Energy Norms
Optimal Left and Right Additive Schwarz Preconditioning for Minimal Residual Methods with Euclidean and Energy Norms Marcus Sarkis Worcester Polytechnic Inst., Mass. and IMPA, Rio de Janeiro and Daniel
More informationRemarks on the analysis of finite element methods on a Shishkin mesh: are Scott-Zhang interpolants applicable?
Remarks on the analysis of finite element methods on a Shishkin mesh: are Scott-Zhang interpolants applicable? Thomas Apel, Hans-G. Roos 22.7.2008 Abstract In the first part of the paper we discuss minimal
More informationParameter robust methods for second-order complex-valued reaction-diffusion equations
Postgraduate Modelling Research Group, 10 November 2017 Parameter robust methods for second-order complex-valued reaction-diffusion equations Faiza Alssaedi Supervisor: Niall Madden School of Mathematics,
More information9.1 Preconditioned Krylov Subspace Methods
Chapter 9 PRECONDITIONING 9.1 Preconditioned Krylov Subspace Methods 9.2 Preconditioned Conjugate Gradient 9.3 Preconditioned Generalized Minimal Residual 9.4 Relaxation Method Preconditioners 9.5 Incomplete
More informationSolving Symmetric Indefinite Systems with Symmetric Positive Definite Preconditioners
Solving Symmetric Indefinite Systems with Symmetric Positive Definite Preconditioners Eugene Vecharynski 1 Andrew Knyazev 2 1 Department of Computer Science and Engineering University of Minnesota 2 Department
More informationLab 1: Iterative Methods for Solving Linear Systems
Lab 1: Iterative Methods for Solving Linear Systems January 22, 2017 Introduction Many real world applications require the solution to very large and sparse linear systems where direct methods such as
More informationFast Iterative Solution of Saddle Point Problems
Michele Benzi Department of Mathematics and Computer Science Emory University Atlanta, GA Acknowledgments NSF (Computational Mathematics) Maxim Olshanskii (Mech-Math, Moscow State U.) Zhen Wang (PhD student,
More informationCholesky factorisations of linear systems coming from a finite difference method applied to singularly perturbed problems
Cholesky factorisations of linear systems coming from a finite difference method applied to singularly perturbed problems Thái Anh Nhan and Niall Madden The Boundary and Interior Layers - Computational
More informationNumerical behavior of inexact linear solvers
Numerical behavior of inexact linear solvers Miro Rozložník joint results with Zhong-zhi Bai and Pavel Jiránek Institute of Computer Science, Czech Academy of Sciences, Prague, Czech Republic The fourth
More informationOUTLINE ffl CFD: elliptic pde's! Ax = b ffl Basic iterative methods ffl Krylov subspace methods ffl Preconditioning techniques: Iterative methods ILU
Preconditioning Techniques for Solving Large Sparse Linear Systems Arnold Reusken Institut für Geometrie und Praktische Mathematik RWTH-Aachen OUTLINE ffl CFD: elliptic pde's! Ax = b ffl Basic iterative
More informationMultigrid absolute value preconditioning
Multigrid absolute value preconditioning Eugene Vecharynski 1 Andrew Knyazev 2 (speaker) 1 Department of Computer Science and Engineering University of Minnesota 2 Department of Mathematical and Statistical
More informationGoal. Robust A Posteriori Error Estimates for Stabilized Finite Element Discretizations of Non-Stationary Convection-Diffusion Problems.
Robust A Posteriori Error Estimates for Stabilized Finite Element s of Non-Stationary Convection-Diffusion Problems L. Tobiska and R. Verfürth Universität Magdeburg Ruhr-Universität Bochum www.ruhr-uni-bochum.de/num
More informationLecture 3: Inexact inverse iteration with preconditioning
Lecture 3: Department of Mathematical Sciences CLAPDE, Durham, July 2008 Joint work with M. Freitag (Bath), and M. Robbé & M. Sadkane (Brest) 1 Introduction 2 Preconditioned GMRES for Inverse Power Method
More informationPreface to the Second Edition. Preface to the First Edition
n page v Preface to the Second Edition Preface to the First Edition xiii xvii 1 Background in Linear Algebra 1 1.1 Matrices................................. 1 1.2 Square Matrices and Eigenvalues....................
More informationAMS526: Numerical Analysis I (Numerical Linear Algebra) Lecture 23: GMRES and Other Krylov Subspace Methods; Preconditioning
AMS526: Numerical Analysis I (Numerical Linear Algebra) Lecture 23: GMRES and Other Krylov Subspace Methods; Preconditioning Xiangmin Jiao SUNY Stony Brook Xiangmin Jiao Numerical Analysis I 1 / 18 Outline
More informationThe Lanczos and conjugate gradient algorithms
The Lanczos and conjugate gradient algorithms Gérard MEURANT October, 2008 1 The Lanczos algorithm 2 The Lanczos algorithm in finite precision 3 The nonsymmetric Lanczos algorithm 4 The Golub Kahan bidiagonalization
More informationNumerical Linear Algebra And Its Applications
Numerical Linear Algebra And Its Applications Xiao-Qing JIN 1 Yi-Min WEI 2 August 29, 2008 1 Department of Mathematics, University of Macau, Macau, P. R. China. 2 Department of Mathematics, Fudan University,
More informationON ORTHOGONAL REDUCTION TO HESSENBERG FORM WITH SMALL BANDWIDTH
ON ORTHOGONAL REDUCTION TO HESSENBERG FORM WITH SMALL BANDWIDTH V. FABER, J. LIESEN, AND P. TICHÝ Abstract. Numerous algorithms in numerical linear algebra are based on the reduction of a given matrix
More informationKey words. GMRES method, convergence bounds, worst-case GMRES, ideal GMRES, field of values
THE FIELD OF VALUES BOUNDS ON IDEAL GMRES JÖRG LIESEN AND PETR TICHÝ 27.03.2018) Abstract. A widely known result of Elman, and its improvements due to Starke, Eiermann and Ernst, gives a bound on the worst-case
More informationSummary of Iterative Methods for Non-symmetric Linear Equations That Are Related to the Conjugate Gradient (CG) Method
Summary of Iterative Methods for Non-symmetric Linear Equations That Are Related to the Conjugate Gradient (CG) Method Leslie Foster 11-5-2012 We will discuss the FOM (full orthogonalization method), CG,
More informationLinear Algebra and Matrix Inversion
Jim Lambers MAT 46/56 Spring Semester 29- Lecture 2 Notes These notes correspond to Section 63 in the text Linear Algebra and Matrix Inversion Vector Spaces and Linear Transformations Matrices are much
More informationDomain decomposition on different levels of the Jacobi-Davidson method
hapter 5 Domain decomposition on different levels of the Jacobi-Davidson method Abstract Most computational work of Jacobi-Davidson [46], an iterative method suitable for computing solutions of large dimensional
More informationPreconditioning for Nonsymmetry and Time-dependence
Preconditioning for Nonsymmetry and Time-dependence Andy Wathen Oxford University, UK joint work with Jen Pestana and Elle McDonald Jeju, Korea, 2015 p.1/24 Iterative methods For self-adjoint problems/symmetric
More informationAn exponentially graded mesh for singularly perturbed problems
An exponentially graded mesh for singularly perturbed problems Christos Xenophontos Department of Mathematics and Statistics University of Cyprus joint work with P. Constantinou (UCY), S. Franz (TU Dresden)
More informationNUMERICAL STUDY OF CONVECTION DIFFUSION PROBLEM IN TWO- DIMENSIONAL SPACE
IJRRAS 5 () November NUMERICAL STUDY OF CONVECTION DIFFUSION PROBLEM IN TWO- DIMENSIONAL SPACE. K. Sharath Babu,* & N. Srinivasacharyulu Research Scholar, Department of Mathematics, National Institute
More informationLecture 18 Classical Iterative Methods
Lecture 18 Classical Iterative Methods MIT 18.335J / 6.337J Introduction to Numerical Methods Per-Olof Persson November 14, 2006 1 Iterative Methods for Linear Systems Direct methods for solving Ax = b,
More informationSPLINE COLLOCATION METHOD FOR SINGULAR PERTURBATION PROBLEM. Mirjana Stojanović University of Novi Sad, Yugoslavia
GLASNIK MATEMATIČKI Vol. 37(57(2002, 393 403 SPLINE COLLOCATION METHOD FOR SINGULAR PERTURBATION PROBLEM Mirjana Stojanović University of Novi Sad, Yugoslavia Abstract. We introduce piecewise interpolating
More informationCS137 Introduction to Scientific Computing Winter Quarter 2004 Solutions to Homework #3
CS137 Introduction to Scientific Computing Winter Quarter 2004 Solutions to Homework #3 Felix Kwok February 27, 2004 Written Problems 1. (Heath E3.10) Let B be an n n matrix, and assume that B is both
More informationDepartment of Computer Science, University of Illinois at Urbana-Champaign
Department of Computer Science, University of Illinois at Urbana-Champaign Probing for Schur Complements and Preconditioning Generalized Saddle-Point Problems Eric de Sturler, sturler@cs.uiuc.edu, http://www-faculty.cs.uiuc.edu/~sturler
More informationMTH 215: Introduction to Linear Algebra
MTH 215: Introduction to Linear Algebra Lecture 5 Jonathan A. Chávez Casillas 1 1 University of Rhode Island Department of Mathematics September 20, 2017 1 LU Factorization 2 3 4 Triangular Matrices Definition
More informationSingularly Perturbed Partial Differential Equations
WDS'9 Proceedings of Contributed Papers, Part I, 54 59, 29. ISN 978-8-7378--9 MTFYZPRESS Singularly Perturbed Partial Differential Equations J. Lamač Charles University, Faculty of Mathematics and Physics,
More informationInexactness and flexibility in linear Krylov solvers
Inexactness and flexibility in linear Krylov solvers Luc Giraud ENSEEIHT (N7) - IRIT, Toulouse Matrix Analysis and Applications CIRM Luminy - October 15-19, 2007 in honor of Gérard Meurant for his 60 th
More informationJURJEN DUINTJER TEBBENS
Proceedings of ALGORITMY 005 pp. 40 49 GMRES ACCELERATION ANALYSIS FOR A CONVECTION DIFFUSION MODEL PROBLEM JURJEN DUINTJER TEBBENS Abstract. When we apply the GMRES method [4] to linear systems arising
More informationA Robust Preconditioner for the Hessian System in Elliptic Optimal Control Problems
A Robust Preconditioner for the Hessian System in Elliptic Optimal Control Problems Etereldes Gonçalves 1, Tarek P. Mathew 1, Markus Sarkis 1,2, and Christian E. Schaerer 1 1 Instituto de Matemática Pura
More informationStabilization and Acceleration of Algebraic Multigrid Method
Stabilization and Acceleration of Algebraic Multigrid Method Recursive Projection Algorithm A. Jemcov J.P. Maruszewski Fluent Inc. October 24, 2006 Outline 1 Need for Algorithm Stabilization and Acceleration
More informationLecture 11: CMSC 878R/AMSC698R. Iterative Methods An introduction. Outline. Inverse, LU decomposition, Cholesky, SVD, etc.
Lecture 11: CMSC 878R/AMSC698R Iterative Methods An introduction Outline Direct Solution of Linear Systems Inverse, LU decomposition, Cholesky, SVD, etc. Iterative methods for linear systems Why? Matrix
More informationSome recent results on positivity-preserving discretisations for the convection-diffusion equation
Some recent results on positivity-preserving discretisations for the convection-diffusion equation Gabriel R. Barrenechea 1, Volker John 2 & Petr Knobloch 3 1 Department of Mathematics and Statistics,
More informationAlgebra C Numerical Linear Algebra Sample Exam Problems
Algebra C Numerical Linear Algebra Sample Exam Problems Notation. Denote by V a finite-dimensional Hilbert space with inner product (, ) and corresponding norm. The abbreviation SPD is used for symmetric
More informationAn algebraic perspective on integer sparse recovery
An algebraic perspective on integer sparse recovery Lenny Fukshansky Claremont McKenna College (joint work with Deanna Needell and Benny Sudakov) Combinatorics Seminar USC October 31, 2018 From Wikipedia:
More informationPreconditioned inverse iteration and shift-invert Arnoldi method
Preconditioned inverse iteration and shift-invert Arnoldi method Melina Freitag Department of Mathematical Sciences University of Bath CSC Seminar Max-Planck-Institute for Dynamics of Complex Technical
More informationA Robust Preconditioned Iterative Method for the Navier-Stokes Equations with High Reynolds Numbers
Applied and Computational Mathematics 2017; 6(4): 202-207 http://www.sciencepublishinggroup.com/j/acm doi: 10.11648/j.acm.20170604.18 ISSN: 2328-5605 (Print); ISSN: 2328-5613 (Online) A Robust Preconditioned
More informationInexact inverse iteration with preconditioning
Department of Mathematical Sciences Computational Methods with Applications Harrachov, Czech Republic 24th August 2007 (joint work with M. Robbé and M. Sadkane (Brest)) 1 Introduction 2 Preconditioned
More informationOn the Superlinear Convergence of MINRES. Valeria Simoncini and Daniel B. Szyld. Report January 2012
On the Superlinear Convergence of MINRES Valeria Simoncini and Daniel B. Szyld Report 12-01-11 January 2012 This report is available in the World Wide Web at http://www.math.temple.edu/~szyld 0 Chapter
More informationFast solvers for steady incompressible flow
ICFD 25 p.1/21 Fast solvers for steady incompressible flow Andy Wathen Oxford University wathen@comlab.ox.ac.uk http://web.comlab.ox.ac.uk/~wathen/ Joint work with: Howard Elman (University of Maryland,
More informationChapter 7 Iterative Techniques in Matrix Algebra
Chapter 7 Iterative Techniques in Matrix Algebra Per-Olof Persson persson@berkeley.edu Department of Mathematics University of California, Berkeley Math 128B Numerical Analysis Vector Norms Definition
More informationITERATIVE METHODS FOR SPARSE LINEAR SYSTEMS
ITERATIVE METHODS FOR SPARSE LINEAR SYSTEMS YOUSEF SAAD University of Minnesota PWS PUBLISHING COMPANY I(T)P An International Thomson Publishing Company BOSTON ALBANY BONN CINCINNATI DETROIT LONDON MADRID
More informationIndefinite and physics-based preconditioning
Indefinite and physics-based preconditioning Jed Brown VAW, ETH Zürich 2009-01-29 Newton iteration Standard form of a nonlinear system F (u) 0 Iteration Solve: Update: J(ũ)u F (ũ) ũ + ũ + u Example (p-bratu)
More information2. Linear algebra. matrices and vectors. linear equations. range and nullspace of matrices. function of vectors, gradient and Hessian
FE661 - Statistical Methods for Financial Engineering 2. Linear algebra Jitkomut Songsiri matrices and vectors linear equations range and nullspace of matrices function of vectors, gradient and Hessian
More informationCONVERGENCE BOUNDS FOR PRECONDITIONED GMRES USING ELEMENT-BY-ELEMENT ESTIMATES OF THE FIELD OF VALUES
European Conference on Computational Fluid Dynamics ECCOMAS CFD 2006 P. Wesseling, E. Oñate and J. Périaux (Eds) c TU Delft, The Netherlands, 2006 CONVERGENCE BOUNDS FOR PRECONDITIONED GMRES USING ELEMENT-BY-ELEMENT
More informationDELFT UNIVERSITY OF TECHNOLOGY
DELFT UNIVERSITY OF TECHNOLOGY REPORT 16-02 The Induced Dimension Reduction method applied to convection-diffusion-reaction problems R. Astudillo and M. B. van Gijzen ISSN 1389-6520 Reports of the Delft
More informationKey words. conjugate gradients, normwise backward error, incremental norm estimation.
Proceedings of ALGORITMY 2016 pp. 323 332 ON ERROR ESTIMATION IN THE CONJUGATE GRADIENT METHOD: NORMWISE BACKWARD ERROR PETR TICHÝ Abstract. Using an idea of Duff and Vömel [BIT, 42 (2002), pp. 300 322
More informationConjugate gradient method. Descent method. Conjugate search direction. Conjugate Gradient Algorithm (294)
Conjugate gradient method Descent method Hestenes, Stiefel 1952 For A N N SPD In exact arithmetic, solves in N steps In real arithmetic No guaranteed stopping Often converges in many fewer than N steps
More informationLecture 10: Finite Differences for ODEs & Nonlinear Equations
Lecture 10: Finite Differences for ODEs & Nonlinear Equations J.K. Ryan@tudelft.nl WI3097TU Delft Institute of Applied Mathematics Delft University of Technology 21 November 2012 () Finite Differences
More informationIntroduction. Chapter One
Chapter One Introduction The aim of this book is to describe and explain the beautiful mathematical relationships between matrices, moments, orthogonal polynomials, quadrature rules and the Lanczos and
More informationA High-Performance Parallel Hybrid Method for Large Sparse Linear Systems
Outline A High-Performance Parallel Hybrid Method for Large Sparse Linear Systems Azzam Haidar CERFACS, Toulouse joint work with Luc Giraud (N7-IRIT, France) and Layne Watson (Virginia Polytechnic Institute,
More informationReview Questions REVIEW QUESTIONS 71
REVIEW QUESTIONS 71 MATLAB, is [42]. For a comprehensive treatment of error analysis and perturbation theory for linear systems and many other problems in linear algebra, see [126, 241]. An overview of
More informationA fast and well-conditioned spectral method: The US method
A fast and well-conditioned spectral method: The US method Alex Townsend University of Oxford Leslie Fox Prize, 24th of June 23 Partially based on: S. Olver & T., A fast and well-conditioned spectral method,
More informationA COMPUTATIONAL TECHNIQUE FOR SOLVING BOUNDARY VALUE PROBLEM WITH TWO SMALL PARAMETERS
Electronic Journal of Differential Equations, Vol. 2013 (2013), No. 30, pp. 1 10. ISSN: 1072-6691. URL: http://ejde.math.txstate.edu or http://ejde.math.unt.edu ftp ejde.math.txstate.edu A COMPUTATIONAL
More informationEfficient Estimation of the A-norm of the Error in the Preconditioned Conjugate Gradient Method
Efficient Estimation of the A-norm of the Error in the Preconditioned Conjugate Gradient Method Zdeněk Strakoš and Petr Tichý Institute of Computer Science AS CR, Technical University of Berlin. Emory
More informationIterative methods for positive definite linear systems with a complex shift
Iterative methods for positive definite linear systems with a complex shift William McLean, University of New South Wales Vidar Thomée, Chalmers University November 4, 2011 Outline 1. Numerical solution
More informationIterative Methods for Linear Systems
Iterative Methods for Linear Systems 1. Introduction: Direct solvers versus iterative solvers In many applications we have to solve a linear system Ax = b with A R n n and b R n given. If n is large the
More informationWe first repeat some well known facts about condition numbers for normwise and componentwise perturbations. Consider the matrix
BIT 39(1), pp. 143 151, 1999 ILL-CONDITIONEDNESS NEEDS NOT BE COMPONENTWISE NEAR TO ILL-POSEDNESS FOR LEAST SQUARES PROBLEMS SIEGFRIED M. RUMP Abstract. The condition number of a problem measures the sensitivity
More informationTopics. The CG Algorithm Algorithmic Options CG s Two Main Convergence Theorems
Topics The CG Algorithm Algorithmic Options CG s Two Main Convergence Theorems What about non-spd systems? Methods requiring small history Methods requiring large history Summary of solvers 1 / 52 Conjugate
More informationParallel Numerics, WT 2016/ Iterative Methods for Sparse Linear Systems of Equations. page 1 of 1
Parallel Numerics, WT 2016/2017 5 Iterative Methods for Sparse Linear Systems of Equations page 1 of 1 Contents 1 Introduction 1.1 Computer Science Aspects 1.2 Numerical Problems 1.3 Graphs 1.4 Loop Manipulations
More informationMIDTERM. b) [2 points] Compute the LU Decomposition A = LU or explain why one does not exist.
MAE 9A / FALL 3 Maurício de Oliveira MIDTERM Instructions: You have 75 minutes This exam is open notes, books No computers, calculators, phones, etc There are 3 questions for a total of 45 points and bonus
More informationAdaptive algebraic multigrid methods in lattice computations
Adaptive algebraic multigrid methods in lattice computations Karsten Kahl Bergische Universität Wuppertal January 8, 2009 Acknowledgements Matthias Bolten, University of Wuppertal Achi Brandt, Weizmann
More informationIterative Solution methods
p. 1/28 TDB NLA Parallel Algorithms for Scientific Computing Iterative Solution methods p. 2/28 TDB NLA Parallel Algorithms for Scientific Computing Basic Iterative Solution methods The ideas to use iterative
More information9. Iterative Methods for Large Linear Systems
EE507 - Computational Techniques for EE Jitkomut Songsiri 9. Iterative Methods for Large Linear Systems introduction splitting method Jacobi method Gauss-Seidel method successive overrelaxation (SOR) 9-1
More informationNewton additive and multiplicative Schwarz iterative methods
IMA Journal of Numerical Analysis (2008) 28, 143 161 doi:10.1093/imanum/drm015 Advance Access publication on July 16, 2007 Newton additive and multiplicative Schwarz iterative methods JOSEP ARNAL, VIOLETA
More informationScientific Computing: Dense Linear Systems
Scientific Computing: Dense Linear Systems Aleksandar Donev Courant Institute, NYU 1 donev@courant.nyu.edu 1 Course MATH-GA.2043 or CSCI-GA.2112, Spring 2012 February 9th, 2012 A. Donev (Courant Institute)
More informationOn deflation and singular symmetric positive semi-definite matrices
Journal of Computational and Applied Mathematics 206 (2007) 603 614 www.elsevier.com/locate/cam On deflation and singular symmetric positive semi-definite matrices J.M. Tang, C. Vuik Faculty of Electrical
More informationSUPERCONVERGENCE OF DG METHOD FOR ONE-DIMENSIONAL SINGULARLY PERTURBED PROBLEMS *1)
Journal of Computational Mathematics, Vol.5, No., 007, 185 00. SUPERCONVERGENCE OF DG METHOD FOR ONE-DIMENSIONAL SINGULARLY PERTURBED PROBLEMS *1) Ziqing Xie (College of Mathematics and Computer Science,
More informationFinite Element Superconvergence Approximation for One-Dimensional Singularly Perturbed Problems
Finite Element Superconvergence Approximation for One-Dimensional Singularly Perturbed Problems Zhimin Zhang Department of Mathematics Wayne State University Detroit, MI 48202 Received 19 July 2000; accepted
More informationLanczos tridigonalization and Golub - Kahan bidiagonalization: Ideas, connections and impact
Lanczos tridigonalization and Golub - Kahan bidiagonalization: Ideas, connections and impact Zdeněk Strakoš Academy of Sciences and Charles University, Prague http://www.cs.cas.cz/ strakos Hong Kong, February
More informationNewton s Method and Efficient, Robust Variants
Newton s Method and Efficient, Robust Variants Philipp Birken University of Kassel (SFB/TRR 30) Soon: University of Lund October 7th 2013 Efficient solution of large systems of non-linear PDEs in science
More informationLinear Algebra Massoud Malek
CSUEB Linear Algebra Massoud Malek Inner Product and Normed Space In all that follows, the n n identity matrix is denoted by I n, the n n zero matrix by Z n, and the zero vector by θ n An inner product
More informationNUMERICAL SOLUTION OF GENERAL SINGULAR PERTURBATION BOUNDARY VALUE PROBLEMS BASED ON ADAPTIVE CUBIC SPLINE
TWMS Jour. Pure Appl. Math., V.3, N.1, 2012, pp.11-21 NUMERICAL SOLUTION OF GENERAL SINGULAR PERTURBATION BOUNDARY VALUE PROBLEMS BASED ON ADAPTIVE CUBIC SPLINE R. MOHAMMADI 1 Abstract. We use adaptive
More informationAMS526: Numerical Analysis I (Numerical Linear Algebra for Computational and Data Sciences)
AMS526: Numerical Analysis I (Numerical Linear Algebra for Computational and Data Sciences) Lecture 1: Course Overview; Matrix Multiplication Xiangmin Jiao Stony Brook University Xiangmin Jiao Numerical
More informationLeast Squares. Tom Lyche. October 26, Centre of Mathematics for Applications, Department of Informatics, University of Oslo
Least Squares Tom Lyche Centre of Mathematics for Applications, Department of Informatics, University of Oslo October 26, 2010 Linear system Linear system Ax = b, A C m,n, b C m, x C n. under-determined
More informationConvergence Behavior of a Two-Level Optimized Schwarz Preconditioner
Convergence Behavior of a Two-Level Optimized Schwarz Preconditioner Olivier Dubois 1 and Martin J. Gander 2 1 IMA, University of Minnesota, 207 Church St. SE, Minneapolis, MN 55455 dubois@ima.umn.edu
More information7.6 The Inverse of a Square Matrix
7.6 The Inverse of a Square Matrix Copyright Cengage Learning. All rights reserved. What You Should Learn Verify that two matrices are inverses of each other. Use Gauss-Jordan elimination to find inverses
More informationOn prescribing Ritz values and GMRES residual norms generated by Arnoldi processes
On prescribing Ritz values and GMRES residual norms generated by Arnoldi processes Jurjen Duintjer Tebbens Institute of Computer Science Academy of Sciences of the Czech Republic joint work with Gérard
More informationIndex. book 2009/5/27 page 121. (Page numbers set in bold type indicate the definition of an entry.)
page 121 Index (Page numbers set in bold type indicate the definition of an entry.) A absolute error...26 componentwise...31 in subtraction...27 normwise...31 angle in least squares problem...98,99 approximation
More informationSome preconditioners for systems of linear inequalities
Some preconditioners for systems of linear inequalities Javier Peña Vera oshchina Negar Soheili June 0, 03 Abstract We show that a combination of two simple preprocessing steps would generally improve
More informationNumerical Methods I Solving Square Linear Systems: GEM and LU factorization
Numerical Methods I Solving Square Linear Systems: GEM and LU factorization Aleksandar Donev Courant Institute, NYU 1 donev@courant.nyu.edu 1 MATH-GA 2011.003 / CSCI-GA 2945.003, Fall 2014 September 18th,
More informationLecture 9: Numerical Linear Algebra Primer (February 11st)
10-725/36-725: Convex Optimization Spring 2015 Lecture 9: Numerical Linear Algebra Primer (February 11st) Lecturer: Ryan Tibshirani Scribes: Avinash Siravuru, Guofan Wu, Maosheng Liu Note: LaTeX template
More informationNormed & Inner Product Vector Spaces
Normed & Inner Product Vector Spaces ECE 174 Introduction to Linear & Nonlinear Optimization Ken Kreutz-Delgado ECE Department, UC San Diego Ken Kreutz-Delgado (UC San Diego) ECE 174 Fall 2016 1 / 27 Normed
More informationShifted Laplace and related preconditioning for the Helmholtz equation
Shifted Laplace and related preconditioning for the Helmholtz equation Ivan Graham and Euan Spence (Bath, UK) Collaborations with: Paul Childs (Schlumberger Gould Research), Martin Gander (Geneva) Douglas
More informationMatrix Algebra. Matrix Algebra. Chapter 8 - S&B
Chapter 8 - S&B Algebraic operations Matrix: The size of a matrix is indicated by the number of its rows and the number of its columns. A matrix with k rows and n columns is called a k n matrix. The number
More informationSuperconvergence analysis of the SDFEM for elliptic problems with characteristic layers
Superconvergence analysis of the SDFEM for elliptic problems with characteristic layers S. Franz, T. Linß, H.-G. Roos Institut für Numerische Mathematik, Technische Universität Dresden, D-01062, Germany
More informationA NEW EFFECTIVE PRECONDITIONED METHOD FOR L-MATRICES
Journal of Mathematical Sciences: Advances and Applications Volume, Number 2, 2008, Pages 3-322 A NEW EFFECTIVE PRECONDITIONED METHOD FOR L-MATRICES Department of Mathematics Taiyuan Normal University
More informationLecture Note 7: Iterative methods for solving linear systems. Xiaoqun Zhang Shanghai Jiao Tong University
Lecture Note 7: Iterative methods for solving linear systems Xiaoqun Zhang Shanghai Jiao Tong University Last updated: December 24, 2014 1.1 Review on linear algebra Norms of vectors and matrices vector
More informationEfficient Solvers for the Navier Stokes Equations in Rotation Form
Efficient Solvers for the Navier Stokes Equations in Rotation Form Computer Research Institute Seminar Purdue University March 4, 2005 Michele Benzi Emory University Atlanta, GA Thanks to: NSF (MPS/Computational
More informationAMS526: Numerical Analysis I (Numerical Linear Algebra for Computational and Data Sciences)
AMS526: Numerical Analysis I (Numerical Linear Algebra for Computational and Data Sciences) Lecture 19: Computing the SVD; Sparse Linear Systems Xiangmin Jiao Stony Brook University Xiangmin Jiao Numerical
More informationThe Deflation Accelerated Schwarz Method for CFD
The Deflation Accelerated Schwarz Method for CFD J. Verkaik 1, C. Vuik 2,, B.D. Paarhuis 1, and A. Twerda 1 1 TNO Science and Industry, Stieltjesweg 1, P.O. Box 155, 2600 AD Delft, The Netherlands 2 Delft
More informationNUMERICAL SOLUTION OF CONVECTION DIFFUSION EQUATIONS USING UPWINDING TECHNIQUES SATISFYING THE DISCRETE MAXIMUM PRINCIPLE
Proceedings of the Czech Japanese Seminar in Applied Mathematics 2005 Kuju Training Center, Oita, Japan, September 15-18, 2005 pp. 69 76 NUMERICAL SOLUTION OF CONVECTION DIFFUSION EQUATIONS USING UPWINDING
More informationAMG for a Peta-scale Navier Stokes Code
AMG for a Peta-scale Navier Stokes Code James Lottes Argonne National Laboratory October 18, 2007 The Challenge Develop an AMG iterative method to solve Poisson 2 u = f discretized on highly irregular
More informationCharles University Faculty of Mathematics and Physics DOCTORAL THESIS. Krylov subspace approximations in linear algebraic problems
Charles University Faculty of Mathematics and Physics DOCTORAL THESIS Iveta Hnětynková Krylov subspace approximations in linear algebraic problems Department of Numerical Mathematics Supervisor: Doc. RNDr.
More informationLecture 15 Newton Method and Self-Concordance. October 23, 2008
Newton Method and Self-Concordance October 23, 2008 Outline Lecture 15 Self-concordance Notion Self-concordant Functions Operations Preserving Self-concordance Properties of Self-concordant Functions Implications
More information