On solving linear systems arising from Shishkin mesh discretizations


 Ashlie Jenkins
 11 months ago
 Views:
Transcription
1 On solving linear systems arising from Shishkin mesh discretizations Petr Tichý Faculty of Mathematics and Physics, Charles University joint work with Carlos Echeverría, Jörg Liesen, and Daniel Szyld October 20, 2016, Prague Seminar of Numerical Mathematics, KNM MFF UK 1
2 Problem formulation Convectiondiffusion boundary value problem ɛ u + α u + β u = f in Ω = (0, 1), u(0) = u 0, u(1) = u 1, α > 0, β 0 constants, the problem is convection dominated 0 < ɛ α. [Stynes, 2005] (Acta Numerica) [Roos, Stynes, and Tobiska, 1996, 2008] (book) [Miller, O Riordan, and Shishkin, 1996] (book) 2
3 Solution and boundary layers ɛ = 0.01, α = 1, β = 0, u(0) = u(1) = 0. There are small subregions where the solution has a large gradient. 3
4 Numerical solution, equidistant mesh Standard techniques: u (ih) u i+1 u i 1 2h, u (ih) u i u i 1. h Unnatural oscillations or cannot resolve the layers. Remedy: stabilization or nonequidistant mesh. We study discretizations for a Shishkin mesh. 4
5 Outline 1 Shishkin mesh and discretization 2 How to solve the linear system? 3 Multiplicative Schwarz method 4 Convergence analysis 5 Schwarz method as a preconditioner 6 Numerical examples 5
6 Shishkin mesh on [0, 1] Piecewise equidistant N even, define the transition point 1 τ and n by { 1 τ min 2, ɛ } α 2 ln N, n N 2. If ɛ α, then 1 τ is close to 1. Next define H and h by H 1 τ n, h τ n and consider the Shishkin mesh, x 0 = 0, x i ih, x n+i x n + ih, i = 1,..., n. 6
7 Discretization on the Shishkin mesh  details For simplicity u(0) = u(1) = 0 The upwind difference scheme is given by ɛ δ 2 xu i + α D x u i + βu i = f i, u 0 = u N = 0, and the central difference scheme by where ɛ δ 2 xu i + αd 0 x + βu i = f i, u 0 = u N = 0, δ 2 xu i = 2u i 1 (H + h)h 2u i Hh + 2u i+1 (H + h)h, i = n, and Dx u i = u i u i 1, 1 i n, D 0 H xu i = u i+1 u i 1 h + H, i = n. surveys [Linss, Stynes, 2001], [Stynes, 2005], [Kopteva, O Riordan, 2010] 7
8 Shishkin mesh and ɛuniform convergence The upwind difference scheme ɛ δ 2 xu i + α D x u i + βu i = f i, u 0 = u N = 0. There exists a constant C such that ( ) ln N u(x i ) u i C, i = 0,..., N, N see, e.g., [Stynes, 2005]. 8
9 Shishkin mesh and ɛuniform convergence The upwind difference scheme ɛ δ 2 xu i + α D x u i + βu i = f i, u 0 = u N = 0. There exists a constant C such that ( ) ln N u(x i ) u i C, i = 0,..., N, N see, e.g., [Stynes, 2005]. The central difference scheme ɛ δ 2 xu i + α D 0 xu i + βu i = f i, u 0 = u N = 0. There exists a constant C such that ( ) ln N 2 u(x i ) u i C, i = 0,..., N, N [Andreev and Kopteva, 1996], a difficult proof, the scheme does not satisfy a discrete maximum principle. [Kopteva and Linss, 2001]. 8
10 Outline 1 Shishkin mesh and discretization 2 How to solve the linear system? 3 Multiplicative Schwarz method 4 Convergence analysis 5 Schwarz method as a preconditioner 6 Numerical examples 9
11 Structure of the matrix A = a H b H c H bh c H a H b H c a b c h a h b h c h bh c h a h 10
12 Entries The upwind scheme c H = ɛ H 2 α H, 2ɛ c = H(H + h) α H, c h = ɛ h 2 α h, a H = 2ɛ H 2 + α H + β, b H = ɛ H 2, a = 2ɛ hh + α H + β, b = 2ɛ h(h + h), a h = 2ɛ h 2 + α h + β, b h = ɛ h 2. The central difference scheme c H = ɛ H 2 α 2H, 2ɛ c = H(H + h) α h + H, c h = ɛ h 2 α 2h, b H = ɛ H 2 + α 2H, a = 2ɛ hh + β, b = 2ɛ h(h + h) + a H = 2ɛ H 2 + β, a h = 2ɛ h 2 + β, b h = ɛ h 2 + α 2h. α h + H, 11
13 Matrix properties nonsymmetric A is Mmatrix for the upwind scheme A is not an Mmatrix for the central difference scheme A is highly nonnormal. Consider ɛu + u = 1, u(0) = 0, u(1) = 0, ɛ = 10 8 and N = 46, and the spectral decomposition A = Y DY 1. 12
14 Matrix properties nonsymmetric A is Mmatrix for the upwind scheme A is not an Mmatrix for the central difference scheme A is highly nonnormal. Consider ɛu + u = 1, u(0) = 0, u(1) = 0, ɛ = 10 8 and N = 46, and the spectral decomposition A = Y DY 1. upwind upwind sc. central central sc. κ(a) κ(y )
15 Solving linear system using GMRES 10 0 Convergence of GMRES for the upwind scheme relative residuial norm GMRES, A original GMRES, A scaled number of iterations 13
16 Outline 1 Shishkin mesh and discretization 2 How to solve the linear system? 3 Multiplicative Schwarz method 4 Convergence analysis 5 Schwarz method as a preconditioner 6 Numerical examples 14
17 Multiplicative Schwarz method Idea of solving Ax = b Given an approximation x (k), then x = x (k) + y and y satisfies Ay = b Ax (k). 15
18 Multiplicative Schwarz method Idea of solving Ax = b Given an approximation x (k), then x = x (k) + y and y satisfies Ay = b Ax (k). [ Restriction operators R 1 = I n 0 ] [, R 2 = 0 I n ]. 15
19 Multiplicative Schwarz method Idea of solving Ax = b Given an approximation x (k), then x = x (k) + y and y satisfies Ay = b Ax (k). [ Restriction operators R 1 = Solve on the first domain I n 0 ] [, R 2 = 0 I n ]. (R 1 AR T 1 ) ỹ = R 1 (b Ax (k) ) and approximate y by prolongation of ỹ, i.e., by R1 T ỹ. Define x (k+ 1 2 ) = x (k) + R T 1 (R 1 AR T 1 ) 1 R 1 (b Ax (k) ). 15
20 Multiplicative Schwarz method Idea of solving Ax = b Given an approximation x (k), then x = x (k) + y and y satisfies Ay = b Ax (k). [ Restriction operators R 1 = Solve on the first domain I n 0 ] [, R 2 = 0 I n ]. (R 1 AR T 1 ) ỹ = R 1 (b Ax (k) ) and approximate y by prolongation of ỹ, i.e., by R1 T ỹ. Define x (k+ 1 2 ) = x (k) + R T 1 (R 1 AR T 1 ) 1 R 1 (b Ax (k) ). Similarly, use x (k+ 1 2 ) on the second domain and prolong, x (k+1) = x (k+ 1 2 ) + R T 2 (R 2 AR T 2 ) 1 R 2 (b Ax (k+ 1 2 )). 15
21 Multiplicative Schwarz method Formalism Define P i = R T i A 1 i R i A, A i R i AR T i, i = 1, 2. The multiplicative Schwarz is the iterative scheme x (k+1) = T x (k) + v, T = (I P 2 )(I P 1 ), where v is defined such that the scheme is consistent. 16
22 Multiplicative Schwarz method Formalism Define P i = R T i A 1 i R i A, A i R i AR T i, i = 1, 2. The multiplicative Schwarz is the iterative scheme x (k+1) = T x (k) + v, T = (I P 2 )(I P 1 ), where v is defined such that the scheme is consistent. Hence, and Is it convergent in our case? x x (k+1) = T k+1 (x x 0 ), x x (k+1) T k+1 x x 0. 16
23 Multiplicative Schwarz method Experiment 10 0 upwind difference scheme 10 2 Relative Residual GMRES Multiplicative Schwarz Iterations 17
24 Outline 1 Shishkin mesh and discretization 2 How to solve the linear system? 3 Multiplicative Schwarz method 4 Convergence analysis 5 Schwarz method as a preconditioner 6 Numerical examples 18
25 Convergence analysis Exploiting the structure x x (k+1) x x 0 T k+1. Using T = (I P 2 )(I P 1 ) we are able to show that t 1. T = t n = t e T n+1.. t N 1 19
26 Convergence analysis Exploiting the structure x x (k+1) x x 0 T k+1. Using T = (I P 2 )(I P 1 ) we are able to show that t 1. T = t n = t e T n+1.. t N 1 Therefore, T 2 = t (e T n+1 t) et n+1 = t n+1 T, and T k+1 = t n+1 k T. How to bound t n+1, and T in a convenient norm ( )? 19
27 Convergence analysis Details A = A H b H c a b c h A h. Let m n 1, ρ t n+1... the convergence factor. Then, ( ) bb H A 1 H ρ = ( ) a cb H A 1 H m,m m,m ( ) cc h A 1 h a bc h ( 1,1 ) A 1 h 1,1. 20
28 Convergence analysis Bounding ( A 1 H )m,m and ( ) A 1 h 1,1 A matrix B = [b i,j ] is called a nonsingular Mmatrix when B is nonsingular, b i,i > 0 for all i, b i,j 0 for all i j, and B 1 0 (elementwise). 21
29 Convergence analysis Bounding ( A 1 H )m,m and ( ) A 1 h 1,1 A matrix B = [b i,j ] is called a nonsingular Mmatrix when B is nonsingular, b i,i > 0 for all i, b i,j 0 for all i j, and B 1 0 (elementwise). If A H and A h are nonsingular Mmatrices, then using [Nabben 1999], ( { } 1 A 1 H )m,m min b H, 1, c H ( A 1 h )1,1 min { 1 } b h, 1 c h. A sufficient condition: The sign conditions & irreducibly diagonal dominant nonsingular Mmatrix. [Meurant, 1996], [Hackbusch, 2010] 21
30 Convergence analysis The upwind scheme The matrices A H and A h are Mmatrices, and we know that e (k+1) e (0) ρ k T. 22
31 Convergence analysis The upwind scheme The matrices A H and A h are Mmatrices, and we know that e (k+1) e (0) ρ k T. Theorem (the upwind scheme) [Echeverría, Liesen, T., Szyld, 2016] For the upwind scheme we have ρ ɛ ɛ + αh ɛ ɛ + α, N and ɛ T ɛ + αh. 22
32 Convergence analysis The central difference scheme A h is still an Mmatrix. If αh > 2ɛ, i.e. b H > 0, then A H is not an Mmatrix... the most common situation from a practical point of view. 23
33 Convergence analysis The central difference scheme A h is still an Mmatrix. If αh > 2ɛ, i.e. b H > 0, then A H is not an Mmatrix... the most common situation from a practical point of view. Recall ( ) bb H A 1 H ρ = ( ) a cb H A 1 H m,m m,m ( ) cc h A 1 h a bc h ( 1,1 ) A 1 h 1,1 How to bound ( A 1 ) H m,m?... results by [Usmani 1994]. 23
34 Convergence analysis The central difference scheme A h is still an Mmatrix. If αh > 2ɛ, i.e. b H > 0, then A H is not an Mmatrix... the most common situation from a practical point of view. Recall ( ) bb H A 1 H ρ = ( ) a cb H A 1 H m,m m,m ( ) cc h A 1 h a bc h ( 1,1 ) A 1 h 1,1 How to bound ( A 1 ) H m,m?... results by [Usmani 1994] We proved: If m = N/2 1 is even, then b H (A 1 H ) m,m 1 b H m ch c H bh + b H m < 2mɛ ɛ + αh. ch 2. 23
35 Convergence analysis The central difference scheme A h is Mmatrix, if αh > 2ɛ, A H is not an Mmatrix. Theorem (the central diff. scheme) [Echeverría, Liesen, T., Szyld, 2016] Let m = N/2 1 be even, and let αh > 2ɛ. For the central differences we have ρ < 2mɛ ɛ + αh 2 < N ɛ ɛ + α, N T < 2. Thus, the error of the multiplicative Schwarz method satisfies e (k+1) ( 2mɛ e (0) < 2 ɛ + αh 2 ) k. 24
36 Remarks on diagonally scaled linear algebraic systems DAx = Db The illconditioning can be avoided by diagonal scaling [Roos 1996]: Ax = b is multiplied from the left by d H I m D = d. d h I m 25
37 Remarks on diagonally scaled linear algebraic systems DAx = Db The illconditioning can be avoided by diagonal scaling [Roos 1996]: Ax = b is multiplied from the left by d H I m D = d. d h I m Such a scaling preserves the Toeplitz structure and the Mmatrix property of the submatrices. Analysis depends on these properties and on ratios between elements in the same row such as b/a and b H /c H. These ratios are invariant under diagonal scaling. 25
38 Remarks on diagonally scaled linear algebraic systems DAx = Db The illconditioning can be avoided by diagonal scaling [Roos 1996]: Ax = b is multiplied from the left by d H I m D = d. d h I m Such a scaling preserves the Toeplitz structure and the Mmatrix property of the submatrices. Analysis depends on these properties and on ratios between elements in the same row such as b/a and b H /c H. These ratios are invariant under diagonal scaling. Consequently, all convergence bounds hold for the multiplicative Schwarz method applied to DAx = Db. 25
39 Outline 1 Shishkin mesh and discretization 2 How to solve the linear system? 3 Multiplicative Schwarz method 4 Convergence analysis 5 Schwarz method as a preconditioner 6 Numerical examples 26
40 Schwarz method as a preconditioner We have consistent scheme x (k+1) = T x (k) + v. Hence, x solves Ax = b and also the preconditioned system (I T )x = v. 27
41 Schwarz method as a preconditioner We have consistent scheme x (k+1) = T x (k) + v. Hence, x solves Ax = b and also the preconditioned system (I T )x = v. We can formally define a preconditioner M such Ax = b M 1 Ax = M 1 b (I T )x = v. Clearly M = A(I T ) 1. 27
42 Schwarz method as a preconditioner We have consistent scheme x (k+1) = T x (k) + v. Hence, x solves Ax = b and also the preconditioned system (I T )x = v. We can formally define a preconditioner M such Ax = b M 1 Ax = M 1 b (I T )x = v. Clearly M = A(I T ) 1. Then x (k+1) = x (k) + (I T )(x x (k) ) = x (k) + M 1 r (k). 27
43 Schwarz method as a preconditioner for GMRES The multiplicative Schwarz method as well as GMRES applied to the preconditioned system obtain their approximations from the same Krylov subspace. In terms of the residual norm, the preconditioned GMRES will always converge faster than the multiplicative Schwarz. 28
44 Schwarz method as a preconditioner for GMRES The multiplicative Schwarz method as well as GMRES applied to the preconditioned system obtain their approximations from the same Krylov subspace. In terms of the residual norm, the preconditioned GMRES will always converge faster than the multiplicative Schwarz. Moreover, in this case, the iteration matrix T has rankone structure, and dim (K k (I T, r 0 )) 2. Therefore, GMRES converges in at most 2 steps,... a motivation for more dimensional cases. 28
45 How to multiply by T Schwarz or preconditioned GMRES only multiply by T, T = (I P 2 )(I P 1 ), P i = R T i (R i AR T i ) 1 R i A, i.e., to solve systems of the form (m = n 1 = N 2 1) A H c b H a [ y1:m y m+1 ] = [ z1:m z m+1 ]. 29
46 How to multiply by T Schwarz or preconditioned GMRES only multiply by T, T = (I P 2 )(I P 1 ), P i = R T i (R i AR T i ) 1 R i A, i.e., to solve systems of the form (m = n 1 = N 2 1) A H c b H a Using the Schur complement, [ y1:m y m+1 ] = [ z1:m z m+1 ( A H b ) Hc a e me T b H m y 1:m = z 1:m z m+1 e m. a H Then apply ShermanMorrison formula. We need only to solve systems with A H (Toeplitz)! ]. 29
47 Outline 1 Shishkin mesh and discretization 2 How to solve the linear system? 3 Multiplicative Schwarz method 4 Convergence analysis 5 Schwarz method as a preconditioner 6 Numerical examples 30
48 Numerical examples Consider i.e. ɛu + u = 1, u(0) = 0, u(1) = 0, α = 1, β = 0, f(x) 1. Choose N = 198, various values of ɛ. upwind central differences ɛ ρ up our bound ρ cd our bound ρ up < ɛ ɛ + αh, ρ cd < 2mɛ ɛ + αh 2. 31
49 Numerical examples Upwind, ɛ = Upwind T=(I P 2 )(I P 1 ) error bound T=(I P 1 )(I P 2 ) error bound Error norms and bounds k 32
50 Numerical examples Upwind, ɛ = Upwind T=(I P 2 )(I P 1 ) error bound T=(I P 1 )(I P 2 ) error bound Error norms and bounds k 33
51 Numerical examples Central differences, ɛ = Central differences T=(I P 2 )(I P 1 ) T=(I P 1 )(I P 2 ) error bound Error norms and bounds k 34
52 Numerical examples Central differences, ɛ = Central differences T=(I P 2 )(I P 1 ) T=(I P 1 )(I P 2 ) error bound Error norms and bounds k 35
53 Conclusions and further work We considered finite difference discretizations of the 1D singularlyperturbed convectiondiffusion equation posed on a Shishkin mesh. 36
54 Conclusions and further work We considered finite difference discretizations of the 1D singularlyperturbed convectiondiffusion equation posed on a Shishkin mesh. The matrices that arise from the upwind and the central difference schemes are nonsymmetric and highly nonnormal. 36
55 Conclusions and further work We considered finite difference discretizations of the 1D singularlyperturbed convectiondiffusion equation posed on a Shishkin mesh. The matrices that arise from the upwind and the central difference schemes are nonsymmetric and highly nonnormal. For the upwind scheme, we proved rapid convergence of the multiplicative Schwarz method in the most relevant case Nɛ < α. 36
56 Conclusions and further work We considered finite difference discretizations of the 1D singularlyperturbed convectiondiffusion equation posed on a Shishkin mesh. The matrices that arise from the upwind and the central difference schemes are nonsymmetric and highly nonnormal. For the upwind scheme, we proved rapid convergence of the multiplicative Schwarz method in the most relevant case Nɛ < α. The convergence for the central difference scheme is slower, but still rapid, when N 2 ɛ < α and if N/2 1 is even. 36
57 Conclusions and further work We considered finite difference discretizations of the 1D singularlyperturbed convectiondiffusion equation posed on a Shishkin mesh. The matrices that arise from the upwind and the central difference schemes are nonsymmetric and highly nonnormal. For the upwind scheme, we proved rapid convergence of the multiplicative Schwarz method in the most relevant case Nɛ < α. The convergence for the central difference scheme is slower, but still rapid, when N 2 ɛ < α and if N/2 1 is even. Thanks to the rankone structure of T, the preconditioned GMRES converges in two steps. 36
58 Conclusions and further work We considered finite difference discretizations of the 1D singularlyperturbed convectiondiffusion equation posed on a Shishkin mesh. The matrices that arise from the upwind and the central difference schemes are nonsymmetric and highly nonnormal. For the upwind scheme, we proved rapid convergence of the multiplicative Schwarz method in the most relevant case Nɛ < α. The convergence for the central difference scheme is slower, but still rapid, when N 2 ɛ < α and if N/2 1 is even. Thanks to the rankone structure of T, the preconditioned GMRES converges in two steps. Inspired by 1D case (preconditioner, lowrank structure), we can study 2D case. 36
59 Related papers C. Echeverría, J. Liesen, P. Tichý, and D. Szyld, [Convergence of the multiplicative Schwarz method for singularly perturbed convectiondiffusion problems discretized on a Shishkin mesh, (2016), in preparation] J. Miller, E. O Riordan, and G. Shishkin, [Fitted Numerical Methods for Singular Perturbation Problems: Error Estimates in the Maximum Norm for Linear Problems in One and Two Dimensions, World Scientific, 1996.] HG. Roos, M. Stynes, L. Tobiska, [Robust Numerical Methods for Singularly Perturbed Differential Equations, second edition, Springer Series in Computational Mathematics, SpringerVerlag, Berlin, 2008, 604 pp.] M. Stynes, [Steadystate convectiondiffusion problems, Acta Numerica, 14 (2005), pp ] Thank you for your attention! 37
Optimal Left and Right Additive Schwarz Preconditioning for Minimal Residual Methods with Euclidean and Energy Norms
Optimal Left and Right Additive Schwarz Preconditioning for Minimal Residual Methods with Euclidean and Energy Norms Marcus Sarkis Worcester Polytechnic Inst., Mass. and IMPA, Rio de Janeiro and Daniel
More informationCholesky factorisations of linear systems coming from a finite difference method applied to singularly perturbed problems
Cholesky factorisations of linear systems coming from a finite difference method applied to singularly perturbed problems Thái Anh Nhan and Niall Madden The Boundary and Interior Layers  Computational
More informationNumerical behavior of inexact linear solvers
Numerical behavior of inexact linear solvers Miro Rozložník joint results with Zhongzhi Bai and Pavel Jiránek Institute of Computer Science, Czech Academy of Sciences, Prague, Czech Republic The fourth
More informationOUTLINE ffl CFD: elliptic pde's! Ax = b ffl Basic iterative methods ffl Krylov subspace methods ffl Preconditioning techniques: Iterative methods ILU
Preconditioning Techniques for Solving Large Sparse Linear Systems Arnold Reusken Institut für Geometrie und Praktische Mathematik RWTHAachen OUTLINE ffl CFD: elliptic pde's! Ax = b ffl Basic iterative
More informationAMS526: Numerical Analysis I (Numerical Linear Algebra) Lecture 23: GMRES and Other Krylov Subspace Methods; Preconditioning
AMS526: Numerical Analysis I (Numerical Linear Algebra) Lecture 23: GMRES and Other Krylov Subspace Methods; Preconditioning Xiangmin Jiao SUNY Stony Brook Xiangmin Jiao Numerical Analysis I 1 / 18 Outline
More informationThe Lanczos and conjugate gradient algorithms
The Lanczos and conjugate gradient algorithms Gérard MEURANT October, 2008 1 The Lanczos algorithm 2 The Lanczos algorithm in finite precision 3 The nonsymmetric Lanczos algorithm 4 The Golub Kahan bidiagonalization
More informationNUMERICAL STUDY OF CONVECTION DIFFUSION PROBLEM IN TWO DIMENSIONAL SPACE
IJRRAS 5 () November NUMERICAL STUDY OF CONVECTION DIFFUSION PROBLEM IN TWO DIMENSIONAL SPACE. K. Sharath Babu,* & N. Srinivasacharyulu Research Scholar, Department of Mathematics, National Institute
More informationSPLINE COLLOCATION METHOD FOR SINGULAR PERTURBATION PROBLEM. Mirjana Stojanović University of Novi Sad, Yugoslavia
GLASNIK MATEMATIČKI Vol. 37(57(2002, 393 403 SPLINE COLLOCATION METHOD FOR SINGULAR PERTURBATION PROBLEM Mirjana Stojanović University of Novi Sad, Yugoslavia Abstract. We introduce piecewise interpolating
More informationDepartment of Computer Science, University of Illinois at UrbanaChampaign
Department of Computer Science, University of Illinois at UrbanaChampaign Probing for Schur Complements and Preconditioning Generalized SaddlePoint Problems Eric de Sturler, sturler@cs.uiuc.edu, http://wwwfaculty.cs.uiuc.edu/~sturler
More informationInexactness and flexibility in linear Krylov solvers
Inexactness and flexibility in linear Krylov solvers Luc Giraud ENSEEIHT (N7)  IRIT, Toulouse Matrix Analysis and Applications CIRM Luminy  October 1519, 2007 in honor of Gérard Meurant for his 60 th
More informationOn the Superlinear Convergence of MINRES. Valeria Simoncini and Daniel B. Szyld. Report January 2012
On the Superlinear Convergence of MINRES Valeria Simoncini and Daniel B. Szyld Report 120111 January 2012 This report is available in the World Wide Web at http://www.math.temple.edu/~szyld 0 Chapter
More informationSome recent results on positivitypreserving discretisations for the convectiondiffusion equation
Some recent results on positivitypreserving discretisations for the convectiondiffusion equation Gabriel R. Barrenechea 1, Volker John 2 & Petr Knobloch 3 1 Department of Mathematics and Statistics,
More informationChapter 7 Iterative Techniques in Matrix Algebra
Chapter 7 Iterative Techniques in Matrix Algebra PerOlof Persson persson@berkeley.edu Department of Mathematics University of California, Berkeley Math 128B Numerical Analysis Vector Norms Definition
More informationConjugate gradient method. Descent method. Conjugate search direction. Conjugate Gradient Algorithm (294)
Conjugate gradient method Descent method Hestenes, Stiefel 1952 For A N N SPD In exact arithmetic, solves in N steps In real arithmetic No guaranteed stopping Often converges in many fewer than N steps
More informationIterative Methods for Linear Systems
Iterative Methods for Linear Systems 1. Introduction: Direct solvers versus iterative solvers In many applications we have to solve a linear system Ax = b with A R n n and b R n given. If n is large the
More informationWe first repeat some well known facts about condition numbers for normwise and componentwise perturbations. Consider the matrix
BIT 39(1), pp. 143 151, 1999 ILLCONDITIONEDNESS NEEDS NOT BE COMPONENTWISE NEAR TO ILLPOSEDNESS FOR LEAST SQUARES PROBLEMS SIEGFRIED M. RUMP Abstract. The condition number of a problem measures the sensitivity
More informationThe antitriangular factorisation of saddle point matrices
The antitriangular factorisation of saddle point matrices J. Pestana and A. J. Wathen August 29, 2013 Abstract Mastronardi and Van Dooren [this journal, 34 (2013) pp. 173 196] recently introduced the block
More informationConvergence Behavior of a TwoLevel Optimized Schwarz Preconditioner
Convergence Behavior of a TwoLevel Optimized Schwarz Preconditioner Olivier Dubois 1 and Martin J. Gander 2 1 IMA, University of Minnesota, 207 Church St. SE, Minneapolis, MN 55455 dubois@ima.umn.edu
More informationLinear Algebra Massoud Malek
CSUEB Linear Algebra Massoud Malek Inner Product and Normed Space In all that follows, the n n identity matrix is denoted by I n, the n n zero matrix by Z n, and the zero vector by θ n An inner product
More informationCharles University Faculty of Mathematics and Physics DOCTORAL THESIS. Krylov subspace approximations in linear algebraic problems
Charles University Faculty of Mathematics and Physics DOCTORAL THESIS Iveta Hnětynková Krylov subspace approximations in linear algebraic problems Department of Numerical Mathematics Supervisor: Doc. RNDr.
More informationLeast Squares. Tom Lyche. October 26, Centre of Mathematics for Applications, Department of Informatics, University of Oslo
Least Squares Tom Lyche Centre of Mathematics for Applications, Department of Informatics, University of Oslo October 26, 2010 Linear system Linear system Ax = b, A C m,n, b C m, x C n. underdetermined
More informationNormed & Inner Product Vector Spaces
Normed & Inner Product Vector Spaces ECE 174 Introduction to Linear & Nonlinear Optimization Ken KreutzDelgado ECE Department, UC San Diego Ken KreutzDelgado (UC San Diego) ECE 174 Fall 2016 1 / 27 Normed
More informationAMG for a Petascale Navier Stokes Code
AMG for a Petascale Navier Stokes Code James Lottes Argonne National Laboratory October 18, 2007 The Challenge Develop an AMG iterative method to solve Poisson 2 u = f discretized on highly irregular
More informationLecture 15 Newton Method and SelfConcordance. October 23, 2008
Newton Method and SelfConcordance October 23, 2008 Outline Lecture 15 Selfconcordance Notion Selfconcordant Functions Operations Preserving Selfconcordance Properties of Selfconcordant Functions Implications
More informationProperties of Matrices and Operations on Matrices
Properties of Matrices and Operations on Matrices A common data structure for statistical analysis is a rectangular array or matris. Rows represent individual observational units, or just observations,
More informationBoundary Value Problems and Iterative Methods for Linear Systems
Boundary Value Problems and Iterative Methods for Linear Systems 1. Equilibrium Problems 1.1. Abstract setting We want to find a displacement u V. Here V is a complete vector space with a norm v V. In
More informationRestricted Additive Schwarz Methods for Markov Chains
NUMERICAL LINEAR ALGEBRA WITH APPLICATIONS Numer. Linear Algebra Appl. 2000; 00:1 23 [Version: 2002/09/18 v1.02] Restricted Additive Schwarz Methods for Markov Chains Michele Benzi Verena Kuhlemann Department
More informationLecture 9: Numerical Linear Algebra Primer (February 11st)
10725/36725: Convex Optimization Spring 2015 Lecture 9: Numerical Linear Algebra Primer (February 11st) Lecturer: Ryan Tibshirani Scribes: Avinash Siravuru, Guofan Wu, Maosheng Liu Note: LaTeX template
More informationNumerical Methods I Solving Square Linear Systems: GEM and LU factorization
Numerical Methods I Solving Square Linear Systems: GEM and LU factorization Aleksandar Donev Courant Institute, NYU 1 donev@courant.nyu.edu 1 MATHGA 2011.003 / CSCIGA 2945.003, Fall 2014 September 18th,
More informationLecture 5. Ch. 5, Norms for vectors and matrices. Norms for vectors and matrices Why?
KTH ROYAL INSTITUTE OF TECHNOLOGY Norms for vectors and matrices Why? Lecture 5 Ch. 5, Norms for vectors and matrices Emil Björnson/Magnus Jansson/Mats Bengtsson April 27, 2016 Problem: Measure size of
More informationCME342 Parallel Methods in Numerical Analysis. Matrix Computation: Iterative Methods II. Sparse Matrixvector Multiplication.
CME342 Parallel Methods in Numerical Analysis Matrix Computation: Iterative Methods II Outline: CG & its parallelization. Sparse Matrixvector Multiplication. 1 Basic iterative methods: Ax = b r = b Ax
More informationA Review of Preconditioning Techniques for Steady Incompressible Flow
Zeist 2009 p. 1/43 A Review of Preconditioning Techniques for Steady Incompressible Flow David Silvester School of Mathematics University of Manchester Zeist 2009 p. 2/43 PDEs Review : 1984 2005 Update
More informationBackground. Background. C. T. Kelley NC State University tim C. T. Kelley Background NCSU, Spring / 58
Background C. T. Kelley NC State University tim kelley@ncsu.edu C. T. Kelley Background NCSU, Spring 2012 1 / 58 Notation vectors, matrices, norms l 1 : max col sum... spectral radius scaled integral norms
More informationMath 240 Calculus III
The Calculus III Summer 2015, Session II Wednesday, July 8, 2015 Agenda 1. of the determinant 2. determinants 3. of determinants What is the determinant? Yesterday: Ax = b has a unique solution when A
More informationAnalysis of a StreamlineDiffusion Finite Element Method on BakhvalovShishkin Mesh for Singularly Perturbed Problem
Numer. Math. Theor. Meth. Appl. Vol. 10, No. 1, pp. 4464 doi: 10.408/nmtma.017.y1306 February 017 Analysis of a StreamlineDiffusion Finite Element Method on BakhvalovShishkin Mesh for Singularly Perturbed
More informationPreconditioning. Noisy, IllConditioned Linear Systems
Preconditioning Noisy, IllConditioned Linear Systems James G. Nagy Emory University Atlanta, GA Outline 1. The Basic Problem 2. Regularization / Iterative Methods 3. Preconditioning 4. Example: Image
More informationReview of some mathematical tools
MATHEMATICAL FOUNDATIONS OF SIGNAL PROCESSING Fall 2016 Benjamín Béjar Haro, Mihailo Kolundžija, Reza Parhizkar, Adam Scholefield Teaching assistants: Golnoosh Elhami, Hanjie Pan Review of some mathematical
More informationSolving Sparse Linear Systems: Iterative methods
Scientific Computing with Case Studies SIAM Press, 2009 http://www.cs.umd.edu/users/oleary/sccs Lecture Notes for Unit VII Sparse Matrix Computations Part 2: Iterative Methods Dianne P. O Leary c 2008,2010
More informationw T 1 w T 2. w T n 0 if i j 1 if i = j
Lyapunov Operator Let A F n n be given, and define a linear operator L A : C n n C n n as L A (X) := A X + XA Suppose A is diagonalizable (what follows can be generalized even if this is not possible 
More informationNumerical Methods for LargeScale Nonlinear Systems
Numerical Methods for LargeScale Nonlinear Systems Handouts by Ronald H.W. Hoppe following the monograph P. Deuflhard Newton Methods for Nonlinear Problems Springer, BerlinHeidelbergNew York, 2004 Num.
More informationHOMEWORK PROBLEMS FROM STRANG S LINEAR ALGEBRA AND ITS APPLICATIONS (4TH EDITION)
HOMEWORK PROBLEMS FROM STRANG S LINEAR ALGEBRA AND ITS APPLICATIONS (4TH EDITION) PROFESSOR STEVEN MILLER: BROWN UNIVERSITY: SPRING 2007 1. CHAPTER 1: MATRICES AND GAUSSIAN ELIMINATION Page 9, # 3: Describe
More informationA matrix is a rectangular array of. objects arranged in rows and columns. The objects are called the entries. is called the size of the matrix, and
Section 5.5. Matrices and Vectors A matrix is a rectangular array of objects arranged in rows and columns. The objects are called the entries. A matrix with m rows and n columns is called an m n matrix.
More informationON BEST APPROXIMATIONS OF POLYNOMIALS IN MATRICES IN THE MATRIX 2NORM
ON BEST APPROXIMATIONS OF POLYNOMIALS IN MATRICES IN THE MATRIX 2NORM JÖRG LIESEN AND PETR TICHÝ Abstract. We show that certain matrix approximation problems in the matrix 2norm have uniquely defined
More informationKarhunenLoève Approximation of Random Fields Using Hierarchical Matrix Techniques
Institut für Numerische Mathematik und Optimierung KarhunenLoève Approximation of Random Fields Using Hierarchical Matrix Techniques Oliver Ernst Computational Methods with Applications Harrachov, CR,
More informationPoisson Equation in 2D
A Parallel Strategy Department of Mathematics and Statistics McMaster University March 31, 2010 Outline Introduction 1 Introduction Motivation Discretization Iterative Methods 2 Additive Schwarz Method
More information12. Perturbed Matrices
MAT334 : Applied Linear Algebra Mike Newman, winter 208 2. Perturbed Matrices motivation We want to solve a system Ax = b in a context where A and b are not known exactly. There might be experimental errors,
More informationA parameter tuning technique of a weighted Jacobitype preconditioner and its application to supernova simulations
A parameter tuning technique of a weighted Jacobitype preconditioner and its application to supernova simulations Akira IMAKURA Center for Computational Sciences, University of Tsukuba Joint work with
More informationON AUGMENTED LAGRANGIAN METHODS FOR SADDLEPOINT LINEAR SYSTEMS WITH SINGULAR OR SEMIDEFINITE (1,1) BLOCKS * 1. Introduction
Journal of Computational Mathematics Vol.xx, No.x, 200x, 1 9. http://www.globalsci.org/jcm doi:10.4208/jcm.1401cr7 ON AUGMENED LAGRANGIAN MEHODS FOR SADDLEPOIN LINEAR SYSEMS WIH SINGULAR OR SEMIDEFINIE
More informationA Residual Inverse Power Method
University of Maryland Institute for Advanced Computer Studies Department of Computer Science College Park TR 2007 09 TR 4854 A Residual Inverse Power Method G. W. Stewart February 2007 ABSTRACT The inverse
More informationIntegration of Rational Functions by Partial Fractions
Integration of Rational Functions by Partial Fractions Part 2: Integrating Rational Functions Rational Functions Recall that a rational function is the quotient of two polynomials. x + 3 x + 2 x + 2 x
More informationCalculus II  Basic Matrix Operations
Calculus II  Basic Matrix Operations Ryan C Daileda Terminology A matrix is a rectangular array of numbers, for example 7,, 7 7 9, or / / /4 / / /4 / / /4 / /6 The numbers in any matrix are called its
More informationIterative Methods for Solving A x = b
Iterative Methods for Solving A x = b A good (free) online source for iterative methods for solving A x = b is given in the description of a set of iterative solvers called templates found at netlib: http
More informationAMS526: Numerical Analysis I (Numerical Linear Algebra)
AMS526: Numerical Analysis I (Numerical Linear Algebra) Lecture 1: Course Overview & MatrixVector Multiplication Xiangmin Jiao SUNY Stony Brook Xiangmin Jiao Numerical Analysis I 1 / 20 Outline 1 Course
More informationFunctional Analysis Review
Outline 9.520: Statistical Learning Theory and Applications February 8, 2010 Outline 1 2 3 4 Vector Space Outline A vector space is a set V with binary operations +: V V V and : R V V such that for all
More informationANALYSIS OF AUGMENTED LAGRANGIANBASED PRECONDITIONERS FOR THE STEADY INCOMPRESSIBLE NAVIERSTOKES EQUATIONS
ANALYSIS OF AUGMENTED LAGRANGIANBASED PRECONDITIONERS FOR THE STEADY INCOMPRESSIBLE NAVIERSTOKES EQUATIONS MICHELE BENZI AND ZHEN WANG Abstract. We analyze a class of modified augmented Lagrangianbased
More informationIntroduction and Stationary Iterative Methods
Introduction and C. T. Kelley NC State University tim kelley@ncsu.edu Research Supported by NSF, DOE, ARO, USACE DTU ITMAN, 2011 Outline Notation and Preliminaries General References What you Should Know
More informationHigh Performance Nonlinear Solvers
What is a nonlinear system? High Performance Nonlinear Solvers Michael McCourt Division Argonne National Laboratory IIT Meshfree Seminar September 19, 2011 Every nonlinear system of equations can be described
More informationUnit 1 Matrices Notes Packet Period: Matrices
Algebra 2/Trig Unit 1 Matrices Notes Packet Name: Period: # Matrices (1) Page 203 204 #11 35 Odd (2) Page 203 204 #12 36 Even (3) Page 211 212 #4 6, 17 33 Odd (4) Page 211 212 #12 34 Even (5) Page 218
More informationPreconditioning for Accurate Solutions of Linear Systems and Eigenvalue Problems
arxiv:1705.04340v1 [math.na] 11 May 2017 Preconditioning for Accurate Solutions of Linear Systems and Eigenvalue Problems Qiang Ye Abstract This paper develops the preconditioning technique as a method
More informationETNA Kent State University
Electronic Transactions on Numerical Analysis. Volume 18, pp. 4964, 2004. Copyright 2004,. ISSN 10689613. ETNA EFFICIENT PRECONDITIONING FOR SEQUENCES OF PARAMETRIC COMPLEX SYMMETRIC LINEAR SYSTEMS D.
More informationETNA Kent State University
Electronic Transactions on Numerical Analysis. Volume 41, pp. 159166, 2014. Copyright 2014,. ISSN 10689613. ETNA MAXMIN AND MINMAX APPROXIMATION PROBLEMS FOR NORMAL MATRICES REVISITED JÖRG LIESEN AND
More informationOn the interplay between discretization and preconditioning of Krylov subspace methods
On the interplay between discretization and preconditioning of Krylov subspace methods Josef Málek and Zdeněk Strakoš Nečas Center for Mathematical Modeling Charles University in Prague and Czech Academy
More informationMatrix Algebra, part 2
Matrix Algebra, part 2 MingChing Luoh 2005.9.12 1 / 38 Diagonalization and Spectral Decomposition of a Matrix Optimization 2 / 38 Diagonalization and Spectral Decomposition of a Matrix Also called Eigenvalues
More informationNotes on Some Methods for Solving Linear Systems
Notes on Some Methods for Solving Linear Systems Dianne P. O Leary, 1983 and 1999 and 2007 September 25, 2007 When the matrix A is symmetric and positive definite, we have a whole new class of algorithms
More informationPreconditioning via Diagonal Scaling
Preconditioning via Diagonal Scaling Reza Takapoui Hamid Javadi June 4, 2014 1 Introduction Interior point methods solve small to medium sized problems to high accuracy in a reasonable amount of time.
More informationSTAT200C: Review of Linear Algebra
Stat200C Instructor: Zhaoxia Yu STAT200C: Review of Linear Algebra 1 Review of Linear Algebra 1.1 Vector Spaces, Rank, Trace, and Linear Equations 1.1.1 Rank and Vector Spaces Definition A vector whose
More informationChapter 7. Linear Algebra: Matrices, Vectors,
Chapter 7. Linear Algebra: Matrices, Vectors, Determinants. Linear Systems Linear algebra includes the theory and application of linear systems of equations, linear transformations, and eigenvalue problems.
More information5. Orthogonal matrices
L Vandenberghe EE133A (Spring 2017) 5 Orthogonal matrices matrices with orthonormal columns orthogonal matrices tall matrices with orthonormal columns complex matrices with orthonormal columns 51 Orthonormal
More informationarxiv: v2 [math.na] 1 Feb 2013
A FRAMEWORK FOR DEFLATED AND AUGMENTED KRYLOV SUBSPACE METHODS ANDRÉ GAUL, MARTIN H. GUTKNECHT, JÖRG LIESEN AND REINHARD NABBEN arxiv:1206.1506v2 [math.na] 1 Feb 2013 Abstract. We consider deflation and
More informationChebyshev semiiteration in Preconditioning
Report no. 08/14 Chebyshev semiiteration in Preconditioning Andrew J. Wathen Oxford University Computing Laboratory Tyrone Rees Oxford University Computing Laboratory Dedicated to Victor Pereyra on his
More information2.29 Numerical Fluid Mechanics Spring 2015 Lecture 9
Spring 2015 Lecture 9 REVIEW Lecture 8: Direct Methods for solving (linear) algebraic equations Gauss Elimination LU decomposition/factorization Error Analysis for Linear Systems and Condition Numbers
More informationBasic Concepts in Linear Algebra
Basic Concepts in Linear Algebra Grady B Wright Department of Mathematics Boise State University February 2, 2015 Grady B Wright Linear Algebra Basics February 2, 2015 1 / 39 Numerical Linear Algebra Linear
More informationOn Hybrid MultigridSchwarz algorithms Sébastien Loisel, Reinhard Nabben, and Daniel B. Szyld Research Report August 2007
On Hybrid MultigridSchwarz algorithms Sébastien Loisel, Reinhard Nabben, and Daniel B. Szyld Research Report 07828 August 2007 This report is available in the World Wide Web at http://www.math.temple.edu/~szyld
More informationA coupled system of singularly perturbed semilinear reactiondiffusion equations
A coupled system of singularly perturbed semilinear reactiondiffusion equations J.L. Gracia, F.J. Lisbona, M. MadauneTort E. O Riordan September 9, 008 Abstract In this paper singularly perturbed semilinear
More informationMath Linear Algebra II. 1. Inner Products and Norms
Math 342  Linear Algebra II Notes 1. Inner Products and Norms One knows from a basic introduction to vectors in R n Math 254 at OSU) that the length of a vector x = x 1 x 2... x n ) T R n, denoted x,
More informationA Method for Constructing Diagonally Dominant Preconditioners based on Jacobi Rotations
A Method for Constructing Diagonally Dominant Preconditioners based on Jacobi Rotations Jin Yun Yuan Plamen Y. Yalamov Abstract A method is presented to make a given matrix strictly diagonally dominant
More informationOrthogonality. 6.1 Orthogonal Vectors and Subspaces. Chapter 6
Chapter 6 Orthogonality 6.1 Orthogonal Vectors and Subspaces Recall that if nonzero vectors x, y R n are linearly independent then the subspace of all vectors αx + βy, α, β R (the space spanned by x and
More informationFrom Direct to Iterative Substructuring: some Parallel Experiences in 2 and 3D
From Direct to Iterative Substructuring: some Parallel Experiences in 2 and 3D Luc Giraud N7IRIT, Toulouse MUMPS Day October 24, 2006, ENSINRIA, Lyon, France Outline 1 General Framework 2 The direct
More informationAggregationbased algebraic multigrid
Aggregationbased algebraic multigrid from theory to fast solvers Yvan Notay Université Libre de Bruxelles Service de Métrologie Nucléaire CEMRACS, Marseille, July 18, 2012 Supported by the Belgian FNRS
More informationMatrices and Vectors. Definition of Matrix. An MxN matrix A is a twodimensional array of numbers A =
30 MATHEMATICS REVIEW G A.1.1 Matrices and Vectors Definition of Matrix. An MxN matrix A is a twodimensional array of numbers A = a 11 a 12... a 1N a 21 a 22... a 2N...... a M1 a M2... a MN A matrix can
More informationA PARAMETER ROBUST NUMERICAL METHOD FOR A TWO DIMENSIONAL REACTIONDIFFUSION PROBLEM
A PARAMETER ROBUST NUMERICAL METHOD FOR A TWO DIMENSIONAL REACTIONDIFFUSION PROBLEM C. CLAVERO, J.L. GRACIA, AND E. O RIORDAN Abstract. In this paper a singularly perturbed reaction diffusion partial
More informationChapter Two: Numerical Methods for Elliptic PDEs. 1 Finite Difference Methods for Elliptic PDEs
Chapter Two: Numerical Methods for Elliptic PDEs Finite Difference Methods for Elliptic PDEs.. Finite difference scheme. We consider a simple example u := subject to Dirichlet boundary conditions ( ) u
More informationScientific Computing: Solving Linear Systems
Scientific Computing: Solving Linear Systems Aleksandar Donev Courant Institute, NYU 1 donev@courant.nyu.edu 1 Course MATHGA.2043 or CSCIGA.2112, Spring 2012 September 17th and 24th, 2015 A. Donev (Courant
More informationA Framework for Analyzing and Constructing HierarchicalType A Posteriori Error Estimators
A Framework for Analyzing and Constructing HierarchicalType A Posteriori Error Estimators Jeff Ovall University of Kentucky Mathematics www.math.uky.edu/ jovall jovall@ms.uky.edu Kentucky Applied and
More informationMatrix Inequalities by Means of Block Matrices 1
Mathematical Inequalities & Applications, Vol. 4, No. 4, 200, pp. 48490. Matrix Inequalities by Means of Block Matrices Fuzhen Zhang 2 Department of Math, Science and Technology Nova Southeastern University,
More informationMath 18, Linear Algebra, Lecture C00, Spring 2017 Review and Practice Problems for Final Exam
Math 8, Linear Algebra, Lecture C, Spring 7 Review and Practice Problems for Final Exam. The augmentedmatrix of a linear system has been transformed by row operations into 5 4 8. Determine if the system
More informationFast matrix algebra for dense matrices with rankdeficient offdiagonal blocks
CHAPTER 2 Fast matrix algebra for dense matrices with rankdeficient offdiagonal blocks Chapter summary: The chapter describes techniques for rapidly performing algebraic operations on dense matrices
More informationOn the Vorobyev method of moments
On the Vorobyev method of moments Zdeněk Strakoš Charles University in Prague and Czech Academy of Sciences http://www.karlin.mff.cuni.cz/ strakos Conference in honor of Volker Mehrmann Berlin, May 2015
More informationCombination Preconditioning of saddlepoint systems for positive definiteness
Combination Preconditioning of saddlepoint systems for positive definiteness Andy Wathen Oxford University, UK joint work with Jen Pestana Eindhoven, 2012 p.1/30 compute iterates with residuals Krylov
More informationSection 3.2. Multiplication of Matrices and Multiplication of Vectors and Matrices
3.2. Multiplication of Matrices and Multiplication of Vectors and Matrices 1 Section 3.2. Multiplication of Matrices and Multiplication of Vectors and Matrices Note. In this section, we define the product
More informationBasic Concepts in Matrix Algebra
Basic Concepts in Matrix Algebra An column array of p elements is called a vector of dimension p and is written as x p 1 = x 1 x 2. x p. The transpose of the column vector x p 1 is row vector x = [x 1
More informationCS100: DISCRETE STRUCTURES. Lecture 3 Matrices Ch 3 Pages:
CS100: DISCRETE STRUCTURES Lecture 3 Matrices Ch 3 Pages: 246262 Matrices 2 Introduction DEFINITION 1: A matrix is a rectangular array of numbers. A matrix with m rows and n columns is called an m x n
More informationPreconditioned GMRES Revisited
Preconditioned GMRES Revisited Roland Herzog Kirk Soodhalter UBC (visiting) RICAM Linz Preconditioning Conference 2017 Vancouver August 01, 2017 Preconditioned GMRES Revisited Vancouver 1 / 32 Table of
More informationIr O D = D = ( ) Section 2.6 Example 1. (Bottom of page 119) dim(v ) = dim(l(v, W )) = dim(v ) dim(f ) = dim(v )
Section 3.2 Theorem 3.6. Let A be an m n matrix of rank r. Then r m, r n, and, by means of a finite number of elementary row and column operations, A can be transformed into the matrix ( ) Ir O D = 1 O
More information1. Introduction. We consider restricted Schwarz methods for the solution of linear systems of the form
SIAM J. NUMER. ANAL. Vol. 40, No. 6, pp. 2318 2336 c 2003 Society for Industrial and Applied Mathematics CONVERGENCE THEORY OF RESTRICTED MULTIPLICATIVE SCHWARZ METHODS REINHARD NABBEN AND DANIEL B. SZYLD
More informationCONVERGENCE ANALYSIS OF FINITE ELEMENT SOLUTION OF ONEDIMENSIONAL SINGULARLY PERTURBED DIFFERENTIAL EQUATIONS ON EQUIDISTRIBUTING MESHES
INTERNATIONAL JOURNAL OF NUMERICAL ANALYSIS AND MODELING Volume, Number, Pages 57 74 c 5 Institute for Scientific Computing and Information CONVERGENCE ANALYSIS OF FINITE ELEMENT SOLUTION OF ONEDIMENSIONAL
More informationAN ITERATION. In part as motivation, we consider an iteration method for solving a system of linear equations which has the form x Ax = b
AN ITERATION In part as motivation, we consider an iteration method for solving a system of linear equations which has the form x Ax = b In this, A is an n n matrix and b R n.systemsof this form arise
More informationTotal least squares. Gérard MEURANT. October, 2008
Total least squares Gérard MEURANT October, 2008 1 Introduction to total least squares 2 Approximation of the TLS secular equation 3 Numerical experiments Introduction to total least squares In least squares
More informationGeneralized Shifted Inverse Iterations on Grassmann Manifolds 1
Proceedings of the Sixteenth International Symposium on Mathematical Networks and Systems (MTNS 2004), Leuven, Belgium Generalized Shifted Inverse Iterations on Grassmann Manifolds 1 J. Jordan α, P.A.
More informationNested Krylov methods for shifted linear systems
Nested Krylov methods for shifted linear systems M. Baumann, and M. B. van Gizen Email: M.M.Baumann@tudelft.nl Delft Institute of Applied Mathematics Delft University of Technology Delft, The Netherlands
More information