Krylov subspace methods from the analytic, application and computational perspective

Similar documents
On the interplay between discretization and preconditioning of Krylov subspace methods

On the Vorobyev method of moments

Matching moments and matrix computations

Moments, Model Reduction and Nonlinearity in Solving Linear Algebraic Problems

Principles and Analysis of Krylov Subspace Methods

Key words. conjugate gradients, normwise backward error, incremental norm estimation.

Composite convergence bounds based on Chebyshev polynomials and finite precision conjugate gradient computations

Introduction. Chapter One

MATHEMATICS. Course Syllabus. Section A: Linear Algebra. Subject Code: MA. Course Structure. Ordinary Differential Equations

The Conjugate Gradient Method

Contribution of Wo¹niakowski, Strako²,... The conjugate gradient method in nite precision computa

Efficient Estimation of the A-norm of the Error in the Preconditioned Conjugate Gradient Method

Numerical Methods in Matrix Computations

Golub-Kahan iterative bidiagonalization and determining the noise level in the data

On the influence of eigenvalues on Bi-CG residual norms

On the accuracy of saddle point solvers

Chebyshev semi-iteration in Preconditioning

Preface to the Second Edition. Preface to the First Edition

Chapter 7 Iterative Techniques in Matrix Algebra

Stabilization and Acceleration of Algebraic Multigrid Method

Sensitivity of Gauss-Christoffel quadrature and sensitivity of Jacobi matrices to small changes of spectral data

Applied Mathematics 205. Unit V: Eigenvalue Problems. Lecturer: Dr. David Knezevic

Preconditioning for Nonsymmetry and Time-dependence

AMS526: Numerical Analysis I (Numerical Linear Algebra) Lecture 23: GMRES and Other Krylov Subspace Methods; Preconditioning

Index. higher order methods, 52 nonlinear, 36 with variable coefficients, 34 Burgers equation, 234 BVP, see boundary value problems

Numerical behavior of inexact linear solvers

M.A. Botchev. September 5, 2014

Index. C 2 ( ), 447 C k [a,b], 37 C0 ( ), 618 ( ), 447 CD 2 CN 2

ETNA Kent State University

Lanczos tridigonalization and Golub - Kahan bidiagonalization: Ideas, connections and impact

Preconditioned inverse iteration and shift-invert Arnoldi method

Introduction to Numerical Analysis

Efficient Estimation of the A-norm of the Error in the Preconditioned Conjugate Gradient Method

Numerical Mathematics

The Lanczos and conjugate gradient algorithms

10.6 ITERATIVE METHODS FOR DISCRETIZED LINEAR EQUATIONS

Preconditioned GMRES Revisited

FEM and sparse linear system solving

O jednom starém článku a jeho mnohých souvislostech On an old article and its many connections

APPLIED NUMERICAL LINEAR ALGEBRA

Topics. The CG Algorithm Algorithmic Options CG s Two Main Convergence Theorems

Optimal Left and Right Additive Schwarz Preconditioning for Minimal Residual Methods with Euclidean and Energy Norms

9.1 Preconditioned Krylov Subspace Methods

Fast Iterative Solution of Saddle Point Problems

Linear Algebra Massoud Malek

ON ORTHOGONAL REDUCTION TO HESSENBERG FORM WITH SMALL BANDWIDTH

Conjugate gradient method. Descent method. Conjugate search direction. Conjugate Gradient Algorithm (294)

Conjugate Gradient Method

Iterative Methods for Linear Systems of Equations

Key words. GMRES method, convergence bounds, worst-case GMRES, ideal GMRES, field of values

Geometric Multigrid Methods

On numerical stability in large scale linear algebraic computations

Applied Linear Algebra

ANY FINITE CONVERGENCE CURVE IS POSSIBLE IN THE INITIAL ITERATIONS OF RESTARTED FOM

Key words. preconditioned conjugate gradient method, saddle point problems, optimal control of PDEs, control and state constraints, multigrid method

A Review of Preconditioning Techniques for Steady Incompressible Flow

A stable variant of Simpler GMRES and GCR

Linear Solvers. Andrew Hazel

A Chebyshev-based two-stage iterative method as an alternative to the direct solution of linear systems

Iterative Methods for Sparse Linear Systems

Lecture 3: Inexact inverse iteration with preconditioning

A Method for Constructing Diagonally Dominant Preconditioners based on Jacobi Rotations

6.4 Krylov Subspaces and Conjugate Gradients

On Numerical Stability in Large Scale Linear Algebraic Computations

1 Conjugate gradients

Multilevel Preconditioning of Graph-Laplacians: Polynomial Approximation of the Pivot Blocks Inverses

Iterative methods for positive definite linear systems with a complex shift

Introduction to Iterative Solvers of Linear Systems

Key words. inf-sup constant, iterative solvers, preconditioning, saddle point problems

Iterative Methods for Linear Systems

Boundary Value Problems and Iterative Methods for Linear Systems

Preliminary/Qualifying Exam in Numerical Analysis (Math 502a) Spring 2012

CLASSICAL ITERATIVE METHODS

Algebraic Multigrid as Solvers and as Preconditioner

Iterative methods for Linear System

Lecture 18 Classical Iterative Methods

ITERATIVE METHODS FOR SPARSE LINEAR SYSTEMS

Notes on PCG for Sparse Linear Systems

Efficient Solvers for Stochastic Finite Element Saddle Point Problems

JURJEN DUINTJER TEBBENS

homogeneous 71 hyperplane 10 hyperplane 34 hyperplane 69 identity map 171 identity map 186 identity map 206 identity matrix 110 identity matrix 45

On a residual-based a posteriori error estimator for the total error

Domain Decomposition Preconditioners for Spectral Nédélec Elements in Two and Three Dimensions

BASIC EXAM ADVANCED CALCULUS/LINEAR ALGEBRA

Iterative Methods and Multigrid

Lab 1: Iterative Methods for Solving Linear Systems

Scientific Computing WS 2017/2018. Lecture 18. Jürgen Fuhrmann Lecture 18 Slide 1

Inexact inverse iteration with preconditioning

Lanczos tridiagonalization, Krylov subspaces and the problem of moments

Algebra C Numerical Linear Algebra Sample Exam Problems

Index. for generalized eigenvalue problem, butterfly form, 211

2.29 Numerical Fluid Mechanics Spring 2015 Lecture 9

CME 302: NUMERICAL LINEAR ALGEBRA FALL 2005/06 LECTURE 0

EECS 275 Matrix Computation

Lecture on: Numerical sparse linear algebra and interpolation spaces. June 3, 2014

LARGE SPARSE EIGENVALUE PROBLEMS. General Tools for Solving Large Eigen-Problems

ITERATIVE METHODS BASED ON KRYLOV SUBSPACES

FINITE-DIMENSIONAL LINEAR ALGEBRA

Stopping criteria for iterations in finite element methods

OUTLINE ffl CFD: elliptic pde's! Ax = b ffl Basic iterative methods ffl Krylov subspace methods ffl Preconditioning techniques: Iterative methods ILU

Transcription:

Krylov subspace methods from the analytic, application and computational perspective Zdeněk Strakoš Charles University in Prague and Czech Academy of Sciences http://www.karlin.mff.cuni.cz/ strakos Rencontres INRIA-JLJJ en calcul scientifique, Paris, April 216

Thanks Results affected by very many authors, and coauthored, in particular, with Josef Málek Tomáš Gergelits Jan Papež Joerg Liesen Z. Strakoš 2

Krylov subspace methods A x = b, x, S n, C n A n x n = b n x n approximates the solution x in x + S n with b Ax n orthogonal to C n S n, C n related to K n (A, r ) span {r, Ar,, A n 1 r } moments r A j r, j =, 1, 2, Z. Strakoš 3

Historical development and context 195 Mathematical foundations of quantum mechanics Hilbert 1926/27, von Neumann 1927/32, Wintner 1929 Krylov subspace methods Lanczos 195/52, Hestenes & Stiefel 1952 Modern numerical analysis von Neumann & Goldstine 1947 Turing 1948 Krylov sequences Gantmacher 1934 Transformation of the characteristic equation Krylov 1931 Representation theorem Riesz 199 Orthogonalisation algorithms for functions and vectors Schmidt 195/7, Szász 191 Jacobi form (or matrix) Hellinger & Toeplitz 1914 Foundations of functional analysis, including continuous spectrum, resolution of unity, self-adjoined operators, Hilbert space Hilbert 196-1912 Analytic theory of continued fractions, Riemann-Stieltjes integral, solution of the moment problem Stieltjes 1894 Continued fractions and Chebyshev inequalities Chebyshev 1855, Markov, Stieltjes 1884 Continued fractions and three-term recurrence for orthogonal polynomials Chebyshev 1855/59 Generalisations of the Gauss quadrature, minimal partial realisation Christoffel 1858/77 Gauss quadrature and orthogonal polynomials Jacobi 1826 Orthogonalisation via the Gramian Gram 1883 Orthogonalisation idea Laplace 182 Minimal polynomial Frobenius 1878 Jordan canonical form Weierstrass 1868, Jordan 187 Diagonalisation of quadratic forms Jacobi 1857 Reduction of bilinear form to tridiagonal form Jacobi 1848 Characteristic equation Cauchy 184 Real symmetric matrices have real eigenvalues, interlacing property Cauchy 1824 Infinite series expansions and continued fractions Euler 1744/48 Three-term recurrences and continued fractions Brouncker, Wallis 165s 165 Gauss quadrature Gauss 1814 Mechanical quadrature Newton, Cotes 172s Secular equation of the moon Lagrange 1774 Z. Strakoš 4

Lanczos, Hestenes and Stiefel Convergence analysis Iterative methods Least squares solutions Optimisation Convex geometry Minimising functionals Approximation theory Orthogonal polynomials Chebyshev, Jacobi and Legendre polynomials Green s function Gibbs oscillation Rayleigh quotients Fourier series Rounding error analysis Numerical analysis Polynomial preconditioning Cost of computations Stopping criteria Cornelius Lanczos An iteration method for the solution of the eigenvalue problem of linear differential and integral operators, 195 Solution of systems of linear equations by minimized iterations, 1952 Chebyshev polynomials in the solution of large-scale linear systems, 1952 Magnus R. Hestenes & Eduard Stiefel Methods of conjugate gradients for solving linear systems, 1952 Floating point computations Data uncertainty Structure and sparsity Gaussian elimination Vandermonde determinant Matrix theory Linear algebra General inner products Cauchy-Schwarz inequality Orthogonalisation Projections Functional analysis Differential and integral operators Liouville-Neumann expansion Trigonometric interpolation Continued fractions Sturm sequences Fredholm problem Gauss-Christoffel quadrature Riemann-Stieltjes integral Dirichlet and Fejér kernel Real analysis Z. Strakoš 5

Operator preconditioning Klawonn (1995, 1996); Arnold, Falk, and Winther (1997, 1997); Steinbach and Wendland (1998); Mc Lean and Tran (1997); Christiansen and Nédélec (2, 2); Powell and Silvester (23); Elman, Silvester, and Wathen (25); Hiptmair (26); Axelsson and Karátson (29); Mardal and Winther (211); Kirby (211); Zulehner (211); Preconditioning Conference 213, Oxford;... Related ideas on spectral equivalence of operators can be found, e.g., in Faber, Manteuffel and Parter (199) with references to D Yakonov (1961) and Gunn(1964, 1965).... Very nice recent work Smears (216). Z. Strakoš 6

Mesh independent condition number R. Hiptmair, CMA (26): There is a continuous operator equation posed in infinite-dimensional spaces that underlines the linear system of equations [... ] awareness of this connection is key to devising efficient solution strategies for the linear systems. Operator preconditioning is a very general recipe [... ]. It is simple to apply, but may not be particularly efficient, because in case of the [ condition number ] bound of Theorem 2.1 is too large, the operator preconditioning offers no hint how to improve the preconditioner. Hence, operator preconditioner may often achieve [... ] the much-vaunted mesh independence of the preconditioner, but it may not perform satisfactorily on a given mesh. Z. Strakoš 7

Linear asymptotic behavior? V. Faber, T. Manteuffel and S. V. Parter, Adv. in Appl. Math. (199): For a fixed h, using a preconditioning strategy based on an equivalent operator may not be superior to classical methods [... ] Equivalence alone is not sufficient for a good preconditioning strategy. One must also choose an equivalent operator for which the bound is small. There is no flaw in the analysis, only a flaw in the conclusions drawn from the analysis [... ] asymptotic estimates ignore the constant multiplier. Methods with similar asymptotic work estimates may behave quite differently in practice. Z. Strakoš 8

Outline 1. Numerical solution of BVPs 2. Operator preconditioning 3. Algebraic preconditioning, discretization, and problem formulation 4. Various comments 5. Conclusions Z. Strakoš 9

1 Notation Let V be an infinite dimensional Hilbert space with the inner product (, ) V : V V R, the associated norm V, V # be the dual space of bounded (continuous) linear functionals on V with the duality pairing, : V # V R. For each f V # there exists a unique τf V such that f, v = (τf, v) V for all v V. In this way the inner product (, ) V determines the Riesz map τ : V # V. Z. Strakoš 1

1 Weak formulation of the BVP, assumptions Let a(, ) = V V R be a bounded and coercive bilinear form. For u V we can write the bounded linear functional a(u, ) on V as Au a(u, ) V #, i.e., Au, v = a(u, v) for all v V. This defines the bounded and coercive operator A : V V #, inf Au, u = α >, A = C. u V, u V =1 The Lax-Milgram theorem ensures that for any b V # there exists a unique solution x V of the problem a(x, v) = b, v for all v V. Z. Strakoš 11

1 Operator problem formulation Equivalently, Ax b, v = for all v V, which can be written as the equation in V #, Ax = b, A : V V #, x V, b V #. We will consider A self-adjoint with respect to the duality pairing,. Z. Strakoš 12

1 Discretization using V h V Let let Φ h = (φ (h) 1,...,φ(h) N ) be a basis of the subspace V h V, Φ # h = (φ(h)# 1,...,φ (h)# N ) be the canonical basis of its dual V # h. The Galerkin discretization then gives A h x h = b h, x h V h, b h V # h, A h : V h V # h. Using the coordinates x h = Φ h x, b h = Φ # h the linear algebraic system b, the discretization results in Ax = b. Z. Strakoš 13

1 Computation Preconditioning needed for accelerating the iterations is then often build up algebraically for the given matrix problem, giving (here illustrated as the left preconditioning) M 1 Ax = M 1 b. Then the CG method is applied to the (symmetrized) preconditioned system, i.e., (PCG) (M-preconditioned CG) is applied to the unpreconditioned system. The schema of the solution process: A, b, A,b preconditioning PCG applied to Ax = b. Z. Strakoš 14

1 This talk presents a bit different view Formulation of the model, discretization and algebraic computation, including the evaluation of the error, stopping criteria for the algebraic solver, adaptivity etc. are very closely related to each other. Z. Strakoš 15

Outline 1. Numerical solution of BVPs 2. Operator preconditioning 3. Algebraic preconditioning, discretization, and problem formulation 4. Various comments 5. Conclusions Z. Strakoš 16

2 Operator formulation of the problem Recall that the inner product (, ) V defines the Riesz map τ. It can be used to transform the equation in V # Ax = b, A : V V #, x V, b V #. into the equation in V τax = τb, τa : V V, x V, τb V, This transformation is called operator preconditioning; see Klawonn (1995,... ), Arnold, Winther et al (1997,... ),... Z. Strakoš 17

2 The mathematically best preconditioning? With the choice of the inner product (, ) V = a(, ) we get i.e., a(u, v) = Au, v = a(τau, v) τ = A 1, and the preconditioned system x = A 1 b. Z. Strakoš 18

2 CG in infinite dimensional Hilbert spaces r = b Ax V #, p = τr V. For n = 1, 2,...,n max α n 1 = r n 1, τr n 1 Ap n 1, p n 1 x n = x n 1 + α n 1 p n 1, stop when the stopping criterion is satisfied r n = r n 1 α n 1 Ap n 1 β n = r n, τr n r n 1, τr n 1 p n = τr n + β n p n 1 Hayes (1954); Vorobyev (1958, 1965); Karush (1952); Stesin (1954) Superlinear convergence for (identity + compact) operators. Here the Riesz map τ indeed serves as the preconditioner. Z. Strakoš 19

2 Discretization of the infinite dimensional CG Using the coordinates in the bases Φ h and Φ # h of V h and V # h respectively, ( V # h = AV h ), f, v v f, (u, v) V v Mu, (M ij ) = ((φ j, φ i ) V ) i,j=1,...,n, Au Au, Au = AΦ h u = Φ # h Au ; (A ij) = (a(φ j, φ i )) i,j=1,...,n, τf M 1 f, τf = τφ # h f = Φ hm 1 f ; we get with b = Φ # h b, x n = Φ h x n, p n = Φ h p n, r n = Φ # h r n the algebraic CG formulation Z. Strakoš 2

2 Galerkin discretization gives matrix CG in V h r = b Ax, solve Mz = r, p = z. For n = 1,...,n max α n 1 = z n 1r n 1 p n 1 Ap n 1 x n = x n 1 + α n 1 p n 1, stop when the stopping criterion is satisfied r n = r n 1 α n 1 Ap n 1 Mz n = r n, solve for z n β n = z nr n z n 1 r n 1 p n = z n + β n p n 1 Günnel, Herzog, Sachs (214); Málek, S (215) Z. Strakoš 21

2 Philosophy of the a-priori robust bounds The bound κ(m 1 A) sup u,v V, u V =1, v V =1 Au, v inf u V, u V =1 Au, u is valid independently of the discretization, see, e.g., Hiptmair (26). If the bound is small enough, then the matter about the rate of convergence is resolved. Z. Strakoš 22

2 Observations Unpreconditioned CG, i.e. M = I, corresponds to the discretization basis Φ orthonormal wrt (, ) V. Orthogonalization of the discretization basis with respect to the given inner product in V will result in the unpreconditioned CG that is applied to the transformed (preconditioned) algebraic system. The resulting orthogonal discretization basis functions do not have local support and the transformed matrix is not sparse. Orthogonalization is not unique. For the same inner product we can get different bases and different discretized systems with exactly the same convergence behaviour. Z. Strakoš 23

Outline 1. Numerical solution of BVPs 2. Operator preconditioning 3. Algebraic preconditioning, discretization, and problem formulation 4. Various comments 5. Conclusions Z. Strakoš 24

3 Algebraic preconditioning? Consider an algebraic preconditioning with the (SPD) preconditioner M = L L = L(QQ ) L Where QQ = Q Q = I. Question: Can any algebraic preconditioning be expressed in the operator preconditioning framework? How does it link with the discretization and the choice of the inner product in V? Z. Strakoš 25

3 Change of the basis and of the inner product Transform the discretization bases Φ = Φ(( LQ) ) 1, Φ# = Φ # LQ. with the change of the inner product in V (recall (u, v) V = v Mu ) (u, v) new,v = ( Φû, Φ v) new,v := v û = v LQQ L u = v L L u = v Mu. The discretized Hilbert space formulation of CG gives the algebraically preconditioned matrix formulation of CG with the preconditioner M (more specifically, it gives the unpreconditioned CG applied to the algebraically preconditioned discretized system). Z. Strakoš 26

3 Sparsity, locality, global transfer of information Sparsity of matrices of the algebraic systems is always presented as an advantage of the FEM discretizations. Sparsity means locality of information in the individual matrix rows/columns. Getting a sufficiently accurate approximation to the solution may then require a substantial global transfer of information over the domain, i.e., a large dimension of the Krylov space. Preconditioning can be interpreted in part as addressing the unwanted consequence of sparsity (locality of the supports of the basis functions). Globally supported basis functions (hierarchical bases preconditioning, DD with coarse space components, multilevel methods, hierarchical grids etc.) can efficiently handle the transfer of global information. Z. Strakoš 27

3 Example - Nonhomogeneous diffusion tensor 1 4 Squared energy norm of the algebraic error 1 3 1 2 P1 FEM ; cond =2.57e+3 P1 FEM ichol ; cond =2.67e+1 P1 FEM lapl ; cond =1.e+2 P1 FEM ichol(1e 2) ; cond =1.72e+ 1 1 1 1 1 1 2 1 3 5 1 15 2 25 3 35 4 PCG convergence: unpreconditioned; ichol (no fill-in); Laplace operator preconditioning; ichol (drop-off tolerance 1e-2). Uniform mesh, condition numbers 2.5e3, 2.6e1, 1.e2, 1.7e. Z. Strakoš 28

3 Transformed basis elements Discretization basis function: P1 FEM; nnz = 1 Discretization basis function: P1 FEM ichol; nnz = 5 1.6.75.5.4.25.2 1 1 1 1 1 1 1 1 Original discretization basis element and its transformation corresponding to the ichol preconditioning. Z. Strakoš 29

3 Transformed basis elements Discretization basis function: P1 FEM lapl; nnz = 225 Discretization basis function: P1 FEM ichol(1e 2); nnz = 214.8.8.6.6.4.4.2.2 1 1 1 1 1 1 1 1 Transformed discretization basis elements corresponding o the lapl (left) and ichol(tol) preconditioning (right). Z. Strakoš 3

Outline 1. Numerical solution of BVPs 2. Operator preconditioning 3. Algebraic preconditioning, discretization, and problem formulation 4. Various comments 5. Conclusions Z. Strakoš 31

4 Model reduction using Krylov subspaces Consider B = τa, z = τb τax, and the Krylov sequence z, z 1 = Bz, z 2 = Bz 1 = B 2 z,..., z n = Bz n 1 = B n z,... Determine a sequence of operators B n defined on the sequence of the nested subspaces V n = span {z,...,z n 1 }, with the projector E n onto V n, B n = E n B E n. Convergence B n B? Z. Strakoš 32

4 Vorobyev moment problem The finite dimensional operators B n can be used to obtain approximate solutions to various linear problems. The choice of z, z 1,... as above gives a sequence of Krylov subspaces that are determined by the operator B and the initial element z. In this way the Vorobyev method of moments gives the Krylov subspace methods. Vorobyev (1958, 1965) covers bounded linear operators, bounded self-adjoint operators and some unbounded extensions. He made links to CG, Lanczos, Stieltjes moment problem, work of Markov, Gauss-Christoffel quadrature... Z. Strakoš 33

4 Conjugate gradient method - first n steps The first n steps of the (infinite or finite dimensional) CG method are given by T n y n = z V e 1, x n = x + Q n y n, x n x V n. Assume an approximation to the the n-th Krylov subspace K n is taken as the finite dimensional discretization subspace V h V in {A, b, x, τ} {τa n : K n K n } PCG with {A h,m h }? Then we get a close to optimal discretization (CG minimizes the energy norm over the discretization subspaces). Z. Strakoš 34

4 Gauss-Christoffel quadrature B x = f ω(λ), F(λ) dω(λ) T n y n = z V e 1 ω (n) (λ), n i=1 ω (n) i F ( θ (n) i ) Using F(λ) = λ 1 gives (assuming coercivity) λu λ L λ 1 dω(λ) = n i=1 ω (n) i ( θ (n) i ) 1 + x x n 2 a f 2 V Stieltjes (1894) and Vorobyev (1958) moment problems for self-adjoint bounded operators reduce to the Gauss-Christoffel quadrature (1814). No one would consider describing it by contraction. Z. Strakoš 35

4 Convergence via distribution functions Consider the (blue) distribution function determined by the operator τ A and the normalized τr. For a given n, find the (red) distribution function with n mass points that matches the maximal number (2n) of the first moments ((τa) l τr, τr ) V, l =, 1, 2, M 1 λ 1 M 2 2... 5 λ 2 M N λ N m 1 m 2 3... µ 1 µ 2 m n µ n Z. Strakoš 36

4 Clusters of eigenvalues mean fast convergence? M j M j 4 single eigenvalue λ j many close eigenvalues λ j1, λ j2,..., λ jl Replacing a single eigenvalue by a tight cluster can make a substantial difference; Greenbaum (1989); Greenbaum, S (1992); Golub, S (1994). If it does not, then it means that CG can not adapt to the problem, and it converges almost linearly. In such cases - is it worth using? Z. Strakoš 37

4 Rounding errors can be an important issue If preconditioning ensures getting an acceptable solution in a very few iterations, then rounding errors are not of concern. However, hard problems do exist. Then rounding errors can not be ignored. Descriptions of Krylov subspace methods that are based on contractions (condition numbers) are, in general, not descriptive. Analogy with a-priori and a-posteriori analysis in numerical PDEs. The power of Krylov subspace methods is in their self-adaptation to the problem! Z. Strakoš 38

4 Bounded invertible operators Consider a bounded linear operator B on a Hilbert space V that has a bounded inversion, and the problem B u = f. Since the identity operator on an infinite dimensional Hilbert space is not compact and BB 1 = I, it follows that B can not be compact. A uniform limit (in norm) of finite dimensional (approximation) operators B n is a compact operator. Results on strong convergence (pointwise limit); for the method of moments see Vorobyev (1958, 1965) B n w B w w V. Z. Strakoš 39

4 Invalid argument Let Z h be a numerical approximation of the bounded operator Z such that, with an appropriate extension, Z Z h = O(h). Then we have [(λ Z) 1 (λ Z h ) 1 ] = O(h) uniformly for λ Γ, where Γ surrounds the spectrum of Z with a distance of order O(h) or more. For any polynomial p p(z) p(z h ) = 1 2πi Γ p(λ)[(λ Z) 1 (λ Z h ) 1 ] dλ, and it seems that one can investigate p(z) instead of p(z h ). But the assumption Z Z h = O(h), h does not hold for any bounded invertible infinite dimensional operator Z. Z. Strakoš 4

4 Any GMRES convergence with any spectrum 1 The spectrum of A is given by {λ 1,...,λ N } and GMRES(A,b) yields residuals with the prescribed nonincreasing sequence (x = ) r r 1 r N 1 > r N =. 2 Let C be the spectral companion matrix, h = (h 1,...,h N ) T, h 2 i = r i 1 2 r i 2, i = 1,...,N. Let R be a nonsingular upper triangular matrix such that Rs = h with s being the first column of C 1, and let W be unitary matrix. Then A = WRCR 1 W and b = Wh. Greenbaum, Ptak, Arioli and S (1994-98); Liesen (1999); Eiermann and Ernst (21); Meurant (212); Meurant and Tebbens (212, 214);... Z. Strakoš 41

4 Interpretation? Given any spectrum and any sequence of the nonincreasing residual norms, this gives a complete parametrization of the set of all GMRES associated matrices and right hand sides. The set of problems for which the distribution of eigenvalues alone does not conform to convergence behaviour is not of measure zero and it is not pathological. Widespread eigenvalues alone can not be identified with poor convergence. Clustered eigenvalues alone can not be identified with fast convergence. Equivalent orthogonal matrices, Greenbaum, S (1994). Pseudospectrum indication! Z. Strakoš 42

4 Convection-diffusion model problem.25.2.15.1.5.5.1.15.2.25.15.2.25.3.35.4.45.5.55.6.65 Quiz: In one case the convergence of GMRES is substantially faster than in the other; for the solution see Liesen, S (25). Z. Strakoš 43

4 Delay of convergence due to inexactness 1 1 1 5 1 1? 1 5 1 1 1 15 2 4 6 8 1 1 15 residual smooth ubound backward error loss of orthogonality approximate solution error 1 2 3 4 5 6 7 8 iteration number Here numerical inexactness due to roundoff. How much may we relax accuracy of the most costly operations without causing an unwanted delay and/or affecting the maximal attainable accuracy? Z. Strakoš 44

4 Stopping criteria? 1.4 1.2 1.8.6.4 4 2 2 x 1 4.2 1 4 1 1.5.5 1 1 1 1 1 Exact solution x (left) and the discretisation error x x h (right) in the L-shape Poisson model problem, linear FEM, adaptive mesh refinement. Quasi equilibrated discretization error over the domain. Z. Strakoš 45

4 L-shape domain, Papež, Liesen, S (214) 4 x 1 4 4 x 1 4 2 2 2 2 4 1 4 1 1 1 1 1 1 1 The algebraic error x h x (n) h (right) even while (left) can dominate the total error x x (n) h x x n A x x h a = (x x h ). Z. Strakoš 46

Outline 1. Numerical solution of BVPs 2. Operator preconditioning 3. Algebraic preconditioning, discretization, and problem formulation 4. Various comments 5. Conclusions Z. Strakoš 47

Conclusions Krylov subspace methods adapt to the problem. Exploiting this adaptation is the key to their efficient use. Individual steps modeling-analysis-discretization-computation should not be considered separately within isolated disciplines. They form a single problem. Operator preconditioning follows this philosophy. Fast HPC computations require handling all involved issues. A posteriori error analysis and stopping criteria are essential... We are grateful for collaboration with Martin on these topics. Z. Strakoš 48

References J. Málek and Z.S., Preconditioning and the Conjugate Gradient Method in the Context of Solving PDEs. SIAM Spotlight Series, SIAM (215) T. Gergelits and Z.S., Composite convergence bounds based on Chebyshev polynomials and finite precision conjugate gradient computations, Numer. Alg. 65, 759-782 (214) J. Papež, J. Liesen and Z.S., Distribution of the discretization and algebraic error in numerical solution of partial differential equations, Linear Alg. Appl. 449, 89-114 (214) J. Liesen and Z.S., Krylov Subspace Methods, Principles and Analysis. Oxford University Press (213) Z. Strakoš 49

Merci! Z. Strakoš 5