Stopping criteria for iterations in finite element methods

Similar documents
Iterative Methods for Linear Systems of Equations

The Lanczos and conjugate gradient algorithms

FEM and sparse linear system solving

Key words. conjugate gradients, normwise backward error, incremental norm estimation.

AMS526: Numerical Analysis I (Numerical Linear Algebra) Lecture 23: GMRES and Other Krylov Subspace Methods; Preconditioning

STOPPING CRITERIA FOR MIXED FINITE ELEMENT PROBLEMS

Summary of Iterative Methods for Non-symmetric Linear Equations That Are Related to the Conjugate Gradient (CG) Method

6.4 Krylov Subspaces and Conjugate Gradients

M.A. Botchev. September 5, 2014

Inexactness and flexibility in linear Krylov solvers

Optimal Left and Right Additive Schwarz Preconditioning for Minimal Residual Methods with Euclidean and Energy Norms

Preface to the Second Edition. Preface to the First Edition

Introduction. Chapter One

Preconditioning for Nonsymmetry and Time-dependence

Approximating the matrix exponential of an advection-diffusion operator using the incomplete orthogonalization method

SOLVING SPARSE LINEAR SYSTEMS OF EQUATIONS. Chao Yang Computational Research Division Lawrence Berkeley National Laboratory Berkeley, CA, USA

On the influence of eigenvalues on Bi-CG residual norms

4.8 Arnoldi Iteration, Krylov Subspaces and GMRES

ON ORTHOGONAL REDUCTION TO HESSENBERG FORM WITH SMALL BANDWIDTH

Iterative methods for Linear System

Krylov subspace projection methods

Numerical Solutions to Partial Differential Equations

Preconditioned inverse iteration and shift-invert Arnoldi method

On the interplay between discretization and preconditioning of Krylov subspace methods

4.6 Iterative Solvers for Linear Systems

OUTLINE ffl CFD: elliptic pde's! Ax = b ffl Basic iterative methods ffl Krylov subspace methods ffl Preconditioning techniques: Iterative methods ILU

Computational Linear Algebra

Iterative methods for Linear System of Equations. Joint Advanced Student School (JASS-2009)

Iterative methods for positive definite linear systems with a complex shift

Krylov Subspace Methods that Are Based on the Minimization of the Residual

Introduction to Iterative Solvers of Linear Systems

ELLIPTIC RECONSTRUCTION AND A POSTERIORI ERROR ESTIMATES FOR PARABOLIC PROBLEMS

The amount of work to construct each new guess from the previous one should be a small multiple of the number of nonzeros in A.

Iterative Methods for Sparse Linear Systems

Numerical behavior of inexact linear solvers

9.1 Preconditioned Krylov Subspace Methods

On the accuracy of saddle point solvers

Iterative Methods for Linear Systems

Solving large sparse Ax = b.

Principles and Analysis of Krylov Subspace Methods

Lecture on: Numerical sparse linear algebra and interpolation spaces. June 3, 2014

Lecture 17 Methods for System of Linear Equations: Part 2. Songting Luo. Department of Mathematics Iowa State University

The quadratic eigenvalue problem (QEP) is to find scalars λ and nonzero vectors u satisfying

Adaptive preconditioners for nonlinear systems of equations

On solving linear systems arising from Shishkin mesh discretizations

Chapter 7 Iterative Techniques in Matrix Algebra

On the Preconditioning of the Block Tridiagonal Linear System of Equations

IDR(s) Master s thesis Goushani Kisoensingh. Supervisor: Gerard L.G. Sleijpen Department of Mathematics Universiteit Utrecht

A short course on: Preconditioned Krylov subspace methods. Yousef Saad University of Minnesota Dept. of Computer Science and Engineering

Total least squares. Gérard MEURANT. October, 2008

Boundary Value Problems and Iterative Methods for Linear Systems

Matrix square root and interpolation spaces

Applied Mathematics 205. Unit V: Eigenvalue Problems. Lecturer: Dr. David Knezevic

1 Conjugate gradients

The Nullspace free eigenvalue problem and the inexact Shift and invert Lanczos method. V. Simoncini. Dipartimento di Matematica, Università di Bologna

Krylov Space Solvers

Solving Symmetric Indefinite Systems with Symmetric Positive Definite Preconditioners

Recent advances in approximation using Krylov subspaces. V. Simoncini. Dipartimento di Matematica, Università di Bologna.

Preconditioned GMRES Revisited

AMS526: Numerical Analysis I (Numerical Linear Algebra)

Iterative Methods and Multigrid

Lecture 9: Krylov Subspace Methods. 2 Derivation of the Conjugate Gradient Algorithm

On the Modification of an Eigenvalue Problem that Preserves an Eigenspace

IDR(s) as a projection method

Lecture Note III: Least-Squares Method

Construction of a New Domain Decomposition Method for the Stokes Equations

Adaptive methods for control problems with finite-dimensional control space

Domain Decomposition Algorithms for an Indefinite Hypersingular Integral Equation in Three Dimensions

Lecture # 20 The Preconditioned Conjugate Gradient Method

In order to solve the linear system KL M N when K is nonsymmetric, we can solve the equivalent system

Chebyshev semi-iteration in Preconditioning

b i (x) u + c(x)u = f in Ω,

INF-SUP CONDITION FOR OPERATOR EQUATIONS

Charles University Faculty of Mathematics and Physics DOCTORAL THESIS. Krylov subspace approximations in linear algebraic problems

A Review of Preconditioning Techniques for Steady Incompressible Flow

On Solving Large Algebraic. Riccati Matrix Equations

The antitriangular factorisation of saddle point matrices

PRECONDITIONING TECHNIQUES FOR NEWTON S METHOD FOR THE INCOMPRESSIBLE NAVIER STOKES EQUATIONS

A Chebyshev-based two-stage iterative method as an alternative to the direct solution of linear systems

Last Time. Social Network Graphs Betweenness. Graph Laplacian. Girvan-Newman Algorithm. Spectral Bisection

A stable variant of Simpler GMRES and GCR

THE solution of the absolute value equation (AVE) of

Linear Solvers. Andrew Hazel

ITERATIVE METHODS FOR SPARSE LINEAR SYSTEMS

Lecture 8 Fast Iterative Methods for linear systems of equations

Preliminary/Qualifying Exam in Numerical Analysis (Math 502a) Spring 2012

Topics. The CG Algorithm Algorithmic Options CG s Two Main Convergence Theorems

Lanczos tridigonalization and Golub - Kahan bidiagonalization: Ideas, connections and impact

LECTURE 1: SOURCES OF ERRORS MATHEMATICAL TOOLS A PRIORI ERROR ESTIMATES. Sergey Korotov,

Contents. Preface... xi. Introduction...

DELFT UNIVERSITY OF TECHNOLOGY

Efficient Estimation of the A-norm of the Error in the Preconditioned Conjugate Gradient Method

On prescribing Ritz values and GMRES residual norms generated by Arnoldi processes

Block Bidiagonal Decomposition and Least Squares Problems

Course Notes: Week 1

Fast iterative solvers

Key words. inf-sup constant, iterative solvers, preconditioning, saddle point problems

Krylov Space Methods. Nonstationary sounds good. Radu Trîmbiţaş ( Babeş-Bolyai University) Krylov Space Methods 1 / 17

S-Step and Communication-Avoiding Iterative Methods

Solving Sparse Linear Systems: Iterative methods

Transcription:

Stopping criteria for iterations in finite element methods M. Arioli, D. Loghin and A. J. Wathen CERFACS Technical Report TR/PA/03/21 Abstract This work extends the results of Arioli [1], [2] on stopping criteria for iterative solution methods for finite element problems to the case of nonsymmetric positive-definite problems. We show that the residual measured in the norm induced by the symmetric part of the inverse of the system matrix is relevant to convergence in a finite element context. We then provide alternative ways of calculating or estimating this quantity and present numerical experiments which validate our criteria. Keywords: finite element method, Krylov methods, energy norm, forward and backward error. 1 Introduction Iterative methods of Krylov subspace type form a well-known and well-researched area in the context of solution methods for large sparse linear systems. In some cases, convergence can be described, in others not. Invariably however, the theoretical and practical convergence criterion is chosen to be the Euclidean norm of the residual, with the ubiquitous exception of the Conjugate Gradient method, where the energy norm lends itself quite naturally to analysis. On the other hand, finite element methods which are an important source of large, sparse linear systems provide a natural norm for convergence. While this fact is well-known and has been noted particularly in the case of symmetric positive-definite problems [6], [7], [11], only recently have there been attempts to relate convergence in the energy norm to the finite element context [1], [2], [13]. In particular, Arioli [1] was the first to consider the original finite element setting to provide convergence criteria for the Conjugate Gradient method. 1

In this work we consider the choice of stopping criteria for nonsymmetric positivedefinite problems. The immediate difficulty encountered is that of defining a suitable norm in which to measure convergence. In the case of symmetric positive-definite problems, the energy or A-norm of the error is equal to the dual norm or A 1 -norm of the residual, which is the quantity that is estimated. In the nonsymmetric case, we show that a useful definition of dual norm is the norm induced by the symmetric part of A 1. We show that one can also work with the norm induced by the inverse of the symmetric part of A for problems which are not too non-normal. The paper is structured as follows. In section 2 we describe the problem setting. In section 3 we derive general stopping criteria while in sections 4 and 5 we present ways of approximating the criteria introduced in the case of GMRES; we also consider the effect of preconditioning and derive the corresponding modified bounds. Finally, in section 6 we investigate the stopping criteria by performing experiments on various discretizations of convection-diffusion problems. 2 Problem description 2.1 Abstract formulation Consider the weak formulation Find u H such that for all v H a(u, v) = f(v), (1) where H is a Hilbert space of functions u defined on a closed subset Ω of R d, with dual H and inner-product norm H, while a(, ) is a nonsymmetric, positive-definite bilinear form on H H and f( ) H is a continuous linear form on H. Existence and uniqueness of solutions to problems of type (1) is guaranteed provided the following conditions hold for all u, v H with constants C 1, C 2 independent of discretization. a(w, v) C 1 w H v H (2a) a(v, v) C 2 v 2 H, (2b) 2

Condition (2b) is often used to replace the weaker (and sufficient) conditions of Babuška ([3]) a(w, v) sup v H\{0} v H C 2 w H, (3a) a(w, v) sup w H\{0} w H C 2 v H ; (3b) this is due to the fact that the weak formulation (1) with a(, ) replaced by its symmetric part is often stable in the sense of Babuška (i.e., satisfies (3)), leading to (2b). 2.2 Finite element approximation An approximation to problem (1) is sought through projection onto a finite-dimensional space H h H; the resulting formulation reads Find u h H h such that for all v h H h a(u h, v h ) = f(v h ). (4) Finite element methods choose H h to be a space of functions v h defined on a subdivision Ω h of Ω into simplices T of diameter h T ; h denotes a piecewise constant function defined on Ω h via h T = h T. Since H h H, (2) are satisfied with constants independent of h; thus, there exists a unique finite element approximation u h. Moreover subtracting (4) from (1) yields the standard orthogonality condition for all v h H h a(u u h, v h ) = 0, (5) which can be used (together with conditions (2)) to derive standard error estimates of the form u u h H C 1 inf u v h H. (6) C 2 v h H h Remark 2.1 Replacing v h with the interpolant of u, I h u, and using interpolation error estimates leads to a priori bounds of the form u u h H C(h)C(u) where C(u) is typically a constant depending only on u and its derivatives. 3

Conditions (2) can also be used in determining a posteriori error bounds. In particular, if we define the functional residual as a linear functional via R(u h ), v := f(v) a(u h, v) = a(u u h, v) v H then dividing by v H and using (2) leads to the following upper and lower bounds on the error where 1 C 1 R(u h ) H u u h H 1 C 2 R(u h ) H (7) R(u h ) H := R(u h ), v sup. v H\{0} v H Alternatively, noting that a H H = C 1 (cf. (2a)) we can rewrite (7) as where BE u u h H u h H =: C 1 C 2 BE (8) BE := R(u h) H u h H a H H (9) Definition 2.1 The quantities F E, BE in (8),(9) are the functional forward and backward error respectively [2]. Remark 2.2 The dual norm of the functional residual, R(u h ) H, is not easy to compute and most a posteriori error bounds are derived as approximations of this quantity. However, in general it is known that R(u h ) H and thus BE are (polynomial) functions of the discretization parameter h and thus far from being close to machine precision. This we will use to advantage in the derivation of stopping criteria. However, we will not be concerned here with the derivation of any bounds but we will assume the following generic bound on the relative error u u h H u h H C(h) (10) where C(h) is available via an a priori or a posteriori error analysis. 4

We end this subsection with some standard notation. Expanding u h in a basis of H h, we can derive a linear system of equations involving the coefficients (u) i, i = 1,..., n of u h in our choice of basis of H h Au = f (11) where n = dim H h and A R n n is a non-singular, generally nonsymmetric, matrix. In fact there is an isomorphism Π h between R n and H h which associates to every vector v R n a function v h H h via Π h v = i=1 v i φ i, where {φ i, i = 1,..., n} form a basis for H h. Henceforth, given a vector v R n we will denote its functional counterpart Π h v by v h. Note also that the above choice of basis defines a norm-matrix H via H ij = ((φ i, φ j )) where ((, )) is the H-inner product. Hence v h H = v H. We will also use the matrix norm H1,H 2 : R n n defined via M H1,H 2 := Mx H2 max x n \{0} x H1 where M R n n and H i R n n, i = 1, 2, 3 are symmetric and positive-definite matrices. We also note here that M H1,H 1 2 Finally, the stability conditions (2) become max max w n \{0} v n \{0} = H 1/2 2 MH1 1. (12) w t Av C 1 w H v H (13a) min v t Av C 2 v 2 H (13b) v n \{0} It is also easy to see that there also exits a constant C 3 C 1 such that max v t Av C 3 v 2 H. (13c) v n \{0} 5

Remark 2.3 In many situations of interest one can have C 3 = C 2. Moreover, if the symmetric part of A is H, then C 2 = C 3 = 1. We now state a result which can be found in [4]. Theorem 2.2 Let M R n n be nonsingular and let H R n n be a symmetric and positive-definite matrix. Then M H,H 1 = max M 1 1 H 1,H = min max w n \{0} v n \{0} max w n \{0} v n \{0} The above result justifies the following definition. Definition 2.3 The H-condition number of a matrix M is w t Mv w H v H, w t Mv w H v H. κ(m, H) κ H (M) := M H,H 1 M 1 H 1,H. Thus the stability conditions (2) simply say that the discrete problem (4) is wellconditioned with respect to the H-norm: and hence for all n A H,H 1 = C 1, A 1 1 H 1,H C 2, (14) κ H (A) C 1 C 2. (15) Finally, we note that when solving Au = f the discrete versions of (8), (9) are 3 Stopping criteria F E := u ũ H ũ H, BE := f Aũ H 1. (16) ũ H A H,H 1 In many large-scale computations the exact solution u of the linear system (11) is out of reach and an iterate ũ is used to approximate the solution. Since we identify ũ with a function ũ h H h, we naturally expect a useful iterate ũ to satisfy an error estimate similar to (10) u ũ h H ũ h C(h), H 6

where C(h) is of the same order as C(h) in (10). Our aim is to derive a sufficient and computable criterion for the above error bound to hold. First, we introduce some notation and useful results. Let M R n n. We denote by H M = (M + M t )/2, S M = (M M t )/2 the symmetric and skew-symmetric parts of M, respectively. Moreover, if H M is positive-definite, it induces a norm which we denote by We first prove the following results. M := HM. Lemma 3.1 Let conditions (13) hold. Then 1 C3 r A r H 1 C2 r A and C2 C 1 C 3 r H 1 r A 1 1 C2 r H 1. Proof See Appendix. Theorem 3.2 Let u be the solution of the weak formulation (1) and let u, u h = Π h u satisfy u u h H Au = f; C(h). u h H Then ũ h = Π h ũ satisfies if u ũ h H ũ h H C(h) = O(C(h)) for some t 0. f Aũ H 1 ũ H h t C(h)C 2, (17) Proof Let r = f Aũ. We have u ũ h H u u h H ũ h H u h H C(h) u h H + u h ũ h H ũ h H ũ h H ( 1 + u h ũ h H ũ h H ) + u h ũ h H ũ h H 7

and since u h ũ h H ũ h H = A 1 r H ũ H A 1 H 1,H r H 1 ũ H 1 C 2 r H 1 ũ H (using (14) ) we get u ũ h H ũ h H C(h)(1 + h t C(h)) + h t C(h) =: C(h). Remark 3.1 The stopping criterion (17) is equivalent to requiring the discrete backward error BE defined in (16) to be of the same order as the functional backward error BE = O(BE). This is also a sufficient condition for the discrete forward error F E corresponding to our iterative solution to have the same order as the functional forward error. In fact, criterion (17) can be replaced with a tighter bound. By Lemma 3.1, A 1 r H 1 C2 A 1 r A 1 C2 A 1 A 1,A r A 1 = 1 C2 r A 1 and thus, we can replace the bound (17) with f Aũ A 1 ũ H h t C(h) C 2. (18) The difference between the stopping criteria (17), (18) is not significant if the H- condition number (15) is not too large. This can be seen from the equivalence between H 1, A 1 provided by Lemma 3.1. In particular, if the symmetric part of A is H, then C 2 = C 3 = 1 and the effective condition number is C 1. A large value of C 1 corresponds to a highly nonsymmetric problem for which the use of criterion (18) rather than (17) may be preferable. We return to this issue in the numerics section. 8

3.1 One more crime In practice, the discretization of the weak formulation (1) is generally done in an approximate fashion, very often due to the computational costs involved. This approximation has been qualified as a variational crime [14], as it leads to a perturbed system (A + A)ũ = f. However, it is known that if the perturbation A is suitably small (usually within the finite element error), then the approximate solution ũ satisfies the same error estimates as the exact solution u [14]. In this context, the proposed stopping criteria represent but another variational crime as the following standard result shows (see [2]; see also [12] for the case when l p norms are employed.) Theorem 3.3 Let ũ satisfy Then there exists A such that and Proof See [2, Thm 1]. f Aũ H 1 ũ H h t C(h)C 2. (A + A)ũ = f A H,H 1 h t C(h)C 2 Remark 3.2 The more general case where the right-hand side f is perturbed is treated in [2]. We do not include the results here since in most engineering applications bounds of type (10) are preferred. The stopping criteria derived above pose the problem of estimating the residual in the H 1 - or A 1 -norms. While this was possible for the symmetric and positive-definite case in a natural way (see [6], [7]), the use of a nonsymmetric iterative method does not allow for the same methodology to be applied. In the remainder of the paper we show how this norm can be estimated using the information contained in the Krylov basis K k. For simplicity, we will consider only the case H = (A+A t )/2 (and thus C 2 = 1), i.e., the case when H defines the so-called energy norm for the problem. In the next section, we show how the norm estimation is achieved for GMRES and FOM. Finally, in the case of central preconditioning with H, these two algorithms reduce to a three-term recurrence which computes directly r k H 1. This will be the subject of section 5. 9

4 Stopping criteria for GMRES and FOM We recall here some of the basic facts and standard notation for the GMRES and FOM algorithms. The methods compute an orthonormal basis of K k ; the basis elements are the columns of V k R n k. This orthonormalization is achieved via an Arnoldi process which yields the factorizations V t k AV k = H k, AV k = V k+1 Hk where H k R k k, H k R k+1 k are upper Hessenberg matrices, with H k being obtained from H k by deleting its last row. In the case of GMRES, a QR-factorization of H k is computed (updated at each step) H k = Q k R k. 4.1 Estimation of r k H 1 This can be done simply via r k H 1 λ 1/2 min (H) rk. Depending on the application, the smallest eigenvalue of H may or may not be estimated with sufficient accuracy. If we do not have such an estimate, we must contend ourselves with estimates provided by the iterative process. In the case of GMRES and FOM this can be achieved as follows. Assuming no breakdown, the method computes the following factorization of A involving an orthonormal matrix V n V t n AV n = H n. Thus, V t nhv n = (H n + H t n)/2 =: H s n and therefore λ min (H) = λ min (H s n). Since in practice we wish to use the algorithm only for a small number of steps k, an estimate of λ min (H) can be taken to be λ min (Hk s ). Unfortunately, this estimate is always an upper bound on λ min (H). In fact, we have the following monotonicity result. Lemma 4.1 Let Hk s = (H k + Hk t )/2. Then λ min (H s k+1) λ min (H s k). 10

Proof λ min (Hk s ) = min q t k H kq k q k k q k 2 q t k = min V k tav kq k q k k q k 2 r t Ar = min r K k Vk tr 2 min r K k+1 r t Ar V t k r 2 Now, any r K k+1 can be written as r = V k+1 q k+1 for some q k+1 R k+1 and hence k+1 AV k+1q k+1 λ min (Hk s ) min q t k+1 V t q k+1 k+1 Vk tv k+1q k+1 2 q t k+1 min H k+1q k+1 q k+1 k+1 q k+1 2 = λ min (H s k+1 ). This result enables us to approximate the stopping criteria as follows. Since by the previous lemma λ min (H s k ) λ min(h) monotonically, there exists a k and a constant C = C (k ) such that λ min (H) C λ min (H s k ) for all k > k. Hence, our stopping criterion becomes r k C λ 1/2 min (Hs k )ht C(h). (19) We recall here that the value of t 0 can be chosen according to application (a larger value, for a more pessimistic convergence criterion). Thus, we only have to compute λ min (H s k ) and estimate C. In practice, the constant C is of order one for small values of k. We investigate this issue in the next section. Remark 4.1 Estimating λ min (Hk s ) can be done easily in the case of the FOM algorithm. However, in the case of GMRES this is not necessarily straightforward, since we do not store H k but the R k factor of the QR-factorization of Hk. In this case, a further approximation could be introduced λ min (H s k ) σ min(h k ) σ min ( H k ) = σ min (R k ) 11

leading to the bound r k C σ 1/2 min (R k)h t C(h), where C is a constant which accounts for both convergence to λ min (Hk s ) and the difference between λ min (Hk s) and σ min(r k ), which cannot be guaranteed to be small and is not known a priori. However, this latter bound is useful in estimating r k A 1. 4.2 Estimation of r k A 1 In this case we proceed similarly r k A 1 r k σ 1/2 min (A). A similar monotonicity result holds for the singular values of H k (cf. [9, Cor. 3.1.3]) σ min ( H k ) σ min ( H k+1 ) and thus there exists a k and a constant c = c (k ) such that σ min (A) c σ min (R k ) for all k > k. Thus the stopping criterion (18) can be replaced with r k c σ 1/2 min (R k)h t C(h). (20) where, as before, c is a constant (of order one) which we need to estimate. Remark 4.2 We note that this criterion can be used both in the case of GMRES and FOM, since in the first case the matrix R k is available and in the second case H k is available (with σ min ( H k ) = σ min (R k )). 4.3 Restarted GMRES/FOM There are many situations where the construction of an orthonormal basis for the Krylov subspace is limited to a small number of vectors. This leads to the restarted versions of GMRES or FOM. From the point of view of the above stopping criteria, this does not pose any major problems we still need to estimate either λ min (H) or σ min (A) and this is done in a similar fashion. Thus, assuming we run the algorithms for m iterations of k steps each, we use the following approximations λ min (H) min 1 i m λ(i) min (Hs k ), σ min(h) min 1 i m σ(i) min (R k) (21) where we denote by λ (i) (H s k ), σ(i) (R k ), the eigenvalues and singular values of the indicated matrices constructed at the ith iteration. 12

4.4 The effect of preconditioning In the case where a preconditioner is used, the Arnoldi algorithm constructs a similar factorization of the preconditioned matrix. We consider here only the case of right preconditioning for which the GMRES/FOM residual remains unchanged. The factorization is AP 1 V k = V k+1 Hk and since r A 1 σ 1/2 min (A) r σ 1/2 min (AP 1 )σ 1/2 (P ) r we can derive a stopping criterion similar to (20) using the approximation σ min (AP 1 ) σ min (R k ) min r k c σ 1/2 min (R k)σ 1/2 min (P )ht C(h). (22) However, this requires the estimation of the smallest singular value of P which may not be easy to achieve. We address this issue in the Section 6. 5 A minimum residual algorithm We have seen that in the case of GMRES estimation of r k H 1 can be done provided the Hessenberg matrix is stored. On the other hand, the more relevant quantity r k A 1 can be estimated quite naturally during the GMRES process. However, there is one situation where working with r k H 1 leads to a three-term recurrence algorithm as well as useful preconditioning. The algorithm solves Au = f by minimizing f Au H 1 over the Krylov space. This is by no means a novel result and has been previously considered by Concus and Golub [5] and Widlund [15] in the context of preconditioning nonsymmetric matrices by their symmetric part. We consider below the version of this algorithm which minimizes the H 1 -norm of the residual, where H is the symmetric and positive-definite part of A. Consider the modified problem Ãũ = f (23) where à = H 1/2 AH 1/2, f = H 1/2 f. Let us consider first the FOM algorithm applied to this system. As before, the residual is orthogonal to the Krylov space r k, q = H 1/2 (f Au k ), q = 0, q K k 13

where } K k = span { r 0, Ã r0,..., Ãk 1 r 0 = H 1/2 K k (H 1 A, H 1 r 0 ). Thus, p K k (H 1 A, H 1 r 0 ) H 1/2 r k, H 1/2 p = r k, p = H 1 r k, p H = 0. In other words, the standard FOM algorithm for (23) is also an orthogonal projection method with respect to the H-inner-product onto K k (H 1 A, H 1 r 0 ). Moreover, the advantage of this formulation is that there exists a three-term recurrence which solves this problem. We summarize this below. Lemma 5.1 Let A have symmetric and positive-definite part H. The FOM algorithm applied to (H 1/2 AH 1/2 )(H 1/2 u) = H 1/2 f. in the Euclidean inner-product is equivalent to the FOM algorithm in the H-inner product applied to H 1 Au = H 1 f. Moreover, the Arnoldi orthogonalization process applied to the normal matrix H 1/2 AH 1/2 yields a factorization Vk t AV k = H k, where (H k ) ij = 0 for all i j > 1. Proof See Appendix. This idea is contained in [15], although the author constructs a different tridiagonalization than that constructed by FOM (Arnoldi). Similarly, using the above result one can modify the standard GMRES algorithm into a three-term recurrence which constructs the solution with the smallest residual over K k as measured in the H 1 -norm. We do not include the details here, but only present in the next section numerical results obtained with this modified version of GMRES. Remark 5.1 The action of the inverse of H as a preconditioner can be relaxed in practice. Indeed, solving to an accuracy of order o(c(h)) (say, h t C(h)) is sufficient for convergence of the algorithm. We explore this issue in the next section. 14

6 Examples In this section we are interested in establishing explicit stopping criteria for the generic example of finite element approximation of the solution of scalar elliptic equations. Let Ω R d with boundary Γ. We will be using the following norms: ( v(x) L 2 (Ω) = v(x) 0 = v(x) 2 dx Ω v(x) L (Ω) = v(x) = ess sup v(x) x Ω v(x) H m (Ω) = v(x) m = m Ω ) 1/2 D v(x) 2 dx 1/2 where D v(x) = v(x) x α 1 1 x α d d = α 1 x 1 α d x d and α = (α 1,..., α d ) is an index of order α = α 1 +... + α d. We also need to define the space H 1 0 (Ω) H 1 with norm 6.1 Elliptic problems 0 (Ω) = { v(x) H 1 (Ω) : v(x) Γ = 0 } v(x) H 1 0 (Ω) = v(x) 1 = =1 Consider the general second-order elliptic problem Ω D v(x) 2 dx (a(x) u) + b(x) u + c(x)u = f in Ω R d (24a) 1/2 u = 0 on Γ. (24b) where the matrix a(x) is symmetric and positive definite for all x Ω, i.e., k 2 (x) ξ 2 ξ t a(x)ξ k 1 (x) ξ 2. 15

for some functions k 1 (x), k 2 (x). We also assume that the coefficients are bounded, i.e., (a) ij, (b) i, c L (Ω), i, j = 1,..., d, and that the following condition holds c(x) 1 b(x) 0 x Ω. 2 The weak formulation seeks a solution u H H 1 0(Ω) such that where a(u, v) = f(v) for all v H0 1 (Ω) (25) a(w, v) = (a w, v) + (b w, v) + (cw, v). It is straightforward to show that a(, ) satisfies the continuity and coercivity conditions (2) with respect to the H 1 0 -norm 1 with constants C 1 = k 1 L (Ω) + b L (Ω) + C(Ω) c L (Ω), C 2 = min x Ω k 2(x), where C(Ω) is a constant of order one which depends only on the domain. Let now H h H be a space of piecewise polynomials defined on a partition T h of Ω into simplices T of diameter h T. As described in section (2.2) the inclusion H h H ensures that the stability conditions (13a) and (13b) are satisfied with the constants C 1, C 2 defined above. Moreover, discretizing (25) as Au = f, the constants C 2, C 3 in (13) are given as follows. If we choose to monitor the error with respect to 1 then C 3 = k 1 L (Ω) + C(Ω) c L (Ω), C 2 = min x Ω k 2(x). However, if we work with the energy norm defined by then C 2 = C 3 = 1. We consider both cases below. 6.2 Numerical experiments w = a(w, w), (26) To illustrate the ideas presented above, we chose to perform experiments on a 2D advection-diffusion problem (c = 0). In particular, we chose to study the robustness 16

2 2 1 1 0 1 0 1 0 1 1 0 1 0 1 1 0 1 (a) ν = 1/10 (b) ν = 1/100 Figure 1: Solution of advection-diffusion problem. of our stopping criteria with respect to the nonsymmetry in the problem. Thus, we solved a test problem for constant diffusivity tensors a(x) = νi, where the diffusion parameter ν toggles the degree of nonsymmetry of the matrices involved. The test problem is thus with ν 2 u + b(x, y) u = f in Ω ( 1, 1) 2 (27a) b(x, y) = u = 0 on Γ, (27b) ( ) 2y(1 x 2 ) 2x(1 y 2 ) and right-hand side f such that the solution u is ν ) u(x, y) = (1 e(x 1)/ + e ( x 1)/ ν (1 + y 2 e(y 1)/ν + e ( 2)/ν 1 + e 2/ ν 1 e 2/ν This choice of solution tries to mimick the behaviour of problems where boundary layers are present (see Fig. 1). We first consider the errors with respect to the H 1 0 -norm. We denote by ui the linear interpolant of the solution at the mesh points. Our numerical results below 17 ).

will display the following estimators and errors: (i) : the exact relative (forward) errors u u k h 1/ u k h 1; (ii) : the exact relative (forward) interpolation errors u I u k h 1/ u k h 1; (iii) : the exact H 1 -norm criterion (17) h t C2 1 rk H 1/ u k H ; (iv) : the exact A 1 -norm criterion (18) h t C 1/2 2 r k A 1/ u k H ; (v) -est: the estimated H 1 -norm criterion (19) h t C2 1 r k λ 1/2 min (Hs k )/ uk H ; (vi) -est: the estimated A 1 -norm criterion (20) h t C 1/2 2 r k σ 1/2 min (Hs k )/ uk H ; (vii) the standard 2-norm stopping criterion r k. We chose the exponent t to be in all cases t = 1/2 and the constants c = C = 1 in (19), (20). 6.3 GMRES without preconditioning We begin with the case of a uniform partition of Ω into squares of size h and linear basis functions. The GMRES convergence curves are displayed in Fig. 2. We see that if we are interested in satisfying a tolerance with respect to the H0 1- norm, then in all cases we have to perform far fewer iterations than with a standard stopping criterion such as the relative Euclidean norm of the residual being brought below (standard threshold). Another remarkable fact is that the criterion (17) based on the dual norm of the residual is an upper bound for the interpolation error. The reason for this is not so surprising since in standard finite element calculations the interpolation error is usually smaller than the error in the energy or related norms (sometimes by a factor of h). Thus, our stopping criterion gives an upper bound on both errors so that the iterates have either achieved the final error or have achieved an error bounded from above by our dual norm estimate. We also note here that the constants c, C are indeed of order one for all values of ν. This robustness also holds with respect to the mesh parameter, although we do not include here experiments to demonstrate this fact. The same convergence curves for the case of restarted GMRES are displayed in Fig. 3. They exhibit indeed the most dramatic difference between convergence in the H 1 - or A 1 -norm and the standard 2-norm criterion. Again, the estimation of the relevant convergence curves based on the approximation (21) works extremely well. Moreover, the difference between the two stopping criteria (17), (18) is negligible. However, this may not always be the case. Indeed, the equivalence between the two norms described by Lemma 3.1 deteriorates if the H-condition number of the problem deteriorates. For our test problem this happens when ν is small and the discretization 18

est est 0 10 20 30 40 50 60 70 80 est est 0 20 40 60 80 100 120 140 (a) ν = 1 (b) ν = 1/10 est est 0 50 100 150 200 250 est est 0 50 100 150 200 250 (c) ν = 1/50 (d) ν = 1/100 Figure 2: Comparison of stopping criteria for GMRES; h = 1/32. is nonuniform. We consider this case below. For small values of ν, the problem becomes more nonsymmetric, with the matrices more nonnormal. At the same time, the finite element error on uniform meshes of squares deteriorates and even becomes of order one. One way to avoid this is to refine the mesh suitably. Given the boundary layers in the solution, we chose an exponential refinement of the meshes. In this case, the parameter h is not defined in (19), (20) we chose h := f M / f, where M = (φ i, φ j ) is the Grammian (mass) matrix with 19

respect to the L 2 (Ω)-inner product (, ). The convergence curves are displayed in Fig. 4. Again, we see that the two norm of interest are approximated well; however, the exact convergence curves in the H 1 - and A 1 -norms are not close in the initial phase of the iterative process, but become almost identical close to the convergence stage. Finally, the case when the H0 1 -norm is replaced with the energy norm (26) simplifies considerably for the advection-diffusion problem under consideration. Indeed, the est est 0 20 40 60 80 100 est est 0 50 100 150 200 250 300 350 400 (a) ν = 1 (b) ν = 1/10 est est 0 200 400 600 800 1000 1200 1400 1600 est est 0 200 400 600 800 1000 1200 1400 1600 (c) ν = 1/50 (d) ν = 1/100 Figure 3: Comparison of stopping criteria for GMRES(50); h = 1/32. 20

energy norm is nothing but a scaling by ν of the H0 1 -norm. Hence, all convergence curves displayed in the previous figures are scaled accordingly except for the relative Euclidean norm which stays the same. Thus, by varying ν, we can obtain GMRES convergence curves based on the Euclidean residual which can be either above or below the relevant convergence curve for the finite element error. For this reason we do not display all the performance again, but emphasise the fact that the standard stopping criterion used in iterative solvers has no relation to the actual convergence of quantities of interest in the original problem. 6.4 Preconditioned GMRES We turn now to the case where preconditioning is employed to speed up the iteration process. As specified in section 4, we consider only the case of right preconditioning, which has the advantage of preserving the residual, a property which enabled us to derive the stopping criterion (22). However, the use of this stopping criterion requires the estimation of the smallest singular value of our preconditioner P. In some cases this estimation can be performed cheaply, but in general it may be quite difficult to provide this information. The approximation we use is described below. All preconditioning techniques require the solution of a linear system involving the preconditioner matrix P : P z = v. est est 0 100 200 300 400 500 600 700 800 est est 0 100 200 300 400 500 600 700 800 (a) ν = (b) ν = Figure 4: Comparison of stopping criteria for GMRES exponentially-stretched mesh. 21

est 0 5 10 15 20 25 est 0 5 10 15 20 25 30 (a) ν = 1 (b) ν = 1/10 est 0 5 10 15 20 25 est 0 5 10 15 20 25 (c) ν = 1/50 (d) ν = 1/100 Figure 5: Comparison of stopping criteria for GMRES with ILU(0) preconditioning; h = 1/32. Since ( σ min (P ) = σmax(p 1 1 ) = max v n \{0} ) P 1 1 v v we choose to approximate σ min (P ) via ( ) z k 1 v k σ min (P ) max = min k v k k z k 22

where z k = P 1 v k. In the case of GMRES, the vector v k is the vector generated by the Arnoldi process, so that v k = 1. The performance of GMRES with ILU preconditioning is displayed in Fig. 5. While the number of iterations is greatly reduced, the convergence behaviour is similar to the unpreconditioned case. Moreover, the approximation of the residual A 1 -norm described above appears to work extremely well. However, in general we expect overor under-estimation to occur, in which case alternative methods for the estimation of the smallest singular value of P may have to be employed. 6.5 Three-term GMRES We end this section with numerical results obtained with the minimum residual algorithm based on a three-term recurrence described in section 4. We recall here that this is essentially the GMRES algorithm implemented in the H-norm with leftpreconditioner H. The norm of the modified residual in this method is the quantity we seek, r k H 1. The results are displayed in Fig. 6. As before, our stopping criterion (17) provides an upper bound for the convergence of quantities of interest, such as H0-norm 1 of the error or the interpolation error. More remarkable, though, is the fact that in this case the solver yields iterates whose 2-norm residual traces closely the convergence curves of interest. This is a phenomenon also noticed in the case of a similar GMRES implementation used for the solution of flow problems [10]. The same experiments were run with inexact implementation of the preconditioner H. More precisely, we solved systems with H using CG with an incomplete Cholesky preconditioner and a stopping criterion as described in Arioli [1]; the tolerance was chosen to be of order h 5/2, which for this problem is h 1/2 less than the order of the interpolation error. The results are displayed in Fig. 7. We see indeed that our criterion is an upper bound for both the finite element error u u k h 1 and the interpolation error u I u k h 1. Moreover, the inexact solves do not affect the convergence curve in the regime where it is relevant. 6.6 Other iterative methods In order to test further the relevance of the A 1 - and H 1 -norms of the residual, we ran experiments with BICGSTAB, QMR and CGS. The results for the case of discretization on uniform meshes are displayed in Figs 8, 9, 10. We again see that in all cases the two residuals provide upper bounds for the energy norm of the error 23

0 20 40 60 80 100 0 50 100 150 200 (a) ν = 1/50 (b) ν = 1/100 Figure 6: Comparison of stopping criteria for H-norm minimum residual algorithm: exact preconditioning. 0 20 40 60 80 100 120 140 0 50 100 150 200 250 (a) ν = 1/50 (b) ν = 1/100 Figure 7: Comparison of stopping criteria for H-norm minimum residual algorithm: inexact preconditioning. and interpolation error. In particular, the A 1 -norm provides again the closest approximation to the these quantities, with the H 1 -norm bound possibly deteriorating for more nonsymmetric problems, as Fig. 9 shows. As for the 2-norm of the resid- 24

0 50 100 150 200 250 300 0 100 200 300 400 500 (a) ν = 1/50 (b) ν = 1/100 Figure 8: Comparison of stopping criteria for BICGSTAB: h=1/32. 0 50 100 150 200 250 300 0 50 100 150 200 250 300 (a) ν = 1/50 (b) ν = 1/100 Figure 9: Comparison of stopping criteria for QMR: h=1/32. ual, the behaviour oscillates between the smooth, relevant convergence curve of QMR to the oscillating, large residuals exhibited by CGS. However, the issue of dynamic estimation of the A 1 - and H 1 -norms is not as straightforward as in the case of GMRES. 25

10 10 10 8 10 6 10 4 10 2 0 50 100 150 200 250 300 350 10 10 10 8 10 6 10 4 10 2 0 50 100 150 200 250 300 350 400 (a) ν = 1/50 (b) ν = 1/100 Figure 10: Comparison of stopping criteria for CGS: h=1/32. 7 Conclusion The message of this paper is simple: do not accurately compute the solution of an inaccurate problem. This was highlighted already in [1] for the case of symmetric and positive-definite problems our contribution here was the generalization to the case of nonsymmetric problems. The proposed stopping criteria require the calculation of the residual in a norm related to the problem formulation. We demonstrated that the suggested criteria are relevant to convergence in the energy-norm (or equivalent norms) while at the same time highlighting the fact that the standard criterion based on the Euclidean norm of the residual has no relevance to the quantities of interest and is in general wasteful. Further generalizations of these ideas include the case of indefinite problems and mixed finite element discretizations of systems of partial differential equations, where the use of mixed norms in which to measure convergence arises quite naturally. We hope to address some of these issues in a future paper. Acknowledgements. We thank Serge Gratton for useful discussions and comments. 8 Appendix Proof of Lemma 3.1. 26

We need to show 1 C3 r A r H 1 C2 r A and C2 C 1 C 3 r H 1 r A 1 1 C2 r H 1. The first equivalence is just a restating of the discrete stability conditions (13b), (13c). For the second we have r t A 1 r r t H 1 r σ 1 (H 1/2 A 1 H 1/2 ) = σn 1 (H 1/2 AH 1/2 ) ( x t H 1/2 AH 1/2 x min x n \{0} x t x ( y t ) 1 Ay = min y n \{0} y t Hy C 1 2 ) 1 by using (13b). Finally, since we have C 2 rt Ar r t Hr = rt H A r r t Hr C 3, r t A 1 r r t H 1 r = rt A 1 r r t H 1 A r rt 1 HA r r t H 1 r C 1 3 min r n \{0} r t A 1 r r t H 1 A r C 1 3 min y n \{0} y t à 1 y y t y, where à = I + Ñ, Ñ = H 1/2 A S A H 1/2 A. Since à (and thus à 1 ) is a normal matrix, its field of values is the convex hull of its eigenvalues ([8, p. 11]). Hence, y t à 1 y min y n \{0} y t y = min k 1 Re = min λ k (Ã) Re 1 k 1 + λ k (Ñ) = 1 λk max k (Ã) 2 and since H 1/2 AH 1/2 = A H,H 1 = C 1 (cf. (12), (14)) we get max k λ k (Ã) 1/2 H A AH 1/2 A H 1/2 AH 1/2 κ 2 (H 1/2 A H 1/2 ) and the result follows. 27

Proof of Lemma 5.1. Consider the two equivalent linear systems Ãũ = f, Âu = ˆf where à = H 1/2 AH 1/2, ũ = H 1/2 u, f = H 1/2 f, ÂH 1 A, ˆf = H 1 f. The first part of the Lemma follows from the equivalence of the Arnoldi algorithms below. Arnoldi in (, ) ṽ 1 := r 0 / r 0 do j = 1, 2,..., m h ij = (Ãṽ j, ṽ i ), 1 i j w j = Ãṽ j j h i ij ṽ i if h j+1,j = w j = 0 stop ṽ j+1 = w j / h j+1,j end do Arnoldi in (, ) H ˆv 1 := ˆr 0 / ˆr 0 H do j = 1, 2,..., m ĥ ij = (ˆv j, ˆv i ) H, 1 i j ŵ j = ˆv j j i ĥij ˆv i if ĥj+1,j = ŵ j H = 0 stop ˆv j+1 = ŵ j /ĥj+1,j end do where ṽ i = H 1/2ˆv i, w i = H 1/2 ŵ i, r 0 = H 1/2 r 0, ˆr 0 = H 1 r 0, r 0 = f Au 0 for some initial guess u 0. In particular, they yield the same Hessenberg matrices since h ij = (Ãṽ j, ṽ i ) = (H 1/2 Aˆv j, H 1/2ˆv i ) = (H 1 Aˆv j, ˆv i ) H = ĥij. For the second part, we work with the Arnoldi algorithm in the Euclidean innerproduct and system matrix à = I + N, where N = N t is skew-symmetric. For ease of presentation we drop the s. We now prove by induction on j that for all i j 2 h ij = vi(i t + N)v j = 0. We first note that (i) h ij = 1 if i = j, (ii) h ij = vi tnv j if i < j and (iii) vi tn 3 v i = 0, since N is skew-symmetric. Since w 1 = Nv 1 and w 2 = Nv 2 h 12 v 1 we have using (i)-(iii) h 13 =v t 1 Nv 3 = vt 1 Nw 2 h 32 = vt 1 N(Nv 2 h 12 v 1 ) h 32 = vt 1 N 2 v 2 h 32 = vt 1 N 3 v 1 h 32 h 21 =0 and the first inductive step holds. Assume now that for all i j 2, h ij = v t i Nv j = 0. Then (iv) w j = Nv j h j 1,j v j 1. Hence, for all i j 2, 0 = h i,i 1 v t i Nv j = v t j N t w i 1 = v t j N t Nv i 1 = v t j N 2 v i 1 28

i.e., we have (v) v t j N 2 v i = 0 for all i j 2. We now prove that h i,j+1 = 0 for all i j 1. We have using (iv) h i,j+1 = vt i Nw j = vt i N(Nv j h j 1,j v j 1 ) h j+1,j h j+1,j = vt i N 2 v j h j 1,j v t i Nv j 1 h j+1,j. If i j 3, by the inductive hypothesis v t i Nv j 1 = 0 and by (v) v t i N 2 v j = 0 and hence h i,j+1 = 0. If i = j 2 then h j 2,j+1 = 0 also since v t j 2 N 2 v j h j 1,j v t j 2 Nv j 1 = v t j 2 N 2 v j + h j 1,j h j 1,j 2 = 0 because h j 1,j = v t j 1Nv j = vt j N t w j 2 h j 1,j 2 = vt j N 2 v j 2 h j 1,j 2. Finally, if i = j 1, h j 1,j+1 = v j 1 N 2 v j /h j+1,j = 0 since v j 1 N 2 v j = 0 for all j 2. This use prove again by induction. Assuming v j 1 N 2 v j = 0, we have using (iv), (iii) v j N 2 v j+1 = v jn 2 w j h j+1,j = v jn 2 (Nv j h j 1,j v j 1 ) h j+1,j = 0. The result follows by noting that v 1 N 2 v 2 = v 1 N 2 w 1 /h 21 = v 1 N 3 v 1 /h 21 = 0. References [1] M. Arioli. A stopping criterion for the conjugate gradient algorithm in a finite element method framework. Technical Report IAN-1179, Istituto di Analisi Numerica, Pavia, 2000. [2] M. Arioli, E. Noulard, and A. Russo. Stopping criteria for iterative methods: applications to PDEs. Calcolo, 38:97 112, 2001. [3] I. Babuška. Error bounds for finite element methods. Numer. Math., 16:322 333, 1971. [4] F. Brezzi and K. J. Bathe. A discourse on the stability conditions for mixed finite element formulations. Comp. Meth. Appl. Mech. Engrg., 82:27 57, 1990. 29

[5] P. Concus and G. H. Golub. A generalized conjugate gradient method for nonsymmetric systems of linear equations. In R. Glowinski and J. L. Lions, editors, Proc. Second Internat. Symp. on Computing Methods in Applied Sciences and Engineering, volume 134 of Lecture Notes in Economics and Mathematical Systems, Berlin, 1976. Springer Verlag. [6] G. H. Golub and G. Meurant. Matrices, moments and quadrature II; how to compute the norm of the error in iterative methods. BIT, 37(3):687 705, 1997. [7] G. H. Golub and Z. Strakoš. Estimates in quadratic formulas. Numer. Algorithms, 8:2 4, 1994. [8] R. A. Horn and C. R. Johnson. Matrix Analysis. Cambridge University Press, 1985. [9] R. A. Horn and C. R. Johnson. Topics in Matrix Analysis. Cambridge University Press, 1991. [10] D. Loghin and A. J. Wathen. Analysis of block preconditioners for saddle-point problems. Technical Report 13, Oxford University Computing Laboratory, 2002. Submitted to SISC. [11] G. Meurant. Numerical experiments in computing bounds for the norm of the error in the preconditioned conjugate gradient algorithm. Numer. Algoithms, 22:353 365, 1999. [12] J. L. Rigal and J. Gaches. On the compatibility of a given solution with the data of a linear system. J. Assoc. Comput. Mach., 14(3):543 548, 1967. [13] G. Starke. Field-of-values analysis of preconditioned iterative methods for nonsymmetric elliptic problems. Numer. Math., 78:103 117, 1997. [14] W. G. Strang and G. J. Fix. An Analysis of the Finite Element Method. Prentice- Hall, 1973. [15] O. Widlund. A Lanczos method for a class of nonsymmetric systems of linear equations. SIAM J. Numer. Anal., 15(4), 1978. 30