RESIDUAL SMOOTHING AND PEAK/PLATEAU BEHAVIOR IN KRYLOV SUBSPACE METHODS

Similar documents
Comparison of Fixed Point Methods and Krylov Subspace Methods Solving Convection-Diffusion Equations

ETNA Kent State University

Recent computational developments in Krylov Subspace Methods for linear systems. Valeria Simoncini and Daniel B. Szyld

DELFT UNIVERSITY OF TECHNOLOGY

Summary of Iterative Methods for Non-symmetric Linear Equations That Are Related to the Conjugate Gradient (CG) Method

A stable variant of Simpler GMRES and GCR

Key words. linear equations, polynomial preconditioning, nonsymmetric Lanczos, BiCGStab, IDR

AMS526: Numerical Analysis I (Numerical Linear Algebra) Lecture 23: GMRES and Other Krylov Subspace Methods; Preconditioning

Contribution of Wo¹niakowski, Strako²,... The conjugate gradient method in nite precision computa

Computational Linear Algebra

A short course on: Preconditioned Krylov subspace methods. Yousef Saad University of Minnesota Dept. of Computer Science and Engineering

Conjugate gradient method. Descent method. Conjugate search direction. Conjugate Gradient Algorithm (294)

AMS526: Numerical Analysis I (Numerical Linear Algebra)

On the accuracy of saddle point solvers

On the influence of eigenvalues on Bi-CG residual norms

The amount of work to construct each new guess from the previous one should be a small multiple of the number of nonzeros in A.

Abstract. It is well-known that Bi-CG can be adapted so that hybrid methods with computational complexity

Topics. The CG Algorithm Algorithmic Options CG s Two Main Convergence Theorems

The Lanczos and conjugate gradient algorithms

On the Superlinear Convergence of MINRES. Valeria Simoncini and Daniel B. Szyld. Report January 2012

CG VERSUS MINRES: AN EMPIRICAL COMPARISON

arxiv: v2 [math.na] 1 Sep 2016

Variants of BiCGSafe method using shadow three-term recurrence

Krylov Space Solvers

ANY FINITE CONVERGENCE CURVE IS POSSIBLE IN THE INITIAL ITERATIONS OF RESTARTED FOM

Computational Linear Algebra

Updating the QR decomposition of block tridiagonal and block Hessenberg matrices generated by block Krylov space methods

ON ORTHOGONAL REDUCTION TO HESSENBERG FORM WITH SMALL BANDWIDTH

Iterative Methods for Sparse Linear Systems

Since the early 1800s, researchers have

Formulation of a Preconditioned Algorithm for the Conjugate Gradient Squared Method in Accordance with Its Logical Structure

Further experiences with GMRESR

A general Krylov method for solving symmetric systems of linear equations

Research Article Some Generalizations and Modifications of Iterative Methods for Solving Large Sparse Symmetric Indefinite Linear Systems

Residual iterative schemes for largescale linear systems

4.8 Arnoldi Iteration, Krylov Subspaces and GMRES

Iterative Methods for Linear Systems of Equations

GMRES ON (NEARLY) SINGULAR SYSTEMS

ETNA Kent State University

Solving Sparse Linear Systems: Iterative methods

Solving Sparse Linear Systems: Iterative methods

CG VERSUS MINRES: AN EMPIRICAL COMPARISON

Jordan Journal of Mathematics and Statistics (JJMS) 5(3), 2012, pp A NEW ITERATIVE METHOD FOR SOLVING LINEAR SYSTEMS OF EQUATIONS

BiCGCR2: A new extension of conjugate residual method for solving non-hermitian linear systems

A DISSERTATION. Extensions of the Conjugate Residual Method. by Tomohiro Sogabe. Presented to

PROJECTED GMRES AND ITS VARIANTS

Dedicated to Lothar Reichel on the occasion of his 60th birthday

Solving large sparse Ax = b.

On the Preconditioning of the Block Tridiagonal Linear System of Equations

Key words. conjugate gradients, normwise backward error, incremental norm estimation.

KRYLOV SUBSPACE ITERATION

Lecture 8 Fast Iterative Methods for linear systems of equations

Contents. Preface... xi. Introduction...

EECS 275 Matrix Computation

FEM and sparse linear system solving

Generalized MINRES or Generalized LSQR?

Performance Evaluation of GPBiCGSafe Method without Reverse-Ordered Recurrence for Realistic Problems

IDR(s) Master s thesis Goushani Kisoensingh. Supervisor: Gerard L.G. Sleijpen Department of Mathematics Universiteit Utrecht

ANALYSIS OF THE FINITE PRECISION BI-CONJUGATE GRADIENT ALGORITHM FOR NONSYMMETRIC LINEAR SYSTEMS

HOW TO MAKE SIMPLER GMRES AND GCR MORE STABLE

Tikhonov Regularization of Large Symmetric Problems

Preface to the Second Edition. Preface to the First Edition

Chebyshev semi-iteration in Preconditioning

CONTROLLING INNER ITERATIONS IN THE JACOBI DAVIDSON METHOD

Finite-choice algorithm optimization in Conjugate Gradients

arxiv: v2 [math.na] 1 Feb 2013

IDR(s) A family of simple and fast algorithms for solving large nonsymmetric systems of linear equations

ON THE CONNECTION BETWEEN THE CONJUGATE GRADIENT METHOD AND QUASI-NEWTON METHODS ON QUADRATIC PROBLEMS

Preconditioning for Nonsymmetry and Time-dependence

Fast iterative solvers

ITERATIVE METHODS FOR SPARSE LINEAR SYSTEMS

Krylov Space Methods. Nonstationary sounds good. Radu Trîmbiţaş ( Babeş-Bolyai University) Krylov Space Methods 1 / 17

Krylov subspace iterative methods for nonsymmetric discrete ill-posed problems in image restoration

Iterative methods for Linear System of Equations. Joint Advanced Student School (JASS-2009)

AMS Mathematics Subject Classification : 65F10,65F50. Key words and phrases: ILUS factorization, preconditioning, Schur complement, 1.

Controlling inner iterations in the Jacobi Davidson method

Iterative methods for Linear System

Applied Mathematics 205. Unit V: Eigenvalue Problems. Lecturer: Dr. David Knezevic

Reduced Synchronization Overhead on. December 3, Abstract. The standard formulation of the conjugate gradient algorithm involves

CG Type Algorithm for Indefinite Linear Systems 1. Conjugate Gradient Type Methods for Solving Symmetric, Indenite Linear Systems.

A globally and R-linearly convergent hybrid HS and PRP method and its inexact version with applications

SOLVING HERMITIAN POSITIVE DEFINITE SYSTEMS USING INDEFINITE INCOMPLETE FACTORIZATIONS

CONTROLLING INNER ITERATIONS IN THE JACOBI DAVIDSON METHOD

The solution of the discretized incompressible Navier-Stokes equations with iterative methods

On prescribing Ritz values and GMRES residual norms generated by Arnoldi processes

A QR-decomposition of block tridiagonal matrices generated by the block Lanczos process

CME342 Parallel Methods in Numerical Analysis. Matrix Computation: Iterative Methods II. Sparse Matrix-vector Multiplication.

Key words. linear equations, deflation, eigenvalues, Lanczos, conjugate gradient, QCD, multiple right-hand sides, symmetric, Hermitian

The rate of convergence of the GMRES method

using (13) (Neumaier) / using (15) / standard / using

lib- Axkl[ min lib- Axll, RESIDUAL SMOOTHING TECHNIQUES FOR ITERATIVE METHODS*

CONVERGENCE BOUNDS FOR PRECONDITIONED GMRES USING ELEMENT-BY-ELEMENT ESTIMATES OF THE FIELD OF VALUES

DELFT UNIVERSITY OF TECHNOLOGY

1 Conjugate gradients

Iterative Solvers in the Finite Element Solution of Transient Heat Conduction

Course Notes: Week 4

c 2008 Society for Industrial and Applied Mathematics

Krylov Subspace Methods that Are Based on the Minimization of the Residual

A PRECONDITIONER FOR THE HELMHOLTZ EQUATION WITH PERFECTLY MATCHED LAYER

GMRESR: A family of nested GMRES methods

Transcription:

RESIDUAL SMOOTHING AND PEAK/PLATEAU BEHAVIOR IN KRYLOV SUBSPACE METHODS HOMER F. WALKER Abstract. Recent results on residual smoothing are reviewed, and it is observed that certain of these are equivalent to results obtained by different means that relate peas and plateaus in residual norm sequences produced by certain pairs of Krylov subspace methods. Key words. residual smoothing techniques, iterative linear algebra methods, Krylov subspace methods, orthogonal and minimal residual methods, conjugate gradient and conjugate residual methods, SYMMLQ and MINRES methods, Arnoldi and GMRES methods, ORTHORES and OR- THODIR methods, biconjugate gradient (BCG) and conjugate gradient squared (CGS) methods, Bi-CGSTAB methods, quasi-minimal residual (QMR) methods AMS subject classifications. 65F0. Introduction. Let {x } be a sequence of approximate solutions of a linear system Ax b, A IR n n, and let {r b Ax } be the associated sequence of residuals. We consider the following general residual smoothing technique: (.) y 0 x 0, s 0 r 0, } y y +η (x y ) s s +η (r s ),,. The name derives from the fact that if the original residual sequence {r } is irregularly behaved, then the parameters η can be chosen to produce {y } with a more smoothly behaved residual sequence {s b Ay }. More general forms of residual smoothing are considered by Brezinsi and Redivo-Zaglia [], but (.) will suffice here. A Krylov subspace method for solving Ax b begins with some x 0 and, at the th step, determines an iterate x x 0 + z through a correction z in the th Krylov subspace K span{r 0,Ar 0,...,A r 0 }. The purpose of this brief exposition is to point out the equivalence of certain results for residual smoothing techniques(.) to observations that relate peas and plateaus in residual norm sequences produced by certain pairs of Krylov subspace methods. By a pea, we mean a mared increase in the residual norms followed by a decrease; by a plateau, we mean a period of very slow decrease or stagnation. Some results quotedherecanbephrasedmoregenerally; wehavetriedtostictothemostbasicand important cases for the sae of brevity and clarity. Also, our interest is in theoretical properties, and we do not consider issues having to do with finite-precision arithmetic. Special Issue on Iterative Methods for Linear Equations, Applied Numerical Mathematics, 9 (995), pp. 79 86. Department of Mathematics and Statistics, Utah State University, Logan, UT 843-3900, USA. This wor was supported in part by United States Air Force Office of Scientific Research Grant AFOSR-9-094, United States Department of Energy Grants DE-FG0-9ER536 and DE-FG03-94ER5, and National Science Foundation Grant DMS-94007, all with Utah State University.

. Minimal residual smoothing and pairs of orthogonal and minimal residual methods. A natural choice of η in (.) is that for which s is minimal. This η is easily characterized by (.) η arg min s +η(r s ) st (r s ) η r s. The resulting smoothing technique, which we refer to as minimal residual smoothing, was introduced by Schönauer [4] and studied extensively by Weiss [7]. Note that it clearly gives s s and s r for each. We summarize as follows: Algorithm MRS: Minimal residual smoothing [4], [7] Initialize: Set s 0 r 0 and y 0 x 0. Iterate: For,,, do: Compute x and r. Compute η s T (r s )/ r s. Set s s +η (r s ) and y y +η (x y ). When the residuals r are mutually orthogonal, i.e., when r T i r j 0 for i j, Weiss [7] has shown a number of important results for {s } generated by Algorithm MRS (see also Weiss [8]). Slightly specialized or adapted versions of those most useful here are the following: (.) s r j r i r i, (.3) s r j, (.4) s min η,,η s 0 + η i (r i s i ). i Using the orthogonality of the residuals r, one can establish (.)-(.4) by induction on. Weiss [8] has termed (.3) Kirchoff s rule because of its similarity to the rule of that name for determining resistance in parallel electrical circuits. We note for later use that (.3) is equivalent to (.5) r s ( s / s ). These results have particularly important application to pairs of orthogonal residual and minimal residual Krylov subspace methods. Recall that, at the th step, an orthogonalresidualmethodproducesaresidualr OR thatisorthogonaltok (provided It has been noted in [0] that the straightforward implementation of (.) using η defined by (.) may suffer numerical inaccuracy; however, this is not important here. See [0] for an alternative formulation that may have better numerical performance.

it is possible to do so) while a minimal residual method produces a residual r MR that has minimal norm over all corrections in K. It would be inappropriate to review all of the many implementations of these methods for special and general A, but we note several particularly well-nown orthogonal/minimal residual method pairs: the conjugate gradient (Hestenes and Stiefel [9]) and conjugate residual (Stiefel [5]) methods for symmetric positive-definite A; SYMMLQ and MINRES (Paige and Saunders []) for symmetric indefinite A; ORTHORES and ORTHODIR (Young and Jea [9]) for general A; and, also for general A, the Arnoldi method (or full orthogonalization method) (Saad []) and the generalized minimal residual (GMRES) method (Saad and Schultz [3]). Weiss [7] has shown that applying minimal residual smoothing to an orthogonal residual method results in a minimal residual method. Indeed, the residuals r OR are clearly mutually orthogonal; furthermore, the vectors rj OR rj MR, j,...,, can be seen to form a basis of A(K ) for each. Then the result follows inductively from (.4) above. Specializing (.) and (.3) to this case yields the following theorem. Theorem. ([7]). The residuals produced by orthogonal and minimal residual methods satisfy (.6) r MR r OR j ri OR ri OR, (.7) r MR rj OR. Note that (.6) expresses r MR as a convex combination of r0 OR,..., r OR, in which each ri OR is weighted in proportion to the reciprocal of the square of its norm. Furthermore, rearranging (.7) in the manner of (.5) gives the following corollary. Corollary.. The norms of the residuals produced by orthogonal and minimal residual methods satisfy (.8) r OR r MR ( r MR / r MR ). This corollary is an extension to general orthogonal/minimal residual methods of a result of Brown [, Th. 5.] relating the performance of the Arnoldi and GMRES methods. The result of [, Th. 5.] is given in a somewhat different form but has been rephrased in essentially the form (.8) by Cullum and Greenbaum [4, Th. 3]. This result provides the basis of observations by Brown [] and Cullum and Greenbaum [4], [5] that, taen together, in effect correlate peas in { r OR } with plateaus in { r MR }. Equations (.7) and (.8) are equivalent but give somewhat different perspectives on how { r MR } and { r OR } are related. Equation (.7) maes plain the global dependenceof r MR on r0 OR,..., r OR ; notethat{ r MR }issmallifandonly if at least one of r0 OR,..., r OR is small. Equation (.8) brings out the local dependence of r OR on r MR and r MR. We always have r OR r MR, of course, and (.8) further shows that equality holds if and only if r OR r MR 0, i.e., both methods have reached the solution. Before the solution is reached, we 3

have from (.8) that the factor by which r OR exceeds r MR is determined by r MR / r MR, which measures the progress made by the minimal residual method at the current step. In particular, (.8) shows clearly that, as observed by Brown [] and Cullum and Greenbaum [4] for the Arnoldi and GMRES methods, if one method performs either very well or very poorly, then so does the other. 3. Quasi-minimal residual smoothing and certain pairs of Lanczosbased methods. Another ind of smoothing of the form (.), called quasi-minimal residual smoothing, has been considered by Zhou and Waler [0]. The name derives from the fact that when this smoothing is applied to the residuals and iterates produced by the biconjugate gradient (BCG) method (Lanczos [0], Fletcher [6]), then the resulting smoothed residuals and iterates are just those of the quasi-minimal residual (QMR) method of Freund and Nachtigal [8]; see [0, 3]. We will discuss this relationship between BCG and QMR in more depth after formulating quasi-minimal residual smoothing and reviewing some of its general properties. The general formulation is as follows 3 : Algorithm QMRS: Quasi-minimal residual smoothing [0] Initialize: Set s 0 r 0, y 0 x 0, and τ 0 r 0. Iterate: For,,, do: Compute x and r. Define τ by /τ /τ +/ r. Set s s + τ r (r s ) and y y + τ r (x y ). The following results for {s } and {τ } produced by Algorithm QMRS are easily shown by induction ([0, 3]): (3.) s r j r i r i, (3.) τ r j. Note that (3.) is the same as (.), although now the residuals r are not mutually orthogonal, and that (3.) is an analogue of (.3), in which τ plays the role of a quasi-residual norm. Also, (3.) is equivalent to (3.3) r τ (τ /τ ), which is an analogue of (.5). Here, we mean the basic QMR method without loo-ahead. 3 It is noted in [0] that this formulation may have numerical weanesses similar to those of Algorithm MRS. See [0] for an alternative formulation that avoids these weanesses. 4

From (3.) and (3.), one obtains ([0, 3]) s (3.4) +τ r BCG j + r j Thus s is bounded by the square root of the harmonic mean of r 0,..., r. To specialize these results to BCG and QMR, let {r BCG } and {r QMR } denote the residual sequences generated by those methods and suppose that Algorithm QMRS is applied with {r } {r BCG }. By the results of Zhou and Waler [0, 3] cited earlier, we have {s } {r QMR }. It is further shown in [0, 3] that (3.5) τ min r 0 e H y, y IR in which e (,0,...,0) T IR + and H IR (+) is the tridiagonal matrix satisfying AV V + H, where, for each, V IR n is the matrix the columns of which are the first normalized Lanczos vectors. Then (3.4) becomes (3.6) r QMR +τ +. r BCG j for τ given by (3.5). The bound r QMR +τ has been given by Freund and Nachtigal [8, Prop. 4.] and is used as a preliminary termination criterion in QMR. From (3.), (3.), and (3.5), we have the following analogue of Theorem.. Theorem 3. ([0]). The residuals produced by BCG and QMR satisfy (3.7) r QMR r BCG j ri BCG ri BCG, (3.8) τ rj BCG, where τ satisfies (3.5). In analogy with (.6), (3.7) shows that r QMR is a convex combination of r0 BCG,..., r BCG, in which each ri BCG is weighted in proportion to the reciprocal of the square of its norm. This gives considerable insight into why the QMR residual norms tend to decrease fairly smoothly, if not monotonically. Indeed, if r BCG is small for some, then r BCG is given large weight in (3.7) and r QMR is small. If subsequent BCG residual norms increase, then the effect of this increase is mollified by relatively small weights in (3.7) and any increase in r QMR is correspondingly small. Rearranging (3.8) in the manner of (3.3) gives the following corollary: Corollary 3.. The norms of the BCG residuals satisfy (3.9) r BCG τ, (τ /τ ) where τ satisfies (3.5). 5

This result has been given by Cullum and Greenbaum [4, Th. 6] using closely related results of Freund and Nachtigal [8]. It is used in [4] to correlate peas in { r BCG } with plateaus in the quasi-residual norm sequence {τ }. Although the quasi-residual norms are not the true QMR residual norms (we have only the bound (3.6) in general), it has been experimentally observed by Cullum and Greenbaum [4] that the quasi-residual norms tend to be good approximations of the true QMR residual norms. Thus (3.8) and (3.9) indicate approximate relationships (3.0) (3.) r QMR r BCG ( rj BCG, r QMR r QMR / r QMR ) that suggest correlations of peas in { r BCG } with plateaus in { r QMR }. Such correlations are clearly seen in experiments by Cullum and Greenbaum [4] and Zhou and Waler [0]. Equations (3.8)-(3.9) and approximations (3.0)-(3.) give perspectives on the relationship of { r BCG } to {τ } and { r BCG } similar to those given by (.7)-(.8) on the relationship of { r OR } to { r MR }; see the last paragraph of. Algorithm QMRS can be used to derive a QMR-type method from any given method. Subsequently, (3.)-(3.3) and the bound (3.4) can be used to draw relationships between the residuals of the given method and the residuals and quasi-residual norms of the resulting QMR-type method. In particular, these relationships can be used to correlate peas in the residual norms of the given method with plateaus in the residual and quasi-residual norms of the QMR-type method as in the case of BCG and QMR above. We mention two examples from Zhou and Waler [0, 3] to which these observations are relevant. In these examples, application of Algorithm QMRS to certain methods results in QMR-type methods that have been introduced elsewhere, and the above observations apply. First, the TFQMR method of Freund [7] can be obtained by applying Algorithm QMRS to the residuals, iterates, and certain intermediate quantitites generated by CGS. Second, the QMRCGSTAB method of Chan et al. [3] can be similarly obtained by applying Algorithm QMRS to the residuals, iterates, and certain other quantitites generated by the Bi-CGSTAB method of Van der Vorst [6]. 4. Minimal residual smoothing for non-orthogonal residuals. We conclude by briefly revisiting minimal residual smoothing. Weiss [8, Th. ] has observed that for arbitrary (not necessarily mutually orthogonal) {r }, the bound (3.4), with the quasi-residual norm τ given by (3.), holds for {s } generated by Algorithm MRS as well as by Algorithm QMRS. (This springs directly from the fact that s generated by Algorithm MRS satisfies s min,..., r i ; see [8, Th. ].) Furthermore, since the quasi-residual norms satisfy (3.3), we have the same correlation of peas in { r } with plateaus in {τ } as in the case of Algorithm QMRS, which suggests a correlation of peas in { r } with plateaus in { s } as in that case. In fact, it is seen in a number of experiments reported by Zhou and Waler [0] that Algorithms MRS and QMRS give fairly similar results when applied to the residuals and iterates produced by BCG and CGS. In particular, peas in { r } are strongly correlated 6

with plateaus in { s } when either Algorithm MRS or Algorithm QMRS is applied to BCG or CGS. REFERENCES [] C. Brezinsi and M. Redivo-Zaglia, Hybrid procedures for solving systems of linear equations, Numer. Math., 67 (994), pp. 9. [] P. N. Brown, A theoretical comparison of the Arnoldi and GMRES algorithms, SIAM J. Sci. Stat. Comput., 0 (99), pp. 58 78. [3] T. F. Chan, E. Gallopoulos, V. Simoncini, T. Szeto, and C. H. Tong, A quasi-minimal residual variant of the Bi-CGSTAB algorithm for nonsymmetric systems, SIAM J. Sci. Comput., 5 (994), pp. 338 347. [4] J. K. Cullum and A. Greenbaum, Residual relationships within three pairs of iterative algorithms for solving Ax b, Tech. Report RC 867, IBM T. J. Watson Research Center, Yortown Heights, New Yor, January 993. To appear in SIAM J. Matrix Anal. Appl. [5], Peas, plateaus, numerical instabilities in Galerin and minimal residual pairs of methods for solving Ax b. preprint, 994. [6] R. Fletcher, Conjugate gradient methods for indefinite systems, in Numerical Analysis Dundee 975, Lecture Notes in Mathematics 506, Springer-Verlag, Berlin, Heidelberg, New Yor, 976, pp. 73 89. [7] R. W. Freund, A transpose-free quasi-minimal residual algorithm for non-hermitian linear systems, SIAM J. Sci. Comput., 4 (993), pp. 470 48. [8] R. W. Freund and N. M. Nachtigal, QMR: a quasi-minimal residual method for non- Hermitian linear systems, Numerische Mathemati, 60 (99), pp. 35 339. [9] M. Hestenes and E. Stiefel, Methods of conjugate gradients for solving linear systems, J. Res. National Bureau of Standards, 49 (95), pp. 409 436. [0] C. Lanczos, Solution of systems of linear equations by minimized iterations, J. Res. Nat. Bur. Stand., 49 (95), pp. 33 53. [] C. C. Paige and M. A. Saunders, Solution of sparse indefinite systems of linear equations, SIAM J. Numer. Anal., (975), pp. 67 64. [] Y. Saad, Krylov subspace methods for solving large unsymmetric linear systems, Math. Comp., 37 (98), pp. 05 6. [3] Y. Saad and M. H. Schultz, GMRES: a generalized minimal residual method for solving nonsymmetric linear systems, SIAM J. Sci. Stat. Comput., 7 (986), pp. 856 869. [4] W. Schönauer, Scientific Computing on Vector Computers, North Holland, Amsterdam, New Yor, Oxford, Toyo, 987. [5] E. Stiefel, Relaxationsmethoden bester Strategie zur Lösung linearer Gleichungssysteme, Comm. Math. Helv., 9 (955), pp. 57 79. [6] H. A. van der Vorst, Bi-CGSTAB: a fast and smoothly converging variant of Bi-CG for the solution of nonsymmetric linear systems, SIAM J. Sci. Stat. Comput., 3 (99), pp. 63 644. [7] R. Weiss, Convergence Behavior of Generalized Conjugate Gradient Methods, PhD thesis, University of Karlsruhe, 990. [8], Relations between smoothing and QMR, tech. report, Universität Karlsruhe Numeriforschung für Supercomputer, 993. Submitted to Numerische Mathemati. [9] D. M. Young and K. C. Jea, Generalized conjugate gradient acceleration of nonsymmetrizable iterative methods, Lin. Alg. Appl., 34 (980), pp. 59 94. [0] L. Zhou and H. F. Waler, Residual smoothing techniques for iterative methods, SIAM J. Sci. Comput., 5 (994), pp. 97 3. 7