Non-stationary extremal eigenvalue approximations in iterative solutions of linear systems and estimators for relative error

Similar documents
AN ITERATIVE METHOD WITH ERROR ESTIMATORS

The Lanczos and conjugate gradient algorithms

Key words. conjugate gradients, normwise backward error, incremental norm estimation.

Introduction. Chapter One

ESTIMATES OF THE TRACE OF THE INVERSE OF A SYMMETRIC MATRIX USING THE MODIFIED CHEBYSHEV ALGORITHM

Iterative Methods for Sparse Linear Systems

4.8 Arnoldi Iteration, Krylov Subspaces and GMRES

Tikhonov Regularization of Large Symmetric Problems

Sensitivity of Gauss-Christoffel quadrature and sensitivity of Jacobi matrices to small changes of spectral data

Preliminary/Qualifying Exam in Numerical Analysis (Math 502a) Spring 2012

AMS526: Numerical Analysis I (Numerical Linear Algebra)

Contribution of Wo¹niakowski, Strako²,... The conjugate gradient method in nite precision computa

Lecture 9: Krylov Subspace Methods. 2 Derivation of the Conjugate Gradient Algorithm

The Conjugate Gradient Method

Solving discrete ill posed problems with Tikhonov regularization and generalized cross validation

Matrices, Moments and Quadrature, cont d

Golub-Kahan iterative bidiagonalization and determining the noise level in the data

FEM and sparse linear system solving

ON THE QR ITERATIONS OF REAL MATRICES

Course Notes: Week 1

Orthogonal polynomials

Total least squares. Gérard MEURANT. October, 2008

1 Conjugate gradients

Applied Mathematics 205. Unit V: Eigenvalue Problems. Lecturer: Dr. David Knezevic

STEEPEST DESCENT AND CONJUGATE GRADIENT METHODS WITH VARIABLE PRECONDITIONING

DELFT UNIVERSITY OF TECHNOLOGY

Iterative methods for Linear System

Krylov Subspace Methods that Are Based on the Minimization of the Residual

Iterative Methods for Solving A x = b

Krylov Space Methods. Nonstationary sounds good. Radu Trîmbiţaş ( Babeş-Bolyai University) Krylov Space Methods 1 / 17

Introduction to Iterative Solvers of Linear Systems

AMS526: Numerical Analysis I (Numerical Linear Algebra for Computational and Data Sciences)

Implicitly Defined High-Order Operator Splittings for Parabolic and Hyperbolic Variable-Coefficient PDE Using Modified Moments

Last Time. Social Network Graphs Betweenness. Graph Laplacian. Girvan-Newman Algorithm. Spectral Bisection

Math Introduction to Numerical Analysis - Class Notes. Fernando Guevara Vasquez. Version Date: January 17, 2012.

Conjugate Gradient Method

Numerical Methods in Matrix Computations

Conjugate gradient method. Descent method. Conjugate search direction. Conjugate Gradient Algorithm (294)

Matrices and Moments: Least Squares Problems

Chapter 7 Iterative Techniques in Matrix Algebra

Iterative solution methods and their rate of convergence

Topics. The CG Algorithm Algorithmic Options CG s Two Main Convergence Theorems

22.4. Numerical Determination of Eigenvalues and Eigenvectors. Introduction. Prerequisites. Learning Outcomes

Notes on PCG for Sparse Linear Systems

Lecture 17 Methods for System of Linear Equations: Part 2. Songting Luo. Department of Mathematics Iowa State University

Block Bidiagonal Decomposition and Least Squares Problems

Some minimization problems

Lecture 18 Classical Iterative Methods

Summary of Iterative Methods for Non-symmetric Linear Equations That Are Related to the Conjugate Gradient (CG) Method

Greedy Tikhonov regularization for large linear ill-posed problems

Composite convergence bounds based on Chebyshev polynomials and finite precision conjugate gradient computations

Positive Denite Matrix. Ya Yan Lu 1. Department of Mathematics. City University of Hong Kong. Kowloon, Hong Kong. Abstract

AMS526: Numerical Analysis I (Numerical Linear Algebra) Lecture 23: GMRES and Other Krylov Subspace Methods; Preconditioning

A Method for Constructing Diagonally Dominant Preconditioners based on Jacobi Rotations

Iterative methods for Linear System of Equations. Joint Advanced Student School (JASS-2009)

Numerical Methods - Numerical Linear Algebra

Moments, Model Reduction and Nonlinearity in Solving Linear Algebraic Problems

Numerical Methods I Non-Square and Sparse Linear Systems

A Spectral Time-Domain Method for Computational Electrodynamics

ON ORTHOGONAL REDUCTION TO HESSENBERG FORM WITH SMALL BANDWIDTH

A fast randomized algorithm for overdetermined linear least-squares regression

Linear Solvers. Andrew Hazel

The structure of iterative methods for symmetric linear discrete ill-posed problems

Notes on Some Methods for Solving Linear Systems

MAT 585: Johnson-Lindenstrauss, Group testing, and Compressed Sensing

On the eigenvalues of specially low-rank perturbed matrices

1 Extrapolation: A Hint of Things to Come

Lecture 11: CMSC 878R/AMSC698R. Iterative Methods An introduction. Outline. Inverse, LU decomposition, Cholesky, SVD, etc.

Algebra C Numerical Linear Algebra Sample Exam Problems

CME342 Parallel Methods in Numerical Analysis. Matrix Computation: Iterative Methods II. Sparse Matrix-vector Multiplication.

EECS 275 Matrix Computation

15 Singular Value Decomposition

Krylov Subspace Spectral Methods for the Time-Dependent Schrödinger Equation with Non-Smooth Potentials

Lanczos tridigonalization and Golub - Kahan bidiagonalization: Ideas, connections and impact

Preconditioned conjugate gradient algorithms with column scaling

Roundoff Error. Monday, August 29, 11

SOLVING SPARSE LINEAR SYSTEMS OF EQUATIONS. Chao Yang Computational Research Division Lawrence Berkeley National Laboratory Berkeley, CA, USA

Class notes: Approximation

Large-scale eigenvalue problems

Lecture Note 7: Iterative methods for solving linear systems. Xiaoqun Zhang Shanghai Jiao Tong University

REORTHOGONALIZATION FOR GOLUB KAHAN LANCZOS BIDIAGONAL REDUCTION: PART II SINGULAR VECTORS

PETROV-GALERKIN METHODS

Efficient smoothers for all-at-once multigrid methods for Poisson and Stokes control problems

APPLIED NUMERICAL LINEAR ALGEBRA

In this section again we shall assume that the matrix A is m m, real and symmetric.

Low Rank Approximation Lecture 7. Daniel Kressner Chair for Numerical Algorithms and HPC Institute of Mathematics, EPFL

The restarted QR-algorithm for eigenvalue computation of structured matrices

Iterative Algorithm for Computing the Eigenvalues

The Nearest Doubly Stochastic Matrix to a Real Matrix with the same First Moment

6.4 Krylov Subspaces and Conjugate Gradients

Part 3: Trust-region methods for unconstrained optimization. Nick Gould (RAL)

c 2005 Society for Industrial and Applied Mathematics

Performance Evaluation of GPBiCGSafe Method without Reverse-Ordered Recurrence for Realistic Problems

Section 4.5 Eigenvalues of Symmetric Tridiagonal Matrices

MAA507, Power method, QR-method and sparse matrix representation.

Algorithms that use the Arnoldi Basis

Eigenvalue and Eigenvector Problems

Iterative Methods. Splitting Methods

Some definitions. Math 1080: Numerical Linear Algebra Chapter 5, Solving Ax = b by Optimization. A-inner product. Important facts

Lecture 9 Approximations of Laplace s Equation, Finite Element Method. Mathématiques appliquées (MATH0504-1) B. Dewals, C.

Transcription:

on-stationary extremal eigenvalue approximations in iterative solutions of linear systems and estimators for relative error Divya Anand Subba and Murugesan Venatapathi* Supercomputer Education and Research Center, Indian Institute of Science, Bangalore 5600, India * Corresponding author: murugesh@serc.iisc.ernet.in Abstract on-stationary approximations of the final value of a converging sequence are discussed, and we show that extremal eigenvalues can be reasonably estimated from the CG iterates without much computation at all. We introduce estimators of relative error for conjugate gradient (CG) methods that adopt past wor on computationally efficient bounds of the absolute errors using quadrature formulas. he evaluation of the Gauss quadrature based estimates though, depends on a priori nowledge of extremal eigenvalues; and the upper bounds in particular that are useful as a stopping criterion fail in the absence of a reasonable underestimate of smallest eigenvalue. Estimators for relative errors in A-norm and their extension to errors in l norm are presented with numerical results. Estimating the relative error from the residue in an iterative solution is required for efficient solution of a large problem with even a moderately high condition. Specifically, in a problem of solving for vector x in Ax=b, the uncertainty between the strict upper bound in relative error [κ r i / b ] and its strict lower bound [ r i /(κ b )] is a factor of κ (given residue r i = b-ax i is the residual vector at i th iteration and κ the condition number of the square matrix A). Keywords: linear systems, extremal eigenvalues, iterative methods, error estimation.. Introduction In numerical models of linear systems, the stopping criterion of an iterative solution is typically based on the relative residue at an iteration; for the problem of solving for vector x in Ax=b, this is given by r i / b where r i = b-ax i is the residual vector at i th iteration. But the error in x is the indicator of the convergence behavior and the accuracy of the solution obtained; the relative error in x preferred typically when the unnown x varies significantly by the problem. Specifically for a square matrix A, it can be easily shown that the relative error is bounded such

that κ A r i / b x-x i / x r i /(κ A b ) where κ A is the condition number of A. If one defines a condition of the specific problem κ (A,b) (i.e.) given a matrix A and the vector b, then x-x i / x κ (A,b) r i / b as x i x. Computation of the condition numbers κ A or κ (A,b) are arithmetic operations of O( 3 ) for a general matrix and thus prohibitive when the dimension of matrix is large [-3]; also the above loose bounds of the relative error are not useful as an efficient stopping criterion even if the condition of the matrix κ A is nown, but κ A >>. As an example, figure () shows the relative residue and the actual relative errors for two different vectors b and a symmetric positive definite (SPD) matrix A using the conjugate gradient algorithm. In one case, the relative residue is a few orders of magnitude smaller than the relative error, and alternately larger in the other case. A stopping criteria based on the residue alone can thus be optimistic and erroneous, or conservative and inefficient, depending on the condition of the specific problem κ (A,b), whether pre-conditioned or otherwise. he larger the problem and its condition number, more useful is the estimate of the relative error in enforcing a stopping criterion that is both cost effective but satisfying the required accuracy of solution. Here we extend the wor on estimates of absolute error in CG iterates using quadrature rules [4, 5] that demand negligible computation, to include relative errors, specifically to the realistic scenario where the extremal eigenvalues are unnown. An efficient non-stationary estimation of the extremal eigenvalues in O() operations at every iteration of the CG algorithm is suggested; here we use the iterative algorithm solving the problem to efficiently evaluate extremal eigenvalues as well. Other methods of estimating extremal eigenvalues lie probabilistic bounds using a Krylov subspace of vector b [6] are either dependent on vector b or require arithmetic operations >O( ) on their own. Also there exist anti-gauss quadrature based methods for estimation of errors that do not need the smallest eigenvalue, but negligible computation and upper bounds are not guaranteed [7, 8]. he property of A-orthogonality of the CG increments is used to extend absolute error estimates to relative errors and numerical estimates of the relative error in A-norm are shown. he above method is extended to estimating the relative errors in l norm; this is significant for applications where the A-norm/energy-norm is not an appropriate criterion for stopping.

0 0 0 relative residue () relative error () relative residue () relative error () 0 - Error / Residue 0-4 0-6 0-8 0-0 0 50 00 50 00 50 300 umber of iterations Figure : Relative errors ( x-x i A / x A ) and residues ( r i A / b A ) for two different vectors b and a SPD matrix A (bcsst05; math.nist.gov/matrixmaret). Case : vector b is a sum of canonical basis (all s) for generality and Case : vector b is eigenvector of smallest eigenvalue with a small (< 0.0) random perturbation of its values; κ A = 48, =53.. Mathematical preliminaries o use quadrature formulas in estimation of the absolute error, the quadratic form of the square of error in A-norm, x-x A = r A - r, can be represented by a Riemann Stieltjes integral as below. 3

when AA ε A = A A = IAr [, ] = ra r= rqλ Qr n b = q Λ q= λ q = λ dα( λ) where α is piecewise costant and defined i= i i a αλ ( ) = 0 if λ a = λ Eq. () i = q if λ λ < λ j= n j= j i i+ = q if λ b= λ j n Here a, b (unlie vector b) are the smallest and largest eigenvalues of A respectively. his allows us to use the Gauss, Gauss-Radau and Gauss-Lobatto formulas for a function f given by b M a f( λ) dα( λ) = w f( t ) + v f( z ) + R[ f] where i i j j i= j= ( + M) b M f ( η) R[ f] = ( λ zj)[ ( λ ti)] dα( λ), a< η < b ( + M)! a j= i= Eq. () he above approximations of the integral when M=0 is called the Gauss formula, and when M=, Gauss-Radau and Gauss-Lobatto respectively. When M=, z =a or z =b (for upper or lower bounds respectively) and when M=, z =a, z =b (for an upper bound). he nodes t and z can be obtained by a polynomial decomposition of the integral in terms of p i (λ); moreover a set of orthogonal polynomials provides a 3-term recursion relationship for easy evaluations. his means the recurrence coefficients can be represented in a matrix of symmetric tri-diagonal form as in Eq. (3); the crucial observation being that these can be trivially extracted from the CG iterates, resulting in negligible addition of computation over the iterative solution. In more generality, the CG algorithm can be described as a minimization of the polynomial relation x x = min p ( A )( x x ). A p o A 4

b { 0 when i j Given pi( λ) pj( λ) dα( λ) = { when i= j a γ p ( λ) = ( λ ω ) p ( λ) + γ p ( λ), i=,,... when normalized such that i i i i i i dα =, p ( λ) =, p ( λ) = 0. λp ( λ) = C P ( λ) + γ p ( λ) e, where e = [ 00,,... ], P ( λ) = [ p ( λ) p ( λ)... p ( λ)] C 0 ω γ 0 0 γ ω. 0 = 0.. γ 0 0 γ ω Eq. (3) hus the eigenvalues of C are the nodes of the Gauss formula, and the C R, C L (for Radau and Lobatto formulas) can be computed from C as shown elsewhere [5, 9, 0]. he above coefficients can be computed from the CG coefficients α and β as below. β β ω = + and γ = where β0 = 0, α =. α α α where α r Ar =, x = x + α d d Ad r r r = r α Ad, β =, d = r + β d r r Eq. (4) It is efficient to compute the A-norm of the error using an approximation for C - (,) as shown before [9]. A o x x r [ C = (, ) C (, )] Eq. (5) his algorithm [5] depends on the estimates of the extremal eigenvalues, and the discussed results have been limited to absolute errors and a priori nown extremal eigenvalues (a= λ min,b= λ max ). We find that the upper bounds estimated to be moderately sensitive to approximations of a (shown in figure ()); the lower bounds and the Gauss rule estimates are not useful as stopping criterion in generality, and so are not discussed further in this wor. In fact, even a small overestimation of the smallest eigenvalue can mae the upper bounds lower than the actual error (or result in complex valued estimated norms), thus failing, while any large 5

underestimation maes the bounds too conservative to be helpful. hough figure () shows the sensitivity of the upper bound for a vector b given as a sum of canonical basis for generality, the sensitivity to the estimate of smallest eigenvalue in specific cases can be far greater than shown there. 0 0 a = 0.3λ min a = 0.46λ min a = 0.9λ min a = λ min Relative error a =.05λ min 0 - Error Bounds 0-0 -3 0 40 60 80 00 0 40 60 umbers of itertions Figure : Sensitivity of the Gauss-Lobatto upper bounds of error (in A-norm) on the smallest eigenvalue a of SPD matrix A (bcsst05; math.nist.gov/matrixmaret). Failure of the estimated bound is apparent when a > λ min. Vector b is a sum of canonical basis (all s) to maintain generality and κ A = 48, =53. 3. Methods We begin with showing that the eigenvalues of the tri-diagonal matrix J constructed using CG iterates converge to those of A as. Later we show that non-stationary convergence estimations using the smallest eigenvalue of J can lower bind the smallest eigenvalue a closely even if <<; in fact with negligible computation; moreover, J can be stored efficiently as a sparse matrix. One could include inverse iterations for the tri-diagonal J at every CG iteration, nevertheless an approximation of a even in the same order of magnitude may mean one has to wait until ~ for many matrices with large condition numbers. We suggest efficient non-stationary approximations of a, b using the extremal eigenvalues of J even as <<. We also extend this method to estimation of the l norm of the errors and the errors in 6

relative terms. It is nown that all symmetric matrices have a tri-diagonal symmetric form and this can be shown using orthogonal transformations. heorem: Any symmetric matrix A has an orthogonal transformation to a tri-diagonal symmetric matrix J Proof: Let Q a be the orthogonal projection of a the first column of A such that a Q e i = 0 if i > where e i denote the canonical basis vectors. Here orthogonality of Q implies a i Q Q a j = a i a j for i, j=, where is the dimension of A. Q can be readily obtained by a Gram-Shmidt orthogonalization of a with all e i for i >, or a Householder s reflection of a over e, e. Let a be the second column of the symmetric matrix (Q A Q ) and Q 3 a be the orthogonal projection of a such that a Q 3 e i = 0 if i >3 (i=.). Continuing further in the same fashion to the orthogonal projection Q - a - any symmetric matrix A can be transformed into a symmetric tri-diagonal matrix J = Q AQ = Q - Q 3 Q A Q Q 3 Q -. he above can also be shown using orthogonal transformations of the matrix A. Let A =Q A - Q be an orthogonal transformation such that A j, the j th column of A satisfies e i A j=0 for all j and all i > j+. he resulting A n- =J is the tri-diagonal matrix when A 0 =A is the given symmetric matrix and Q are the Householder s reflectors where I 0 uu Q = and Hn In, u A' e A', H = = 0 n uu Eq. (6) where A = [ A, A,... A ] and note H = I. ' +, +, n, Corollary: Given (J-μ i I)q i = (Q AQ-μ i I)q i =0 and (A-λ i I)p i = 0 where μ i, q i are eigenvalues and vectors of J; λ i, p i are eigenvalues and vectors of A. hen it is evident that μ i =λ i and also, q i = Q p i when λ i 0. Remar: In more generality for a matrix A, if A A = AA, an eigenvalue revealing diagonal transformation A = QΛQ exists and that is not proved here. he above orthogonal transformation of matrix A into a tri-diagonal form is typically achieved by the Lanczos algorithm; it has a three-term recursion representation (lie orthogonal polynomials) resulting in easier computation of a matrix V including the first Lanczos vectors and transformed matrix J. 7

Moreover, Lanczos description helps in highlighting the equivalence of C derived from the CG iterates as in Eq. (3), and the orthogonal transformations A as below. Given A ω γ 0 0 J γ ω. 0 = and J = 0 A 0 n 0.. γ 0 0 γ ω V AV = J and AV = V J + γ v e + + J AV = V+ J ' where J ' = γ + e Eq. (7) Also nown J λp( λ) = P+ ( λ) J ' where J ' = γ + e where P ( λ) = [ p ( λ) p ( λ)... p ( λ)] are a set of orthogonal polynomials such that β β ω = + and γ = where β0 = 0, α =. α α α and αβ, are CG iterates as in Eq.(4) Eq. (8) hus when we use CG, we have all the elements of J, a partial orthogonal transformation of A at the th iteration. he largest eigenvalue of J is nown to converge to that of A quicly, but it is the smallest eigenvalue that may need almost iterations to converge completely. he largest eigenvalue of J is a very good approximation of the required largest eigenvalue b even as <<, and so we will not discuss it further. In general, most iterative solutions of linear systems can be considered effective only when the number of iterations taen to solve a large problem is much less than. We compute the extremal eigenvalues of J and subsequently underestimate the smallest eigenvalue of A even as << using a non-stationary function, with the nowledge that the extremal eigenvalues of J converge to that of A with increasing. We can evaluate the largest and smallest eigenvalues of J in < O( ) operations using methods (available in LAPACK libraries) that are not discussed here. If f is the smallest eigenvalue of J, it is guaranteed that 0 < f + f where f + is the smallest eigenvalue of J +, given the matrix A is positive definite. Any other correlation between f and f + is not nown to satisfy the generality of spectral properties required of any symmetric 8

positive definite matrix A. But it is observed for all matrices studied that the finite difference of the sequence of the smallest eigenvalues of J is in fact nominally decrease with (as defined by β < in the statement below). Moreover, any sequence with a nominally decreasing finite difference can be represented by a sequence with piecewise strictly decreasing finite differences. Hence for sequences with decreasing finite differences, we can find an appropriate non-stationary function with some rigor that is computationally trivial to update. We need a loo-ahead underestimate of smallest eigenvalue of A that modulates f by a factor between and 0, depending on the convergence rate of f. he exponential function (with exponent 0) is then a very suitable loo-ahead to lower bind the smallest eigenvalue of A using the eigenvalue convergence of J. (ote: In the case of a continuous function or a discrete sequence, unless f ( s) or f ( z) the Laplace transform of the function or the Z-transform of the sequence can be well approximated in the limits s 0, z, the final value theorem is not useful). Statement: Let f be any real valued sequence of numbers f f > 0 for = and f f f f + = β < (defined for all f - f 0 in the sequence). f is a sequence of positive real numbers for integers = such that both its value and the nominal finite difference of the sequence are not increasing with. Given f = a, find the positive valued function g in f e g ( fi,, ) = ã where i= < such that () ã a and () = = ( a ã ) ( f a) = δ << For any arbitrary sequence f, simultaneous enforcement of both conditions (, ) can not be guaranteed for all strictly; then minimizing δ in () while satisfying () for almost all is a reasonable objective for the required lower bound estimate of f = a. Proposition: f f+ Let f f + > 0 and f f for =. 9

hen a function g ( fi,, ) fe = f exp( α [ ]) = a ensures there exists α such that = = ( a ã ) ( f a) = δ << when f = a. Proof: We strictly enforce f on the final value of the sequence approximated ã = f e g ( fi,, ) for any sequence f such that ã = f = afor =. his implies g = 0 and this is easily done for all f when g( f,, ) = α ( f) F (, ) where i=, a possible separation of variables for i i function g, and F(, ) d [ ] to the first order. he choice for the multiplicative factor d can either be a constant such as -, a variable such as -, or even a random variable. If F(, ) = [ ]/, ã is a very sensitive to α when <<, but becomes stiff as /, and on the other hand if F(, ) = [ ] / it is relatively stiff throughout; we present results of both cases. hen it is evident that when a = f exp( α [ ] / ), for any f, f α = ln( ) and a = a δ = 0. f i But to find α given only f, i <, we can minimize the error between the nonstationary estimates of the current value using the previous values of the sequence, given by ( f exp( α [ i] / ) f ). When a strictly decreasing finite difference of the sequence is not i guaranteed and it has decreasing finite differences only piecewise, simultaneous enforcement of both conditions, of statement () are not possible for all strictly. his is because then any estimate of α will not monotonically reduce with, and a bound on δ is difficult for such an arbitrary sequence. In such cases, an estimate of ã can be highly oscillatory because of the large - that is to be looed ahead when <<<. he logarithmic scaling in the regression, (i.e.) the approximations of ln( f / f ) for all i such that i+m>, can not only reduce large such variations in i ã but result in a much cheaper computation as Eq. (9) in an estimation α for the underestimate of f = a. he computational cost of the regression in evaluating α is only O(m) where the last m points in the sequence are used to minimize the error in exponential fits of sequence f. his helps in limiting the total computation of the proposed error estimators to much less than an iteration of hosting CG algorithm which is O( ), and thus a recursive estimate of 0

extremal eigenvalues can be evaluated in all iterations of CG. he following is then a trivial update of the estimate of the smallest eigenvalue of matrix A. d (ln( f / f ) α [ i] / ) i= m i i fi ln i= m f α = and a = f exp( α [ ]) i ( ) i= m dα = 0 Eq. (9) hough Figure (3) plotted in logarithmic scale might mislead one to conclude that these loo-ahead underestimates are useful only for the initial iterations, it should be reminded that even a small but continued overestimate of smallest eigenvalue a can fail the upper bounds on error (as shown in figure ()). Hence these exponential lower bounds of the λ min are useful for tight and assured upper bounds of relative error even as by ensuring a tight underestimate of the smallest eigenvalue of A. Also, iterative solutions of linear systems are considered effective for a problem only when a stopping / < 0., when an underestimate of λ min is critical for the estimate of upper bound of errors.

0 6 (a) 0 4 λ min of J F=[-]/ λ min of A F=[-]/ 0 Eigen Value 0 0 0-0 -4 0 50 00 50 00 50 300 350 400 umber of iterations 0 6 (b) 0 5 0 4 λ min of A λ min of J F=[-]/ F=[-]/ 0 3 Eigen value 0 0 0 0 0-50 00 50 00 50 300 350 umber of Iterations Figure 3: λ min estimates using the proposed extremal eigenvalue approximation for SPD matrices (math.nist.gov/matrixmaret) along with CG: (a) bcsstm9; κ A = 33734, =87 and (b) bcsstm0; κ A = 55380, =485 when vector b is a sum of canonical basis for generality and m = 0.

4. umerical results and Discussion he estimated extremal eigenvalues are then introduced into the quadrature based estimates for both relative errors in l norm and A-norm. As mentioned before, the l norm estimates are not guaranteed to be upper bounds but nevertheless are close enough to the actual relative errors to be used as a stopping criterion. Actual errors are evaluated using the final x computed after iterations and x at th iteration. his is valid when the relative residues iterate to r / b <<, indicating convergence of the CG algorithm. Once the A-norm of the absolute error is estimated, the A-orthogonality of the CG iterates can be invoed to estimate the errors in relative terms. x A = x - x + x A and thus x A = ( x - x A + x A ) / when <x-x, Ax > = 0; hence x - x A / x A can be well approximated. he evaluation of bounds of l norm of the error, x-x = r A - r using the quadrature rules involves additional difficulties; especially in ascertaining if the evaluations are upper or the lower bounds []. he l norm of the absolute error in terms of the matrix J is given by Eq. (0) and approximations of the first term is especially non-trivial. Unlie there where Gauss rules were used, we have used the extremal eigenvalue approximations in estimating Gauss-Radau lower bounds of ε ; this used in Eq. (0) ensures we have an upper bound of l A norm most liely. he estimates of relative error in these cases is well approximated by x-x / x xx / x when it is <<, in spite of oscillations in x, and we show a few examples in figure (5) that use the extremal eigenvalue approximations in the computation. ε [ e J e ] = ro [ e J e] [ e J e] ε Eq. (0) A [ e J e ] 3

(a) 0 0 0-0 -4 Error 0-6 0-8 Relative Error in A-norm Upper bound using λ min approx. 0-0 0-00 00 300 400 500 600 700 umber of iterations 0 0 (b) 0-0 -4 Error 0-6 0-8 Relative Error in A-norm Upper bounds using λ min approx. 0-0 0 50 00 50 00 50 300 350 400 umber of iterations Figure 4: Gauss-Lobatto upper bounds of relative errors in A-norm using the proposed extremal eigenvalue approximation for SPD matrices: (a) bcsstm9; κ A = 33734, =87 and (b) bcsstm0; κ A = 55380, =485 (math.nist.gov/matrixmaret) when vector b is a sum of canonical basis for generality. 4

0 0 (a) 0-0 -4 Error Bounds 0-6 0-8 0-0 Relative Error in l norm Upper bounds using λ min approx. 0-0 -4 00 00 300 400 500 600 700 umber of iterations (b) 0 0 0 - Error Bounds 0-4 0-6 0-8 0-0 Relative Error in -norm Upper bounds using λ min approx. 0-0 50 00 50 00 50 300 350 400 umber of iterations Figure 5: Estimates of l norm of relative error using the proposed extremal eigenvalue approximation and the Gauss-Radau rules for SPD matrices: (a) bcsstm9; κ A = 33734, =87 and (b) bcsstm0; κ A = 55380, =485 (math.nist.gov/matrixmaret) when vector b is a sum of canonical basis for generality. 5. Conclusions he significance of error estimators for even problems with a moderately high condition number (>0 ) is evident and was emphasized by a few general examples in the introduction. A 5

stopping criterion by the residue can be orders of magnitude (~κ) conservative or optimistic and thus computationally either very costly or error prone for large problems. hough results of error estimation without a priori nowledge of eigenvalues have been shown here for SPD matrices and the CG algorithm, other iterative algorithms for generalized matrices can as well include such eigenvalue approximations and error estimations. here can be matrices of large condition numbers and clustered eigenvalues where Lanczos algorithm may need more than iterations due to finite numerical precision (in converging to an tri-diagonal matrix that includes all eigenvalues of A); the methods presented here are not in anyway limited by such matrices and the number of iterations of the CG algorithm. Acnowledgements DAS thans IISc Mathematics Initiative (IMI) for the generous support of its graduate students. Both DAS and MV than Prof. Srinivas alabatulla for his mentorship of DAS. Reference:. icholas J. Higham, A survey of condition number estimation for triangular matrices, SIAM Review 9 (987) 575-596.. John D. Dixon, Estimating extremal eigenvalues and condition numbers of matrices, SIAM J. umer. Anal. 0 (983) 8-84. 3. A. K. Cline, C. B. Moler, G. W. Stewart and J. H. Wilinson, An estimate for the condition number of a matrix, SIAM J. umer. Anal. 6 (979) 368-375. 4. G. Dahlquist, S.C. Eisenstat and G.H. Golub, Bounds for the error of linear systems of equations using the theory of moments, J. Math. Anal. Appl. 37 (97) 5 66. 5. G.H. Golub and G. Meurant, Matrices, moments and quadrature with applications, Princeton University Press, 00. 6. J. Kuczyisi and H. Woniaowsi, Probabilistic bounds on the extremal eigenvalues and condition number by the Lanczos algorithm, SIAM J. Matrix Anal. Appl.5 (994) 67-69. 7. D. Calvetti, S. Morigi, L. Reichel, F. Sgallari, An iterative method with error estimators, J Comput. Appl. Math. 7 (00) 93-9. 8. D. Calvetti, S. Morigi, L. Reichel, F. Sgallari, Computable error bounds and estimates for the conjugate gradient method, umer. Algorithms 5 (000) 75-88. 9. G. Meurant, he computation of bounds for the norm of the error in the conjugate gradient algorithm, umer. Algorithms 6 (997) 77 87. 6

0. G. Meurant and ichy, On computing quadrature-based bounds for the A-norm of the error in conjugate gradients, umer. Algorithms (0); DOI: 0.007/s075-0-959-9. G. Meurant, Estimates of the l norm of the error in the conjugate gradient algorithm, umer. Algorithms 40 (005) 57 69. 7