I.C.F. Ipsen small magnitude. Our goal is to give some intuition for what the bounds mean and why they hold. Suppose you have to compute an eigenvalue

Similar documents
deviation of D and D from similarity (Theorem 6.). The bound is tight when the perturbation is a similarity transformation D = D? or when ^ = 0. From

2 and bound the error in the ith eigenvector in terms of the relative gap, min j6=i j i? jj j i j j 1=2 : In general, this theory usually restricts H

RITZ VALUE BOUNDS THAT EXPLOIT QUASI-SPARSITY

or H = UU = nx i=1 i u i u i ; where H is a non-singular Hermitian matrix of order n, = diag( i ) is a diagonal matrix whose diagonal elements are the

DS-GA 1002 Lecture notes 0 Fall Linear Algebra. These notes provide a review of basic concepts in linear algebra.

Linear Algebra March 16, 2019

Institute for Advanced Computer Studies. Department of Computer Science. On the Perturbation of. LU and Cholesky Factors. G. W.

Introduction to Numerical Linear Algebra II

Class notes: Approximation

then kaxk 1 = j a ij x j j ja ij jjx j j: Changing the order of summation, we can separate the summands, kaxk 1 ja ij jjx j j: let then c = max 1jn ja

14 Singular Value Decomposition

1 Linear Algebra Problems

216 S. Chandrasearan and I.C.F. Isen Our results dier from those of Sun [14] in two asects: we assume that comuted eigenvalues or singular values are

Rank, Trace, Determinant, Transpose an Inverse of a Matrix Let A be an n n square matrix: A = a11 a1 a1n a1 a an a n1 a n a nn nn where is the jth col

October 25, 2013 INNER PRODUCT SPACES

Numerical Analysis Lecture Notes

Matrix Theory, Math6304 Lecture Notes from October 25, 2012

Numerical Linear Algebra

linearly indepedent eigenvectors as the multiplicity of the root, but in general there may be no more than one. For further discussion, assume matrice

CS 246 Review of Linear Algebra 01/17/19

Some inequalities for sum and product of positive semide nite matrices

Linear Algebra (part 1) : Vector Spaces (by Evan Dummit, 2017, v. 1.07) 1.1 The Formal Denition of a Vector Space

Preliminary/Qualifying Exam in Numerical Analysis (Math 502a) Spring 2012

The Eigenvalue Problem: Perturbation Theory

The Lanczos and conjugate gradient algorithms

Numerical Methods - Numerical Linear Algebra

Math 408 Advanced Linear Algebra

Elementary linear algebra

Matrices. Chapter What is a Matrix? We review the basic matrix operations. An array of numbers a a 1n A = a m1...

MATRICES WITH POSITIVE DEFINITE HERMITIAN PART: INEQUALITIES AND LINEAR SYSTEMS ROY MATHIAS

Lecture notes: Applied linear algebra Part 1. Version 2

UvA-DARE (Digital Academic Repository) Matrix perturbations: bounding and computing eigenvalues Reis da Silva, R.J. Link to publication

Linear Algebra. Solving Linear Systems. Copyright 2005, W.R. Winfrey

GAUSSIAN ELIMINATION AND LU DECOMPOSITION (SUPPLEMENT FOR MA511)

What is it we are looking for in these algorithms? We want algorithms that are

Last Time. Social Network Graphs Betweenness. Graph Laplacian. Girvan-Newman Algorithm. Spectral Bisection

the Unitary Polar Factor æ Ren-Cang Li P.O. Box 2008, Bldg 6012

Notes on the matrix exponential

Fundamentals of Engineering Analysis (650163)

1 Vectors. Notes for Bindel, Spring 2017 Numerical Analysis (CS 4220)

Applied Linear Algebra

Chapter 2: Matrix Algebra

Linear Algebra: Characteristic Value Problem

Vector Space Basics. 1 Abstract Vector Spaces. 1. (commutativity of vector addition) u + v = v + u. 2. (associativity of vector addition)

Institute for Advanced Computer Studies. Department of Computer Science. On the Adjoint Matrix. G. W. Stewart y ABSTRACT

6.4 Krylov Subspaces and Conjugate Gradients

Linear Algebra in Actuarial Science: Slides to the lecture

Equality: Two matrices A and B are equal, i.e., A = B if A and B have the same order and the entries of A and B are the same.

Linear algebra 2. Yoav Zemel. March 1, 2012

Chapter 3 Least Squares Solution of y = A x 3.1 Introduction We turn to a problem that is dual to the overconstrained estimation problems considered s

STABILITY OF INVARIANT SUBSPACES OF COMMUTING MATRICES We obtain some further results for pairs of commuting matrices. We show that a pair of commutin

Ritz Value Bounds That Exploit Quasi-Sparsity

Math 108b: Notes on the Spectral Theorem

4 Linear Algebra Review

Index. book 2009/5/27 page 121. (Page numbers set in bold type indicate the definition of an entry.)

7 Principal Component Analysis

13-2 Text: 28-30; AB: 1.3.3, 3.2.3, 3.4.2, 3.5, 3.6.2; GvL Eigen2

the Unitary Polar Factor æ Ren-Cang Li Department of Mathematics University of California at Berkeley Berkeley, California 94720

Quantum Computing Lecture 2. Review of Linear Algebra

CS 6820 Fall 2014 Lectures, October 3-20, 2014

KEYWORDS. Numerical methods, generalized singular values, products of matrices, quotients of matrices. Introduction The two basic unitary decompositio

UMIACS-TR July CS-TR 2494 Revised January An Updating Algorithm for. Subspace Tracking. G. W. Stewart. abstract

c 1997 Society for Industrial and Applied Mathematics Vol. 39, No. 2, pp , June

MAT 2037 LINEAR ALGEBRA I web:

1 Multiply Eq. E i by λ 0: (λe i ) (E i ) 2 Multiply Eq. E j by λ and add to Eq. E i : (E i + λe j ) (E i )

Linear Algebra. Min Yan

On the Quadratic Convergence of the. Falk-Langemeyer Method. Ivan Slapnicar. Vjeran Hari y. Abstract

Eigenvectors and Hermitian Operators

5.6. PSEUDOINVERSES 101. A H w.

CHARACTERIZATIONS. is pd/psd. Possible for all pd/psd matrices! Generating a pd/psd matrix: Choose any B Mn, then

Linear algebra for computational statistics

Linear Algebra, part 3 QR and SVD

ELEMENTARY LINEAR ALGEBRA WITH APPLICATIONS. 1. Linear Equations and Matrices

Matrix Inequalities by Means of Block Matrices 1

Lecture 3: QR-Factorization

QUASI-UNIFORMLY POSITIVE OPERATORS IN KREIN SPACE. Denitizable operators in Krein spaces have spectral properties similar to those

STAT 309: MATHEMATICAL COMPUTATIONS I FALL 2018 LECTURE 9

Chapter 3 Transformations

[Disclaimer: This is not a complete list of everything you need to know, just some of the topics that gave people difficulty.]

STAT 309: MATHEMATICAL COMPUTATIONS I FALL 2018 LECTURE 13

Institute for Advanced Computer Studies. Department of Computer Science. Iterative methods for solving Ax = b. GMRES/FOM versus QMR/BiCG

Linear Algebra Review

forms Christopher Engström November 14, 2014 MAA704: Matrix factorization and canonical forms Matrix properties Matrix factorization Canonical forms

G1110 & 852G1 Numerical Linear Algebra

Math 407: Linear Optimization

Math Linear Algebra Final Exam Review Sheet

Linear Regression and Its Applications

IV. Matrix Approximation using Least-Squares

Multivariate Statistical Analysis

Chapter 7. Linear Algebra: Matrices, Vectors,

Proposition 42. Let M be an m n matrix. Then (32) N (M M)=N (M) (33) R(MM )=R(M)

Convexity of the Joint Numerical Range

AMS526: Numerical Analysis I (Numerical Linear Algebra)

Solution Set 7, Fall '12

Linear Algebra. Linear Equations and Matrices. Copyright 2005, W.R. Winfrey

EIGENVALUE PROBLEMS. EIGENVALUE PROBLEMS p. 1/4

Large Scale Data Analysis Using Deep Learning

Eigenvalue Problems. Eigenvalue problems occur in many areas of science and engineering, such as structural analysis

EE731 Lecture Notes: Matrix Computations for Signal Processing

Scientific Computing: An Introductory Survey

Transcription:

Acta Numerica (998), pp. { Relative Perturbation Results for Matrix Eigenvalues and Singular Values Ilse C.F. Ipsen Department of Mathematics North Carolina State University Raleigh, NC 7695-85, USA ipsen@math.ncsu.edu http://www4.ncsu.edu/~ipsen/info.html It used to be good enough to bound absolute errors of matrix eigenvalues and singular values. Not anymore. Now it is fashionable to bound relative errors. We present a collection of relative perturbation results which have emerged during the past ten years. No need to throw away all those absolute error bounds, though. Deep down the derivation of many relative bounds can be based on absolute bounds. This means, relative bounds are not always better. They may just be better sometimes { and exactly when depends on the perturbation. CONTENTS Introduction Additive Perturbations for Eigenvalues 6 3 Additive Perturbations for Singular Values 4 Some Applications of Additive Perturbations 7 5 Multiplicative Perturbations for Eigenvalues 33 6 Multiplicative Perturbations for Singular Values 39 7 Some Applications of Multiplicative Perturbations 4 8 The End 47 References 48. Introduction Are you concerned about accuracy of matrix eigenvalues or singular values, especially the ones close to zero? If so, this paper is for you! We present error bounds for eigenvalues and singular values that can be much tighter than the traditional bounds, especially when these values have Work supported in part by grant CCR-949 from the National Science Foundation.

I.C.F. Ipsen small magnitude. Our goal is to give some intuition for what the bounds mean and why they hold. Suppose you have to compute an eigenvalue of a complex square matrix A. Numerical software usually produces a number ^ that is not the desired eigenvalue. So you ask yourself how far away is ^ from an eigenvalue of A? If ^ was produced by a reliable (i.e. backward stable) numerical method, there is a round o error analysis to assure you that ^ is an eigenvalue of a nearby matrix A + E, where E is small in some sense. Then you can use perturbation theory to estimate the error in ^. For instance, when A is diagonalisable the Bauer-Fike theorem bounds the absolute distance between ^ and a closest eigenvalue of A by j? ^j (X) kek; (.) where (X) kxk kx? k is the condition number of an eigenvector matrix X of A. The quantity j?^j represents an absolute error. Traditional perturbation theory assesses the quality of a perturbed eigenvalue by bounding absolute errors. However there are practical situations where small eigenvalues have physical meaning and should be determined to high relative accuracy. Such situations include computing modes of vibration in a nite element context, and computing energy levels in quantum mechanical systems (Demmel, Gu, Eisenstat, Slapnicar, Veselic and Drmac 997, x). Absolute error bounds cannot cope with relative accuracy, especially when confronted with small eigenvalues or singular values. The following section explains why... Why Absolute Bounds Don't Do the Job If we want relative accuracy, we need relative error bounds. The simplest way to generate a relative error bound is to divide an absolute error bound by an eigenvalue. For instance, dividing the absolute error bound (:) by a non-zero eigenvalue produces the relative error bound j? ^j jj kek jj : Unlike the absolute bound, though, the relative bound depends on. This has several disadvantages. First, each eigenvalue has a dierent relative bound. Second, the relative bound is smaller for eigenvalues that are large in magnitude than for those that are small in magnitude. Third, the relative bound can be pessimistic for eigenvalues of small magnitude, as the following example illustrates.

Example. Relative Perturbation Results 3 The mere act of storing a diagonal matrix A = B @... in oating point arithmetic produces a perturbed matrix A + E = B @ ( + ) n C A C... A ; n ( + n ) where j i j and > reects the machine accuracy. According to the absolute perturbation bound (:), the error in an eigenvalue ^ of A + E is bounded by min i j i? ^j kek = max k j k k j max j k j: k This bound is realistic for eigenvalues of largest magnitude: If ^ is closest to an eigenvalue max of largest magnitude among all eigenvalues of A, then j max? ^j j max j : Since the relative error in all eigenvalues does not exceed, the bound is tight in this case. However the bound is too pessimistic for eigenvalues of smallest magnitude: If ^ is closest to an eigenvalue min of smallest magnitude among all eigenvalues of A, then j min? ^j j min j j maxj j min j : The bound is much larger than when the magnitude of the eigenvalues varies widely. Since the relative error does not exceed, the bound is not tight. There are algorithms whose relative error bounds do not depend on the eigenvalues. These algorithms compute all eigenvalues or singular values to high relative accuracy, even those of small magnitude: the dqds algorithm for singular values of bidiagonal matrices (Fernando and Parlett 994, Parlett 995), for instance, as well as Jacobi methods for eigenvalues of symmetric positive-denite matrices and for singular values (Demmel 997, x5.4.3), (Mathias 995a). Absolute perturbation bounds cannot account for this phenomenon. Absolute error bounds are well suited for describing the accuracy of xed point arithmetic. But xed point arithmetic has been replaced by oating point arithmetic, especially on general purpose machines where many

4 I.C.F. Ipsen eigenvalue and singular value computations are carried out nowadays. The accuracy of oating point arithmetic is best described by relative errors. In the absence of underow and overow, a number is represented as a oating point number ^ ( + ); where j j ; and > reects the machine accuracy. In IEEE arithmetic, for instance,?7 in single precision and?6 in double precision. Therefore the accuracy of oating point arithmetic can be described by relative error bounds of the form j^? j jj or j^? j j^j : Absolute error bounds cannot model this situation. And even if you never require high relative accuracy from your small eigenvalues or singular values, you can still prot from it. It turns out that intermediate quantities computed to high relative accuracy can sometimes speed up subsequent computations. For instance, computing eigenvalues of a real, symmetric, tridiagonal matrix to high relative accuracy can accelerate eigenvector computations because the time consuming process of orthogonalising eigenvectors can be shortened or even avoided (Dhillon, Fann and Parlett 997). Now that we have established the need for `genuine' relative error bounds beyond any shadow of a doubt, it's time to nd out what kind of relative bounds are out there... Overview Relative error bounds have been derived in the context of two dierent perturbation models: Additive perturbations (xx, 3, 4) represent the perturbed matrix as A + E. Multiplicative perturbations (xx5, 6, 7) represent the perturbed matrix as D AD, where D and D are non-singular matrices. The traditional absolute error bounds are derived in the context of additive perturbations. We group the bounds for eigenvalues (x, x5) and for singular values (x3, x6) according to a loose order of increasing specialisation: Bauer-Fike-type: Two-norm bounds on the distance between a perturbed eigenvalue and a closest exact eigenvalue. Homan-Wielandt-type: Frobenius norm bounds on the sum (of squares) of all distances between

Relative Perturbation Results 5 perturbed eigenvalues and corresponding exact eigenvalues, where perturbed and exact eigenvalues are paired up in a one-to-one fashion. Similar for singular values. Weyl-type: Two norm bounds on the largest distance between a perturbed eigenvalue and the corresponding exact eigenvalue, where the ith largest perturbed eigenvalue is paired up with the ith largest exact eigenvalue. Similar for singular values. There are several dierent ways to normalise an absolute error j? ^j and turn it into a relative error. We present bounds for the following relative error measures j? ^j ; jj j? q ^j ; jj j^j j? qjj ^j ; p + j^j p where p is an integer. For instance, the traditional relative error j? ^j=jj can be larger or smaller than the second error measure, while it is never smaller than the third. Detailed relationships among the dierent measures are discussed in (Li 994a, x). Since the measures are essentially proportional to each other we disregard any dierences among them. Sections xx4 and 7 discuss applications of additive and multiplicative perturbations..3. Notation We use two norms: the two-norm kak = max x6= kaxk kxk ; where kxk = p x x; and the superscript denotes the conjugate transpose; and the Frobenius norm kak F = s X i;j p ja ij j ; where a ij are the elements of the matrix A. The identity matrix of order n is B I =. @.. C A = ( e : : : e n ) with columns e i. For a matrix A we denote by range(a) its column space, A? its inverse and A y its Moore-Penrose inverse. The absolute value matrix jaj has elements ja ij j. A matrix inequality of the form jaj jbj is meant element-wise, i.e. ja ij j jb ij j for all i and j.

6 I.C.F. Ipsen. Additive Perturbations for Eigenvalues Let A be a complex square matrix. We want to bound the absolute and relative errors in the eigenvalues of the perturbed matrix A + E. In the process we show that relative error bounds are as natural as absolute error bounds, and that many relative bounds are implied by absolute bounds... Bauer-Fike-Type Bounds for Diagonalisable Matrices The Bauer-Fike theorem bounds the distance between an eigenvalue of A+E and a closest eigenvalue of A. The matrix A must be diagonalisable, while A + E does not have to be. Let A = XX? be an eigendecomposition of A, where = B @... and i are the eigenvalues of A. Let ^ be an eigenvalue of A + E. The Bauer-Fike Theorem for the two-norm (Bauer and Fike 96, Theorem IIIa) bounds the absolute error, n C A ; min i j i? ^j (X) kek: (.) The relative version of the Bauer-Fike Theorem below requires in addition that A be non-singular. Theorem. If A is diagonalisable and non-singular then min i where (X) kxk kx? k. j i? ^j j i j (X) ka? Ek; Proof. (Eisenstat and Ipsen 997, Corollary 3.) The idea is to `divide' by the eigenvalues of A and apply the absolute error bound (:). Write (A + E)^x = ^^x as (^A?? A? E)^x = ^x: This means, is an eigenvalue of ^A?? A? E. The matrix ^A? has the same eigenvector matrix as A and its eigenvalues are ^= i. Apply the Bauer- Fike Theorem (:) to ^A? and to the perturbed matrix ^A?? A? E. If we interpret the amplier (X) as a condition number for the eigenvalues of A then absolute and relative error bounds have the same condition

Relative Perturbation Results 7 number. This means in the eyes of the Bauer-Fike Theorem an eigenvalue is as sensitive in the absolute sense as it is in the relative sense. A comparison of the absolute bound (:) and the relative bound in Theorem. shows that the absolute error is bounded in terms of the absolute perturbation E, while the relative error is bounded in terms of the relative perturbation A? E. But A? E is not the only way to express a relative perturbation. Why keep A? on the left of E? Why not move it to the right, or distribute it on both sides of E? Splitting A = A A and sandwiching E between the two factors, like A? EA?, results in undreamt-of possibilities for relative perturbations. Theorem. Let A be diagonalisable and non-singular. If A = A A where A and A commute then min i j i? ^j j i j (X) ka? EA? k: Proof. (Eisenstat and Ipsen 997, Corollary 3.4) The idea is to apply Theorem. to the similarity transformations A AA? and A (A + E)A? : Fortunately, similarity transformations preserve eigenvalues. And the commutativity of A and A prevents the similarity from changing A, A A A? = A (A A ) A? = A A = A A = A: Therefore we retain the condition number of the original eigenvector matrix X. When A = A and A = I, Theorem. reduces to the relative bound in Theorem.. Setting A = I and A = A gives (Eisenstat and Ipsen 997, Corollary 3.5) min i j i? ^j j i j (X) kea? k: This bound includes as a special case (Veselic and Slapnicar 993, Theorem 3.7). Another popular choice for A and A is a square root A = of A. In this case Theorem. gives (Eisenstat and Ipsen 997, Corollary 3.6) min i j i? ^j j i j (X) ka?= EA?= k:.. Bauer-Fike-Type Bounds for Normal Matrices Normal matrices have unitary eigenvector matrices but, in contrast to Hermitian or real symmetric matrices, their eigenvalues are not necessarily real.

8 I.C.F. Ipsen Normal matrices include Hermitian, skew-hermitian, real symmetric, real skew-symmetric, diagonal, unitary and real orthogonal matrices. Since the condition number of a unitary eigenvector matrix equals one, the Bauer-Fike theorem applied to a normal matrix A simplies. The absolute error bound (:) becomes min j i? ^j kek; i while the corresponding relative bound requires again that A be non-singular. Theorem.3 Let A be normal and non-singular. If A = A A, where A and A commute, then j i? min ^j ka? i j i j EA? k: Proof. Follows immediately from Theorem.. Therefore eigenvalues of normal matrices are well-conditioned, in the absolute as well as in many relative senses. The relative bound in Theorem.3 is tight for diagonal matrices A and component-wise perturbations E, like those in Example.. For the relative bound to remain in eect, A and A have to commute. Our choices for `commuting factorisations' have been so far: (A ; A ) = (A; I); (A ; A ) = (I; A); (A ; A ) = (A = ; A = ): But since A is normal there is another commuting factorisation: the polar factorisation. Every square matrix A has a polar factorisation A = HU, where H (AA ) = is Hermitian positive-semidenite and U is unitary (Horn and Johnson 985, Theorem 7.3.). The matrix H is always unique, while U is only unique when A is non-singular. In particular, when A is Hermitian positive-denite H = A and U is the identity. We use the fact that polar factors of normal non-singular matrices commute in the following sense (Eisenstat and Ipsen 997, Lemma 4.) HU = UH = H = UH = : Theorem.4 If A is normal and non-singular, with Hermitian positivedenite polar factor H then min i j i? ^j j i j kh?= EH?= k: Proof. (Eisenstat and Ipsen 997, Theorem 4.3) Since A = H = UH =, we can set A = H = U and A = H = in Theo-

Relative Perturbation Results 9 rem.3 to get ka? EA? k = ku H?= EH?= k = kh?= EH?= k: Therefore the eigenvalues of a normal matrix have the same relative error bound as the eigenvalues of its positive-denite polar factor. This suggests that the eigenvalues of a normal matrix are as well conditioned as the eigenvalues of its positive-denite polar factor. More generally we conclude that eigenvalues of normal matrices are no more sensitive than eigenvalues of Hermitian positive-denite matrices..3. Homan-Wielandt-Type Bounds for Diagonalisable Matrices The Homan-Wielandt theorem establishes a one-to-one pairing between all eigenvalues of A and A + E and bounds the sum of all pairwise distances in the Frobenius norm. This requires not only A but also A + E to be diagonalisable. Let A and A + E be diagonalisable matrices with eigendecompositions A = XX? and A + E = ^X ^ ^X?, respectively. The eigenvalues are = B @... n C A ; ^ = B @ ^... The extension of the Homan-Wielandt theorem from normal to diagonalisable matrices (Elsner and Friedland 995, Theorem 3.) bounds the absolute error, vu u t X n i= ^ n C A : j i? ^ (i) j ( ^X) (X) kekf (.) for some permutation. The condition numbers (X) and ( ^X) are expressed in the two-norm to make the bound tighter, since the two-norm never exceeds the Frobenius norm. We can obtain a relative Homan-Wielandt-type bound from a stronger version of (:) that deals with eigenvalues of matrix products. To this end write the perturbed matrix as AC + E, where C must have the same eigenvector matrix as AC + E. The bound (:) is the special case where C = I. The eigendecomposition of C is C = ^X? ^X? ; where? = B @... n C A :

I.C.F. Ipsen The eigendecompositions of A and the perturbed matrix remain the same, A = XX? ; AC + E = ^X ^ ^X? : The stronger Homan-Wielandt-type bound below bounds the sum of squares of absolute errors in the products of the eigenvalues of A and C. Lemma. If A, C and AC + E are diagonalisable then there exists a permutation such that vu u t X n i= j i (i)? ^ (i) j ( ^X) (X) kekf : Proof. (Eisenstat and Ipsen 997, Theorem 6.) Now we are ready for the relative bound. The stronger absolute bound in Lemma. implies a relative version of the original Homan-Wielandt-type bound (:), provided A is non-singular. Theorem.5 Let A and A + E be diagonalisable. If A is non-singular then there exists a permutation so that vu ux t n j i? ^! (i) j ( ^X) (X) ka? Ek F : j i j i= Proof. (Eisenstat and Ipsen 997, Corollary 6.) Since A? (A + E)? A? E = I we can set A A? ; C A + E; E?A? E: Then A is diagonalisable with eigenvector matrix X and eigenvalues? i ; C is diagonalisable with eigenvector matrix ^X and eigenvalues ^i ; and AC + E = ^XI ^X? is diagonalisable, where the eigenvalues are and one can choose ^X as an eigenvector matrix. Applying Lemma. to A, C and E gives nx i= j? i ^ (i)? j ( ^X) (X) ka? Ek F :.4. Homan-Wielandt-Type Bounds for Hermitian Matrices When A and A+E are Hermitian the permutation in the Homan-Wielandt Theorem is the identity, provided exact and perturbed eigenvalues are numbered as n : : : ; ^n : : : ^ :

Relative Perturbation Results The Homan-Wielandt Theorem for Hermitian matrices (Bhatia 997, Exercise III.6.5) (Lowner 934) bounds the absolute error by vu u t X n i= j i? ^ i j kek F : (.3) The relative bound below requires in addition that A and A + E be positivedenite. Theorem.6 Proof. vu u t n X i= If A and A + E be Hermitian positive-denite then j i? ^ i j i^i ka?= EA?= (I + A?= EA?= )?= k F : (Li 994a, Theorem 3.), (Li and Mathias 997, Proposition 3.4') As consequence a small ka?= EA?= k F guarantees a small eigenvalue error. If one does not mind dealing with majorisation theory one can derive bounds that are stronger than Theorem.6 and hold for any unitarily invariant norm (Li and Mathias 997, Proposition 3.4, (3.9))..5. Weyl-Type Bounds Weyl's perturbation theorem (Bhatia 997, Corollary III..6) bounds the worst distance between the ith eigenvalues of Hermitian matrices A and A + E in the two-norm, max j i? ^ i j kek: (.4) in The absolute bound (:4) implies a relative bound, provided that A is positive-denite. There is no restriction on E other than being Hermitian. Theorem.7 then Let A and A+E be Hermitian. If A is also positive-denite j i? max ^ i j ka?= EA?= k: in j i j Proof. (Eisenstat and Ipsen 997, Corollary 5.), (Mathias 994, Theorem.3) We reproduce the proof from (Eisenstat and Ipsen 997) because it explains how the absolute bound (:4) implies the relative bound. Fix an index i. Let ^x be an eigenvector of A + E associated with ^ i, i.e. (A + E)^x = ^ i^x: Multiplying (^ i I? E)^x = A^x by A?= on both sides gives ( A + E) z = z;

I.C.F. Ipsen where A ^ i A? ; E?A?= EA?= ; z A =^x: Hence is an eigenvalue of A + E. We show that it is actually the (n? i + )st eigenvalue and argue as in the proof of (Eisenstat and Ipsen 995, Theorem.). Since ^ i is the ith eigenvalue of A + E, must be the ith eigenvalue of (A + E)? ^ i I = A = (I? A? E) A = : But this is a congruence transformation because square-roots of positivedenite matrices are Hermitian. Congruence transformations preserve the inertia. Hence is the ith eigenvalue of I? A? E, and is the (n? i + )st eigenvalue of A + E. Applying Weyl's Theorem (:4) to A and A + E gives max jn ^ i? j k Ek = ka?= EA?= k; n?j+ where j are the eigenvalues of A + E. When j = n? i +, then j = and we get the desired bound. The following example illustrates what the relative bound in Theorem.7 looks like when E is a component-wise relative perturbation. Example. (Mathias 994, pages 6, 7) Let's rst subject a single diagonal element of a Hermitian positive-denite matrix A to a component-wise relative perturbation. Say, a jj is perturbed to a jj ( + ). The perturbed matrix is A + E, where E = e j e T j and e j is the jth column of the identity matrix. Then ka?= EA?= k = jj ka?= e j e T j A?= k = jj je T j A?= A?= e j j = jj (A? ) jj ; where (A? ) jj is the jth diagonal element of A? (which is positive since A is positive-denite). The relative error bound in Theorem.7 is j i? max ^ i j jj (A? ) jj : in j i j This means, a small relative error in a diagonal element of a Hermitian positive-denite matrix causes only a small relative error in the eigenvalues if the corresponding diagonal element of the inverse is not much larger than one. Next we'll subject a pair of o-diagonal elements to a component-wise relative perturbation. Say, a jk and a kj are perturbed to a jk ( + ) and a kj (+), respectively. The perturbed matrix is A+E, where E = (e j e T k +

e j e T k ). In this case we get Relative Perturbation Results 3 ka?= EA?= k jj q (A? ) jj (A? ) kk : The relative error bound in Theorem.7 becomes j i? max ^ i j jj in j i j q (A? ) jj (A? ) kk : This means, a small relative error in a pair of o-diagonal elements of a Hermitian positive-denite matrix causes only a small relative error in the eigenvalues if the product of the corresponding diagonal elements in the inverse is not much larger than one. The bound below, for a dierent error measure, is similar to the Frobenius norm bound in Theorem.6. Theorem.8 If A and A + E are Hermitian positive-denite then j i? max ^ i j ka in q?= EA?= (I + A?= EA?= )?= k: i^i Proof. (Li 994a, Theorem 3.), (Li and Mathias 997, Proposition 3.4').6. Weyl-Type Bounds for More Restrictive Perturbations It is possible to get a Weyl-type bound for eigenvalues of Hermitian matrices without ocially asking for positive-deniteness. The price to be paid, however, is a severe restriction on E to prevent perturbed eigenvalues from switching sign. Theorem.9 then Let A and A + E be Hermitian. If < l u and l x Ax x Ex u x Ax for all x l i ^ i? i u i ; i n: Proof. This is a consequence of (Barlow and Demmel 99, Lemma ). The Minimax Principle for eigenvalues of Hermitian matrices (Bhatia 997, Corollary III..) implies i = x Ax max min dim(s)=i xs x x = min x Ax xs x x ; for some subspace S of dimension i. Then ^ i = max dim(s)=i min xs x (A + E)x x x min xs x (A + E)x x x = x (A + E)x x x

4 I.C.F. Ipsen for some x S. The above expression for i and the assumption imply x (A + E)x x x x Ax min xs x x + x Ax l x x i + l i ; where we have also used the fact l >. Hence l i ^ i? i. The upper bound is proved similarly using the characterisation i = x Ax min max dim(s)=n?i+ xs x x : Therefore the relative error in the eigenvalues lies in the same interval as the relative perturbation. The relative error bound in Theorem.9 implies ( + l ) i ^ i ( + u ) i ; where l and u are positive. Hence ^ i has the same sign as i, and j^ i j > j i j. Thus the restriction on the perturbation is strong enough that it not only forces A and A+E to have the same inertia, but it also pushes the perturbed eigenvalues farther from zero than the exact eigenvalues. The restriction on the perturbation in the following bound is slightly weaker. It uses the polar factor technology from Theorem.4. Theorem. Let A and A + E be Hermitian. If H is the positivesemidenite polar factor of A and if for some < < then jx Exj x Hx j i? ^ i j j i j: for all x Proof. This is a consequence of (Veselic and Slapnicar 993, Theorem.). The assumption implies x (A? H)x x (A + E)x x (A + H)x: If A = XX is an eigendecomposition of A then, because A is Hermitian, the polar factor is H = XjjX, where jj is the matrix whose elements are the absolute values of. Hence A and H have the same eigenvectors. A min-max argument as in the proof of Theorem.9 establishes i? j i j ^ i i + j i j: Therefore the relative error in the eigenvalues is small if the relative perturbation with regard to the polar factor is small. Since < < the relative error bound implies that ^ i has the same sign as i. Hence the assumptions in Theorem. ensure that A + E has the same inertia as A.

Relative Perturbation Results 5 Theorem. applies to component-wise relative perturbations. Matrix inequalities below of the form jej jaj are to be interpreted element-wise. Corollary. Let A and A + E be Hermitian, and jej jaj for some >. If H is the positive-denite polar factor of A and for some > then jxj jajjxj x Hx j i? ^ i j j i j: for all x Proof. This is a consequence of (Veselic and Slapnicar 993, Theorem.). We merely need to verify that the assumptions of Theorem. hold, jx Exj jxj jejjxj jxj jajjxj x Hx:.7. Congruence Transformations for Positive-Denite Matrices All the bounds we have presented so far for positive-denite matrices contain the term A?= EA?=. Since A is Hermitian positive-denite, A = is Hermitian, which makes A?= EA?= Hermitian. This in turn implies that the two-norm and Frobenius norm of A?= EA?= are invariant under congruence transformations. We say that two square matrices A and M are congruent if A = D MD for some non-singular matrix D. If A is Hermitian positive-denite, so is M because congruence transformations preserve the inertia. We start out by showing that the bound in Theorem.7 is invariant under congruence transformations. Corollary. Let A and be Hermitian positive-denite and A + E Hermitian. If A = D MD and A+E = D (M +F )D, where D is non-singular, then j i? max ^ i j km?= F M?= k: in j i j Proof. (Mathias 994, Theorem.4) Start with the bound in Theorem.7, j i? max ^ i j ka?= EA?= k: in j i j Positive-deniteness is essential here. Since A is Hermitian positive-denite, it has a Hermitian square-root A =. Hence A?= EA?= is Hermitian. This implies that the norm is an eigenvalue, ka?= EA?= k = max jn j j(a?= EA?= )j:

6 I.C.F. Ipsen Now comes the trick. Since eigenvalues are preserved under similarity transformations, we can reorder the matrices in a circular fashion until all grading matrices have cancelled each other out, j (A?= EA?= ) = j (A? E) = j (D? M? F D) = j (M? F ) At last recover the norm, = j (M?= F M?= ): max jn j j(m?= F M?= )j = km?= F M?= k: Corollary. extends (Barlow and Demmel 99, Theorem ) and (Demmel and Veselic 99, Theorem.3) to a larger class of matrices. It suggests that the eigenvalues of A + E have the same error bound as the eigenvalues of M + F. We can interpret this to mean that the eigenvalues of a Hermitian positive-denite matrix behave as well as the eigenvalues of any matrix congruent to A. The example below illustrates this. Example. (Demmel and Veselic 99, page ) The matrix A = @ 4 9 9 9 9 9 9 is symmetric positive-denite with eigenvalues (to six decimal places) A : 4 ; 9:9 9 ; 9:888? : If we write A = DMD, where M = @ : : : : : : A ; D = @ then the eigenvalues of M are (to six decimal places) A ; 9:? ; 9:? ; :: Corollary. implies that the widely varying eigenvalues of A, and in particular the very small ones, are as impervious to changes in M as the uniformly sized eigenvalues of M. Already thirty years ago structural engineers considered congruence transformations like the one above where D is diagonal and all diagonal elements of M are equal to one (Rosano, Glouderman and Levy 968, pages 4, 5). They observed that such an equilibration `reduce[s] the ratio of extreme eigenvalues' (Rosano et al. 968, page 45), and that `equilibration is of major importance in measurement of matrix conditioning' (Rosano et al. 968, page 59).

Relative Perturbation Results 7 From the circular reordering argument in the proof of Corollary. it also follows that the other bounds for positive-denite matrices are invariant under congruences. One bound is Theorem.8. Corollary.3 Let A and A + E be Hermitian positive-denite. If A = D MD and A + E = D (M + F )D, where D is non-singular then j i? max ^ i j km in q?= F M?= (I + M?= F M?= )?= k: i^i Proof. (Li 994a, Theorem 3.), (Li and Mathias 997, Proposition 3.4') The other bound that is also invariant under congruences is the Frobenius norm bound Theorem.6. Corollary.4 Let A and A + E be Hermitian positive-denite. If A = D MD and A + E = D (M + F )D, where D is non-singular then vu u t n X i= j i? ^ i j i^i km?= F M?= (I + M?= F M?= )?= k F : Proof. (Li 994a, Theorem 3.), (Li and Mathias 997, Proposition 3.4') Since the Frobenius norm sums up squares of eigenvalues, the bound from Theorem.8 can be written as ka?= EA?= (I + A?= EA?= )?= k F = nx i= i + i where i are the eigenvalues of the Hermitian matrix A?= EA?=. The circular reordering argument from the proof of Corollary. implies that i are also the eigenvalues of M?= F M?=. One may wonder what's so interesting about congruence transformations. One can use congruence transformations to pull the grading out of a matrix (Barlow and Demmel 99, x), (Demmel and Veselic 99, x, x.), (Mathias 995a). Consider the matrix A in Example.. It has elements of widely varying magnitude that decrease from top to bottom. The diagonal matrix D removes the grading and produces a matrix M, where M D? AD?, all of whose elements have about the same order of magnitude and all of whose eigenvalues are of about the same size. More generally we say that a Hermitian positive-denite matrix A is graded, or scaled, if A = DMD and the eigenvalues of M vary much less in magnitude than the eigenvalues of A (Mathias 995a, x).

8 I.C.F. Ipsen.8. Congruence Transformations for Indenite Matrices Since the application of congruence transformations is not restricted to Hermitian positive-denite matrices, we may as well try to nd out whether indenite matrices are invariant under congruences. It turns out that the resulting error bounds are weaker than the ones for positive-denite matrices because they require stronger assumptions. If we are a little sneaky (by extracting the congruence from the polar factor rather than the matrix proper) then the bound for normal matrices in Theorem.4 becomes invariant under congruences. Corollary.5 Let A be normal and non-singular, with Hermitian positive-denite polar factor H. If D is non-singular and E = DE D ; H = DM D then j i? min ^j km?= E M?= k: i j i j Proof. (Eisenstat and Ipsen 997, Corollary 4.4) This means the error bound for eigenvalues of a normal matrix is the same as the error bound for eigenvalues of the best scaled version of its positive-denite polar factor. Let's return to Weyl-type bounds, but now under the condition that the congruence transformation is real diagonal. Theorem. leads to a bound that is essentially scaling invariant. It is similar to the one above, in the sense that the scaling matrix is extracted from its positive-denite polar factor. However now the perturbations are restricted to be component-wise relative. Corollary.6 Let A = DMD be non-singular Hermitian, where D is diagonal with positive diagonal elements, and let H = DM D be the positivedenite polar factor of A. If A + E is Hermitian and jej jaj for some > then j i? max ^ i j k jmj k km? k: in j i j Proof. This is a consequence of (Veselic and Slapnicar 993, Theorem.3). We use variational inequalities to show that the assumptions of Corollary. are fullled. Since D is positive-denite, jxj jaj jxj = jxj D jmj Djxj k jmj k x D x for all x: Variational inequalities imply x D x km? k x Hx:

Therefore Relative Perturbation Results 9 jxj jaj jxj x Hx for all x; where k jmj k km? k. Now apply Corollary.. The amplier in the bound, k jmj k km? k, is almost like the condition number of the absolute value matrix jm j or almost like the condition number of the scaled polar factor M. Therefore the relative error in the eigenvalues of a Hermitian matrix is small if the polar factor and the absolute value of the matrix are well-scaled. The following bound is similar in the sense that it applies to a column scaling of a Hermitian matrix A = M D. In contrast to Corollary.6, however, the scaled matrix M is in general not Hermitian anymore; and the inverse of the scaled matrix, M?, now appears in the bound rather than the inverse of the scaled polar factor, M?. Corollary.7 Let A = MD be non-singular Hermitian and D diagonal with positive diagonal elements. If A + E is Hermitian, and jej jaj for some > then j i? max ^ i j k jmj k km? k: in j i j Proof. (Veselic and Slapnicar 993, Theorem 3.6) First we take care of the scaling matrix D. The technology of previous proofs requires that D appear on both sides of the matrix. That's why we consider A = A A = DM MD. The component-wise perturbation implies jx E xj jxj jaj jxj: Proceeding as in the proof of Corollary.6 gives Hence jxj jaj jxj k jmj k km? k x A x for all x: jx E xj x A x; where k jmj k km? k. Now that we got rid of D, we need to undo the squares. In order to take the positive square-root without losing the monotonicity we need positive-denite matrices under the squares. Polar factors do the job. If H and H E are the Hermitian positive-denite polar factors of A and E, respectively then Therefore x E x = x H Ex; x A x = x H x: x H Ex x H x for all x:

I.C.F. Ipsen Now comes the trick. Because H E and H are Hermitian positive-denite we can apply the fact that the square-root is operator-monotone (Bhatia 997, Proposition V.I.8) and conclude x H E x x Hx: Since jx Exj x H E x, Theorem. applies. The next bound, the last one in this section, applies to general perturbations. Compared to other bounds it severely constrains the size of the perturbation by forcing it to be smaller than any eigenvalue of any principal submatrix. Theorem. Let A = DMD and A + E = D(M + F )D be real, symmetric and D a real non-singular diagonal matrix. Among all eigenvalues of principal submatrices of M, let be the smallest in magnitude. If kf k < jj then jj? kf k?kf k ^ i? i jj? kf k kf k jj i (jj? kf k) ; i n: Proof. (Gu and Eisenstat 993, Corollary 5).9. Ritz Values Ritz values are `optimal' approximations to eigenvalues of Hermitian matrices. Let A be a Hermitian matrix of order n and Q a matrix with m orthonormal columns. Then W Q AQ is a matrix of order m whose eigenvalues ^ : : : ^ m ; are called Ritz values of A (Parlett 98, x.3). The corresponding residual is R AQ? QW. Ritz values are optimal in the following sense. Given Q, the norm of R can only increase if we replace W by another matrix, i.e. (Parlett 98, Theorem -4-5), krk = kaq? QW k kaq? QCk for all matrices C of order m. Moreover one can always nd m eigenvalues of A that are within absolute distance krk of the Ritz values (Parlett 98, Theorem -5-), max jm j (j)? ^ j j krk for some permutation. Unfortunately the corresponding relative error bounds are not as simple. They are expressed in terms of angles between the relevant subspaces range(q), range(aq), and range(a? Q). Let = be the maximal

Relative Perturbation Results principal angle between range(q) and range(aq), and = be the maximal principal angle between range(aq) and range(a? Q). If A is non-singular Hermitian then there exists a permu- Theorem. tation so that j (i)? ^ i j max sin + tan : im j (i) j Proof. (Drmac 996a, Theorem 3, Proposition 5) In order to exhibit the connection to previous results we sketch the idea for the proof. First express the Ritz values as an additive perturbation. To this end dene the Hermitian perturbation E?(RQ + QR ): Then Q is an invariant subspace of A + E, (A + E)Q = QW; and the eigenvalues of W are eigenvalues of A + E. Now proceed as in the proof of Corollary.7 and look at the squares, x E x = x A A? E EA? Ax kea? k x A x for all x: Undo the squares using polar factors and the operator-monotonicity of the square root, and apply Theorem.. Hence the eigenvalues : : : n of A + E satisfy j i? i j max kea? k: in j i j Let ^ : : : ^ m be those i that are also eigenvalues of W ; and let be a permutation that numbers the eigenvalues of A corresponding to i rst. Then j (i)? ^ i j max kea? k: im j (i) j We still have to worry about kea? k. Write?EA? = (I? QQ )AQQ A? + QQ (I? AQQ A? ): Here QQ is the orthogonal projector onto range(q), while AQQ A? is the oblique projector onto range(aq) along range(q A? ). This expression for EA? appears in (Drmac 996a, Theorem 3). It can be bounded above by sin + tan. Therefore the relative error in the Ritz values of W = Q AQ is small if both subspace angles and are small. Things simplify when the matrix A is also positive-denite because there is only one angle to deal with.

I.C.F. Ipsen Let A be Hermitian positive-denite with Cholesky factorisation A = LL. Let = be the maximal principal angle between range(l Q) and range(l? Q). Theorem.3 If A is Hermitian positive-denite and if sin < then there exists a permutation so that j (i)? ^ i j max im j (i) j Proof. (Drmac 996a, Theorem 6) sin? sin : Theorem.3 can be extended to semi-denite matrices (Drmac and Hari 997). 3. Additive Perturbations for Singular Values Let B be a complex matrix. We want to estimate the absolute and the relative errors in the singular values of the perturbed matrix B + F. For deniteness we assume that B is tall and skinny, i.e. B is m n with m n (if this is not the case just consider B ). Perturbation bounds for singular values are usually derived by rst converting the singular value problem to an Hermitian eigenvalue problem. 3.. Converting Singular Values to Eigenvalues The singular value decomposition of a m n matrix B, m n, is B = U V ; where the left singular vector matrix U and the right singular vector matrix V are unitary matrices of order m and n, respectively. The non-negative diagonal matrix of order n contains the singular values i of B, where = B @... n : : : n : There are two popular ways to convert a singular value problem to an eigenvalue problem. The eigenvalues of C A ; m n m B A n B

are Relative Perturbation Results 3 ; : : : ; n ;? ; : : : ;? n ; ; : : : ; : {z } m?n Therefore the singular values of B are the n largest eigenvalues of A (Horn and Johnson 985, Theorem 7.3.7). The eigenvalues of B B are ; : : : ; n: Therefore the singular values of B are the positive square-roots of the eigenvalues of B B (Horn and Johnson 985, Lemma 7.3.). Since singular values are eigenvalues of a Hermitian matrix, they are wellconditioned in the absolute sense. 3.. Homan-Wielandt-Type Bounds We bound the sum of squares of all distances between the ith exact and perturbed singular values in terms of the Frobenius norm. The singular values of B and B + F are, respectively, : : : n ; ^ : : : ^ n : Converting the singular value problem to an eigenvalue problem a la x3. and applying the Homan-Wielandt Theorem for Hermitian matrices (:3) leads immediately to the absolute error bound vu u t X n i= j i? ^ i j kf k F : The relative bound below requires in addition that both matrices be nonsingular. Theorem 3. vu u t n X i= Let B and B + F be non-singular. If kf B? k < then j i? ^ i j i^ i k(i + F B? )? (I + F B? )? k F : Therefore the error in the singular values is small if I + F B? is close to being unitary (or orthogonal). This is case when B + F = (I + F B? ) B is more or less a unitary transformation away from B. Proof. (Li 994a, Theorem 4.3). 3.3. Weyl-Type Bounds We bound the worst-case distance between the ith exact and perturbed singular values in terms of the two-norm.

4 I.C.F. Ipsen The absolute error bound is an immediate consequence of Weyl's Perturbation Theorem (:4) j i? ^ i j kf k; i n: The corresponding relative bound below restricts the range of F but not its size. Here B y is the Moore-Penrose inverse of B. Theorem 3. If range(b + F ) range(b) then If range((b + F ) ) range(b ) then j i? ^ i j i kb y F k; i n: j i? ^ i j i kf B y k; i n: Proof. (Di Lena, Peluso and Piazza 993, Theorem.) We prove the rst bound, the proof for the second one is similar. Life would be easy if we could pull B out of F, say if F = BC for some matrix C. Then we could write B + F = B(I + C) and apply the sum and product inequalities for singular values (Horn and Johnson 985, page 43) to get the relative bound j i? ^ i j i kck: It turns out that the range condition is exactly what is needed to pull B out of F. This is because range(b + F ) range(b) implies F = BC for some C. This allows us to write F = BC = BB y B C = BB y F: Consequently, setting C B y F gives the desired result. When B has full column rank the second range condition in Theorem 3. is automatically satised. Corollary 3. If B has full column rank then j i? ^ i j i kf B y k; i n: Proof. (Di Lena et al. 993, Remark.) If B has full column rank n then its rows span n-space. Hence range((b + F ) ) range(b ) for any F, and the second relative bound in Theorem 3. holds. Therefore singular values of full-rank matrices are well-conditioned in the absolute as well as relative sense. This may sound implausible at rst, in particular when B has full rank while B +F is rank-decient. In this case all singular values of B are non-zero while at least one singular value of B + F is zero. Hence a zero singular value of B + F must have relative error equal to one. How can the singular values of B be well-conditioned? The answer

Relative Perturbation Results 5 is that in this case the relative perturbation kb y F k is large. The following example illustrates that kb y F k is large when B and B + F dier in rank. Example 3. Let B = @ A ; F = @? A ; where 6=. The rank of B is two, while the rank of B + F is one. The relative error in the singular value ^ = of B + F is equal to one because j? j=jj =. Since B y = we get kb y F k =. Corollary 3. gives min i! ; B y F = j i? ^j i : ;? Therefore Corollary 3. is tight for the zero singular values of B + F. Corollary 3. extends (Demmel and Veselic 99, Lemma.) to matrices that do not necessarily have full rank (Di Lena et al. 993, Remark.). When B is non-singular Corollary 3. implies that both range conditions in Theorem 3. hold automatically. Corollary 3. If B is non-singular then j i? ^ i j max minfkb? F k; kf B? kg: in i The following bound, for a dierent error measure, is similar to the Frobenius norm bound Theorem 3.. It requires that both B and B + F be non-singular. Theorem 3.3 Let B and B + F be non-singular. Then j i? ^ i j max p in i^ i k(i + F B? )? (I + F B? )? k: Proof. (Li 994a, Theorem 4.3). As in Theorem 3., the error in the singular values is small if I + F B? is close to being unitary (or orthogonal). This is case when B + F = (I + F B? ) B is more or less a unitary transformation away from B.

6 I.C.F. Ipsen 3.4. Congruence Transformations We start with one-sided grading of matrices B with full-column rank. That is, B = CD or B = DC, where D is non-singular. In the event D is diagonal, CD represents a column scaling while DC represents a row scaling. All relative singular value bounds presented so far are invariant under onesided grading from the appropriate side. That's because the bounds contain terms of the form B y F or F B y. Consider B y F, for instance. Grading from the right is sandwiched in the middle, between B y and F, and therefore cancels out. Let's rst look at the Homan-Wielandt-type bound for graded matrices, which follows directly from Theorem 3.. Corollary 3.3 Let B = CD and B + F = (C + G)D be non-singular. If kgc? k < then vu u t n X i= j i? ^ i j i^ i k(i + GC? )? (I + GC? )? k F : Proof. (Li 994a, Theorem 4.3) The grading is sandwiched in the middle of the relative perturbation F B?, and cancels out, F B? = (GD) (CD)? = G DD? C? = GC? : Moving right along to the two-norm, we see that Corollary 3. is invariant under grading from the right. Corollary 3.4 then If B = CD has full column rank, and if B +F = (C +G)D j i? ^ i j i kgc y k; i n: Proof. (Di Lena et al. 993, Remark.) Full column rank is needed to extract the grading matrix from the inverse, B y = (CD) y = D y C y = D? C y : Therefore the relative error in the singular values of B +F is small if there is a grading matrix D that causes the relative perturbation of the graded matrices kgc y k to be small. For instance, suppose B = CD has columns whose norms vary widely while the columns of C are almost orthonormal. If the perturbed matrix is scaled in the same way then the error bound in Corollary 3.4 ignores the scaling and acts as if it saw the well-behaved

Relative Perturbation Results 7 matrices C and C + G. Corollary 3.4 extends (Demmel and Veselic 99, Theorem.4) to a larger class of matrices. The other two-norm bound, Theorem 3.3, is also invariant under grading. Corollary 3.5 Let B = CD and B + F = (C + G)D be non-singular. If kgc? k < then j i? ^ i j max p in i^ i k(i + GC? )? (I + GC? )? k: Proof. (Li 994a, Theorem 4.3) Finally we present the only bound that is invariant under grading from both sides. It requires that the grading matrices be real diagonal; and it restricts the size of the perturbation more severely than the other bounds. Theorem 3.4 Let B = D l CD r and B + F = D l (C + G)D r be real symmetric, where D l and D r are real non-singular diagonal matrices. Among the singular values of all square submatrices of B, let be the smallest one. If kgk < then? kgk?kgk ^ i? i kgk i Proof. (Gu and Eisenstat 993, Corollary )? kgk (? kgk) i n: 4. Some Applications of Additive Perturbations We discuss Jacobi's method for computing singular values and eigenvalues, and deation of triangular and bidiagonal matrices. 4.. Jacobi's Method for Singular Values Jacobi's method is generally viewed as a method that computes eigenvalues and singular values to optimal accuracy. It was Jacobi's method that rst attracted attention to invariance of eigenvalue and singular values error bounds under congruence (Demmel and Veselic 99, Mathias 995a, Rosano et al. 968). We give a very intuitive plausibility argument, shoving many subtleties under the rug, to explain the high accuracy and invariance under grading of Jacobi's method. Our discussion runs along the lines of (Demmel 997, x5.4.3) and (Mathias 995a, x, 3). Other detailed accounts can be found in (Demmel and Veselic 99), (Drmac 996b). An attempt at a geometric interpretion of Jacobi's high accuracy is made in (Rosano et al. 968, pages 45-6). A one-sided Jacobi method computes the singular values of a tall and skinny matrix by applying a sequence of orthogonal transformations on the right side of the matrix. The duty of each orthogonal transformation is

8 I.C.F. Ipsen to orthogonalise two columns of the matrix. The method stops once all columns are suciently orthogonal to each other. At this point the singular values are approximated by the column norms, i.e. the Euclidian lengths of the columns. For simplicity assume that B is a real non-singular matrix of order n. Let D be a row scaling of B, i.e. B = DC, where D is diagonal. We show that the one-sided Jacobi method ignores the row scaling. When Jacobi applies an orthogonal transformation Q to B the outcome in oating point arithmetic is BQ + F. Corollary 3. implies that the singular values ^ i of BQ + F satisfy j i? ^ i j max kc? Gk kc? k kgk; in i where F = DG. Let's bound the squared error kgk. Round-o error analysis tells us that the error in the ith row is ke T i F k i ke T i Bk + O( ); where i depends on the matrix size n and > reects the machine accuracy. Now the crucial observation is that the orthogonal transformations happen on one side of the matrix and the scaling on the other side. Because Q operates on columns it does not mix up dierent rows and therefore preserves the row-scaling. This means we can pull the ith diagonal element of D out of e T i F, ke T i F k i ke T i Bk = i ke T i (DC)k = jd ii j i ke T i Ck: This gives a bound for the ith row of G, ke T i Gk = jd ii j? ke T i F k i ke T i Ck + O( ): The total error is therefore bounded by kgk kck + O( ); where depends on n. Therefore the error bound for the singular values of BQ + F is independent of the row-scaling, j i? ^ i j max (C) + O( ): in i This means Jacobi's method produces singular values of B but acts as if it saw C instead. That's good, particularly if D manages to pull out all the grading. Then all singular values of C have about the same magnitude and (C) is close to one. Therefore the above bound (C) tends to be on the order of machine accuracy, implying that the relative error in the singular values is on the order of machine accuracy.

Relative Perturbation Results 9 The argument is more complicated when the orthogonal transformations are applied on the same side as the scaling matrix. Fortunately the resulting error bounds do not tend to be much weaker (Mathias 995a, x4). 4.. Jacobi's Method for Eigenvalues A two-sided Jacobi method computes eigenvalues of a real symmetric positive-denite matrix by applying a sequence of orthogonal similarity transformations to the matrix. An orthogonal similarity transformation operates on two rows, i and j, and two columns, i and j, to zero out elements (i; j) and (j; i). The method stops once all o-diagonal elements are suciently small. At this point the eigenvalues are approximated by the diagonal elements. Let A be real symmetric positive-denite of order n, and A = DMD, where D is a non-singular diagonal matrix. The Jacobi method computes the eigenvalues of a matrix A + E. According to Corollary., the error bound for the eigenvalues ^ i of A + E is j i? max ^ i j km?= F M?= k km? k kf k; (4.) in j i j where E = DF D. One can show that the error is bounded by kf k kmk + O( ); where depends on n and > reects the machine accuracy. Therefore j i? max ^ i j (M) + O( ): in j i j This means, the relative error in the eigenvalues is small, provided the amplier (M) is small. The amplier (M) can be minimised via an appropriate choice of the scaling matrix D. If D = B @ p a... p ann then all diagonal elements of M are equal to one. Therefore (van der Sluis 969, Theorem 4.) (M) n min (SAS); S where the minimum ranges over all non-singular diagonal matrices S. This means, a diagonal scaling that makes all diagonal elements the same gives the minimal condition number among all diagonal scalings (up to a factor of matrix size n). We claimed above that the error kf k in (4:) is small. Let's examine in C A

3 I.C.F. Ipsen more detail why. The error F comes about because of oating point arithmetic and because of the fact that eigenvalues are approximated by diagonal elements when the o-diagonal elements are small but not necessarily zero. Let's ignore the round-o error, and let's ask why ignoring small o-diagonal elements results in a small kf k. The trick here is to be clever about what it means to be `small enough'. Suppose the Jacobi method has produced a matrix A whose o-diagonal elements are small compared to the corresponding diagonal elements, ja ij j p a ii a jj : This implies jm ij j where m ij are the elements of the graded matrix M = D? AD? and D is the above grading matrix with p a ii on the diagonal. Since the diagonal elements of M are equal to one, we can write M = I +F, where F contains all the o-diagonal elements of M and kf k (n? ): Therefore kf k is small, and (4:) implies that the error in the eigenvalues is bounded by j i? max ^ i j km? k (n? ): in j i j Furthermore one can bound km? k in terms of, km? k = k(i + F )? k Replacing this in the error bound gives j i? max ^ i j in j i j? kf k? (n? ) : (n? )? (n? ) : Therefore ignoring small o-diagonal elements produces a small relative error. The preceding arguments illustrate that Jacobi's method views a matrix in the best possible light, i.e. in its optimally scaled version. Therefore eigenvalues produced by Jacobi's method tend to have relative accuracy close to machine precision. This is as accurate as it gets. In this sense Jacobi's method is considered optimally accurate. One way to implement a two-sided Jacobi method is to apply a one-sided method to a Cholesky factor (Barlow and Demmel 99, Mathias 996). Let A be a Hermitian positive-denite matrix with Cholesky decomposition A = L L. The squares of the singular values of L are the eigenvalues of A. The singular values of L can be computed by the one-sided Jacobi method from x4.. The preliminary Cholesky factorisation does not harm the accuracy of the eigenvalues. Here is why. The computed Cholesky factor