deviation of D and D from similarity (Theorem 6.). The bound is tight when the perturbation is a similarity transformation D = D? or when ^ = 0. From

Similar documents
or H = UU = nx i=1 i u i u i ; where H is a non-singular Hermitian matrix of order n, = diag( i ) is a diagonal matrix whose diagonal elements are the

RITZ VALUE BOUNDS THAT EXPLOIT QUASI-SPARSITY

2 and bound the error in the ith eigenvector in terms of the relative gap, min j6=i j i? jj j i j j 1=2 : In general, this theory usually restricts H

PROOF OF TWO MATRIX THEOREMS VIA TRIANGULAR FACTORIZATIONS ROY MATHIAS

the Unitary Polar Factor æ Ren-Cang Li P.O. Box 2008, Bldg 6012

I.C.F. Ipsen small magnitude. Our goal is to give some intuition for what the bounds mean and why they hold. Suppose you have to compute an eigenvalue

Index. for generalized eigenvalue problem, butterfly form, 211

216 S. Chandrasearan and I.C.F. Isen Our results dier from those of Sun [14] in two asects: we assume that comuted eigenvalues or singular values are

STABILITY OF INVARIANT SUBSPACES OF COMMUTING MATRICES We obtain some further results for pairs of commuting matrices. We show that a pair of commutin

Perturbation results for nearly uncoupled Markov. chains with applications to iterative methods. Jesse L. Barlow. December 9, 1992.

Numerical Methods in Matrix Computations

A Note on Eigenvalues of Perturbed Hermitian Matrices

Matrix Algorithms. Volume II: Eigensystems. G. W. Stewart H1HJ1L. University of Maryland College Park, Maryland

1 Quasi-definite matrix

The Lanczos and conjugate gradient algorithms

Iterative Algorithm for Computing the Eigenvalues

Index. book 2009/5/27 page 121. (Page numbers set in bold type indicate the definition of an entry.)

Principal Angles Between Subspaces and Their Tangents

Abstract. Expressions and bounds are derived for the residual norm in GMRES. It is shown

Chapter 3 Transformations

Department of Mathematics Technical Report May 2000 ABSTRACT. for any matrix norm that is reduced by a pinching. In addition to known

ETNA Kent State University

Yimin Wei a,b,,1, Xiezhang Li c,2, Fanbin Bu d, Fuzhen Zhang e. Abstract

Key words. multiplicative perturbation, relative perturbation theory, relative distance, eigenvalue, singular value, graded matrix

Contents. Preface for the Instructor. Preface for the Student. xvii. Acknowledgments. 1 Vector Spaces 1 1.A R n and C n 2

Linear Algebra: Matrix Eigenvalue Problems

S.F. Xu (Department of Mathematics, Peking University, Beijing)

The Eigenvalue Problem: Perturbation Theory

Matrix Mathematics. Theory, Facts, and Formulas with Application to Linear Systems Theory. Dennis S. Bernstein

A fast randomized algorithm for overdetermined linear least-squares regression

Majorization for Changes in Ritz Values and Canonical Angles Between Subspaces (Part I and Part II)

Linear Algebra using Dirac Notation: Pt. 2

LOW-RANK MATRIX APPROXIMATIONS DO NOT NEED A SINGULAR VALUE GAP

Off-diagonal perturbation, first-order approximation and quadratic residual bounds for matrix eigenvalue problems

On Computing Accurate Singular. Values and Eigenvalues of. Matrices with Acyclic Graphs. James W. Demmel. William Gragg.

Numerical Methods for Solving Large Scale Eigenvalue Problems

Numerical Methods - Numerical Linear Algebra

Quantum Computing Lecture 2. Review of Linear Algebra

Lecture notes: Applied linear algebra Part 2. Version 1

Characterization of half-radial matrices

ANGLES BETWEEN SUBSPACES AND THE RAYLEIGH-RITZ METHOD. Peizhen Zhu. M.S., University of Colorado Denver, A thesis submitted to the

Spectral inequalities and equalities involving products of matrices

Math 102, Winter Final Exam Review. Chapter 1. Matrices and Gaussian Elimination

Bounding the Spectrum of Large Hermitian Matrices

DEN: Linear algebra numerical view (GEM: Gauss elimination method for reducing a full rank matrix to upper-triangular

UMIACS-TR July CS-TR 2721 Revised March Perturbation Theory for. Rectangular Matrix Pencils. G. W. Stewart.

MATH 423 Linear Algebra II Lecture 33: Diagonalization of normal operators.

DAVIS WIELANDT SHELLS OF NORMAL OPERATORS

Ole Christensen 3. October 20, Abstract. We point out some connections between the existing theories for

Computation of eigenvalues and singular values Recall that your solutions to these questions will not be collected or evaluated.

Singular Value Decomposition and Polar Form

Institute for Advanced Computer Studies. Department of Computer Science. On the Perturbation of. LU and Cholesky Factors. G. W.

Is there a Small Skew Cayley Transform with Zero Diagonal?

Ritz Value Bounds That Exploit Quasi-Sparsity

MATRIX AND LINEAR ALGEBR A Aided with MATLAB

Foundations of Matrix Analysis

Lecture Notes for Inf-Mat 3350/4350, Tom Lyche

Positive Denite Matrix. Ya Yan Lu 1. Department of Mathematics. City University of Hong Kong. Kowloon, Hong Kong. Abstract

Applied Mathematics 205. Unit V: Eigenvalue Problems. Lecturer: Dr. David Knezevic

GENERALIZED FINITE ALGORITHMS FOR CONSTRUCTING HERMITIAN MATRICES WITH PRESCRIBED DIAGONAL AND SPECTRUM

Lecture notes: Applied linear algebra Part 1. Version 2

Institute for Advanced Computer Studies. Department of Computer Science. On the Adjoint Matrix. G. W. Stewart y ABSTRACT

Numerical Methods I Eigenvalue Problems

Eigenvalues and Eigenvectors

MATH 425-Spring 2010 HOMEWORK ASSIGNMENTS

Multiplicative Perturbation Analysis for QR Factorizations

Introduction to Numerical Linear Algebra II

A NOTE ON MATRIX REFINEMENT EQUATIONS. Abstract. Renement equations involving matrix masks are receiving a lot of attention these days.

KEYWORDS. Numerical methods, generalized singular values, products of matrices, quotients of matrices. Introduction The two basic unitary decompositio

Although the rich theory of dierential equations can often be put to use, the dynamics of many of the proposed dierential systems are not completely u

SUMMARY OF MATH 1600

Convexity of the Joint Numerical Range

Mathematical foundations - linear algebra

Computing Eigenvalues and/or Eigenvectors;Part 2, The Power method and QR-algorithm

ETNA Kent State University

On Orthogonal Block Elimination. Christian Bischof and Xiaobai Sun. Argonne, IL Argonne Preprint MCS-P

First, we review some important facts on the location of eigenvalues of matrices.

Maximizing the numerical radii of matrices by permuting their entries

Preliminary/Qualifying Exam in Numerical Analysis (Math 502a) Spring 2012

linearly indepedent eigenvectors as the multiplicity of the root, but in general there may be no more than one. For further discussion, assume matrice

Institute for Advanced Computer Studies. Department of Computer Science. Two Algorithms for the The Ecient Computation of

then kaxk 1 = j a ij x j j ja ij jjx j j: Changing the order of summation, we can separate the summands, kaxk 1 ja ij jjx j j: let then c = max 1jn ja

Interlacing Inequalities for Totally Nonnegative Matrices

Applied Linear Algebra

Linear Algebra 2 Spectral Notes

Ir O D = D = ( ) Section 2.6 Example 1. (Bottom of page 119) dim(v ) = dim(l(v, W )) = dim(v ) dim(f ) = dim(v )

Detailed Proof of The PerronFrobenius Theorem

APPENDIX A. Background Mathematics. A.1 Linear Algebra. Vector algebra. Let x denote the n-dimensional column vector with components x 1 x 2.

A HARMONIC RESTARTED ARNOLDI ALGORITHM FOR CALCULATING EIGENVALUES AND DETERMINING MULTIPLICITY

arxiv: v3 [math.ra] 22 Aug 2014

MATH36001 Generalized Inverses and the SVD 2015

Direct methods for symmetric eigenvalue problems

Interpolating the arithmetic geometric mean inequality and its operator version

AN INVERSE EIGENVALUE PROBLEM AND AN ASSOCIATED APPROXIMATION PROBLEM FOR GENERALIZED K-CENTROHERMITIAN MATRICES

arxiv: v1 [math.na] 1 Sep 2018

The Eigenvalue Shift Technique and Its Eigenstructure Analysis of a Matrix

APPLIED NUMERICAL LINEAR ALGEBRA

QUASI-UNIFORMLY POSITIVE OPERATORS IN KREIN SPACE. Denitizable operators in Krein spaces have spectral properties similar to those

NEW ESTIMATES FOR RITZ VECTORS

Scientific Computing: An Introductory Survey

Transcription:

RELATIVE PERTURBATION RESULTS FOR EIGENVALUES AND EIGENVECTORS OF DIAGONALISABLE MATRICES STANLEY C. EISENSTAT AND ILSE C. F. IPSEN y Abstract. Let ^ and ^x be a perturbed eigenpair of a diagonalisable matrix A. The problem is to bound the error in ^ and ^x. We present one absolute perturbation bound and two relative perturbation bounds. The absolute perturbation bound implies that the condition number for ^x is the norm of an orthogonal projection of the reduced resolvent at ^. This condition number can be a lot less pessimistic than the traditional one, which is derived from a rst-order analysis. A further upper bound leads to an extension of Davis and Kahan's sin Theorem from Hermitian to diagonalisable matrices. The two relative perturbation bounds assume that ^ and ^x are an exact eigenpair of a perturbed matrix D AD, where D and D are non-singular, but D AD is not necessarily diagonalisable. We derive a bound on the relative error in ^ and a sin theorem based on a relative eigenvalue separation. The perturbation bounds contain both the deviation of D and D from similarity and the deviation of D from identity. Key words. eigenvalue, eigenvector, condition number, relative error, diagonalisable matrix, reduced resolvent, angle between subspaces AMS subject classication. 5A8, 5A4, 5A60, 65F5, 65F35, 65G99. Introduction. The protagonist is a complex, diagonalisable matrix A with eigendecomposition A = XX? ; where is a diagonal matrix whose diagonal elements i are the eigenvalues of A. The complex number ^, the `perturbed eigenvalue', approximates an eigenvalue of A; and the unit vector ^x, the `perturbed eigenvector', approximates an eigenvector. The problem is to bound the error in ^ and ^x. Our results extend the work in [9, 8] and are stronger and simpler than the results in [4, 5]. First we derive an upper bound on sin, where is the angle between ^x and the eigenspace corresponding to those eigenvalues closest to ^ (Theorem 4.). Our `condition number' is the norm of an orthogonal projection of the reduced resolvent, which can be arbitrarily much smaller than the norm of the reduced resolvent that appears in a rst-order analysis [3, Denition 3.3.]. An upper bound on this condition number leads to an extension of Davis and Kahan's sin Theorem [9, x6], [0, x] from Hermitian to diagonalisable matrices (Corollary 4.3). Compared to the Hermitian case, the new bound contains the additional factor, the condition number of a subset of the left eigenvectors. This is an exact rather than a rst-order bound and holds without any assumptions on the perturbations. Next we derive two relative perturbation bounds based on a multiplicative perturbation model. We assume that ^ and ^x are an exact eigenpair of a perturbed matrix D AD, where D and D are non-singular matrices and D AD need not be diagonalisable. We prove a bound on the relative error in ^ that contains the Department of Computer Science, Yale University, P. O. Box 0885, New Haven, CT 0650-885 (eisenstat-stan@cs.yale.edu). The research of this author was supported in part by NSF grant CCR-94009. y Center for Research in Scientic Computation, Department of Mathematics, North Carolina State University, P. O. Box 805, Raleigh, NC 7695-805 (ipsen@math.ncsu.edu). The research of this author was supported in part by NSF grants CCR-90853 and CCR-94009.

deviation of D and D from similarity (Theorem 6.). The bound is tight when the perturbation is a similarity transformation D = D? or when ^ = 0. From the absolute sin theorem described above we derive a relative bound on sin that consists of two summands: the deviation of D from identity; and the deviation of D and D from similarity, amplied by and a relative eigenvalue separation (Theorem 7.). We conclude that the eigenvectors of diagonalisable matrices are wellconditioned when the perturbation is caused by a similarity transformation; and that the null vectors of diagonalisable matrices are well-conditioned when the perturbation is multiplicative (Corollary 7.3). Our relative perturbation bounds are no stronger than the traditional absolute perturbation bounds because we use the traditional bounds to derive them. Notation. k k denotes the two-norm. If V is a matrix of full column rank, then is the Moore Penrose-inverse of V and is its two-norm condition number. V y (V V )? V (V ) kv k kv y k. Related Work. In 959 Ostrowski gave the rst relative perturbation bounds for eigenvalues by considering multiplicative perturbations D AD of Hermitian matrices A [8],[, Theorem 4.5.9]. He bounded the ratio of exact and perturbed eigenvalues in terms of the smallest and largest eigenvalues of D D, min (D D) i (A) i (D AD) max (D D) i (A): About thirty years later bounds on the ratio of exact and perturbed eigenvalues appeared for dierent types of structured matrices, but without any reference to Ostrowski's bounds. Inspired by Kahan's work in 966 [3], Demmel and Kahan [] and Deift, Demmel, Li, and Tomei [] derive relative perturbation bounds for eigenvalues of real symmetric tridiagonal matrices with zero diagonal. Demmel and Gragg [4] extend these bounds to real symmetric positive-denite matrices with acyclic graphs. Deift et al. [] also derive bounds on the error angles of eigenvectors in terms of a relative eigenvalue separation. Subsequent work on perturbation theory has extended these results to larger classes of matrices. These bounds are derived for two dierent perturbation models: component-wise relative perturbations and multiplicative perturbations. In the context of (a superset of) component-wise relative perturbations, Barlow and Demmel [] derive bounds for real symmetric scaled diagonally dominant matrices; and Demmel and Veselic [3] derive bounds for real symmetric positive-denite matrices. Mathias [6] unies the results by Demmel and Veselic [3] with the results by Barlow and Demmel [] for positive-denite matrices. Veselic and Slapnicar [33, 37] extend these bounds to indenite Hermitian matrices. In addition they prove a relative Gerschgorin theorem for diagonalisable matrices. Pietzsch [9] adapts the eigenvalue bounds in [, 3, 33, 37] to real skew-symmetric matrices A =?A T by observing that ia =?ia T is Hermitian. Drmac [5, 6] gives residual bounds for Hermitian matrices. In the context of multiplicative perturbations of Hermitian matrices, Eisenstat and Ipsen [9, 8] bound the relative error in the eigenvalues and the error angle

between subspaces. Similarly, Li [4, 5] derives bounds for eigenvalues of diagonalisable matrices and for eigenvectors of Hermitian matrices. About ten years earlier, Price [30] had already recommended multiplicative perturbations in the form of matrix exponentials to bound the absolute error in matrix transformations. Applications include the computation of eigenvalues of symmetric matrices by Jacobi's method [30, x3] and the solution of bordered systems of linear equations []. Finally, Drmac and Hari [5, 6, 7] derive relative perturbation bounds for Hermitian matrices when the perturbed eigenvalues are Ritz values. 3. Setting the Stage. To establish a correspondence between perturbed and exact quantities, we partition the eigenvalues of A so that contains all eigenvalues closest to ^ and contains the remaining eigenvalues; i.e., where = k? ^Ik = min i j i? ^j; and k? ^Ik < =k(? ^I)? k: Thus k? ^Ik represents the absolute error in ^. The `absolute gap' absgap =k(? ^I)? k = min j( ) ii? ^j i is the absolute separation of ^ from the eigenvalues of. It approximates the absolute separation of the desired eigenvalues from the remaining eigenvalues. We partition the eigenvectors conformally with, X = ( X X ) and X? = Y Y ; where denotes the conjugate transpose. Then ^x can be regarded as approximating a unit vector in range(x ). Since consists of all eigenvalues on a circle of minimal radius around ^, in general range(x ) is associated with several distinct, possibly multiple eigenvalues. To determine how well the unit vector ^x approximates a vector in the right eigenspace range(x ), we measure the angle between ^x and range(x ). More precisely, let 0 = be the largest principal angle between range(^x) and range(x ) [0, x.4.3]. Then sin is obtained from the orthogonal projection P I? (X y ) X : of ^x onto range(x )?, the orthogonal complement of range(x ). Theorem 3. ([9, x6], [0, p. 0], [36, (.3)]). sin = kp ^xk: 3 When A is normal then range(x ) = range(x )? and X has orthonormal columns. Hence P = X X and sin = kx X ^xk = kx ^xk:

4 When A is non-normal, X may not be the orthogonal complement of X, and the columns of X may not be orthonormal. Our perturbation bounds for eigenvectors of diagonalisable matrices are derived by bounding sin from above. The following result expresses P in terms of a basis Y for the left eigenspace range(y ). Lemma 3.. The orthogonal projector P satises P = (Y y ) Y : Proof. X? X = I implies range(y ) = range(x )?. Thus P = I? (X y ) X = (Y y ) Y : In general the orthogonal projector P (oblique) spectral projector in Lemma 3. is not identical to the Z X Y : Although both projectors share the same null space, range(x ), their ranges are dierent. One projects on the left eigenspace, while the other projects on the right one: range(p ) = range(y ) and range(z) = range(x ): 4. A sin Theorem for Diagonalisable Matrices. The traditional perturbation bound for the eigenvectors of a Hermitian matrix is Davis and Kahan's sin Theorem [9, x6], [0, x], sin krk=absgap; where r (A? ^I)^x is the residual. That is, for a Hermitian matrix the condition number of the perturbed eigenvector is inversely proportional to the separation between ^ and the remaining eigenvalues. We extend this result to diagonalisable matrices by rst deriving a sin theorem whose condition number is an orthogonal projection of the `reduced resolvent at ^', (see [7, x.], [3, xiii.3.]). Theorem 4.. The angle satises S(^) X (? ^I)? Y sin kp S(^)k krk: Proof. As our goal is a residual bound for the angle, we write r = (A? ^I)^x? = X ^I? ^I X?^x = X? ^I? ^I ^c^s ; where ^c Y ^x and ^s Y ^x. Multiplying both sides by (? ^I)? Y gives ^s = (? ^I)? Y r:

5 Then Lemma 3. and Y X = I imply P ^x = (Y y ) Y ^x = (Y y )^s = (Y y ) Y X (? ^I)? Y r = P S(^)r: Finally, Theorem 3. implies sin = kp ^xk = kp S(^)rk kp S(^)k krk: When A is normal then P = X X and X has orthonormal columns. Hence kp S(^)k = ks(^)k = =absgap; and Theorem 4. reduces to Davis and Kahan's sin Theorem. Moreover, it is not dicult to extend Theorem 4. to eigenspaces of larger dimension. Let's compare kp S(^)k to a similar condition number: A rst-order perturbation analysis for arbitary matrices yields as condition number for ^x the norm of the reduced resolvent, ks(^)k [3, Denition III.3.]. When A is diagonalisable and ^ is simple, an expansion of S(^) gives information about the sensitivity of the eigenspace corresponding to ^x [0, x7..4], [3, xiii.3.]. The following example shows that our condition number kp S(^)k can be much less pessimistic than ks(^)k. Remark. kp S(^)k can be much smaller than ks(^)k. Since P is an orthogonal projector, kp k and kp S(^)k ks(^)k. Consider the case = I. Then the reduced resolvent is a multiple of the spectral projector Z so that S(^) =? ^ Z and P S(^) =? ^ P: Thus ks(^)k = j? ^j kzk and kp S(^)k = j? ^j : Let X = QL be a QL decomposition of X, where Q is unitary and L is lower triangular. Partitioning Q and L conformally with X, L Q = ( Q Q ) and L = ; L L gives Thus and X = Q L and Y = (?L? L L? L? ) Q : Z = X Y = Q (?L L? I ) Q kzk = k (?L L? I ) k kl L? k: Choosing L and L appropriately makes kzk, and hence ks(^)k, arbitrarily large while kp S(^)k remains constant.

6 To extract the eigenvalue separation from kp S(^)k, we derive an upper bound. Corollary 4.. The orthogonal projection of the reduced resolvent satises where (Y ) : Proof. The result follows from kp S(^)k =absgap kp S(^)k = k(y y ) Y X (? ^I)? Y k since Y X = I. Corollary 4.3. The angle satises = k(y y ) (? ^I)? Y k k(y y ) k k(? ^I)? k ky k sin krk=absgap: The worse conditioned the left eigenvectors Y and hence X?, the larger is likely to be. If Y has orthonormal columns, which occurs when A is normal, then = and Theorem 4.3 reduces to Davis and Kahan's sin Theorem. Although Davis and Kahan considered only Hermitian matrices, the ideas for this extension to diagonalisable matrices are already present in [0, Theorem 6.] (essentially their factor = corresponds to our above). Related Work. Bhatia, Davis and McIntosh [4] prove sin theorems in the more general context of normal operators on Hilbert spaces. Their bounds are of the same form as Corollary 4.3, sin c ka ^X? ^XBk=absgap; where A and B are normal operators and ^X represents a perturbed subspace of any dimension. Determining the value of the positive constant c amounts to solving a minimisation problem for functions in L [3]. The existing results cited below hold in the general situation when the dimension of the perturbed subspace is arbitrary. Here we restrict the discussion to our context, which means exact and the perturbed eigenspace have the same dimension. Hence X consists of a single unit vector, say x. For diagonalisable matrices Varah [36, Theorem.] shows that if krk is suciently small then sin (X) krk=absgap: For general, possibly defective matrices, Stewart [34, Theorem 4.], [35, Theorem 4.] derives a tan bound for the case when ^ is a Rayleigh quotient and krk is suciently small. And Ruhe [3, Corollary ] bounds the sine of the angle between ^x and a singular vector associated with the smallest singular value of A q (here absgap is replaced by the gap n?? n between the two smallest singular values of A). Instead of sin one can also bound kx? ^xk. Such bounds are slightly stronger because they imply sin theorems: kx? ^xk = sin cos sin = sin :

For instance, Bohte's bound [5, ()] and its a posteriori version [6, (.)] are of the form kx? ^xk (X) krk=absgap: For normal matrices Wilkinson [38, (3.54.4)] shows that kx? ^xk krk absgap s + krk ; absgap where is some scalar with jj =. Since = for normal matrices, Corollary 4.3 is stronger than Wilkinson's bound. 5. Multiplicative Perturbations. Assume that the perturbed quantities ^ and ^x are an exact eigenvalue and eigenvector of a matrix D AD, where D and D are non-singular matrices; i.e., (D AD ) ^x = ^^x: The matrix D AD is not required to be diagonalisable. When D = D? then D AD is a similarity transformation of A, which means that A and D AD have the same eigenvalues. When D = D then D AD is a congruence transformation of A, which means for Hermitian A that A and D AD have the same inertia. The derivation of relative perturbation bounds for such multiplicative perturbations is based on the following idea. From (D AD )^x = ^^x follows Setting z D ^x=kd ^xk, we have Thus the residual for ^ and z is A(D ^x) = Az = ^D? D? (D ^x): ^D? D? z: f Az? ^z = ^ (D? D?? I)z = ^ (D?? D )^x=kd ^xk: The upper bounds kfk j^j ki? D? D? k and kfk j^j kd?? D k=kd ^xk consist of two parts: the factor j^j, which is responsible for the perturbation bound being relative, and the deviation of D and D from similarity. The factor kd??d k is an absolute deviation from similarity, while the factor ki? D? D?k constitutes a relative deviation because is a dierence relative to D. I? D? D? = (D? D? )D? 7

8 6. Eigenvalues of Multiplicative Perturbations. We now bound the relative error in ^ in terms of the relative deviation from similarity of the matrices D and D. Theorem 6.. The perturbed eigenvalue ^ satises k? ^Ik j^j (X) ki? D? D? k: Proof. Apply the Bauer-Fike Theorem with residual bound [, Theorem IIIa], [7, x3..7] to the residual f for ^ and z to get k? ^Ik = min j i? ^j (X) kfk i and bound kfk as above. In the case of a similarity transformation D = D?, the eigenvalues are preserved and the bound above is zero, which is tight. When ^ = 0, A itself must be singular and the bound is zero, which is again tight. It is not clear how to extend Theorem 6. to general, possibly defective matrices and end up with pleasant bounds. For instance, the generalisation of the Bauer-Fike Theorem [8, Theorem 3A] gives an upper bound on the relative error that contains a factor =j^j (p?)=p, where p is the index of nilpotency of the strictly upper triangular matrix in a Schur decomposition of A. Related Work. Veselic and Slapnicar [37] prove a relative Gerschgorin theorem for perturbed matrices A + A with kaxk kaxk for any x. They show that the eigenvalues ^ of A + A lie in the union of disks [37, Theorem 3.7] j i? ^j j i j (X) : Li [4] establishes a one-to-one pairing between the exact and perturbed eigenvalues and bounds the errors in all eigenvalue pairs simultaneously: Theorem 6. (Li [4, Theorems 6. and 6.4]). Let D AD be diagonalisable with eigenvalues ^ i eigenvector matrix ^X. Then there exists a permutation such that vu u t X i j i? ^ (i) j i + ^ (i) (X) ( ^X) min qki? D k F + ki? D? k F q ; ki? D? k F + ki? D k F : If in addition all i and all ^ i are nonnegative then for any p-norm, p, max i q j i? ^ (i) j q q i + ^ q (i) (X) ( ^X) min n q s q ki? D k s p + ki? D? s ks p ; ki? D? ks p + ki? D k s p where s = q=(q? ). In contrast to Theorem 6., these bounds are non-zero for nontrivial similarity transformations D = D? 6= I. o ;

In the special case D = I or D = I, Li [4, Theorem 6.6] proves a variation of Theorem 6. where the upper bound contains the exact eigenvalue instead of the perturbed one. The following two bounds can be interpreted as a relative version of Weyl's theorem [, Theorem 4.3.]. They follow from the relative bounds by Ostrowski [8], [, Theorem 4.5.9] mentioned in the beginning of x. Theorem 6.3 (Eisenstat and Ipsen [9, Theorem.]). Let A be Hermitian with i in increasing order, let D be non-singular, and let ^i be the eigenvalues of D AD in increasing order. Then j^ i? i j j i j ki? D Dk for all i: 9 Corollary 6.4 (Li [4, Theorem 7.]). Under the assumptions of Theorem 6.3, j i? ^ i j j^ i j ki? D? D? k for all i: Proof. Reverse the r^oles of A and D AD in Theorem 6.3. This bound is stronger than the specialisation of Theorem 6. to Hermitian matrices because it also establishes a one-to-one correspondence between perturbed and exact eigenvalues. 7. Eigenvectors of Multiplicative Perturbations. We use the absolute perturbation bound in Corollary 4.3 to derive bounds on sin in terms of a `relative gap', i.e., the relative separation of ^ from the remaining eigenvalues. We dene the relative gap via its inverse, =relgap j^j=absgap 0; to ensure that the denition is valid even when ^ = 0. Since multiplicative perturbations lead to a residual for z = D ^x=kd ^xk rather than for ^x, a straightforward application of Corollary 4.3 results in a bound on the angle between z and range(x ) rather than the desired bound on. The following lemma shows how to adjust this bound. Lemma 7.. Let 0 = be the largest principal angle between z and range(x ). Then and sin kd ^xk sin + kd? Ik; sin sin + kd? Ik: Proof. To derive the rst bound write P ^x = P D ^x + P (I? D )^x = kd ^xk P z + P (I? D )^x; and note that sin = kp ^xk and sin = kp zk:

0 The second bound follows from the rst if kd ^xk, so assume that kd ^xk >. Then Note that sin = kp ^xk kp zk + kp (z? ^x)k sin + kz? ^xk: k(d? I)^xk = k(d ^x? z) + (z? ^x)k Since ^x and z are unit vectors, Thus = kz? ^xk + kd ^x? zk + Re (D ^x? z) (z? ^x): Re (D ^x? z) (z? ^x) = (kd ^xk? ) Re (? z ^x) 0: k(d? I)^xk kz? ^xk ; and the result follows from the inequality k(d? I)^xk kd? Ik. In general any bound on sin derived from Lemma 7. must consist of two summands. The theorem below bounds the rst summand by the (absolute or relative) deviation of D and D from a similarity transformation, amplied by and by the relative eigenvalue separation; and bounds the second summand by the (absolute and relative) deviation of the similarity transformation from the identity. Theorem 7.. Let Then kd?? D k; ki? D? D? k; and (Y ): sin minf ; g relgap + ki? D k; where if kd? k and if kd k. Proof. Apply Corollary 4.3 to the residual f for ^ and z to get sin kfk=absgap: To derive the bound with, bound kfk by kfk j^j kd?? D k=kd ^xk = j^j =kd ^xk and use the rst bound in Lemma 7.. To derive the bound with, bound kfk by kfk j^j ki? D? D? k = j^j and use the second bound in Lemma 7.. Finally, kd k implies while kd? k implies = kd? D? k ki? D? D? kd kk ; = ki? D? D? k kd? D? k kd? k :

In some cases the rst summand can be omitted. Corollary 7.3. If D = D? or ^ = 0, then sin = kp (I? D )^xk ki? D k: Proof. Consider the case D = D?. Then D AD ^x = ^^x implies AD^x = ^D^x, i.e., ^ and D^x are an exact eigenpair of A. According to the partitioning dened in x3, we must have = ^I, so D ^x must be in range(x ) = range(y )?. Thus P D ^x = 0 and we have sin = kp ^xk = kp (D ^x? ^x)k ki? D k: Next consider the case ^ = 0. Then D AD ^x = 0^x implies that D? AD ^x = 0 ^x, since D and D are non-singular. Hence ^ and ^x are an exact eigenpair of a similarity transformation of A, and we are back to the rst case. For similarity transformations D = D?, the bounds in Corollary 7.3 are essentially tight. The error angle is bounded by the relative deviation of D from identity, without any amplication by or by a relative gap. Therefore we can say that the eigenvectors of diagonalisable matrices are well-conditioned when the perturbation is caused by a similarity transformation. Moreover, when ^ = 0, the bounds in Corollary 7.3 are again essentially tight. This means that the null vectors of diagonalisable matrices are also well-conditioned when the perturbation is multiplicative. Related Work. Eisenstat and Ipsen [9, 8] bound sin for Hermitian matrices. The result below corresponds to a specialisation of Theorem 7.. Corollary 7.4 (Eisenstat and Ipsen [8, Theorem.]). Let A be Hermitian with i in increasing order, let D be non-singular, and let ^ i be the eigenvalues of D AD in increasing order. Also let 0 i = be the largest principal angle between an eigenvector associated with ^ i and the eigenspace associated with i. Then sin i ki? D? D? k relgap i + ki? Dk; where relgap i j^i j min j 6= i j j? ^ i j = max j 6= i j^i j j j? ^ i j : When the perturbed eigenspace ^X and the exact eigenspace X have the same dimension k, a similar bound holds for the largest principal angle k sin k between the two spaces. Theorem 7.5 (Eisenstat and Ipsen [8, Variant of Theorem 3.]). Let A be Hermitian, let D be non-singular, and let the k columns of ^X be an orthonormal basis for the eigenspace associated with the eigenvalues ^ ^ k of D AD. Let the eigenvalues of A be ordered such that i is the l i th eigenvalue of A whenever ^ i is the l i th eigenvalue of D AD for i k. Then k sin k p ki? D? D? k k + relgap ki? Dk ;

where sin P ^X; and relgap max max j^ i j ik j>k j j? ^ i j : Li [5] derives the following bound on eigenspace perturbations for Hermitian matrices. Theorem 7.6 (Li [5, Theorem 3.]). Under the assumptions of Theorem 7.5, where k sin k F k(d? D? )X k F relgap 0 + k(i? D )X k F ; relgap max max j^ j j 0 ik j>k j i? ^ j j : It is not clear what the relation between Theorems 7.5 or 7.6 is, and whether one is stronger than the other. Acknowledgement. We thank Axel Ruhe and Kresimir Veselic for helpful discussions, and Ren-Cang Li and the referees for suggestions that improved the paper. REFERENCES [] J. Barlow and J. W. Demmel, Computing accurate eigensystems of scaled diagonally dominant matrices, SIAM J. Numer. Anal., 7 (990), pp. 76{9. [] F. Bauer and C. Fike, Norms and exclusion theorems, Numer. Math., (960), pp. 37{4. [3] R. Bhatia, C. Davis, and P. Koosis, An extremal problem in Fourier analysis with applications in operator theory, J. Funct. Anal., 8 (989), pp. 38{50. [4] R. Bhatia, C. Davis, and A. McIntosh, Perturbation of spectral subspaces and solution of linear operator equations, Linear Algebra Appl., 5/53 (983), pp. 45{57. [5] Z. Bohte, A posteriori error bounds for eigensystems of matrices, in Numerical Methods and Approximation Theory III, University of Nis, Yugoslavia, 988, pp. 45{67. [6], Computable error bounds for approximate eigensystems of matrices, in VII Conference on Applied Mathematics, R. Scitovski, ed., University of Osijek, Croatia, 990, pp. 9{37. [7] F. Chatelin, Valeurs Propres de Matrices, Masson, Paris, 986. [8] K. Chu, Generalization of the Bauer-Fike theorem, Numer. Math., 49 (986), pp. 685{9. [9] C. Davis and W. Kahan, Some new bounds on perturbation of subspaces, Bull. Amer. Math. Soc., 75 (969), pp. 863{8. [0], The rotation of eigenvectors by a perturbation, III, SIAM J. Numer. Anal., 7 (970), pp. {46. [] P. Deift, J. W. Demmel, L.-C. Li, and C. Tomei, The bidiagonal singular value decomposition and Hamiltonian mechanics, SIAM J. Numer. Anal., 8 (99), pp. 463{56. [] J. Demmel and W. Kahan, Accurate singular values of bidiagonal matrices, SIAM J. Sci. Stat. Comput., (990), pp. 873{9. [3] J. Demmel and K. Veselic, Jacobi's method is more accurate than QR, SIAM J. Matrix Anal. Appl., 3 (99), pp. 04{45. [4] J. W. Demmel and W. Gragg, On computing accurate singular values and eigenvalues of matrices with acyclic graphs, Lin. Alg. Applics., 85 (993), pp. 03{7. [5] Z. Drmac, Computing the Singular and the Generalized Singular Values, PhD thesis, Fachbereich Mathematik, Fernuniversitat Gesamthochschule Hagen, Germany, 994.

[6], On relative residual bounds for the eigenvalues of a Hermitian matrix, Linear Algebra Appl., 44 (996), pp. 55{64. [7] Z. Drmac and V. Hari, Relative a posteriori residual bounds for the eigenvalues of a Hermitian matrix, tech. rep., Department of Mathematics, University of Hagen, Germany, 994. [8] S. Eisenstat and I. Ipsen, Relative perturbation bounds for eigenspaces and singular vector subspaces, in Applied Linear Algebra, SIAM, 994, pp. 6{5. [9], Relative perturbation techniques for singular value problems, SIAM J. Numer. Anal., 3 (995), pp. 97{88. [0] G. Golub and C. van Loan, Matrix Computations, The Johns Hopkins Press, Baltimore, second ed., 989. [] W. Govaerts and J. Pryce, Block elimination with one renement solves bordered linear systems accurately, BIT, 30 (990), pp. 490{507. [] R. A. Horn and C. R. Johnson, Matrix Analysis, Cambridge University Press, Cambridge, 985. [3] W. Kahan, Accurate eigenvalues of a symmetric tridiagonal matrix, Technical Report No. CS-4, Department of Computer Science, Stanford University, July 966 (revised June 968). [4] R. Li, Relative perturbation theory: (I), eigenvalue variations, LAPACK working note 84, revised, Computer Science Department, University of Tennessee, Knoxville, January 996. [5], Relative perturbation theory: (II), eigenspace variations, LAPACK working note 85, revised, Computer Science Department, University of Tennessee, Knoxville, April 996. [6] R. Mathias, Spectral perturbation bounds for graded positive denite matrices, technical report, Department of Mathematics, College of William and Mary, Williamsburg, VA, May 994. [7] J. Ortega, Numerical Analysis, A Second Course, Academic Press, New York, 97. [8] A. Ostrowski, A quantitative formulation of Sylvester's law of inertia, Proc. Nat. Acad,. Sci., 45 (959), pp. 740{4. [9] E. Pietzsch, Genaue Eigenwertberechnung Nichtsingularer Schiefsymmetrischer Matrizen, PhD thesis, Fachbereich Mathematik, Fernuniversitat Gesamthochschule Hagen, Germany, 993. [30] J. Pryce, Multiplicative error analysis of matrix transformation algorithms, IMA Journal of Numerical Analysis, 5 (985), pp. 437{45. [3] A. Ruhe, Perturbation bounds for means of eigenvalues and invariant subspaces, Bit, 0 (970), pp. 343{54. [3] Y. Saad, Numerical Methods for Large Eigenvalue Problems, Manchester University Press, New York, 99. [33] I. Slapnicar, Accurate Symmetric Eigenreduction by a Jacobi Method, PhD thesis, Fachbereich Mathematik, Fernuniversitat Gesamthochschule Hagen, Germany, 99. [34] G. Stewart, Error bounds for approximate invariant subspaces, of closed linear operators, SIAM J. Numer. Anal., 8 (97), pp. 796{808. [35], Error and perturbation bounds for subspaces associated with certain eigenvalue problems, SIAM Review, 5 (973), pp. 77{64. [36] J. Varah, Computing invariant subspaces of a general matrix when the eigensystem is poorly conditioned, Math. Comp., 4 (970), pp. 37{49. [37] K. Veselic and I. Slapnicar, Floating-point perturbations of Hermitian matrices, Linear Algebra Appl., 95 (993), pp. 8{6. [38] J. Wilkinson, The Algebraic Eigenvalue Problem, Oxford University Press, Oxford, 965. 3