c 2009 Society for Industrial and Applied Mathematics

Similar documents
Arnoldi Methods in SLEPc

On the Ritz values of normal matrices

A HARMONIC RESTARTED ARNOLDI ALGORITHM FOR CALCULATING EIGENVALUES AND DETERMINING MULTIPLICITY

Computation of eigenvalues and singular values Recall that your solutions to these questions will not be collected or evaluated.

Last Time. Social Network Graphs Betweenness. Graph Laplacian. Girvan-Newman Algorithm. Spectral Bisection

ECS231 Handout Subspace projection methods for Solving Large-Scale Eigenvalue Problems. Part I: Review of basic theory of eigenvalue problems

On prescribing Ritz values and GMRES residual norms generated by Arnoldi processes

Math 504 (Fall 2011) 1. (*) Consider the matrices

PERTURBED ARNOLDI FOR COMPUTING MULTIPLE EIGENVALUES

Numerical Methods I Eigenvalue Problems

On the influence of eigenvalues on Bi-CG residual norms

Iterative methods for symmetric eigenvalue problems

Approximating the matrix exponential of an advection-diffusion operator using the incomplete orthogonalization method

Matrix Algorithms. Volume II: Eigensystems. G. W. Stewart H1HJ1L. University of Maryland College Park, Maryland

ON ORTHOGONAL REDUCTION TO HESSENBERG FORM WITH SMALL BANDWIDTH

Eigenvalue Problems. Eigenvalue problems occur in many areas of science and engineering, such as structural analysis

Scientific Computing: An Introductory Survey

AMS526: Numerical Analysis I (Numerical Linear Algebra)

DELFT UNIVERSITY OF TECHNOLOGY

The Lanczos and conjugate gradient algorithms

Applied Mathematics 205. Unit V: Eigenvalue Problems. Lecturer: Dr. David Knezevic

Index. for generalized eigenvalue problem, butterfly form, 211

Solution of eigenvalue problems. Subspace iteration, The symmetric Lanczos algorithm. Harmonic Ritz values, Jacobi-Davidson s method

4.8 Arnoldi Iteration, Krylov Subspaces and GMRES

Preconditioned inverse iteration and shift-invert Arnoldi method

Solution of eigenvalue problems. Subspace iteration, The symmetric Lanczos algorithm. Harmonic Ritz values, Jacobi-Davidson s method

GMRES ON (NEARLY) SINGULAR SYSTEMS

Math 405: Numerical Methods for Differential Equations 2016 W1 Topics 10: Matrix Eigenvalues and the Symmetric QR Algorithm

Augmented GMRES-type methods

The quadratic eigenvalue problem (QEP) is to find scalars λ and nonzero vectors u satisfying

Krylov Subspaces. Lab 1. The Arnoldi Iteration

EIGIFP: A MATLAB Program for Solving Large Symmetric Generalized Eigenvalue Problems

LARGE SPARSE EIGENVALUE PROBLEMS. General Tools for Solving Large Eigen-Problems

ABSTRACT. Professor G.W. Stewart

LARGE SPARSE EIGENVALUE PROBLEMS

Course Notes: Week 1

Convergence Theory for Iterative Eigensolvers

EVALUATION OF ACCELERATION TECHNIQUES FOR THE RESTARTED ARNOLDI METHOD

Computation of Smallest Eigenvalues using Spectral Schur Complements

1 Extrapolation: A Hint of Things to Come

Krylov subspace projection methods

Key words. Krylov subspace methods, Arnoldi algorithm, eigenvalue computations, containment gap, pseudospectra

THE STABILITY OF GMRES CONVERGENCE, WITH APPLICATION TO APPROXIMATE DEFLATION PRECONDITIONING. 16 September 2011

I. Multiple Choice Questions (Answer any eight)

EIGENVALUE PROBLEMS. EIGENVALUE PROBLEMS p. 1/4

An Asynchronous Algorithm on NetSolve Global Computing System

Application of Lanczos and Schur vectors in structural dynamics

Solving large sparse eigenvalue problems

Finding Rightmost Eigenvalues of Large, Sparse, Nonsymmetric Parameterized Eigenvalue Problems

EECS 275 Matrix Computation

Chasing the Bulge. Sebastian Gant 5/19/ The Reduction to Hessenberg Form 3

Block Bidiagonal Decomposition and Least Squares Problems

Orthogonal iteration to QR

Convergence of Polynomial Restart Krylov Methods for Eigenvalue Computations

RITZ VALUE BOUNDS THAT EXPLOIT QUASI-SPARSITY

Krylov Subspaces. The order-n Krylov subspace of A generated by x is

MULTIGRID ARNOLDI FOR EIGENVALUES

The restarted QR-algorithm for eigenvalue computation of structured matrices

ON RESTARTING THE ARNOLDI METHOD FOR LARGE NONSYMMETRIC EIGENVALUE PROBLEMS

Key words. linear equations, polynomial preconditioning, nonsymmetric Lanczos, BiCGStab, IDR

Numerical Methods in Matrix Computations

APPLIED NUMERICAL LINEAR ALGEBRA

5.3 The Power Method Approximation of the Eigenvalue of Largest Module

Key words. conjugate gradients, normwise backward error, incremental norm estimation.

AMS526: Numerical Analysis I (Numerical Linear Algebra) Lecture 23: GMRES and Other Krylov Subspace Methods; Preconditioning

On the solution of large Sylvester-observer equations

Eigenvalue Problems CHAPTER 1 : PRELIMINARIES

Krylov Space Methods. Nonstationary sounds good. Radu Trîmbiţaş ( Babeş-Bolyai University) Krylov Space Methods 1 / 17

On the loss of orthogonality in the Gram-Schmidt orthogonalization process

THE QR ALGORITHM REVISITED

On Solving Large Algebraic. Riccati Matrix Equations

Eigenvalue Problems and Singular Value Decomposition

LINEAR ALGEBRA 1, 2012-I PARTIAL EXAM 3 SOLUTIONS TO PRACTICE PROBLEMS

ARPACK. Dick Kachuma & Alex Prideaux. November 3, Oxford University Computing Laboratory

MATH 31 - ADDITIONAL PRACTICE PROBLEMS FOR FINAL

Preliminary/Qualifying Exam in Numerical Analysis (Math 502a) Spring 2012

Krylov Subspace Methods for Large/Sparse Eigenvalue Problems

MATH 5524 MATRIX THEORY Problem Set 5

Matrices, Moments and Quadrature, cont d

Charles University Faculty of Mathematics and Physics DOCTORAL THESIS. Krylov subspace approximations in linear algebraic problems

Bounding the Spectrum of Large Hermitian Matrices

11.0 Introduction. An N N matrix A is said to have an eigenvector x and corresponding eigenvalue λ if. A x = λx (11.0.1)

1. Introduction. In this paper we consider the large and sparse eigenvalue problem. Ax = λx (1.1) T (λ)x = 0 (1.2)

Partial Eigenvalue Assignment in Linear Systems: Existence, Uniqueness and Numerical Solution

c 2006 Society for Industrial and Applied Mathematics

MATH 1120 (LINEAR ALGEBRA 1), FINAL EXAM FALL 2011 SOLUTIONS TO PRACTICE VERSION

Algorithms that use the Arnoldi Basis

Bindel, Fall 2016 Matrix Computations (CS 6210) Notes for

Check that your exam contains 30 multiple-choice questions, numbered sequentially.

of dimension n 1 n 2, one defines the matrix determinants

Notes on Eigenvalues, Singular Values and QR

Pseudospectra and Nonnormal Dynamical Systems

Diagonalizing Matrices

Exponentials of Symmetric Matrices through Tridiagonal Reductions

Chapter 6: Orthogonality

Eigenvalue and Eigenvector Problems

Key words. cyclic subspaces, Krylov subspaces, orthogonal bases, orthogonalization, short recurrences, normal matrices.

eciency. In many instances, the cost of computing a few matrix-vector products

7. Symmetric Matrices and Quadratic Forms

Rational Krylov methods for linear and nonlinear eigenvalue problems

Transcription:

SIAM J. MATRIX ANAL. APPL. Vol. 3, No., pp. c 29 Society for Industrial and Applied Mathematics THE ARNOLDI EIGENVALUE ITERATION WITH EXACT SHIFTS CAN FAIL MARK EMBREE Abstract. The restarted Arnoldi algorithm, implemented in the ARPACK software library and MATLAB s eigs command, is among the most common means of computing select eigenvalues and eigenvectors of a large, sparse matrix. To assist convergence, a starting vector is repeatedly refined via the application of automatically constructed polynomial filters whose roots are known as exact shifts. Though Sorensen proved the success of this procedure under mild hypotheses for Hermitian matrices, a convergence proof for the non-hermitian case has remained elusive. The present note describes a class of examples for which the algorithm fails in the strongest possible sense; that is, the polynomial filter used to restart the iteration deflates the eigenspace one is attempting to compute. Key words. implicitly restarted Arnoldi algorithm, Krylov Schur algorithm, eigenvalues, exact shifts, ARPACK, eigs AMS subject classifications. 5A8, 5A6, 65F5 DOI..37/666997. Setting. Large-scale matrix eigenvalue problems typically derive from applications in which one seeks only some small subset of the spectrum, such as the largest magnitude or rightmost eigenvalues for stability analysis of dynamical systems. Given amatrixa C n n, one can approximate such eigenvalues by projecting A onto an appropriate low-dimensional subspace and then solving a small eigenvalue problem with a dense method such as the QR algorithm. The Arnoldi method [, 3] orthogonally projects A onto the Krylov subspace K k (A, v) :=span{v, Av,...,A k v} generated by the starting vector v C n. This subspace the span of iterates of the power method often produces eigenvalue estimates that fall on the periphery of the spectrum [6], though the quality of such approximations depends on the distribution of the eigenvalues of A, the angles between associated eigenvectors, and the starting vector v. Provided that dim K k+ (A, v) =k +, by the end of the kth iteration the Arnoldi algorithm has produced orthonormal vectors v,...,v k+ such that span{v,...,v j } = K j (A, v) for all j =,...,k +. Accumulating v,...,v k into the columns of V k C n k, the algebra affecting the construction of thebasiscanbesummarizedinthearnoldi factorization () AV k = V k H k + h k+,k v k+ e k, where e k denotes the last column of the k k identity matrix and H k = V k AV k is an upper Hessenberg matrix. The eigenvalues of H k,theritz values, approximate Received by the editors September 5, 26; accepted for publication by F. Dopico November 2, 26; published electronically February 27, 29. Supported by Department of Energy grant DE- FG3-2ER2553 and National Science Foundation grant DMS-CAREER-449973. http://www.siam.org/journals/simax/3-/6699.html Department of Computational and Applied Mathematics, Rice University, 6 Main Street, MS 34, Houston, TX 775 892 (embree@rice.edu).

2 MARK EMBREE eigenvalues of A in the sense that if H k u = θu, then A(V k u) θ(v k u) h k+,k e ku. These Ritz values must fall within the numerical range (or field of values) ofa, (2) W (A) :={z C : z = x Ax for some x C n with x =}, since θ =(V k u) A(V k u). The numerical range of a normal matrix is the convex hull of its spectrum [5]; for nonnormal matrices, the numerical range may contain points far from any eigenvalue (see, e.g., [7, Ch. 7]), a fact critical to the examples we shall present. Given the Arnoldi algorithm s broad utility, its nontrivial convergence behavior has attracted considerable attention [3, 4, 6, 7, 8, 3]. The computational cost of enlarging the Krylov subspace in the Arnoldi algorithm grows with the subspace dimension, k, and for large problems storage of the basis vectors alone becomes burdensome. Unfortunately, in many cases practical values of k fail to produce sufficiently accurate eigenvalue estimates. A simple solution is to restart the iteration, using information culled from the factorization () to refine the starting vector v in a manner that enriches components in the direction of desired eigenvectors while damping unwanted eigenvectors. This process is most commonly implemented through a polynomial filter: the starting vector for the new Arnoldi factorization takes the form v + = ψ(a)v for some polynomial ψ, generally of low degree. Strategies for constructing ψ vary. Saad advocated a choice designed to produce v + as a linear combination of the desired Ritz vectors [3]. Later methods chose ψ to be small over an adaptively identified region of the complex plane containing only the unwanted eigenvalues, but such procedures proved difficult to reliably automate. Sorensen proposed a less elaborate strategy: take ψ to be a polynomial whose roots match the undesired Ritz values [4]. More precisely, suppose we seek m eigenvalues and have already built an Arnoldi factorization of dimension k = m + p for some p>. Sort the eigenvalues of H k into the m possessing the desired trait (e.g., those rightmost or largest in magnitude), labeled θ,...,θ m, and the remaining p values θ m+,...,θ m+p, which we implicitly assume are coarse approximations to eigenvalues we do not seek. Set ψ(z) = p j= (z θ m+j), so that v + = p (A θ m+j I)v. j= These roots of ψ, the undesired Ritz values, are called exact shifts. Such shifts form an essential part of the ARPACK library [9], and their efficacy is responsible in no small part for the popularity enjoyed by that software and MATLAB s subordinate eigs command. (In the same paper that advocated exact shifts, Sorensen also proposed a robust method of implicitly restarting the Arnoldi algorithm using the roots of an arbitrary filter polynomial [4]. An alternative implementation, Stewart s Krylov Schur algorithm, requires the use of exact shifts [5].) Numerous computational examples demonstrate the success of exact shifts; see, e.g., [3, 4, 9]. A convergence proof would require that one make precise the notion that the p Ritz values used as shifts approximate the unwanted eigenvalues. For Hermitian

ARNOLDI WITH EXACT SHIFTS CAN FAIL 3 A an appeal to the interlacing theorem [] suffices. Labeling the eigenvalues of A as λ λ 2 λ n, there can be at most one Ritz value in the interval (λ,λ 2 ), two Ritz values in (λ 2,λ 3 ), and so on. This interlacing forms the basis of Sorensen s proof that for a Hermitian matrix the restarted Arnoldi algorithm converges to extreme eigenvalues. In the years since the introduction of exact shifts, a convergence proof for non-hermitian matrices has remained elusive. Several results ensure convergence provided the exact shifts satisfy an appropriate distribution [3, 4, 8], but conditions guaranteeing that exact shifts exhibit such behavior have not yet been established. Indeed, few fine results about the Ritz values of non-hermitian matrices are known. The purpose of this note is to provide a counterexample to the conjecture that the restarted Arnoldi algorithm with exact shifts must converge under hypotheses resembling those sufficient for the Hermitian case. Our examples illustrate the failure of the algorithm in the strongest sense, in that the filter polynomial ψ exactly deflates a sought-after (perfectly conditioned) eigenvalue. Throughout we assume that all computations are performed in exact arithmetic and in particular that eigenvalues of the upper Hessenberg matrix H k are determined exactly. 2. Simple example. We begin with a small example that clearly demonstrates how restarting with exact shifts can produce catastrophic results. Suppose we seek the largest magnitude (and rightmost) eigenvalue λ = and associated eigenvector u =[] T of the matrix A = 6 2 2 We denote the bottom-right 3 3 submatrix of A by D = 6 2 2. The desired eigenvalue λ =ofa is far from the rest of the spectrum, σ(d) ={}, and is also fairly well separated [6, p. 46]:. sep(, D) = (I D) =.837... In the language of pseudospectra [7], this implies σ ε (D) for all ε<.837..., where (3) σ ε (D) ={z C : z σ(d + E) forsomee C n n with E <ε} = {z C : (zi D) >ε }. Furthermore, the eigenvector u is orthogonal to the complementary invariant subspace associated with the zero eigenvalue; that is, λ = is perfectly conditioned [2]. Sorensen s theorem [4, Thm. 5.9] imposes several reasonable caveats: the starting vector cannot be deficient in any of the sought-after eigenvectors, and the iteration must not come arbitrarily close to lucky breakdown ; i.e., all the subdiagonal entries in H k must be bounded away from zero, so none of the intermediate Krylov subspaces fall too close to an invariant subspace.

4 MARK EMBREE Apply the restarted Arnoldi method to A with the starting vector v =[] T. We seek the m = eigenvalue and will build the Krylov subspace out to dimension k = m + p = 2 before restarting. Two steps of the Arnoldi algorithm produce the orthonormal basis vectors v = 2 and upper Hessenberg matrix H 2 =[v v 2 ] A[v v 2 ]=, v 2 = 35 7 3 9 7 [ ] 7/4 3 35/4. 35/4 5/4 The characteristic polynomial for H 2 is det(λi H 2 )=(λ 7/4)(λ 5/4) 3 35 4 4 =(λ )(λ 2), and hence the Ritz values are θ =2andθ 2 = (ordered by decreasing magnitude). The smaller magnitude (and leftmost) of these is chosen as the exact shift; hence the filter polynomial ψ(z) =z produces the new starting vector v + = ψ(a)v = Note that v + has no component in the desired eigenvector u that is, v + is contained in the invariant subspace associated with the zero eigenvalue, and hence the eigenvalue λ = will exert no influence upon further Arnoldi iterations. One might suspect that this starting vector has been specially engineered to yield this behavior. While this is true of the last three components of v, it is curious that the phenomenon persists even when v is arbitrarily enriched in the eigenvector we wish to compute. If we set v =[α ] T for any α R, then two steps of the Arnoldi algorithm yield the upper Hessenberg matrix H 2 = with characteristic polynomial 3. 6+α 2 3α 3 3+α 2 (3 + α 2 )(24 + α 2 ) /2 (24 + α 2 ) /2 3+2α 2 3+α 2 3+α 2 det(λi H 2 )=λ 2 3λ +2=(λ )(λ 2). The Ritz values θ =2andθ 2 = are independent of the bias of the starting vector v toward the desired eigenvector. As demonstrated in the next section, this behavior is an instance of a broader phenomenon that will facilitate the design of larger examples.

ARNOLDI WITH EXACT SHIFTS CAN FAIL 5 3. General construction. While the orthogonal basis for the Krylov subspace and the resulting Arnoldi factorization () have computational advantages, other bases for K k (A, v) are often better suited to analysis. Among the choices available (see [5, Thm. 2.2]), we shall use a decomposition introduced by Ruhe [2, sect. 2]. Suppose dim K k (A, v) =k. Arrange the Krylov basis vectors v, Av,..., A k v into the columns of K k = [ v Av A k v ], and define the companion matrix C k = c c 2.... c k Ck k for constants c,...,c k ; unspecified entries equal zero. A direct calculation reveals (4) AK k K k C k = re k for the vector k r := A k v c j A j v K k+ (A, v). j= Now since Ran K j =RanV j = K j (A, v), j =,...,k, thereexistssomeinvertibler k C k k such that K k = V k R k. Recalling that H k = V k AV k, pre- and postmultiply (4) by V k H k R k C k R k = V k re k R k. and R k to obtain If c,...,c k are chosen such that r K k (A, v) =RanV k, we conclude that H k = R k C k R k ; i.e., the companion matrix C k and the upper Hessenberg matrix H k are similar and thus have the same eigenvalues. This development facilitates the following result. Theorem 3.. Let θ,...,θ k denote the Ritz values produced by k l steps of the Arnoldi algorithm applied to the matrix D C l l with starting vector w C l,where dim K k (D, w) =k. For any T C m m, m k, withspectrumσ(t) {θ,...,θ k } (respecting multiplicity) and any x C m, k steps of the Arnoldi process applied to the matrix and starting vector A = [ T D produce the identical Ritz values θ,...,θ k. ] [ ] x, v = w

6 MARK EMBREE Proof. To write a Krylov factorization for K k (D, w), we define c K k = [ w Dw D k w ] c 2, C k =..... Ck k, c k with coefficients c,...,c k chosen, e.g., via the Gram Schmidt process, to ensure k r := D k w c j D j w is orthogonal to K k (D, w). This choice gives σ(c k )={θ,...,θ k }.Nowwith K k = [ v Av A k v ], observe that AK k K k C k = re k, j= where r := A k v k c j A j v = j= [ T k x k j= c jt j ] x D k w k j= c. jd j w Since σ(t) {θ,...,θ k } (respecting multiplicity), the polynomial k k z k c j z j = (z θ j ) j= j= annihilates T, leaving r = [ r ]. It follows that r is orthogonal to K k (A, v), and hence the Ritz values for A and v drawn from this subspace must be θ,...,θ k. As Ritz values are invariant to unitary similarity transformations, the matrix A and v in the theorem could be replaced by Q AQ and Q v for any unitary matrix Q C n n. Of fundamental importance is the fact that the invariant subspace associated with σ(t) be orthogonal to the one associated with σ(d). The matrix T need not be normal nor even diagonalizable. This result (which relates to the example of maximal Krylov subspaces for derogatory matrices given in [3, p. 8]) suggests a procedure for manufacturing examples like the one presented in section 2. The following approach presumes that one seeks the largest magnitude eigenvalues of a matrix (as is typical, e.g., after performing a shift-invert or Cayley transformation of the original matrix []).. Find a matrix D and starting vector w such that the Ritz values produced by k 2m steps of the Arnoldi process, when ordered by decreasing magnitude, satisfy (5) θ m > θ m+ θ 2m > max λ σ(d) λ.

ARNOLDI WITH EXACT SHIFTS CAN FAIL 7 2. Construct a matrix T C m m with eigenvalues θ m+,...,θ 2m ;takex C m to be any vector. 3. Build the matrix and starting vector A = [ T D ], v = [ x w ]. We wish to compute the largest magnitude eigenvalues of A, which by (5) are the eigenvalues θ m+,...,θ 2m of T. Apply k steps of the Arnoldi process, resulting in the same Ritz values obtained in step. 4. The exact shift procedure will select Ritz values θ m+,...,θ k as the roots of the filter polynomial. These shifts include the eigenvalues of A we seek, which are thus deflated from v and hence cannot be recovered by the restarted iteration. Examples of this sort rely on Ritz values that fall well beyond all the eigenvalues of D. If that matrix is normal, then all its Ritz values would be contained in the convex hull of its spectrum. Hence no normal D will give Ritz values larger in magnitude than (or further to the right of) its eigenvalues, and consequently no such matrix is suitable for use in the above construction. Nonnormality thus plays a central role in examples of this form. Both the example of section 2 and the one we shall next present have been designed so that the eigenvalues of interest are not unduly influenced by this nonnormality and hence would likely be meaningful in applications (e.g., D may be far from normal, but the desired eigenvalues of A are well separated from D). As in [4, section 4.4], nonnormality associated with undesired eigenvalues complicates convergence to desired, ideally conditioned eigenvalues. 4. Larger example. Next we follow the procedure just described to produce a scenario that perhaps appears less contrived than the 4 4 example of section 2. To begin, consider the upper triangular matrix D = Λ + βs γ,where with s Λ = diag(λ,...,λ n ), S = λ j = α +...... sn 2α(j ), s j = j n n for parameters α, β, γ. The spectrum of this matrix is uniformly distributed over the interval [ α, α]. For α> the matrix is diagonalizable, with β and γ controlling the conditioning of the eigenvalues. Qualitatively, the growth of s j with j causes eigenvalues on the right end of the spectrum to exhibit greater sensitivity than those on the left. Increasing β magnifies the ill-conditioning throughout the spectrum, while increasing γ improves the conditioning, most acutely for the leftmost eigenvalues. In the example that follows, D has dimension n = with α =/2 andβ = γ = 4. Figure (a) shows the spectrum, numerical range (2), and ε-pseudospectra (3) of D. The matrix exhibits a moderate departure from normality, with the right half of the spectrum especially sensitive. Figure (b) shows k = Ritz values for D with the starting vector w = [,..., ] T. If we seek m = 5 eigenvalues, then the exact shift strategy would choose,

8 MARK EMBREE 3 2 (a) 3 2 (b) 2 2 3 2 2 3 3 2 2 3 3 3 2 (c) 2 (d) 2 2 3 2 2 3 3 2 2 3 Fig.. Illustration of the example discussed in section 4. (a) The eigenvalues (dots covering the segment [ /2, /2]), ε-pseudospectra (boundaries shown as solid lines for ε =, 3,..., 9 ), and the numerical range (boundary shown as a dashed line) for D. (b) The eigenvalues of D with the k =Ritz values for D and w (crosses); the gray circles indicate the magnitude of the fifth and sixth largest Ritz values. (c) The eigenvalues of A. (d) The eigenvalues of A (dots) with k =Ritz values for A and v (circles and crosses); the smallest five in magnitude (circles) match the largest magnitude eigenvalues of A and will be used as exact shifts. the k m = 5 smallest magnitude Ritz values as roots of the filter polynomial; these Ritz values are the leftmost ones shown in the figure. Comparing this with Figure (a), notice that these Ritz values are outside the ε-pseudospectra of A for all ε 3 ; equivalently, the sep of all these Ritz values from D is larger than 3. We now follow the recipe outlined above to obtain an A and v for which these shifts would be catastrophic. Let T C 5 5 be a diagonal matrix whose eigenvalues equal those five smallest magnitude Ritz values, and set A to be the 5 5 matrix A = diag(t, D); the eigenvalues of A are shown in Figure (c). By design, the eigenvalues of T are the largest magnitude eigenvalues of A, and they all fall beyond the ε = 3 pseudospectrum of D. Now compute k = Ritz values for A with v = [x T w T ] T for any choice of x, shown in Figure (d). (The computation to produce this illustration used x = [,..., ] T, but that choice has no influence on the figure.) As ensured by the theorem,

ARNOLDI WITH EXACT SHIFTS CAN FAIL 9 these Ritz values are identical to those obtained from D and w. To compute the largest magnitude eigenvalues of A, the exact shift strategy picks as shifts the five smallest Ritz values which coincide with the five largest magnitude eigenvalues of A. Again, the exact shift procedure will deflate precisely those eigenvalues we wish to compute. 5. Discussion. Our aim here has been to address a theoretical question concerning the convergence of the restarted Arnoldi algorithm with exact shifts. In no way do we suggest that the behavior our examples exhibit is commonplace. The exact shift procedure remains the most robust general-purpose method for restarting the Arnoldi method, endorsed by years of successful computation. Our constructions rely on special choices for the starting vector (the components in w) and a fixed number of steps; by changing w, or increasing or decreasing k, onemaywellconvergetothe correct eigenvalues without incident. Even with problematic w and k, the rounding errors that occur in practical computations can allow the desired eigenvectors to emerge after numerous additional iterations. (Indeed, when applied to the example in section 4, eigs eventually finds the five desired eigenvalues.) On those occasions when ARPACK fails to converge, the culprit is likely more mundane than the extreme failure exhibited here. For example, the requested eigenvalues may form part of a tight cluster, in which case any shift procedure would struggle to develop, in a tractable number of iterations, a polynomial filter that is large on the desired eigenvalues while being small on those nearby undesired eigenvalues. Such behavior is explained by the conventional restarted Arnoldi convergence analysis cited in the introduction. The mere existence of examples for which the restarted Arnoldi algorithm with exact shifts deflates the desired eigenvalues underscores the need for a deeper understanding of the behavior of Ritz values of non-hermitian matrices. Though these eigenvalue estimates must fall within the numerical range, little else is known, deterministically or stochastically, about their distribution. Progress on this important problem could provide a foundation for a robust convergence theory for the restarted Arnoldi algorithm and illuminate many other corners of iterative linear algebra. Acknowledgments. I thank Rich Lehoucq for drawing my attention to Ruhe s paper [2], and Howard Elman for his kind hospitality during a visit to the University of Maryland, where this work was begun. The pseudospectra in the figure were computed using the EigTool package of T. G. Wright [8]. REFERENCES [] W. E. Arnoldi, The principle of minimized iterations in the solution of the matrix eigenvalue problem, Quart. Appl. Math., 9 (95), pp. 7 29. [2] Z. Bai, J. Demmel, and A. McKenney, On computing condition numbers for the nonsymmetric eigenproblem, ACM Trans. Math. Software, 9 (993), pp. 22 223. [3] C. Beattie, M. Embree, and J. Rossi, Convergence of restarted Krylov subspaces to invariant subspaces, SIAM J. Matrix Anal. Appl., 25 (24), pp. 74 9. [4] C. A. Beattie, M. Embree, and D. C. Sorensen, Convergence of polynomial restart Krylov methods for eigenvalue computations, SIAM Review, 47 (25), pp. 492 55. [5] R. A. Horn and C. R. Johnson, Topics in Matrix Analysis, Cambridge University Press, Cambridge, UK, 99. [6] Z. Jia, The convergence of generalized Lanczos methods for large unsymmetric eigenproblems, SIAM J. Matrix Anal. Appl., 6 (995), pp. 843 862. [7] L. Knizhnerman, Error bounds for the Arnoldi method: A set of extreme eigenpairs, Linear Algebra Appl., 296 (999), pp. 9 2. [8] R. B. Lehoucq, Implicitly restarted Arnoldi methods and subspace iteration, SIAMJ. Matrix Anal. Appl., 23 (2), pp. 55 562.

MARK EMBREE [9] R. B. Lehoucq, D. C. Sorensen, and C. Yang, ARPACK Users Guide: Solution of Large- Scale Eigenvalue Problems with Implicitly Restarted Arnoldi Methods, Software Environ. Tools 6, SIAM, Philadelphia, 998. [] K. Meerbergen, A. Spence, and D. Roose, Shift-invert and Cayley transforms for detection of rightmost eigenvalues of nonsymmetric matrices, BIT, 34 (994), pp. 49 423. [] B. N. Parlett, The Symmetric Eigenvalue Problem, Prentice Hall, Englewood Cliffs, NJ, 98. [2] A. Ruhe, Rational Krylov sequence methods for eigenvalue computation, Linear Algebra Appl., 58 (984), pp. 39 45. [3] Y. Saad, Variations on Arnoldi s method for computing eigenelements of large unsymmetric matrices, Linear Algebra Appl., 34 (98), pp. 269 295. [4] D. C. Sorensen, Implicit application of polynomial filters in a k-step Arnoldi method, SIAM J. Matrix Anal. Appl., 3 (992), pp. 357 385. [5] G. W. Stewart, A Krylov Schur algorithm for large eigenproblems, SIAMJ. MatrixAnal. Appl., 23 (2), pp. 6 64. [6] G. W. Stewart, Matrix Algorithms Volume II: Eigensystems, SIAM, Philadelphia, 2. [7] L. N. Trefethen and M. Embree, Spectra and Pseudospectra: The Behavior of Nonnormal Matrices and Operators, Princeton University Press, Princeton, NJ, 25. [8] T. G. Wright, EigTool, 22. Software available at http://www.comlab.ox.ac.uk/ pseudospectra/eigtool.