NOTICE WARNING CONCERNING COPYRIGHT RESTRICTIONS: The copyright law of the United States (title 17, U.S. Code) governs the making of photocopies or

Similar documents
Examples include: (a) the Lorenz system for climate and weather modeling (b) the Hodgkin-Huxley system for neuron modeling

Synopsis of Numerical Linear Algebra

MAT Linear Algebra Collection of sample exams

Linear Algebra Practice Problems

There are six more problems on the next two pages

The Eigenvalue Problem: Perturbation Theory

Foundations of Matrix Analysis

Linear Algebra: Matrix Eigenvalue Problems

Implicit Functions, Curves and Surfaces

arxiv: v1 [math.na] 6 Aug 2010

Spheres with maximum inner diameter

MASSACHUSETTS INSTITUTE OF TECHNOLOGY Chemistry 5.76 Revised February, 1982 NOTES ON MATRIX METHODS

IMPORTANT DEFINITIONS AND THEOREMS REFERENCE SHEET

Chap 3. Linear Algebra

The goal of this chapter is to study linear systems of ordinary differential equations: dt,..., dx ) T

POSITIVE SEMIDEFINITE INTERVALS FOR MATRIX PENCILS

IMPORTANT DEFINITIONS AND THEOREMS REFERENCE SHEET

1. If 1, ω, ω 2, -----, ω 9 are the 10 th roots of unity, then (1 + ω) (1 + ω 2 ) (1 + ω 9 ) is A) 1 B) 1 C) 10 D) 0

MATH 220 FINAL EXAMINATION December 13, Name ID # Section #

, b = 0. (2) 1 2 The eigenvectors of A corresponding to the eigenvalues λ 1 = 1, λ 2 = 3 are

DISCUSSION CLASS OF DAX IS ON 22ND MARCH, TIME : 9-12 BRING ALL YOUR DOUBTS [STRAIGHT OBJECTIVE TYPE]

Properties of Matrices and Operations on Matrices

The Generalized Eigenvalue-Eigenvector Problem or How to Simultaneously Diagonalize Two Quadratic Forms

LINEAR ALGEBRA QUESTION BANK

Chapter 3 Transformations

1 Matrices and vector spaces

1. Linear systems of equations. Chapters 7-8: Linear Algebra. Solution(s) of a linear system of equations (continued)

EIGENVALUES AND EIGENVECTORS 3

Lecture 10 - Eigenvalues problem

Quadratic forms. Here. Thus symmetric matrices are diagonalizable, and the diagonalization can be performed by means of an orthogonal matrix.

Math 12 Pre-Calculus Midterm Review (Chapters 1 6)

LECTURE NOTES ELEMENTARY NUMERICAL METHODS. Eusebius Doedel

Chapter 7. Canonical Forms. 7.1 Eigenvalues and Eigenvectors

NATIONAL BOARD FOR HIGHER MATHEMATICS. M. A. and M.Sc. Scholarship Test. September 24, Time Allowed: 150 Minutes Maximum Marks: 30

Tutorial on Principal Component Analysis

5.) For each of the given sets of vectors, determine whether or not the set spans R 3. Give reasons for your answers.

14.2 QR Factorization with Column Pivoting

Remark By definition, an eigenvector must be a nonzero vector, but eigenvalue could be zero.

Eigenvalues and Eigenvectors

Math Linear Algebra Final Exam Review Sheet

NOTICE WARNING CONCERNING COPYRIGHT RESTRICTIONS: The copyright law of the United States (title 17, U.S. Code) governs the making of photocopies or

Solutions Parabola Volume 49, Issue 2 (2013)

1. Let m 1 and n 1 be two natural numbers such that m > n. Which of the following is/are true?

MATH 423 Linear Algebra II Lecture 33: Diagonalization of normal operators.

TEST CODE: MMA (Objective type) 2015 SYLLABUS

Eigenvalues and Eigenvectors

An Extrapolated Gauss-Seidel Iteration

1.1 Single Variable Calculus versus Multivariable Calculus Rectangular Coordinate Systems... 4

Fall 2016 MATH*1160 Final Exam

Mathematics 1 Lecture Notes Chapter 1 Algebra Review

A Note on Eigenvalues of Perturbed Hermitian Matrices

LINEAR SYSTEMS AND MATRICES

Cheat Sheet for MATH461

Algebraic. techniques1

EK102 Linear Algebra PRACTICE PROBLEMS for Final Exam Spring 2016

Solutions to Final Exam 2011 (Total: 100 pts)

Analysis-3 lecture schemes

Math Camp II. Basic Linear Algebra. Yiqing Xu. Aug 26, 2014 MIT

EXERCISE SET 5.1. = (kx + kx + k, ky + ky + k ) = (kx + kx + 1, ky + ky + 1) = ((k + )x + 1, (k + )y + 1)

Week Quadratic forms. Principal axes theorem. Text reference: this material corresponds to parts of sections 5.5, 8.2,

EE731 Lecture Notes: Matrix Computations for Signal Processing

Chapter 8. Rigid transformations

Remarks on Definitions

Linear Algebra, part 2 Eigenvalues, eigenvectors and least squares solutions

Solutions: Problem Set 3 Math 201B, Winter 2007

TABLE OF CONTENTS. Our aim is to give people math skillss in a very simple way Raymond J. Page 2 of 29

Matrix Vector Products

MA 265 FINAL EXAM Fall 2012

Applied Linear Algebra in Geoscience Using MATLAB

EXERCISES ON DETERMINANTS, EIGENVALUES AND EIGENVECTORS. 1. Determinants

1 Linear Algebra Problems

4.6 Bases and Dimension

MATH 3321 Sample Questions for Exam 3. 3y y, C = Perform the indicated operations, if possible: (a) AC (b) AB (c) B + AC (d) CBA

= F (b) F (a) F (x i ) F (x i+1 ). a x 0 x 1 x n b i

3. Identify and find the general solution of each of the following first order differential equations.

Matrix Algorithms. Volume II: Eigensystems. G. W. Stewart H1HJ1L. University of Maryland College Park, Maryland

June 2011 PURDUE UNIVERSITY Study Guide for the Credit Exam in (MA 262) Linear Algebra and Differential Equations

List of Symbols, Notations and Data

Matrices and Vectors. Definition of Matrix. An MxN matrix A is a two-dimensional array of numbers A =

Total 170. Name. Final Examination M340L-CS

EIGENVALUE PROBLEMS. Background on eigenvalues/ eigenvectors / decompositions. Perturbation analysis, condition numbers..

Reconstructing an Ellipsoid from its Perspective Projection onto a Plane

T k b p M r will so ordered by Ike one who quits squuv. fe2m per year, or year, jo ad vaoce. Pleaie and THE ALTO SOLO

FACULTY OF MATHEMATICAL STUDIES MATHEMATICS FOR PART I ENGINEERING. Lectures

1 Linear transformations; the basics

Mathematics 1EM/1ES/1FM/1FS Notes, weeks 18-23

MATH 583A REVIEW SESSION #1

TEST CODE: MIII (Objective type) 2010 SYLLABUS

Name: AK-Nummer: Ergänzungsprüfung January 29, 2016

NOTES on LINEAR ALGEBRA 1

Jim Lambers MAT 610 Summer Session Lecture 2 Notes

Remark 1 By definition, an eigenvector must be a nonzero vector, but eigenvalue could be zero.

Math Ordinary Differential Equations

Error Analysis for Solving a Set of Linear Equations

Review of Integration Techniques

DIAGONALIZATION. In order to see the implications of this definition, let us consider the following example Example 1. Consider the matrix

CHAPTER 3. Matrix Eigenvalue Problems

2 Eigenvectors and Eigenvalues in abstract spaces.

2. Every linear system with the same number of equations as unknowns has a unique solution.

Chapter 3. Linear and Nonlinear Systems

Transcription:

NOTICE WARNING CONCERNING COPYRIGHT RESTRICTIONS: The copyright law of the United States (title 17, U.S. Code) governs the making of photocopies or other reproductions of copyrighted material. Any copying of this document without permission ofits author may be prohibited by law.

GEbSCHGORIN THEORY FOR THE GENERALIZED EIGENVALUE PROBLEM Ax = \Bx G. W. Stewart Department of Mathematics Department of Computer Science Carnegie-MelIon University Pittsburgh, Pa. October, 1973 This work was supported in part by the Office of Naval Research under Contract No, N00014-67-A-0314-0018.

ABSTRACT A generalization of Gerschgorin's theorem is developed for the eigenvalue problem Ax = \Bx and is applied to obtain perturbation bounds for multiple eigenvalues. The results are interpreted in terms of the chordal metric on the Riemann sphere, which is especially convenient for treating infinite eigenvalues. i

1. Introduction The object of this paper is to develop a perturbation theory for the generalized eigenvalue problem Ax = \Bx that parallels the perturbation theory developed by Wilkinson [4, Ch. 2] for the ordinary eigenvalue problem. The theory rests on a generalization of the Gerschgorin theorem, whose results are interpreted in terms of the chordal metric in the Riemann sphere. This approach has the advantage that it deals readily with multiple and infinite eigenvalues. Throughout this paper we shall identify the (possibly infinite) eigenvalue X = a/b with the point in the projective complex line defined by [a,0] = {(a,b) t (0,0): Q/B = X% We shall equip the projective complex line with the metric x defined by The number X([osB],[a!,B f ]) is the chordal distance between the two points X - o/b and X 1 = a'/b 1 when they are projected in the usual way onto the Riemann sphere. By abuse of notation we shall let x(x,x f ) denote this distance, so that in this context x Is the chordal metric for the set of complex numbers [1, p. 81]. The justification of the use of the chordal metric lies in the simplicity of the final results, in particular in the uniform treatment of large and small eigenvalues. To illustrate this, we shall begin with an application to first order perturbation theory for a simple eigenvalue.

-2- Let A, B, E, and F be square nxn matrices with complex elements and let X be a simple eigenvalue of the problem Ax = XBx with eigenvector x normalized so that F-x 12 = L e t y be the left eigenvector corresponding to X, again normalized so that j yj 2 a 0«L e t a s y H Ax, 0 «y^x, so that the point is identified with X. It has been shown by the author [3] that for sufficiently small E, and F, there is an eigenvalue X 1 satisfying (A+E)x! «X f (B+F)x that can be identified with la + y H Ex, B + y H Fx], 2 2 except for terms of order E and F. Thus in our approach the sensitivity of X to perturbations in A and B will be measured by X(X,X') =X([a,B], [«+ y H Ex, 0 + 7 H Fx]) = InAx-By^xJ J\ a\ Z +101 ' l J\ cffy H Ex 1 2 +1 P+y H Fx 1 2 To obtain meaningful approximate perturbation bounds, let 9 = tan" 1 1 X, v = 7TCV 2 + E 2, and e = ^y H Ex + y H Fx 2 Then c/v B sin 8 and P/v = cos 0. Hence we have the bound 0.1) XCXA')^ 7^ 0 0 8 9 1 1,.:^ 1^ 9 1 '

-3- which is accurate up to terras of order e^. In the terminology of numerical analysis, the bound (1.1) says that v ^ is a condition number for the eigenvalue X, in the sense that it measures how perturbations in A and B will affect X. If v is small compared with E and F, then one can expect large changes in X. It should be noted that illconditioned eigenvalues need not be large, and conversely a large eigenvalue need not be ill-conditioned. For example, the bound (1.1) can be quite small even when (3=0 and hence X = 00, which illustrates the utility of the chordal metric in dealing with this problem. The factors cos 9 and sin 0 appearing in (1.1) are a little unusual, but they make sense. For example when 8=0, the eigenvalue X is zero and the matrix A is singular. The disappearance of the term sin 9 y**fx in (1.1) then says that perturbations in B cannot affect the singularity of A. Although the above bound is quite satisfactory for practical work, it is interesting to relate v to the condition number for the ordinary eigenvalue problem. This may be done as follows. Suppose Bx ^ 0 (if Bx = 0, use Ax in what follows). Define - sec ^(y,bx) and p=v Ax 2 + Bx 2. Then -1 H When B = I, the number n is the secant of the angle between the left and right eigenvectors corresponding to X, which is the condition number for the ordinary

-4- eigenvalue problem [4, Ch. 2]. The number p measures how nearly x is an approximate null vector of both A and B (when Ax = Bx = 0, det(a-\b) s 0 and any number \ is an eigenvalue). When B = I and \ <> 1, p rjz and the bound becomes essentially the bound for the ordinary eigenvalue problem. If x ^ 1, then the bound (1.1) deviates from the usual bound because of the distorting effects of the chordal metric outside the unit circle. The above theory has two drawbacks. First, although the bound is asymptotically accurate, the theory does not provide a bound on the remainder, nor does it specify where it is applicable. Second, the theory is not applicable to multiple eigenvalues. It is thè object of the next two sections to remedy these defects. In Section 2 we shall develop a generalization of Gerschgorin 1 s theorem and in Section 3 apply it to develop a perturbation theory for multiple eigenvalues. Throughout the paper we shall use Householder's notational conventions [2] # We shall use the symbols * (p = 1,2, ) to denote the usual Holder vector norms.

-5-2. Gerschgorin Theory Let A and B be matrices of order n. Set a i = (a iv'"> ai,i-t a i,i+1'"- a i,n ); that is a. is the vector formed from the i-th row of A by deleting its i-th component. Define the vectors b^ similarly. The following theorem generalizes the Gerschgorin exclusion theorem [2, p. 65 ff]. Theorem 2.1. Let X be an eigenvalue of the problem Ax = XBx. Then X lies in the union of the regions defined by & - {[or.. + p.. + B%: x <: 1}, L i 11 1 ' K ll i 11 11 J 9 oo (i = 1,2,...,n) Proof: Let Ax = XBx and suppose that the i-th component of x is largest in absolute value. Since x ^ 0, we may assume without loss of generality that =! Form x* from x by deleting its i-th component. Then x] ^ 1 and otti + a H x = X(P ti + B^T), which says that X if., m In stating Theorem 2.1 we have used the usual identification of the complex projective line with the Riemann sphere. The proof of the theorem in addition exhibits the region in which X must lie; namely the region corresponding to a maximal component at the eigenvector. When B = I, the regions become the disks in the complex plane defined by

-6- [ - {X: X-FLF LT * HaJI,}, which are the usual Gerschgorin disks. As they are defined, the sets fif^ are not convenient to work with. Instead we shall work with neighborhoods, defined in terms of the chordal metric, that are generally larger. Specifically, we have \ot..b. - B..a.x 1 ii i ii i 1 x 11V?-»»'?! I. HP., where a[ ± = max{0, 0 F LI - a i 1} and = max{0, - b. I,}. Hence if we set the eigenvalues of Ax = XBx lie in the union of the Note that if B = I and the elements of A are less than unity, then the^!^ give a fair approximation of the usual Gerschgorin disks. The utility of the Gerschgorin theorem in its applications to the eigenvalue problem is enhanced by the fact that we can often localize a specific

-7- number *of eigenvalues in a union of some of the The same is true of our generalization. Theorgm 2.2. If the union of k at the Gerschgorin regions is disjoint from the remaining regions and is not the entire sphere, then exactly k eigenvalues lie in the union. Proof. Suppose, without loss of generality, that the regions in question are Write A = A 1 + A 2 where A 1 = diag(^ 1,of 22 >» a nn^ a n d sl mi l a *"ly write B - B 1 + B 2 where B- «diag(b,g ). For 0 ^ e ^ 1 set i ii zz nn A(e) = A + ea 2 and B(e) - B 1 + eb 2, and let (G) be the corresponding Gerschgorin regions, which of course satisfy C fe^f. Now the hypotheses of the theorem insure that the characteristic polynomial p (X) = det(a(e) - XB(e)) cannot vanish identically; for otherwise 6 n any number X would be an eigenvalue and [^J would be the entire sphere. i=l 1 This means that the eigenvalues of A(«)x = XB(e)x must vary continuously in the sphere, or, what is equivalent, there are continuous functions X I (C) (i 88 l,2,...,n) such that X-j (e),x 2 (e),...,x n (c) comprise all the

-8- eigenvalues of A(e)x = \B(e)x. We may choose X.^0) = a L I/P I I«T H E N S I N C E X.(0) U ^((0) (i - 1,2,...,k) and \J #(«) 0 (J E)^ J0, we must 1 j-1 J k j=l J j-k+1 3 have X t ( D U (i = k+1,,n) for (i = 1,2,...,k) and X^l) U ) for

-9-3. Perturbation Theory In [4] Wilkinson has applied the Gerschgorin theorem to produce a perturbation theory for the eigenvalue problem. The heart of his theory is a technique in which off diagonal elements of order e are reduced to order 2 e by diagonal similarity transformations. In this section we shall show that the same technique can be applied to the generalized eigenvalue problem. Since it is the technique rather than its specific applications that is of chief interest, we shall confine ourselves to one of the several cases treated by Wilkinson; the case where the problem has a complete set of eigenvectors. Specifically we shall be concerned with the case where there are nonsingular matrices X and Y such that H Y AX = diagi^,of 2,..,a n ) and Y H BX = diag( 1,l 2,...,e n ). For definiteness we shall take X and Y to have columns of 2-norm unity. Of course the eigenvalues of Ax = \Bx are \^ = (i = l,2,...,n). We shall suppose that the first p eigenvalues are equal and distinct from the remaining q = n-p eigenvalues, and we shall apply the Gerschgorin theory of the last section to obtain perturbation bounds for this set of multiple eigenvalues. Let E and F be matrices of order n and let e be an upper bound on the elements of the matrices Y^EX and Y^X. Then Y H (A+E)X and Y H (B+F)X have the generic forms (n = 5)

-10- CH+C e e e e <*2 + e e e ^+6 e e E e e p2+e e e C e e P3+e e e e e e & 4 +E e V e e e a^+e e e e e P5+e If e is small enough, the first p Gerschgorin regions will be disjoint from the others, and their union will contain exactly p eigenvalues which are necessarily near X-j However we can obtain an even better result. Let T ^ 0 be given. The eigenvalues of (A+E)X - X(B+F)X will be unchanged H H if the last q columns of Y (A+E)X and Y (B+F)X are multiplied by T. If this is done the new matrice Y (A+E)X has the form (n=5, p=2) /a^+e e T6 T6 T6 \ e a 2 +E TC TC T6 \ e e T(Q? 3 +6) Te Te \ e / and Y H (B+F)X has a similar form. We shall attempt to choose T ^ 1 so that the first p Gerschgorin regions are disjoint. Let 6 = min{x(x I AJ): 1 = 1*2,...,p; J 65 P+1 Let and v = min{v.: i = l,2,...,n}.

11 Then the first p Gerschgorin regions are contained in the disk whose center is Xj and whose chordal radius is M + ^ <y<tq> «, J****, v ± v i ^(p+tq)e v-js"(p+tq)c Likewise the remaining q regions will be contained in disks whose centers are X. and whose chordal radius is 1 JG(P+TQ)C 9 Tv ^(p Kq)e Thus if we can find T satisfying 0 < T < 1 such that ^(P+TCJ)6 <rv ^(p+tq)e ^6 2 the appropriate Gerschgorin regions will be disjoint. This will surely be true if n n ^ _ J2 ne (2+6) ^. Thus we have shown that if (3.1) is satisfied, there are exactly p eigenvalues in the region defined by x(x r x) W 2 > ^ ) e, 1 v ^(p h q)e where v = min{v^: i = l,2,...,p). For small c this radius is asymptotic to pe - 9 V which shows that ^p/v is a condition number for* the multiple eigenvalue X-j

-12- When X 1 is simple and (3.1) is satisfied, then the perturbed eigenvalue must lie in the disk defined by /c^+yft^ \ ( of 1 l+ P 1 l+2c)tqc This bound approaches zero quadratically with 6, which makes rigorous the observations of Section 1. The treatment sketched above of course does not exhaust all possible cases. Various defective cases can be treated by applying the Gerschgorin theorem to canonical forms, as has been done by Wilkinson for the ordinary eigenvalue problem. Alternatively one can use the techniques of [3] to split off a set of multiple eigenvalues and treat these separately by means of the Gerschgorin theorem. With either approach, the generalized Gerschgorin theorem is required to deal with multiple eigenvalues, and the use of the chordal metric simplifies the final results.

-13-4. A cknowledgmen t s Part of this work was done while I was a visitor at the IBM Thomas J. Watson Research Center. I should also like to thank W. Kahan for his comments. References 1. C. Caratheodory, Theory of Functions, v. I, Chelsea, 1958, New York. 2. A. S. Householder, The Theory of Matrices in Numerical Analysis, Blaisdell, 1964, New York. 3. G. W. Stewart, Error and perturbation bounds for subspaces associated with certain eigenvalue problems, to appear SIAM Review. 4. J. H. Wilkinson, The Algebraic Eigenvalue Problem, Clarendon, 1965, Oxford.