Gershgorin s Circle Theorem for Estimating the Eigenvalues of a Matrix with Known Error Bounds

Similar documents
Linear Algebra. Matrices Operations. Consider, for example, a system of equations such as x + 2y z + 4w = 0, 3x 4y + 2z 6w = 0, x 3y 2z + w = 0.

CS 246 Review of Linear Algebra 01/17/19

Dominant Eigenvalue of a Sudoku Submatrix

4. Determinants.

c c c c c c c c c c a 3x3 matrix C= has a determinant determined by

Linear Algebra Primer

Up to this point, our main theoretical tools for finding eigenvalues without using det{a λi} = 0 have been the trace and determinant formulas

Chapter 3. Determinants and Eigenvalues

Eigenvalues and Eigenvectors

Linear Algebra Primer

MTH 464: Computational Linear Algebra

Econ Slides from Lecture 7

Introduction to Matrices

MATH 240 Spring, Chapter 1: Linear Equations and Matrices

Chapters 5 & 6: Theory Review: Solutions Math 308 F Spring 2015

Some Notes on Linear Algebra

Recall : Eigenvalues and Eigenvectors

MAC Module 12 Eigenvalues and Eigenvectors. Learning Objectives. Upon completing this module, you should be able to:

MAC Module 12 Eigenvalues and Eigenvectors

Lecture 10 - Eigenvalues problem

MATRIX ALGEBRA AND SYSTEMS OF EQUATIONS. + + x 1 x 2. x n 8 (4) 3 4 2

Notes on the Matrix-Tree theorem and Cayley s tree enumerator

Knowledge Discovery and Data Mining 1 (VO) ( )

Determinants by Cofactor Expansion (III)

Chapter 2:Determinants. Section 2.1: Determinants by cofactor expansion

Linear Algebra: Lecture Notes. Dr Rachel Quinlan School of Mathematics, Statistics and Applied Mathematics NUI Galway

Math 1553, Introduction to Linear Algebra

Fall 2016 MATH*1160 Final Exam

A note on polar decomposition based Geršgorin-type sets

MAC Module 3 Determinants. Learning Objectives. Upon completing this module, you should be able to:

Math Linear Algebra Final Exam Review Sheet

a 11 a 12 a 11 a 12 a 13 a 21 a 22 a 23 . a 31 a 32 a 33 a 12 a 21 a 23 a 31 a = = = = 12

Numerical Linear Algebra Homework Assignment - Week 2

Basic Concepts in Linear Algebra

Lecture Summaries for Linear Algebra M51A

Math 240 Calculus III

Linear Systems and Matrices

Conceptual Questions for Review

Review of Basic Concepts in Linear Algebra

ELE/MCE 503 Linear Algebra Facts Fall 2018

Matrices and Linear Algebra

Chapter 2. Matrix Arithmetic. Chapter 2

Eigenvalues and Eigenvectors. Review: Invertibility. Eigenvalues and Eigenvectors. The Finite Dimensional Case. January 18, 2018

JUST THE MATHS SLIDES NUMBER 9.6. MATRICES 6 (Eigenvalues and eigenvectors) A.J.Hobson

3 Matrix Algebra. 3.1 Operations on matrices

Math 308 Midterm Answers and Comments July 18, Part A. Short answer questions

Elementary maths for GMT

[Disclaimer: This is not a complete list of everything you need to know, just some of the topics that gave people difficulty.]

Lemma 8: Suppose the N by N matrix A has the following block upper triangular form:

IMPORTANT DEFINITIONS AND THEOREMS REFERENCE SHEET

Lecture 1 Systems of Linear Equations and Matrices

k is a product of elementary matrices.

Math Camp Lecture 4: Linear Algebra. Xiao Yu Wang. Aug 2010 MIT. Xiao Yu Wang (MIT) Math Camp /10 1 / 88

Linear Algebra Highlights

3 (Maths) Linear Algebra

Introduction to Quantitative Techniques for MSc Programmes SCHOOL OF ECONOMICS, MATHEMATICS AND STATISTICS MALET STREET LONDON WC1E 7HX

IMPORTANT DEFINITIONS AND THEOREMS REFERENCE SHEET

Math 520 Exam 2 Topic Outline Sections 1 3 (Xiao/Dumas/Liaw) Spring 2008

Final Review Sheet. B = (1, 1 + 3x, 1 + x 2 ) then 2 + 3x + 6x 2

First, we review some important facts on the location of eigenvalues of matrices.

A Review of Matrix Analysis

7.5 Operations with Matrices. Copyright Cengage Learning. All rights reserved.

and let s calculate the image of some vectors under the transformation T.

1 Multiply Eq. E i by λ 0: (λe i ) (E i ) 2 Multiply Eq. E j by λ and add to Eq. E i : (E i + λe j ) (E i )

Chapter 5. Linear Algebra. A linear (algebraic) equation in. unknowns, x 1, x 2,..., x n, is. an equation of the form

Lecture Notes in Linear Algebra

Extra Problems for Math 2050 Linear Algebra I

Solution Set 7, Fall '12

Determinants. Chia-Ping Chen. Linear Algebra. Professor Department of Computer Science and Engineering National Sun Yat-sen University 1/40

Ir O D = D = ( ) Section 2.6 Example 1. (Bottom of page 119) dim(v ) = dim(l(v, W )) = dim(v ) dim(f ) = dim(v )

Quick Tour of Linear Algebra and Graph Theory

Topic 15 Notes Jeremy Orloff

MAC Module 2 Systems of Linear Equations and Matrices II. Learning Objectives. Upon completing this module, you should be able to :

M. Matrices and Linear Algebra

Homework 1 Elena Davidson (B) (C) (D) (E) (F) (G) (H) (I)

Equality: Two matrices A and B are equal, i.e., A = B if A and B have the same order and the entries of A and B are the same.

7.3. Determinants. Introduction. Prerequisites. Learning Outcomes

Math 343 Midterm I Fall 2006 sections 002 and 003 Instructor: Scott Glasgow

6 EIGENVALUES AND EIGENVECTORS

MATRICES. knowledge on matrices Knowledge on matrix operations. Matrix as a tool of solving linear equations with two or three unknowns.

Matrix decompositions

Matrix Operations. Linear Combination Vector Algebra Angle Between Vectors Projections and Reflections Equality of matrices, Augmented Matrix

Math 4A Notes. Written by Victoria Kala Last updated June 11, 2017

Linear Algebra March 16, 2019

Multiplying matrices by diagonal matrices is faster than usual matrix multiplication.

Linear Algebra in Actuarial Science: Slides to the lecture

Matrix decompositions

REVIEW FOR EXAM II. The exam covers sections , the part of 3.7 on Markov chains, and

A matrix is a rectangular array of. objects arranged in rows and columns. The objects are called the entries. is called the size of the matrix, and

Math 489AB Exercises for Chapter 2 Fall Section 2.3

APPENDIX: MATHEMATICAL INDUCTION AND OTHER FORMS OF PROOF

Chapter 4 - MATRIX ALGEBRA. ... a 2j... a 2n. a i1 a i2... a ij... a in

a s 1.3 Matrix Multiplication. Know how to multiply two matrices and be able to write down the formula

Chapter SSM: Linear Algebra Section Fails to be invertible; since det = 6 6 = Invertible; since det = = 2.

LINEAR ALGEBRA SUMMARY SHEET.

Solving a system by back-substitution, checking consistency of a system (no rows of the form

Linear Algebra Practice Problems

A matrix is a rectangular array of. objects arranged in rows and columns. The objects are called the entries. is called the size of the matrix, and

CHAPTER 3. Matrix Eigenvalue Problems

MATH 310, REVIEW SHEET 2

Transcription:

Gershgorin s Circle Theorem for Estimating the Eigenvalues of a Matrix with Known Error Bounds Author: David Marquis Advisors: Professor Hans De Moor Dr. Kathryn Porter Reader: Dr. Michael Nathanson May 15, 2016

Figure 1: Semyon Aranovich Gershgorin Abstract Square matrices are characterized by their eigenvalues, so we would like to be able to produce the eigenvalues for any square matrix. Cofactor expansion is an algorithm used to produce what is known as the characteristic polynomial of a matrix A, the roots of which are the eigenvalues of A. However, the algorithm for cofactor expansion is computationally intensive and as the output is an n th degree polynomial, we do not have any guaranteed method to find the roots for large matrices. Semyon Aranovich Gershgorin (1901-1933), was a mathematician who worked mainly in the study of differential equations. Among his accomplishments were a degree from the Petrograd Technological Institute, a professorship at the Leningrad Institute of Mechanical Engineering, and he became head of the Division of Mechanics at the Turbine Institute. In 1931, he authored a paper titled Über die Abgrenzung der Eigenwerte einer Matrix in which he gave the proof of what has come to be called the Gershgorin Circle Theorem with provided a simple method to estimate the eigenvalues of a complex matrix with known error bounding. Introduction In linear algebra, linear systems of equations in the form of matrices are studied. Definition: A matrix is an n m array of elements from a vector space. In this essay, we

will only consider matrices with entries from the complex numbers. An example of a matrix is i 0 87 12 12 3 + 2i 8i 1 0.11 54 45 3 i Matrices can easily represent a system of n equations in m variables. A matrix that has the same number of rows and columns, is called a square matrix. One way we can study such a matrix is to find its eigenvalues and eigenvectors. Definition: An eigenvector is a column vector which when multiplied on the left by a matrix results in a scalar multiple of the original vector or Ax = λx. Although the equation A 0 = λ 0 is true for all A and λ, we define 0 to not be an eigenvector for any matrix. Definition: The value λ is what is known as the eigenvalue associated with the eigenvector x. The matrix ] A = [ 3 2 1 2 has eigenvalues λ 1 = 4 and λ 2 = 1. The corresponding eigenvectors are x 1 = [ ] 1 x 2 = and all scalar multiples of either. This is because 1 [ ] [ ] [ ] [ ] 3 2 2 8 2 = = 4 1 2 1 4 1 [ 2 1 ] and and [ 3 2 1 2 Thus Ax 1 = λ 1 x 1 and Ax 2 = λ 2 x 2. ] [ 1 1 ] = [ 1 1 ] [ 1 = 1 1 ] Eigenvalues and eigenvectors characterize a matrix. Another related value associated with a matrix is its determinant. Although for this paper, we won t need to examine the determinants of any matrices, we will need to know the algorithm for finding determinants as it leads us to an algorithm for finding eigenvalues. The general method for finding the determinant of a matrix is called cofactor expansion. For a matrix R = [r i,j ], the minor M k,l of the entry r k,l is the determinant of the matrix that

is left after removing the k th row and l th column from R. The cofactor C k,l of r k,l is then ( 1) k+l M k,l. If the cofactor produced then requires one to evaluate the determinant of a matrix one dimension smaller than the previous matrix, we then perform cofactor expansion of that minor. We continue this process until all of the determinants are of 2 2 matrices. The formula to evaluate the determinant of a 2 2 matrix is [ ] a b = ad bc c d By selecting a row or column of R and for that row or column, summing the product of each entry with its cofactor, we find the determinant of R. The equation Ax = λx is equivalent to the equation Ax = λix where I, the identity matrix, is the matrix with the same dimension as A whose entries are 1 s along the main diagonal and 0 elsewhere. Then, we can subtract Ax from both sides and factor x from both terms to get (λi A)x = 0 Because x 0, (λi A) must be singular, so det(λi A) = 0 Solving for which values of λ satisfy the above equation produces the eigenvalues of A. The eigenvalues in certain matrices are very easy to determine. Definition: An upper triangular matrix is a square matrix with all entries below the main diagonal equal to zero. To illustrate, if you replace each? with any complex number, then the following matrix will be upper triangular????? 0???? 0 0??? 0 0 0?? 0 0 0 0? Theorem: The eigenvalues of a triangular matrix (upper, lower or diagonal) are exactly the entries on the main diagonal.

This is because, for a matrix [a i,j ] if you imagine performing cofactor expansion on the first column of (λi A), there will only be one non-zero term to sum and it will be (λ a 1,1 ) times the determinant of the minor. The determinant of that minor is then (λ a 2,2 ) times the determinant of its minor, et cetera until we get the factored characteristic polynomial (λ a 1,1 )(λ a 2,2 )...(λ a n,n ) = 0. Thus the eigenvalues of a triangular matrix are the values along its main diagonal. Example 1: Determine the eigenvalues of the matrix 2 3 + i 8 0 0 0.5 0 0 9 Because this is an upper triangular matrix, we can easily tell that the eigenvalues are 2, 0, and 9. Example 2: Determine the eigenvalues of the matrix 10 2 3 0 11 1 0 1 13 It is incorrect to claim that the eigenvalues are 10, 11, & 13 because this matrix has non-zero entries both above and below the main diagonal, that claim would be false. We have to find the solutions to the characteristic equation det(λi A) = 0 or λ 10 2 3 0 λ 11 1 0 1 λ 13 = 0 Performing cofactor expansion on the first column we obtain which when factored is (λ 10)((λ 11)(λ 13) ((1)( 1))) = 0 (λ 10)(λ 2 24λ + 144) = 0 (λ 10)(λ 12)(λ 12) = 0 and finally tells us that the eigenvalues of this matrix are 10 and 12. Calculating eigenvalues in this way is computationally intensive even for a small matrix, so as the size of the matrix becomes very large, a significant amount of processing power is necessary to find the eigenvalues.

Notice also that for the matrix 10 2 3 0 11 1 0 0 13 (notice that a 3,2 was changed from a 1 to a 0) we can see that the eigenvalues are 10, 11, and 13. So it would seem that small changes to the values of the entries may only result in small changes to the eigenvalues. From this observation, back in 1931, Semyon Aronovich Gershgorin devised a method to estimate the eigenvalues of a square matrix with a known error bound. Gershgorin s Circle Theorem The concept of the Gershgorin Circle Theorem is that one can take the diagonal entries of an n n matrix as the coordinates in the complex plane. These points then act as the centers of n discs which have radii of the sum of the magnitudes of the n 1 other entries from the same row. Then, all of the eigenvalues of this matrix will lie within the union of these discs. We now state the theorem formally preceded by a definition. Definition: If A is an n n matrix with complex entries a i,j then r i (A) = i j a i,j is defined as the sum of the magnitudes of the non-diagonal entries of the i th row. Then a Gershgorin disc is the disc D(a i,i, r i (A)) centered at a ii on the complex plane with radius r i (A). Theorem: Every eigenvalue of a matrix lies within at least one Gershgorin disc. Proof (Varga 2010). Let λ be an eigenvalue of a matrix A and let x = (x j ) be its corresponding non-zero eigenvector. Since x is an eigenvector of A, we can rewrite Ax = λx using sigma notation as inner product of each row of A with x: a ij x j = λx i Subtracting a ii x i from both sides splits the sum: j

a ij x j = λx i a ii x i Due to the fact that x 0, we know there is some k N such that 0 < x k = max{ x i : i N} With this k, we see i j which is the same as Using the triangle inequality, a ki x i = λx k i N a ki x i = (λ a k,k )x k i j λ a k,k x k = i j a ai x i i j a ki x k = x k r k (A) where i j a ki x k = x k r k (A) comes from the definition of a Gershgorin Disc. Finally, since x k is greater than 0 we can divide by x k, to get λ a kk r k (A) As an example, consider the graph of the Gershgorin discs for our sample matrix 10 2 3 A = 0 11 1 0 1 13 Since the entries on the main diagonal are 10+0i, 11+0i, and 13+0i, those will be the centers of the three discs in the complex plane. Those three discs will have radii of 2 + 3 = 5, 0 + 1 = 1, and 0 + 1 = 1 respectively see Figure 2. Notice in Figure 2, that the two smaller discs are contained within the larger disc. This leads us to an important corollary to Gershgorin s Circle Theorem.

Figure 2: The Gershgorin discs for matrix A Corollary 1: If Ŝ is a union of m discs R i such that Ŝ is disjoint from all other discs, then Ŝ contains precisely m eigenvalues (counting multiplicities) of A Proof of Corollary 1: (Ortega 1990). Let A be an n n matrix. Then, we can create a family of matrices A(t) = tb + D where D is the same as A with all the off-diagonal entries reduced to zero and B is the same as A with all the diagonal entries reduced to zero all along the interval 0 t 1. Then, each Gershgorin disc of A(t) will have the same center of a disc from A, but with its radius scaled by t. Now if we let R i (t) be the disc with center a i,i and radius t j i a i,j and define Ŝ = m i=1 R i, we can then set Ŝ(t) = m i=1 R i(t) and S(t) = n i=m+1 R i. From our assumptions, the sets Ŝ and S are disjoint and therefore Ŝ(t) and S(t) are disjoint throughout the interval of t. We can see that Ŝ(0) contains m eigenvalues of A 0. Since Gershgorin s Circle Theorem tells us that all eigenvalues of A t are contained in Ŝ S. However, since Ŝ S = and an eigenvalue cannot jump from one set to the other, Ŝ(0) containing exactly m eigenvalues implies that Ŝ(t) contains exactly m eigenvalues for the entire interval of t. To demonstrate this corollary, see figure 3 which depicts a subset of the Gershgorin discs for some matrix (with the actual locations of the eigenvalues marked with x s).

Notice that of the three discs, the one that is disjoint contains exactly one eigenvalue, yet of the two continuous discs, one contains both of their eigenvalues and the other has none. Refinement of Gershgorin s Circle Theorem Even though we already know a method to precisely calculate the eigenvalues of a matrix, namely cofactor expansion, we may still want to be able to estimate them. Cofactor expansion becomes exponentially more computationally expensive as n increases. This can become cumbersome even when automated, so application of Gershgorin s Theorem may become more useful for many real-world applications. One thing we can do to make our estimations better is to realize that taking the transpose of a square matrix does not change its eigenvalues. Therefore, we can compare the results of using Gershgorin s Circle Theorem on a matrix A with those from A T and use the more favorable results. Example 3: Use Gershgorin s Circle Theorem to find a bound for A T where 10 2 3 A = 0 11 1 0 1 13 The transpose of A is See Figure 4. A T = 10 0 0 2 11 1 3 1 13 Another, less commonly applicable, corollary of Gershgorin s Circle Theorem which relates back to the first corollary is thus: Corollary 2. Any Gershgorin disc with radius 0 is an eigenvalue. Proof of Corollary 2. (Marquis 2016). Suppose A = [a i,j ] is a matrix such that at least one of its Gershgorin discs has radius 0. That means that A has at least one row with nondiagonal entries all equal to zero. Without loss of generality, call this the k th row. If we

Figure 3: Demonstration of the application of corollary 1

Figure 4: The Gershgorin Discs for A T form the matrix (λi A) and perform cofactor expansion along the k th row, we will get (λ a k,k )( 1) k+k M k,k. Thus, (λ a k,k ) is a factor of the characteristic polynomial, so a k,k is a root, so a k,k is an eigenvalue of A. Corollary 2 is obvious for a disc of radius 0, or point-disc, by itself, so corollary 2 is only really useful when a point-disc lies within another disc. In that case, corollary 1 would tell us that the two eigenvalues could be anywhere within the two discs, but corollary 2 guarantees that every disc with radius 0 is an eigenvalue. Notice also that when comparing the Gershgorin Discs for A and A T, there are different bounds on the possible eigenvalues because the discs are not the same unless A is a symmetric matrix. In our examples along the real axis, the discs of A reach from 5 to 15 whereas the discs of A T reach from 8 to 17. Similarly, the upper and lower bounds of the complex components at any value along the real axis vary between the two sets of discs. This leads to the realization that we can bound the eigenvalues by the intersection of the unions of the Gershgorin discs for A and A T i.e. for A = [a i,j ] and A T = [b k,l ], D(ai,i, R i ) D(b k,k, R k ) So, for our example matrix, our bound for the eigenvalues of A becomes the shaded area in Figure 5.

Figure 5: Improved bound using the intersection of the unions of the discs. Final Thoughts Whether we are interested in the eigenvalues of a matrix to learn how a linear system acts, as a stepping stone for finding the eigenvectors, or to study matrices in an abstract sense, we need to have an algorithmic method for finding these eigenvalues. Although cofactor expansion works well on small matrices, pending advances in establishing roots of n th degree polynomials, alternative methods must be used. The power method and inverse power method are popular choices for estimating the eigenvalue with greatest magnitude, but still require iterative processes and, theoretically, could fail if iterated on a vector orthogonal to the dominant eigenvector. The real power of Gershgorin s Circle Theorem lies in how easily it allows approximation of all of the eigenvalues. Summing n 1 terms n times is about as simple of a calculation as can be. In addition, finding the intersection of the union of a matrix A s gershgorin discs with the union of the discs from A T can be easily implemented to improve estimates of any non-symmetric matrix. In practice, though Gershgorin s Circle Theorem is powerful, using other techniques to augment Gershgorin s Theorem can make ever more potent estimations. For instance, one can use Gershgorin s Theorem to estimate the dominant eigenvalue, use that to estimate the corresponding eigenvector, and then use the power method on that to improve the estimates of both. One can also precondition a matrix before applying Gershgorin s Theorem in an attempt to decrease the area of the union of the Gershgorin discs.

References Anton, Howard. Elementary Linear Algebra. Print. O Connor, J. J., and E. F. Robertson. Semyon Aranovich Gershgorin. Gershgorin Biography. Jan. 2012. Web. 12 Apr. 2016. Ortega, James M. Numerical Analysis: A Second Course. Philadelphia: Society for Industrial and Applied Mathematics, 1990. Print. On the Accuracy of the Gershgorin Circle Theorem for Bounding the Spread of a Real Symmetric Matrix by David S. Scott https://stmarys-ca.illiad.oclc.org/illiad/illiad.dll?action%3d10&form%3d Varga, Richard S. Gerssgorin and His Circles. Heidelberg: Springer-Verlag, 2010. Print.