c 1995 Society for Industrial and Applied Mathematics Vol. 37, No. 1, pp , March

Similar documents
Up to this point, our main theoretical tools for finding eigenvalues without using det{a λi} = 0 have been the trace and determinant formulas

CHAPTER 5. Basic Iterative Methods

Math 5630: Iterative Methods for Systems of Equations Hung Phan, UMass Lowell March 22, 2018

Math Introduction to Numerical Analysis - Class Notes. Fernando Guevara Vasquez. Version Date: January 17, 2012.

CAAM 454/554: Stationary Iterative Methods

Department of Mathematics California State University, Los Angeles Master s Degree Comprehensive Examination in. NUMERICAL ANALYSIS Spring 2015

Scientific Computing WS 2018/2019. Lecture 9. Jürgen Fuhrmann Lecture 9 Slide 1

The Solution of Linear Systems AX = B

Algebra C Numerical Linear Algebra Sample Exam Problems

Lecture Note 7: Iterative methods for solving linear systems. Xiaoqun Zhang Shanghai Jiao Tong University

Preliminary/Qualifying Exam in Numerical Analysis (Math 502a) Spring 2012

Numerical Methods - Numerical Linear Algebra

On the Schur Complement of Diagonally Dominant Matrices

Solving Linear Systems

Jordan Journal of Mathematics and Statistics (JJMS) 5(3), 2012, pp A NEW ITERATIVE METHOD FOR SOLVING LINEAR SYSTEMS OF EQUATIONS

AN ELEMENTARY PROOF OF THE SPECTRAL RADIUS FORMULA FOR MATRICES

Computational Methods. Systems of Linear Equations

Notes on Linear Algebra and Matrix Theory

First, we review some important facts on the location of eigenvalues of matrices.

The convergence of stationary iterations with indefinite splitting

LECTURE NOTES ELEMENTARY NUMERICAL METHODS. Eusebius Doedel

Math/Phys/Engr 428, Math 529/Phys 528 Numerical Methods - Summer Homework 3 Due: Tuesday, July 3, 2018

CPE 310: Numerical Analysis for Engineers

A NEW EFFECTIVE PRECONDITIONED METHOD FOR L-MATRICES

Numerical Analysis: Solving Systems of Linear Equations

Iterative Methods for Ax=b

Computational Linear Algebra

On the adjacency matrix of a block graph

ON A HOMOTOPY BASED METHOD FOR SOLVING SYSTEMS OF LINEAR EQUATIONS

Modified Gauss Seidel type methods and Jacobi type methods for Z-matrices

Recall the convention that, for us, all vectors are column vectors.

Linear Algebra. Matrices Operations. Consider, for example, a system of equations such as x + 2y z + 4w = 0, 3x 4y + 2z 6w = 0, x 3y 2z + w = 0.

Ma/CS 6b Class 20: Spectral Graph Theory

BOUNDS OF MODULUS OF EIGENVALUES BASED ON STEIN EQUATION

Z-Pencils. November 20, Abstract

Chapter 7. Canonical Forms. 7.1 Eigenvalues and Eigenvectors

Math 102, Winter Final Exam Review. Chapter 1. Matrices and Gaussian Elimination

Iterative Methods. Splitting Methods

Ma/CS 6b Class 20: Spectral Graph Theory

COURSE Numerical methods for solving linear systems. Practical solving of many problems eventually leads to solving linear systems.

Eventual Linear Ranking Functions

Recall : Eigenvalues and Eigenvectors

Discrete mathematical models in the analysis of splitting iterative methods for linear systems

MAT 1332: CALCULUS FOR LIFE SCIENCES. Contents. 1. Review: Linear Algebra II Vectors and matrices Definition. 1.2.

MATH 5720: Unconstrained Optimization Hung Phan, UMass Lowell September 13, 2018

Computationally, diagonal matrices are the easiest to work with. With this idea in mind, we introduce similarity:

Eigenvalues and Eigenvectors

Generalized line criterion for Gauss-Seidel method

Eigenvalues and Eigenvectors 7.2 Diagonalization

Mathematical Optimisation, Chpt 2: Linear Equations and inequalities

Iterative techniques in matrix algebra

(f + g)(s) = f(s) + g(s) for f, g V, s S (cf)(s) = cf(s) for c F, f V, s S

STAT200C: Review of Linear Algebra

Kernels of Directed Graph Laplacians. J. S. Caughman and J.J.P. Veerman

Markov Chains and Stochastic Sampling

Minimum number of non-zero-entries in a 7 7 stable matrix

Here each term has degree 2 (the sum of exponents is 2 for all summands). A quadratic form of three variables looks as

Next topics: Solving systems of linear equations

10.34: Numerical Methods Applied to Chemical Engineering. Lecture 7: Solutions of nonlinear equations Newton-Raphson method

In particular, if A is a square matrix and λ is one of its eigenvalues, then we can find a non-zero column vector X with

Here is an example of a block diagonal matrix with Jordan Blocks on the diagonal: J

Foundations of Matrix Analysis

Theory of Iterative Methods

On the simultaneous diagonal stability of a pair of positive linear systems

Introduction to Scientific Computing

Solving Linear Systems of Equations

Solving Linear Systems

Lecture Notes: Matrix Inverse. 1 Inverse Definition. 2 Inverse Existence and Uniqueness

Eigenvalue and Eigenvector Problems

Classical iterative methods for linear systems

Convex Functions and Optimization

CS 323: Numerical Analysis and Computing

Chapter 2. Solving Systems of Equations. 2.1 Gaussian elimination

Computing Eigenvalues and/or Eigenvectors;Part 1, Generalities and symmetric matrices

Question: Given an n x n matrix A, how do we find its eigenvalues? Idea: Suppose c is an eigenvalue of A, then what is the determinant of A-cI?

Lecture 7: Positive Semidefinite Matrices

Mathematical Olympiad Training Polynomials

Ir O D = D = ( ) Section 2.6 Example 1. (Bottom of page 119) dim(v ) = dim(l(v, W )) = dim(v ) dim(f ) = dim(v )

Numerical Linear Algebra Homework Assignment - Week 2

1. Nonlinear Equations. This lecture note excerpted parts from Michael Heath and Max Gunzburger. f(x) = 0

LINEAR SYSTEMS (11) Intensive Computation

Bindel, Fall 2016 Matrix Computations (CS 6210) Notes for

4. Determinants.

Chapter 9: Systems of Equations and Inequalities

Detailed Proof of The PerronFrobenius Theorem

OPTIMAL SCALING FOR P -NORMS AND COMPONENTWISE DISTANCE TO SINGULARITY

LINEAR ALGEBRA KNOWLEDGE SURVEY

Linear Algebra Primer

Gershgorin s Circle Theorem for Estimating the Eigenvalues of a Matrix with Known Error Bounds

Goal: to construct some general-purpose algorithms for solving systems of linear Equations

Lecture 18 Classical Iterative Methods

SPRING OF 2008 D. DETERMINANTS

EIGENVALUE PROBLEMS. Background on eigenvalues/ eigenvectors / decompositions. Perturbation analysis, condition numbers..

Lesson 3. Inverse of Matrices by Determinants and Gauss-Jordan Method

G1110 & 852G1 Numerical Linear Algebra

Math Camp Lecture 4: Linear Algebra. Xiao Yu Wang. Aug 2010 MIT. Xiao Yu Wang (MIT) Math Camp /10 1 / 88

10.2 ITERATIVE METHODS FOR SOLVING LINEAR SYSTEMS. The Jacobi Method

Eigenvalues and Eigenvectors

Chapter 12: Iterative Methods

Process Model Formulation and Solution, 3E4

Transcription:

SIAM REVIEW. c 1995 Society for Industrial and Applied Mathematics Vol. 37, No. 1, pp. 93 97, March 1995 008 A UNIFIED PROOF FOR THE CONVERGENCE OF JACOBI AND GAUSS-SEIDEL METHODS * ROBERTO BAGNARA Abstract. We present a new unified proof for the convergence of both the Jacobi and the Gauss Seidel methods for solving systems of linear equations under the criterion of either (a) strict diagonal dominance of the matrix, or (b) diagonal dominance and irreducibility of the matrix. These results are well known. The proof for criterion (a) makes use of Geršgorin s theorem, while the proof for criterion (b) uses Taussky s theorem that extends Geršgorin s work. Hence the topic is interesting for teaching purposes. Key words. Systems of equations, iterative methods. AMS subject classifications. 65-01, 65F10 We present a unified proof for the convergence of both the Jacobi and the Gauss Seidel iterative methods for solving systems of linear equations under the criterion of either (a) strict diagonal dominance of the matrix, or (b) diagonal dominance and irreducibility of the matrix. These results are well known [2]. For each criterion, the proof is unified in that it uses similar arguments for both the methods. The proof is simple because low level details are contained in an easy Lemma and a couple of Observations, while the rest of the analysis is based on elementary properties of matrices and their eigenvalues. We recall three fundamental results about the location of the eigenvalues, known as Geršgorin type theorems [1]. Let A C n n be a complex matrix: the Geršgorin (row) disks of A are given, for i = 1,..., n, by { (1) K i = z C z a ii } a ij. The first Geršgorin theorem says that the eigenvalues of A are contained within the union, K, of the disks, where (2) K = n K i. The second Geršgorin theorem says further that any union of k disks of A not intersecting the remaining (n k) disks, must contain exactly k of the eigenvalues (counting multiplicities). The third Geršgorin type theorem was established by Taussky [1] and says that if A is irreducible and has an eigenvalue λ on the boundary of K, then λ is on the boundary of each of the n Geršgorin disks of A. Notice that since A and A T have exactly the same eigenvalues, analogous results hold when column sums of A, instead of row sums, are used to define the radii of the corresponding (column) disks. *Received by the editors July 21, 1993; accepted for publication (in revised form) July 14, 1994. Dipartimento di Informatica, Università di Pisa, Corso Italia, 40, 56125 Pisa, Italy (bagnara@di.unipi.it). 93

94 roberto bagnara We now define the terms used in criteria (a) and (b): Definition 1. A complex matrix A C n n is reducible iff there exist a permutation matrix Π (i.e., Π is obtained from the identity, I, by a permutation of the rows of I) and an integer k {1,..., n 1} such that ( (3) ΠAΠ T A11 A = 12 0 A 22 where A 11 is k k and A 22 is (n k) (n k). If A is not reducible, than A is said to be irreducible (see [1] for further discussion of irreducibility). Note that premultiplication of A by Π, produces a matrix ΠA that consists of the rows of A in the same order that the rows of I appear in Π. On the other hand, when the columns of I are permuted in that order, then the matrix Π T is produced. Thus, AΠ T is a matrix consisting of the columns of A in that same order. Definition 2. A C n n is said to be strictly diagonally dominant (by rows) iff, for each i = 1,..., n, (4) a ii > a ij ; similarly, A is said to be strictly diagonally dominant (by columns) iff, for each j = 1,..., n, (5) a jj > a ij. i j Definition 3. A C n n is said to be diagonally dominant (by rows) iff, for each i = 1,..., n, (6) a ii ), a ij and s {1,..., n} such that a ss > a sj ; similarly, A is said to be diagonally dominant (by columns) iff, for each j = 1,..., n, (7) a jj a ij and t {1,..., n} such that a tt > i j j s a it. Here are a couple of observations which are useful in proving the convergence theorem: Observation 1. If the matrix, A, is strictly diagonally dominant, then A is nonsingular. Proof. None of the Geršgorin disks of A contain the origin and therefore, by the first Geršgorin theorem, zero is not an eigenvalue of A. Observation 2. If the matrix, A, is irreducible and diagonally dominant, then A is nonsingular. Proof. Otherwise, if A were singular, then one of its eigenvalues would be zero and, by the first Geršgorin theorem, would be contained in at least one of the disks. i t

convergence of jacobi and gauss seidel methods 95 But, by the assumed diagonal dominance, zero cannot be in the interior of any of the disks. Thus, zero would have to be on the boundary of the union, K, of the disks. However, Taussky s theorem would then place zero on the boundary of each of the disks. But, this would be impossible for either of the disks labelled s or t in Definition 3. The Jacobi and Gauss-Seidel iterative methods to solve the system (8) Ax = b where A C n n, x, b C n, are formulated in terms of a decomposition of the matrix A: (9) A = D B C, where the components of D, B, and C are defined by { { aij, if i = j, aij, if i > j, (10) d ij = b 0, if i j; ij = 0, if i j; c ij = { aij, if i < j, 0, if i j. Both iterative methods begin with an arbitrary initial vector x 0 C n and then produce a sequence of vectors, x k, for k = 1, 2, 3,..., by solving either: (11) Jacobi: (12) Gauss Seidel: Dx k+1 = (B + C)x k + b, (D B)x k+1 = Cx k + b, or for k = 0, 1, 2,.... It is clear that, for the Jacobi method to be applicable, D must be invertible. Similarly, the invertibility of (D B) is required in order to apply the Gauss Seidel method. The iteration matrices of the methods are then given, respectively, by (13) Jacobi: (14) Gauss Seidel: J = D 1 (B + C), G = (D B) 1 C. The iterative steps of the methods are then defined, respectively, by (15) Jacobi: (16) Gauss Seidel: x k+1 = Jx k + D 1 b, or x k+1 = Gx k + (D B) 1 b, for k = 0, 1, 2,.... Let ρ[q] denote the spectral radius of the matrix Q, for any Q C n n. The Jacobi method is convergent iff ρ[j] < 1, while the Gauss Seidel method is convergent iff ρ[g] < 1. Theorem 1 (convergence). Let A C n n be decomposed as in (9), and let J and G be defined as in (13) and (14), respectively. Under either the criterion (a): A is strictly diagonally dominant; or the criterion (b): A is diagonally dominant and irreducible, then (i) both D and (D B) are invertible; and (ii) ρ[j] < 1, and ρ[g] < 1. Proof of (i). For criterion (a). We see that a ii 0 for all i = 1,..., n, whence D and (D B) are nonsingular. For criterion (b). We note by Observation 2 that A is nonsingular and hence does not have a zero row. On the other hand, if either D or (D B) were assumed to be singular, then d kk = 0 for some k = 1,..., n. But, then by diagonal dominance we would see that row k of A would be a zero row. This contradiction establishes (i) for criterion (b).

96 roberto bagnara Proof of (ii). In order to show that both spectral radii are less than 1, we define the matrices A J (λ) and A G (λ), where λ C, by (17) (18) A J (λ) = λd B C; A G (λ) = λ(d B) C; and establish the following: Lemma. For each λ C, with λ 1, if A satisfies any of the following properties: (a) A is strictly diagonally dominant (by rows or by columns); (b) A is diagonally dominant (by rows, or by columns); (c) A is irreducible; then both A J (λ) and A G (λ) satisfy the same properties. Proof. (a) Let A be strictly diagonally dominant by rows (the proof for the other case is almost the same). By hypothesis, for each i = 1,..., n, (19) a ii > if λ 1 then, for each i = 1,..., n, a ij, (20) ( ) λa ii = λ a ii > λ a ij = λ a ij + λ λ a ij + = λa ij + a ij + a ij a ij a ij hence the thesis for A G (λ), a ij hence the thesis for A J (λ). (b) Very similar to (a). The only difference is that the hypothesis ensures that strict inequality holds in the disequation marked with ( ) for at least one i {1,..., n}. Weak inequality is guaranteed for all the other cases. (c) Since the three matrices A, A J (λ), and A G (λ), for λ 0, have zero components in exactly the same locations, it follows that if a permutation matrix Π reduces one of these matrices, then Π also reduces the other two matrices. We now resume the proof of (ii), Theorem 1. The eigenvalues λ of J are all and only the solutions of the equation (21) det(λi J) = 0,

convergence of jacobi and gauss seidel methods 97 but, from properties of matrices and determinants we have (22) det(λi J) = det ( λi D 1 (B + C) ) = det ( λd 1 D D 1 (B + C) ) = det ( D 1 (λd B C) ) = det(d 1 ) det(λd B C) = det(d 1 ) det(a J (λ). So, for (21) to hold we must have det(a J (λ) = 0, as we have already shown that D is nonsingular. But, since A J (λ) is nonsingular for λ 1 (by the Lemma and the Observations), it follows that all of the eigenvalues of J are in the interior of the unit circle. Hence, ρ[j] < 1. Similarly, the eigenvalues λ of G are all and only the solutions of the equation (23) det(λi G) = det ( λi (D B) 1 C ) = det ( λ(d B) 1 (D B) (D B) 1 C ) = det ((D B) 1( λ(d B) C )) = det ( (D B) 1) det ( λ(d B) C ) = det ( (D B) 1) det ( A G (λ) ) = 0. Reasoning in the exactly the same way as above we conclude that ρ[g] < 1. Acknowledgements. We wish to thank the anonymous referee for a detailed and careful review of the manuscript, which helped to substantially improve the presentation. REFERENCES [1] P. Lancaster and M. Tismenetsky, The Theory of Matrices, Second Edition with Applications, Academic Press, Orlando, 1985. [2] J. Stoer and R. Bulirsch, Introduction to Numerical Analysis, Springer-Verlag, New York, 1980.