CAAM 454/554: Stationary Iterative Methods

Similar documents
Lecture Note 7: Iterative methods for solving linear systems. Xiaoqun Zhang Shanghai Jiao Tong University

Solving Linear Systems

COURSE Iterative methods for solving linear systems

Goal: to construct some general-purpose algorithms for solving systems of linear Equations

Math Introduction to Numerical Analysis - Class Notes. Fernando Guevara Vasquez. Version Date: January 17, 2012.

CLASSICAL ITERATIVE METHODS

Jordan Journal of Mathematics and Statistics (JJMS) 5(3), 2012, pp A NEW ITERATIVE METHOD FOR SOLVING LINEAR SYSTEMS OF EQUATIONS

Iterative Methods for Solving A x = b

Bindel, Fall 2016 Matrix Computations (CS 6210) Notes for

9. Iterative Methods for Large Linear Systems

Chapter 7 Iterative Techniques in Matrix Algebra

Math 471 (Numerical methods) Chapter 3 (second half). System of equations

Lecture 18 Classical Iterative Methods

Preliminary/Qualifying Exam in Numerical Analysis (Math 502a) Spring 2012

Computational Linear Algebra

CHAPTER 5. Basic Iterative Methods

Numerical Methods - Numerical Linear Algebra

3.2 Iterative Solution Methods for Solving Linear

The Solution of Linear Systems AX = B

Classical iterative methods for linear systems

JACOBI S ITERATION METHOD

Iterative Methods for Ax=b

Math/Phys/Engr 428, Math 529/Phys 528 Numerical Methods - Summer Homework 3 Due: Tuesday, July 3, 2018

The amount of work to construct each new guess from the previous one should be a small multiple of the number of nonzeros in A.

Lab 1: Iterative Methods for Solving Linear Systems

Up to this point, our main theoretical tools for finding eigenvalues without using det{a λi} = 0 have been the trace and determinant formulas

Iterative techniques in matrix algebra

Introduction to Scientific Computing

Math 1080: Numerical Linear Algebra Chapter 4, Iterative Methods

Introduction. Math 1080: Numerical Linear Algebra Chapter 4, Iterative Methods. Example: First Order Richardson. Strategy

4.6 Iterative Solvers for Linear Systems

9.1 Preconditioned Krylov Subspace Methods

Iterative Solution methods

Solving Linear Systems

Math 5630: Iterative Methods for Systems of Equations Hung Phan, UMass Lowell March 22, 2018

EXAMPLES OF CLASSICAL ITERATIVE METHODS

Iterative Methods. Splitting Methods

Here is an example of a block diagonal matrix with Jordan Blocks on the diagonal: J

c 1995 Society for Industrial and Applied Mathematics Vol. 37, No. 1, pp , March

Analysis of Iterative Methods for Solving Sparse Linear Systems C. David Levermore 9 May 2013

Chapter 2. Solving Systems of Equations. 2.1 Gaussian elimination

The convergence of stationary iterations with indefinite splitting

FEM and Sparse Linear System Solving

Computational Economics and Finance

Computational Methods. Systems of Linear Equations

Some definitions. Math 1080: Numerical Linear Algebra Chapter 5, Solving Ax = b by Optimization. A-inner product. Important facts

Notes for CS542G (Iterative Solvers for Linear Systems)

1. Nonlinear Equations. This lecture note excerpted parts from Michael Heath and Max Gunzburger. f(x) = 0

Next topics: Solving systems of linear equations

Partitioned Methods for Multifield Problems

Numerical Programming I (for CSE)

Iterative methods for Linear System

30.5. Iterative Methods for Systems of Equations. Introduction. Prerequisites. Learning Outcomes

Convergence of a Class of Stationary Iterative Methods for Saddle Point Problems

Stabilization and Acceleration of Algebraic Multigrid Method

Scientific Computing WS 2018/2019. Lecture 9. Jürgen Fuhrmann Lecture 9 Slide 1

Review of matrices. Let m, n IN. A rectangle of numbers written like A =

Introduction to Iterative Solvers of Linear Systems

Lecture 11. Fast Linear Solvers: Iterative Methods. J. Chaudhry. Department of Mathematics and Statistics University of New Mexico

Parallel Numerics, WT 2016/ Iterative Methods for Sparse Linear Systems of Equations. page 1 of 1

Background. Background. C. T. Kelley NC State University tim C. T. Kelley Background NCSU, Spring / 58

MA3232 Numerical Analysis Week 9. James Cooley (1926-)

Theory of Iterative Methods

6. Iterative Methods for Linear Systems. The stepwise approach to the solution...

Numerical Analysis: Solutions of System of. Linear Equation. Natasha S. Sharma, PhD

Num. discretization. Numerical simulations. Finite elements Finite volume. Finite diff. Mathematical model A set of ODEs/PDEs. Integral equations

Kasetsart University Workshop. Multigrid methods: An introduction

Computers and Mathematics with Applications. Convergence analysis of the preconditioned Gauss Seidel method for H-matrices

DEN: Linear algebra numerical view (GEM: Gauss elimination method for reducing a full rank matrix to upper-triangular

AMS526: Numerical Analysis I (Numerical Linear Algebra)

COURSE Numerical methods for solving linear systems. Practical solving of many problems eventually leads to solving linear systems.

Recall : Eigenvalues and Eigenvectors

Jae Heon Yun and Yu Du Han

Algebra C Numerical Linear Algebra Sample Exam Problems

Conjugate Gradient (CG) Method

Monte Carlo Method for Finding the Solution of Dirichlet Partial Differential Equations

Math 108b: Notes on the Spectral Theorem

Lecture 4 Basic Iterative Methods I

10.2 ITERATIVE METHODS FOR SOLVING LINEAR SYSTEMS. The Jacobi Method

A generalization of the Gauss-Seidel iteration method for solving absolute value equations

Lecture Notes 6: Dynamic Equations Part C: Linear Difference Equation Systems

Introduction and Stationary Iterative Methods

A NEW EFFECTIVE PRECONDITIONED METHOD FOR L-MATRICES

ON A HOMOTOPY BASED METHOD FOR SOLVING SYSTEMS OF LINEAR EQUATIONS

MAT 610: Numerical Linear Algebra. James V. Lambers

EE5120 Linear Algebra: Tutorial 6, July-Dec Covers sec 4.2, 5.1, 5.2 of GS

Iterative Methods and Multigrid

Chapter 12: Iterative Methods

AN ITERATION. In part as motivation, we consider an iteration method for solving a system of linear equations which has the form x Ax = b

Fixed Points and Contractive Transformations. Ron Goldman Department of Computer Science Rice University

The upper Jacobi and upper Gauss Seidel type iterative methods for preconditioned linear systems

Motivation: Sparse matrices and numerical PDE's

Midterm for Introduction to Numerical Analysis I, AMSC/CMSC 466, on 10/29/2015

Math 577 Assignment 7

Iterative Methods and Multigrid

Solving Linear Systems of Equations

Lecture 1: Entropy, convexity, and matrix scaling CSE 599S: Entropy optimality, Winter 2016 Instructor: James R. Lee Last updated: January 24, 2016

Gauss-Seidel method. Dr. Motilal Panigrahi. Dr. Motilal Panigrahi, Nirma University

Sparse Linear Systems. Iterative Methods for Sparse Linear Systems. Motivation for Studying Sparse Linear Systems. Partial Differential Equations

Boundary Value Problems - Solving 3-D Finite-Difference problems Jacob White

Transcription:

CAAM 454/554: Stationary Iterative Methods Yin Zhang (draft) CAAM, Rice University, Houston, TX 77005 2007, Revised 2010 Abstract Stationary iterative methods for solving systems of linear equations are considered by some as out of date and out of favor, as compared to methods based on Krylov subspace iterations. However, these methods are still useful in many circumstances because they are easier to implement and, more importantly, can be used as pre-conditioners in combination with Krylov-subspace methods. In this note, we briefly introduce the fundamental ideas of stationary iterative methods. 1 Introduction We consider solving a linear system of equations Ax = b, (1) where we will always assume, unless specified otherwise, that A is n by n and real, i.e., A R n n, nonsingular, and the right-hand side (RHS) b R n is nonzero. For any nonsingular matrix Q R n n, one can rewrite the system into an equivalent form: Ax = b x = Mx + c, (2) 1

where, M = I Q 1 A, c = Q 1 b, (3) The equivalence can be easily seen from Mx + c = x + Q 1 (b Ax). (4) Based on the equation x = Mx + c, we can derive a stationary iterative method of the form: given an initial guess x 0 R n, for k = 0, 1,, do x k+1 = Mx k + c, (5) until some stopping criterion is met. Methods of this form are called stationary because we do exactly the same thing at every iteration, which is to multiply the iterate by M and add to it the c vector. Let x be the solution to Ax = b or equivalently, let it satisfy x = Mx + c. (6) Subtracting (6) from (5), we have x k+1 x = M(x k x ). (7) Therefore, for any vector norm, x k+1 x M x k x, (8) where by convention the matrix norm is the one induced by the given vector norm. Recall an induced matrix norm is defined by M := max x 0 Mx x. It is clear from (8) that after every iteration the error, as is measured by the given norm, is at least reduced by a fixed factor of M whenever M is less than one. Therefore, we have a sufficient condition for convergence. 2

Theorem 1. Let x satisfy x = Mx + c. The stationary iterative method (5) converges, i.e., for any initial guess x 0, lim k xk = x, if M < 1 for some induced matrix norm. We note that M < 1 implies that I M = Q 1 A is nonsingular, thus so is A. We still have the freedom to choose the nonsingular matrix Q, hence the iteration matrix M = I Q 1 A. Noting that Mx = x Q 1 (Ax), to make the iterative method practical we need to choose Q so that solving the linear system Qx = d is very inexpensive, much more so than solving the original system. We have two central questions: (1) how to choose the matrix Q, and (2) for given A and Q, whether or not the method converges? 2 Jacobi method From Wikipedia, the free encyclopedia, it reads The Jacobi method is an algorithm in linear algebra for determining the solutions of a system of linear equations with largest absolute values in each row and column dominated by the diagonal element. Each diagonal element is solved for, and an approximate value plugged in. The process is then iterated until it converges. The method is named after German mathematician Carl Gustav Jakob Jacobi. I would not call the above a clear description of the Jacobi method (in fact, some statements are technically wrong), though it at least informs us that Jacobi was a German mathematician. 3

In the Jacobi method, Q is chosen as the diagonal matrix formed by the diagonal of A; that is, Q = D where A ii, i = j, D ij = 0, i j. Therefore, the Jacobi method can be written into the following scheme: (9) x x + D 1 (b Ax). (10) The idea behind the Jacobi method is simple. At each iteration, one solves the i-th equation in Ax = b for a new value of x i, the i-th variable in x, while fixing all the other variables at their values in the prior iteration. In Matlab, (10) could be implemented as r = b A x; x = x + invd. r; where r is the current residual and invd = 1./diag(A). A natural stopping criterion is that the relative residual is less than some appropriately chosen tolerance ɛ > 0: b Ax 1 + b < ɛ, (11) where in the denominator one is added to the norm of b to guard against the case of an excessively small right-hand side b. When is the Jacobi method convergent? Let us inspect the l norm of its corresponding iteration matrix M = I D 1 A (which is the maximum row sum in absolute value). Since the diagonal of M is zero, we have a ij M = max 1 i n a ii, j i where in the summation j is running and i is fixed. Therefore, a ii > j i a ij, i M < 1. (12) 4

A matrix satisfying the left inequality in (12) is called row strictly diagonally dominant. Similarly, a matrix is called column strictly diagonally dominant if for each column the absolute value of the diagonal element is great than the sum of the absolute values of all the off-diagonal elements in that column. Proposition 1. If A is row strictly diagonally dominant, then the Jacobi method converges from any initial guess. This theorem does not say that if a matrix is not strictly diagonally dominant, then the Jacobi method does not converge. It provides a sufficient, but not necessary, condition for convergence. Now what if A is column but not row strictly diagonally dominant? In this case, we can measure the errors in the vector norm D defined by x D := Dx 1, which induces the matrix norm M D D( )D 1 1 since Mx D x D = DMx 1 Dx 1 = DMD 1 (Dx) 1 Dx 1. Observe that DMD 1 = D(I D 1 A)D 1 = I AD 1. Hence M D = I AD 1 a ij 1 = max 1 j n a jj < 1 i j whenever A is column strictly diagonally dominant. As a result, the theorem below follows. Theorem 2. If A is, either row or column, strictly diagonally dominant then the Jacobi method converges from any initial guess. 5

3 Fundamental Theorem of Convergence We already mentioned that Theorem 1 is only a sufficient condition. The following theorem gives a necessary and sufficient condition in terms of the spectral radius of M, defined to be ρ(m) = max{ λ : det(m λi) = 0}, (13) that is, the largest modulus over the set of eigenvalues of M. Theorem 3. The stationary iterative method (5) converges from any initial guess if and only if ρ(m) < 1. (14) To prove the theorem, we will make use of the following lemma. Lemma 1. For any B C n n, where the infimum is taken over all induced matrix norms. ρ(b) = inf B, (15) Proof. Obviously, in any vector norm and for any eigen-pair (λ, x) of B, Bx = λx λ x = Bx B x λ B ρ(b) B. Indeed ρ(b) is a lower bound of B over all induced norms. We prove (15) by showing that for any ɛ > 0, there exists a nonsingular matrix S that defines an induced norm S := S( )S 1 1 such that B S ρ(b) + ɛ. By the well-know Schur s theorem, any square matrix is similar to a upper triangular matrix. Namely, P BP 1 = D + U where D is diagonal and U is strictly upper triangular. Furthermore, let T be the diagonal matrix with for any arbitrary t 0. Then T ii = 1/t i, i = 1, 2,, n, T P BP 1 T 1 = D + T UT 1. 6

The strict upper triangular matrix Û := T UT 1 has elements u ij t j i, j > i, û ij = 0, j i, where the nonzero elements can be made arbitrarily small by choosing t arbitrarily small. Therefore, letting S = T P, we have shown that SBS 1 = D + Û, Û 1 ɛ for some t value chosen sufficiently small (however, Û 0 unless B is diagonalizable). Noting that D is a diagonal matrix with the eigenvalues of B on the diagonal, we see that B S := SBS 1 1 D 1 + Û 1 ρ(b) + ɛ. Since S is an induced norm (please verify), we have proved the lemma. Now we prove Theorem 3. Since ρ(m) < 1 implies M < 1 for some induced norm, the sufficiency of condition (14) for convergence follows directly from Theorem 1. For necessity, let us assume that there exists an eigen-pair (λ, d) of M such that Md = λd and λ 1. Let x 0 = x + d, where x is the solution, so that e 0 = x 0 x = d. Then e k = M k e 0 = λ k e 0 d, implying non-convergence. This establishes the necessity of condition (14) for convergence from any initial point. However, it should be clear from the proof that convergence from some initial guesses is still possible even when ρ(m) 1. 4 Gauss-Seidel Method Another popular stationary iterative method is the Gauss-Seidel (GS) method where Q is chosen to be the lower triangular part, including the diagonal, of A. If one 7

partitions A into three parts: A = D L U where D is the diagonal and L ( U) is the strictly lower (upper) triangular part of A. Then for the GS method Q = D L is a lower triangular matrix (hence Qx = r is easy to solve), and the corresponding iteration matrix M is M = I (D L) 1 A (D L) 1 U. (16) Both the Jacobi and the GS method solves one equation (the i-th) for one variable (the i-th) at a time. The difference is that while the Jacobi method fixes other variables at their prior iteration values, the GS method immediately uses new values once they become available. Therefore, the GS method generally converges faster. Like the Jacobi method, the GS method has guaranteed convergence for strictly diagonally dominant matrices. Theorem 4. If A is, either row or column, strictly diagonally dominant then the Gauss-Seisel method converges from any initial guess. Proof. (i) Row case: In this case, a ii > a ij a ii a ij > a ij. (17) j i j<i j>i We will show ρ(m) < 1. Let (λ, x) be an eigen-pair of M where x is scaled so that x j 1 for j = 1, 2., n and x i = 1 for some index i. From the definition of M, Mx = λx λ(dx Lx) = Ux. Consider the i-th equation on the right, ( λ a ii x i + ) a ij x j = a ij x j. j<i j>i 8

Taking the moduli of both sides and in view of the fact that x i = 1 and x j 1, we have ( λ a ii ) a ij a ij, j<i j>i Hence, in view of (17), λ ( a ij / a ii ) a ij < 1, j>i j<i for all eigenvalues of M, confirming ρ(m) < 1. (ii) Column case: In this case, consider ρ(m) = ρ(m T ) and apply a similarity transformation (D L) T M T (D L) T first. Details are left as an exercise. Unlike the Jacobi method, the Gauss-Seidel method has guaranteed convergence for another class of matrices. Theorem 5. The Gauss-Seidel method converges from any initial guess if A is symmetric positive definite. We sketch a proof as follows (the details are left as an exercise). Consider an eigen-pair of M, (λ, x), and rearrange the equation Mx = λx. We have Ax = (1 λ)qx x Ax = (1 λ)x Qx = (1 λ)x Q x > 0, from which there holds x Dx x Ax = x (Q + Q A)x x Ax Therefore, λ < 1, implying ρ(m) < 1. = 1 λ 2 1 λ 2 > 0. 5 SOR Method SOR stands for Successive Over-Relaxation. It is an extension to the GS method. For SOR, the diagonal is split into two parts and distributed to both the left and the 9

right hand sides. That is, Q = 1 ( ω D L, Q A = U 1 1 ) D. ω Therefore, M = ( ) 1 ( ( 1 ω D L U 1 1 ) ) D, ω or equivalently, M(ω) = (D ωl) 1 (ωu + (1 ω)d). (18) Obviously, ω = 1 gives the GS method. The extra degree of freedom in ω, if appropriately chosen, can generally help reduce the spectral radius of ρ(m(ω)) from the value of ρ(m(1)) for the GS method, though an optimal ω value is generally impossible or impractical to calculate. As an extension to the GS method, it is not surprising that the SOR method converges for symmetric positive definite matrices. Theorem 6. The SOR method converges from any initial guess if A is symmetric positive definite and ω (0, 2). The proof follows from a similar argument as for the GS method and is left as an exercise. In addition, the condition ω (0, 2) is always necessary for convergence from any initial guess. Exercises 1. Prove that for any nonsingular matrix S R n n, S( )S 1 p is an induced matrix norm in R n n for p 1 (where p is the matrix norm induced by the vector p-norm). 2. Prove the convergence of the GS method for column strictly diagonally dominant matrices. 10

3. Prove Theorem 5 in details following the given sketch. 4. Prove Theorem 6. 5. Prove that a necessary condition for SOR to converge is ω (0, 2). (Hint: First show det(m(ω)) = (1 ω) n.) References [1] James M. Ortega Numerical Analysis, A Second Course. Academic Press, 1972. [2] Richard S. Varga. Matrix Iterative Analysis. Prentice-Hall, Englewood, NJ, 1962. [3] David M. Young. Iterative Solution of Large Linear Systems. Academic Press, New York, 1971. 11