Generalized line criterion for Gauss-Seidel method

Similar documents
Numerical Analysis: Solving Systems of Linear Equations

Computational Linear Algebra

Introduction to Scientific Computing

c 1995 Society for Industrial and Applied Mathematics Vol. 37, No. 1, pp , March

CAAM 454/554: Stationary Iterative Methods

CHAPTER 5. Basic Iterative Methods

COURSE Numerical methods for solving linear systems. Practical solving of many problems eventually leads to solving linear systems.

COURSE Iterative methods for solving linear systems

LINEAR SYSTEMS (11) Intensive Computation

Math Introduction to Numerical Analysis - Class Notes. Fernando Guevara Vasquez. Version Date: January 17, 2012.

Review of matrices. Let m, n IN. A rectangle of numbers written like A =

Here is an example of a block diagonal matrix with Jordan Blocks on the diagonal: J

Solution of Linear Equations

Discrete mathematical models in the analysis of splitting iterative methods for linear systems

Classical iterative methods for linear systems

Next topics: Solving systems of linear equations

Algebra C Numerical Linear Algebra Sample Exam Problems

Scientific Computing WS 2018/2019. Lecture 9. Jürgen Fuhrmann Lecture 9 Slide 1

The Solution of Linear Systems AX = B

9. Iterative Methods for Large Linear Systems

Theory of Iterative Methods

MATH 425-Spring 2010 HOMEWORK ASSIGNMENTS

Chap 3. Linear Algebra

ELEMENTARY LINEAR ALGEBRA

Spectral radius, symmetric and positive matrices

Iterative Methods for Smooth Objective Functions

A matrix over a field F is a rectangular array of elements from F. The symbol

Introduction to Iterative Solvers of Linear Systems

Numerical Methods - Numerical Linear Algebra

DEN: Linear algebra numerical view (GEM: Gauss elimination method for reducing a full rank matrix to upper-triangular

Iterative Methods for Solving A x = b

OPTIMAL SCALING FOR P -NORMS AND COMPONENTWISE DISTANCE TO SINGULARITY

Numerical Methods I Non-Square and Sparse Linear Systems

Computational Methods. Systems of Linear Equations

6. Iterative Methods for Linear Systems. The stepwise approach to the solution...

Chapter 7 Iterative Techniques in Matrix Algebra

Mathematical Optimisation, Chpt 2: Linear Equations and inequalities

j=1 u 1jv 1j. 1/ 2 Lemma 1. An orthogonal set of vectors must be linearly independent.

Today s class. Linear Algebraic Equations LU Decomposition. Numerical Methods, Fall 2011 Lecture 8. Prof. Jinbo Bi CSE, UConn

Lecture 7: Positive Semidefinite Matrices

The convergence of stationary iterations with indefinite splitting

Department of Mathematics California State University, Los Angeles Master s Degree Comprehensive Examination in. NUMERICAL ANALYSIS Spring 2015

Preliminary/Qualifying Exam in Numerical Analysis (Math 502a) Spring 2012

LINEAR ALGEBRA 1, 2012-I PARTIAL EXAM 3 SOLUTIONS TO PRACTICE PROBLEMS

Iterative Methods. Splitting Methods

Math 240 Calculus III

Up to this point, our main theoretical tools for finding eigenvalues without using det{a λi} = 0 have been the trace and determinant formulas

a 11 x 1 + a 12 x a 1n x n = b 1 a 21 x 1 + a 22 x a 2n x n = b 2.

LECTURE NOTES ELEMENTARY NUMERICAL METHODS. Eusebius Doedel

Chapter 2. Solving Systems of Equations. 2.1 Gaussian elimination

Solving Linear Systems

Linear Algebra in Actuarial Science: Slides to the lecture

Lecture 18 Classical Iterative Methods

This can be accomplished by left matrix multiplication as follows: I

Chapter 7. Iterative methods for large sparse linear systems. 7.1 Sparse matrix algebra. Large sparse matrices

Math 5630: Iterative Methods for Systems of Equations Hung Phan, UMass Lowell March 22, 2018

Modified Gauss Seidel type methods and Jacobi type methods for Z-matrices

Numerical Programming I (for CSE)

EXAMPLES OF CLASSICAL ITERATIVE METHODS

Linear Algebra. Solving Linear Systems. Copyright 2005, W.R. Winfrey

30.5. Iterative Methods for Systems of Equations. Introduction. Prerequisites. Learning Outcomes

2 Eigenvectors and Eigenvalues in abstract spaces.

1 Last time: least-squares problems

Jordan Journal of Mathematics and Statistics (JJMS) 5(3), 2012, pp A NEW ITERATIVE METHOD FOR SOLVING LINEAR SYSTEMS OF EQUATIONS

Practical Linear Algebra: A Geometry Toolbox

A = 3 1. We conclude that the algebraic multiplicity of the eigenvalues are both one, that is,

1 Number Systems and Errors 1

Numerical Linear Algebra Homework Assignment - Week 2

Formula for the inverse matrix. Cramer s rule. Review: 3 3 determinants can be computed expanding by any row or column

G1110 & 852G1 Numerical Linear Algebra

Linear Algebraic Equations

3 (Maths) Linear Algebra

Process Model Formulation and Solution, 3E4

JACOBI S ITERATION METHOD

MATH 5720: Unconstrained Optimization Hung Phan, UMass Lowell September 13, 2018

Partitioned Methods for Multifield Problems

Solving Linear Systems of Equations

Math 471 (Numerical methods) Chapter 3 (second half). System of equations

PROOF OF TWO MATRIX THEOREMS VIA TRIANGULAR FACTORIZATIONS ROY MATHIAS

Part IB - Easter Term 2003 Numerical Analysis I

OR MSc Maths Revision Course

AN ITERATION. In part as motivation, we consider an iteration method for solving a system of linear equations which has the form x Ax = b

Stabilization and Acceleration of Algebraic Multigrid Method

Linear Algebra I Lecture 8

MATH 221, Spring Homework 10 Solutions

BOUNDS OF MODULUS OF EIGENVALUES BASED ON STEIN EQUATION

In particular, if A is a square matrix and λ is one of its eigenvalues, then we can find a non-zero column vector X with

Math 108b: Notes on the Spectral Theorem

Lecture 2 - Unconstrained Optimization Definition[Global Minimum and Maximum]Let f : S R be defined on a set S R n. Then

Numerical Analysis: Solutions of System of. Linear Equation. Natasha S. Sharma, PhD

Direct Methods for Solving Linear Systems. Simon Fraser University Surrey Campus MACM 316 Spring 2005 Instructor: Ha Le

Theorem A.1. If A is any nonzero m x n matrix, then A is equivalent to a partitioned matrix of the form. k k n-k. m-k k m-k n-k

MAC Module 2 Systems of Linear Equations and Matrices II. Learning Objectives. Upon completing this module, you should be able to :

Symmetric matrices and dot products

Matrix Theory. A.Holst, V.Ufnarovski

A NEW EFFECTIVE PRECONDITIONED METHOD FOR L-MATRICES

Lecture Note 7: Iterative methods for solving linear systems. Xiaoqun Zhang Shanghai Jiao Tong University

Foundations of Matrix Analysis

FEM and Sparse Linear System Solving

Numerical Methods in Matrix Computations

Transcription:

Computational and Applied Mathematics Vol. 22, N. 1, pp. 91 97, 2003 Copyright 2003 SBMAC Generalized line criterion for Gauss-Seidel method M.V.P. GARCIA 1, C. HUMES JR. 2 and J.M. STERN 2 1 Department of Applied Mathematics, University of São Paulo, Brasil 2 Department of Computer Science, University of São Paulo, Brasil E-mail: mane@ime.usp.br / chumes@ime.usp.br / jstern@ime.usp.br Abstract. We present a module based criterion, i.e. a sufficient condition based on the absolute value of the matrix coefficients, for the convergence of Gauss Seidel method (GSM) for a square system of linear algebraic equations, the Generalized Line Criterion (GLC). We prove GLC to be the most general module based criterion and derive, as GLC corollaries, some previously know and also some new criteria for GSM convergence. Although far more general than the previously known results, the proof of GLC is simpler. The results used here are related to recent research in stability of dynamical systems and control of manufacturing systems. Mathematical subject classification: 65F10, 65F35, 15A09. Key words: Gauss-Seidel method, line criterion, stability and control of manufacturing systems. 1 Introduction We present a module based criterion, i.e. a sufficient condition based on the absolute value of the matrix coefficients, for the convergence of Gauss Seidel method (GSM) for a square system of linear algebraic equations, the Generalized Line Criterion (GLC). We prove GLC to be the most general module based criterion and derive, as GLC corollaries, some previously know and also some new criteria for GSM convergence. Although far more general than the previously known results, the proof of GLC is simpler. GLC is shown to be a necessary and sufficient condition of the convergence for a class of matrices defined by its sign structure. #507/00. Received: 8/XI/01.

92 GENERALIZED LINE CRITERION FOR GAUSS-SEIDEL METHOD The results here presented are analogous to important results in the stability of manufacturing systems, where matrices with the sign structure of S(A) are common. These matrices appear in the dynamics of systems operating under generalized Round-Robin policies, either explicitly imposed, as in [5], or induced by polyhedral cones [8], [3], [4]. It is quite natural that convergence and stability results are closely related, as extremely well exploited in [6], [7]. In many instances the GLC, and the associated linear algebra operations, like scaling and permutation, have natural interpretations for system stabilization and control. Therefore, the major motivation for the GLC is to present these results in a self contained form, that is completely in the context of simple computational linear algebra. The simple linear form of the GLC also facilitates the use of several related techniques like optimization, sensitivity, and post-optimal analysis for manufacturing systems. 2 Problem setting In this paper, the following notation is used, 1 is the vector [1, 1,...1] t, the diagonal operator D = diag(d) defines the diagonal matrix D whose diagonal elements are given by vector d, whereas a = diag(a) defines vector a as the main diagonal of matrix A,soD = diag(diag(a)) gives D i,i = A i,i, D i,j = 0,i j. The matrices A and S(A) associated to a given matrix, A, are defined next. Matrices with the sign structure of S(A) appear naturally in the control theory of manufacturing systems [5], [3]. A i,j = Ai,j ; S(A)i,i = Ai,i and S(A) i,j = Ai,j,i j The problem in question is the solution of the linear system of equations Ax = b, where A R n n, b R n and the coefficient matrix has no null diagonal elements, A i,i 0, as usual in GSM. GSM, and several other iterative methods can be written as stationary iterative methods: x k+1 = Bx k + c where B is called the method s iteration matrix. Many of those methods are obtained splitting the coefficient matrix A = M N, M non-singular, and then

M.V.P. GARCIA, C. HUMES JR. and J.M. STERN 93 writing Mx k+1 = Nx k + b, or x k+1 = M 1 Nx k + M 1 b. Splitting the coefficient matrix in its lower, upper triangular and diagonal parts: A = L+D +U, i j L i,j = 0, j i U i,j = 0 and D = diag(diag(a)), GSM is the stationary iterative method with M = (D + L) and N = U, or B = (D + L) 1 U and c = (D + L) 1 b. If the above iteration is performed in parallel, i.e., using at iteration k+1 only x k and not the already available elements of x k+1, we have the Jacobi method, JM, where M = D and N = (L + U),orB = D 1 (L + U)and c = D 1 b. The method converges, for any initial point x 0,iffρ(B) max eig(b) < 1, i.e. iff all eigenvalues of the iteration matrix lie inside the unit circle. This general necessary and sufficient criterion is usually hard to use when asserting convergence, because finding the largest eingenvalue of B is a hard nonlinear problem. However since, for any consistent matrix norm, ρ(b) B, the condition B < 1 is a sufficient convergence criterion. In particular, the infinite norm B max x =1 Bx = max i j Bi,j provides strict diagonally dominance as an easily verified sufficient criterion for JM and GSM convergence [1]. This is known as the Line Criterion (LC). A obbeys LC if S(A)1 > 0. The following lemma will be needed: Lemma 1. The convergence of the iterative method Mx k+1 = Nx k + b implies that A = M N is non-singular. Proof. Assume by contradiction that for some h 0, Ah = 0, so h = M 1 Nh. Then B = M 1 N has 1 as an eingenvalue, and the method can not be convergent [2]. 3 Generalized Line Criterion (GLC) Scaling the equations and variables of a liner system Ax = b is the operation of multiplying the coefficient matrix by left and right non-singular diagonal

94 GENERALIZED LINE CRITERION FOR GAUSS-SEIDEL METHOD matrices obtaining the equivalent system EAWy = Eb. Scaling is widely used in computational linear algebra for pre-conditioning, control floating point error accumulation, etc. The GSM convergence has the interesting property of not being affected by scaling the linear system: Lemma 2. Let E and W be non-singular diagonal matrices. Then the GSM converges on Ax = b iff it converges on EAWy = Eb. Proof. Since the pre-multiplication of A by E does not affect the GSM formulas and considering the bijection Wy = x, the result follows immediately. Let us now define the GLC convergence criterion for the GSM on Ax = b. Generalized Line Criterion (GLC). A obeys GLC iff there is a vector w > 0 such that S(A)w > 0. It is trivial that if A obeys LC then A obeys GLC with w = 1. Moreover, using W = diag(w), if A obeys GLC then AW obeys the line criterion. This observation and lemma 2 lead to: Theorem 1. If A obeys the GLC then the GSM iteration applied to Ax = b converges. Although GLC is a sufficient condition for GSM convergence, it is not a necessary condition as it can be seen by the example: 2 1 1 0 1/2 1/2 A = (L+D+U) = 1 2 1,B = (D+L) 1 U = 0 1/4 3/4, 1 1 2 0 3/8 5/8 where the eigenvalues of B = (D + L) 1 U all lie inside the unit circle although the system does not satisfy GLC, because det(s(a)w) = det(s(a)) = 0. However, if A has a special sign structure, then GLC becomes also a necessary condition: Theorem 2. If A has a sign structure such that, in each line, the diagonal element sign is opposite to the sign of all other elements, then GLC turns out to be a necessary and sufficient condition for GSM convergence.

M.V.P. GARCIA, C. HUMES JR. and J.M. STERN 95 Proof. Using the left scaling E = diag(diag(a)) 1,wehaveS(A) = A, and it suffices to prove the theorem for the case diag(a) = 1 and A i,j 0, i j. Since the sufficiency of GLC has already been established in Theorem 1, we only have to show that if GSM converges for an A such that diag(a) = 1 and A i,j 0, i j, then A obeys GLC. Splitting A = L + U, diag(l) = 1, and we can write L and T = L 1 in product form, L = L 1 L 2...L n 1 and T = T n 1 T n 2...T 1, where L k and T k = (L k ) 1, have the structure: 1 0 1 0 0... 0... L k 1. =,T k 1. =. L k+1,k. L k+1,k.... 0.... 0 0 L n,k 1 0 L n,k 1 The factors in the product form perform elementary row operations, i.e. if y = L k x then y k+h = x k+h + L k k+h xk. Let us now consider b = 1, x 0 0, and the corresponding GSM iteration sequence x k+1 = L 1 Ux k + L 1 1 = L 1 ( U)x k + L 1 1 From the product form of T = L 1 and remembering that L i,j 0 and U i,j 0, it is easy to see that L 1 1 > 0 and L 1 U 0. From the GSM iteration it follows that x 0 0,x k+1 L 1 1 > 0. Assuming GSM convergence and x 0 0, lim k xk = z L 1 1 > 0, and Az = 1 > 0 so A obeys GLC. From theorem 2 we see that GLC is the most general module based criterion of convergence of GSM. Some small variations, including a minor generalization of theorem 1 can be stated, for example:

96 GENERALIZED LINE CRITERION FOR GAUSS-SEIDEL METHOD Corollary 1. If there is a w>0such that S(A) 0 with no more than one coordinate of S(A)w equal to zero, then the GSM converges for A. Proof. If S(A)w > 0 we are in the situation of theorem 1. Otherwise let h be the unique null coordinate of S(A)w. Using S(A) h,h > 0 and the continuity of S(A) : R n R n, there is a sufficiently small ε>0such that S(A)w > 0 for w =[w 1,...w h + ε...w n ] > 0. Corollary 2. If det(s(a)) 0 and if there is a w>0 such that S(A)w 0 then the GSM converges to A. Proof. Since S(A) in non singular let y = S(A) 1 1. By continuity of S(A) : R n R n, there is a sufficiently small ε>0such that w = w + εy > 0 and S(A)w = S(A)w + ε1 > 0. From Theorem 2, any module based convergence criterion for GSM must also assert GLC. The simplest such module based criterion is LC, that is GLC for w = 1. Finally we consider the effect of permutation operations: Lemma 3. (Permutations). Let p and q be permutations of [1,...n], with the corresponding row and column permutation matrices, P i,j = I p(i),j and Q i,j = I i,q(j), so that the permuted matrix PAQhas no null diagonal elements. GSM converges for A iff GSM converges for PAQ. Proof. We only have to remember that any permutation matrix P is orthogonal, i.e. PP t = I, so the 2-norm of any vector v is not changed by the rotation w = Pv, and neither is the spectral radius of the permuted matrix PAQ. For symmetric permutations, PAP t, we have a simple permutation of diagonal elements, i.e. diag(p AP t ) = P diag(a). Again, this special case is of particular interest in manufacturing systems as in [4] and [5]. REFERENCES [1] D.P. Bertsekas and J.N. Tsitsiklis, Parallel and Distributed Computation, Numerical Methods. Prentice-Hall, (1989).

M.V.P. GARCIA, C. HUMES JR. and J.M. STERN 97 [2] G.H. Golub, C.F. van Loan, Matrix Computations, The Johns Hopkins Univ. Press, (1989). [3] C. Humes Jr., L.O. Brandao and M.P. Garcia, A mixed dynamics approach for linear corridor policies (a revisitation of dynamic setup scheduling and flow control in manufacturing systems), Discrete Event Dynamical Systems, 5, 1, 59 82, January (1995). [4] C. Humes Jr. and T.I. Seidman, Some kanban-controlled manufacturing systems: a first stability analysis", IEEE Transactions on Automatic Control, 41, 7, 1013 1018, July (1996). [5] J.R. Perkins, C.Humes Jr. and P.R. Kumar, Distributed control of flexible manufacturing systems: stability and performance, IEEE Transactions on Robotics and Automation, 10, 2, 133 141, April (1994). [6] E.Polak and E. Polak, Optimization: Algorithms and Consistent Approximations, Springer (1997). [7] J. Renegar, A Polynomial-Time Algorithm based on Newton s Method for Linear Programming, Mathematical Programming, 40 (1988), pp. 59 93. [8] A. Sharifnia, M. Caramanis and S. Gershwin, Dynamical setup scheduling and Flow control in Manufacturing Systems, Discret Event Dynamical Systems: Theory and Applications, 1 (1991), pp. 149 175.