Convergence of Asynchronous Matrix Iterations Subject to Diagonal Dominance*

Similar documents
CAAM 454/554: Stationary Iterative Methods

Solving Linear Systems

c 1995 Society for Industrial and Applied Mathematics Vol. 37, No. 1, pp , March

Modified Gauss Seidel type methods and Jacobi type methods for Z-matrices

Generalized line criterion for Gauss-Seidel method

Introduction to Scientific Computing

WE consider finite-state Markov decision processes

CLASSICAL ITERATIVE METHODS

ETNA Kent State University

Value and Policy Iteration

Lecture 18 Classical Iterative Methods

Partitioned Methods for Multifield Problems

The upper Jacobi and upper Gauss Seidel type iterative methods for preconditioned linear systems

Here is an example of a block diagonal matrix with Jordan Blocks on the diagonal: J

A SIMPLIFIED FORM FOR NEARLY REDUCIBLE AND NEARLY DECOMPOSABLE MATRICES D. J. HARTFIEL

Iterative Methods for Ax=b

Computers and Mathematics with Applications. Convergence analysis of the preconditioned Gauss Seidel method for H-matrices

Detailed Proof of The PerronFrobenius Theorem

Solving Linear Systems

ETNA Kent State University

Markov Chains and Stochastic Sampling

Jordan Journal of Mathematics and Statistics (JJMS) 5(3), 2012, pp A NEW ITERATIVE METHOD FOR SOLVING LINEAR SYSTEMS OF EQUATIONS

Linear Operators Preserving the Numerical Range (Radius) on Triangular Matrices

Math/Phys/Engr 428, Math 529/Phys 528 Numerical Methods - Summer Homework 3 Due: Tuesday, July 3, 2018

Scientific Computing WS 2018/2019. Lecture 9. Jürgen Fuhrmann Lecture 9 Slide 1

Iterative Methods. Splitting Methods

Second Order Iterative Techniques for Boundary Value Problems and Fredholm Integral Equations

Bindel, Fall 2016 Matrix Computations (CS 6210) Notes for

Online solution of the average cost Kullback-Leibler optimization problem

DEN: Linear algebra numerical view (GEM: Gauss elimination method for reducing a full rank matrix to upper-triangular

Interlacing Inequalities for Totally Nonnegative Matrices

Lecture Note 7: Iterative methods for solving linear systems. Xiaoqun Zhang Shanghai Jiao Tong University

Asynchronous Parareal in time discretization for partial differential equations

A NEW EFFECTIVE PRECONDITIONED METHOD FOR L-MATRICES

Algebra C Numerical Linear Algebra Sample Exam Problems

642:550, Summer 2004, Supplement 6 The Perron-Frobenius Theorem. Summer 2004

Computational Linear Algebra

JACOBI S ITERATION METHOD

Stochastic Shortest Path Problems

6. Iterative Methods for Linear Systems. The stepwise approach to the solution...

Agreement algorithms for synchronization of clocks in nodes of stochastic networks

Spectral gradient projection method for solving nonlinear monotone equations

Absolute value equations

Background. Background. C. T. Kelley NC State University tim C. T. Kelley Background NCSU, Spring / 58

Numerische Mathematik (g Springer-Verlag1982

Conjugate Gradient (CG) Method

Extended Monotropic Programming and Duality 1

A ROLE FOR DOUBLY STOCHASTIC MATRICES IN GRAPH THEORY

Hessenberg Pairs of Linear Transformations

Alternative Characterization of Ergodicity for Doubly Stochastic Chains

Optimal Preconditioners for Interval Gauss Seidel Methods

CONVERGENCE OF MULTISPLITTING METHOD FOR A SYMMETRIC POSITIVE DEFINITE MATRIX

Projected Equation Methods for Approximate Solution of Large Linear Systems 1

Comparison results between Jacobi and other iterative methods

Iterative Methods for Solving A x = b

Up to this point, our main theoretical tools for finding eigenvalues without using det{a λi} = 0 have been the trace and determinant formulas

Convergence Rate for Consensus with Delays

CHUN-HUA GUO. Key words. matrix equations, minimal nonnegative solution, Markov chains, cyclic reduction, iterative methods, convergence rate

Approximating a single component of the solution to a linear system

Eigenvectors Via Graph Theory

Another algorithm for nonnegative matrices

Journal of Computational and Applied Mathematics. Projected equation methods for approximate solution of large linear systems

Eigenvalue comparisons in graph theory

An Extrapolated Gauss-Seidel Iteration

1 Last time: least-squares problems

A theorethic treatment of the Partitioned Iterative Convergence methods

Computational Economics and Finance

Spectral inequalities and equalities involving products of matrices

On factor width and symmetric H -matrices

The Solution of Linear Systems AX = B

9. Iterative Methods for Large Linear Systems

Relation of Pure Minimum Cost Flow Model to Linear Programming

1 Number Systems and Errors 1

LINEAR SYSTEMS (11) Intensive Computation

CHAPTER 5. Basic Iterative Methods

Spectral Theorem for Self-adjoint Linear Operators

A MULTIGRID ALGORITHM FOR. Richard E. Ewing and Jian Shen. Institute for Scientic Computation. Texas A&M University. College Station, Texas SUMMARY

In particular, if A is a square matrix and λ is one of its eigenvalues, then we can find a non-zero column vector X with

Solving the Hamiltonian Cycle problem using symbolic determinants

Solving Linear Systems of Equations

Some bounds for the spectral radius of the Hadamard product of matrices

Chapter 7 Network Flow Problems, I

Hegselmann-Krause Dynamics: An Upper Bound on Termination Time

The Perron Frobenius theorem and the Hilbert metric

Math Introduction to Numerical Analysis - Class Notes. Fernando Guevara Vasquez. Version Date: January 17, 2012.

Nonnegative and spectral matrix theory Lecture notes

THE PERTURBATION BOUND FOR THE SPECTRAL RADIUS OF A NON-NEGATIVE TENSOR

Affine iterations on nonnegative vectors

The Multi-Path Utility Maximization Problem

In Advances in Neural Information Processing Systems 6. J. D. Cowan, G. Tesauro and. Convergence of Indirect Adaptive. Andrew G.

Q-Learning for Markov Decision Processes*

Data Mining and Analysis: Fundamental Concepts and Algorithms

On the Skeel condition number, growth factor and pivoting strategies for Gaussian elimination

On the simultaneous diagonal stability of a pair of positive linear systems

ON A HOMOTOPY BASED METHOD FOR SOLVING SYSTEMS OF LINEAR EQUATIONS

The convergence of stationary iterations with indefinite splitting

INVERSE SPANNING TREE PROBLEMS: FORMULATIONS AND ALGORITHMS

Fast Linear Iterations for Distributed Averaging 1

Convex Optimization of Graph Laplacian Eigenvalues

Reaching a Consensus in a Dynamically Changing Environment A Graphical Approach

Transcription:

December 27, 1988 LIDS-P-1842 Convergence of Asynchronous Matrix Iterations Subject to Diagonal Dominance* by Paul Tsengt Abstract We consider point Gauss-Seidel and Jacobi matrix iterations subject to diagonal dominance. We give conditions on the relaxation parameter under which the iteration mapping has the maximum norm nonexpansive properties given in [TBT88]. It then follows from a result in [TBT88] that, under these conditions, the associated relaxation iterations converge under asynchronous implementation. Our conditions for convergence improve upon those given by James [Jam73] for the special cases of strict diagonal dominance and irreducible diagonal dominance and synchronous implementation. KEY WORDS: diagonal dominance, relaxation iteration, asynchronous computation * This research is partially supported by the U.S. Army Research Office, contract DAAL03-86-K-0171 (Center for Intelligent Control Systems), and by the National Science Foundation under grant NSF- ECS-8519058. t Laboratory for Information and Decision Systems, Room 35-205, Massachusetts Institute of Technology, Cambridge, MA 02139. Acknowledgement: Thanks are due to Professor J.N. Tsitsiklis for bringing to my attention the reference [Jam73] which motivated this work.

1. Introduction Consider stationary point relaxation methods of the Gauss-Seidel and Jacobi type for solving systems of linear equations [OrR70, Var62, You71]. For the special case where the coefficient matrix is either strictly diagonally dominant or irreducibly diagonally dominant, James [Jam73] gave general conditions on the relaxation parameters under which these methods converge. In this note we expand on the work of James. In particular, we show that these methods converge under broader conditions on the relaxation parameters and for problems where the coefficient matrix is irreducible and diagonally dominant. Moreover, our convergence results carry over to the asynchronous versions of these methods. Our arguments are based on the notion of maximum norm nonexpansive mappings and are fundamentally different from the eigenvalue analysis of James and others. We will use the following definitions (cf. [OrR70]): Definition 1 An nxn complex matrix A = [aij] is diagonally dominant if, for all i, j,ii aijl < laii I, (1.1) and is strictly diagonallv dominant if strict inequality holds in (1.1) for all i. Similarly, A is irreducibly diagonally dominant if it is irreducible, diagonally dominant, and strict inequality holds in (1.1) for at least one i. In what follows, A will denote the coefficient matrix associated with the system of linear equations and n will denote its dimension. For simplicity we will assume that A has unity diagonal entries. The iteration matrix for the stationary point relaxation methods corresponding to A is (cf. [Jam73]) M(a,Q) = (I + aqll)- 1 [(I - () - (1 - a)2l - QU], (1.2) where Q2 is an nxn diagonal matrix with positive diagonal entries, o is a scalar inside [0,1], and L (U) denotes the lower (upper) triangular part of A, i.e. A = L+I+U. [The familiar Jacobi (Gauss-Seidel) iteration is obtained by setting f2 = I and a = 0 (a = 1).] Under

2 appropriate conditions on a and Q, all eigenvalues of M(a,fQ) lie strictly within the unit circle and the stationary relaxation method given by M(a,Q) converges [Jam73] (see also [JaR75, Var76]). However, the circle criterion is not sufficient to guarantee that the asynchronous version of this method converges. Instead, we need the following notions: Definition 2 An nxn complex matrix M = [mij] is nonexpansive if, for all i, Yj Imijl < 1, (1.3) and is contractive if strict inequality holds in (1.3) for all i. M is block irreducibly nonexpansive if it is nonexpansive and, for any SC { 1,2,...,n} such that mij = 0 for all is S and all j S, there exists an se S such that _j Imsjl < 1. M is block irreducibly contractive if it is block irreducibly nonexpansive and strict inequality holds in (1.3) for at least one i. If M(a,fQ) is contractive, then the associated relaxation mapping is a maximum norm contraction and, by a result of Chazan and Miranker [ChM69] (see also [Bau78; Ber83; BeT89, 6.2]), the associated asynchronous method converges. If M(a,CQ) is block irreducibly contractive or is block irreducibly nonexpansive with diagonal entries having positive real parts, then it can be shown that the associated relaxation mapping has the maximum norm nonexpansion properties described in [TBT88] and, by Proposition 2.1 in [TBT88], the associated asynchronous method converges (see also [BeT89, 7.2; LuD86; Tsi84, 4]). [The results in the above references are stated in terms of real matrices, but they can be extended to complex matrices.] Below we give conditions on a and Q2 under which M(xa,Q) satisfies the above convergence criteria when (i) A is strictly diagonally dominant, (ii) A is irreducibly diagonally dominant, and (iii) A is irreducible and diagonally dominant. The first two cases had been analyzed in [Jam73], but our conditions on a and fq for the second case are more general. The third case, to the best of our knowledge, has not been analyzed previously, although it has a number of interesting applications. For example, consider the matrix EDET, where D is a diagonal matrix with positive diagonal entries and E is the node-arc incidence matrix for a connected directed graph. This matrix, which arises in the solution of linear network flow problems by interior point methods [Kar84], can be shown to be irreducible and diagonally dominant [TBT88, 3.2]. For another example, consider the probability transition matrix P for an irreducible Markov chain and suppose that in* is the column vector of invariant probabilities of this

3 Markov chain. Then the coefficient matrix I-PT is irreducible and, upon scaling by the diagonal matrix whose ith component is the ith component of a*, is also diagonally dominant [BeT89, 7.3.2]. [It can be seen that scaling of A by a diagonal matrix with positive diagonal entries results in the scaling of M(a,Q) by the same matrix and that the above criteria for convergence is unaffected by such scaling of M(a,K2).] 2. Convergence Results Let w i denote the ith diagonal entry of l and denote, for each i, li = j<ilaijl, u i = jilaijl, (2.1) From the definition of M(a,9Q) (cf. (1.2)) we have that (I + acql)m(a,q) = I + aml - QA. Hence, if we let mij denote the (i,j)th entry of M(a,Q), then mij + awi (k<i aik mkj) = (a - 1)w i aij, if j <i, (2.2a) mij + awi ('k<i aik mkj) = 1 - wi, if j = i, (2.2b) mij + awi (k<i aik mkj) = -wi aij, if j > i. (2.2c) The above identities will be used frequently in the analysis to follow. We first have the following two main results regarding M(a,QZ) (cf. Theorems 2 and 3 in [Jam73]): Proposition 1 If A is strictly diagonally dominant (i.e., l i + u i < 1 for all i), then, for any ae [0,1] and any Q such that

4 0 < w i < 2/(l+li+ui), i = 1,..., n, (2.3) the matrix M(a,fQ) is contractive. Proof: First we note from the hypotheses of the proposition that wi(l i + ui) + 11 - Wil < 1, i = 1,...,n. (2.4) [To see this, note that if 0 < wi < 1, then (cf. diagonal dominance) wi(li + ui) + I1 - wil = wi(li + ui) + 1 - w i < 1; while if 1 < w i < 2/(l+li+ui), then (cf. (2.3)) wi(li + ui) + I1 - wil = Wi(i + ui) - 1 + w i < 1.] For i = 1, we have from (2.2b)-(2.2c) m 1 l = (1 -wl) mlj = -wl alj, ifj > 1. Hence Y-j Imljl < 11 - W 1 1 + Wl(j>llaljl ) = II-wll+w l u 1 < 1, where the last inequality follows from (2.4). Now suppose that Ij Imijl < i 1, = 1,..., s-l, (2.5) for some integer s > 2. We will show that Xj Imsjl < 1, thus completing our induction. Now, from (2.2a)-(2.2c) we obtain that j Imsjl < aws(xj, k<slaski Imkjl) + (1 - a)ws(j< s l asji) + 11-wsl + Ws(Yj>slasjI) < aws.l s + (1 -C)wS.l s + I1-wsI + Ws-U s

5 = ws(l s + us) + I1 - Ws1 < 1, where the second inequality follows from the inductive hypothesis (2.5) and the third inequality follows from (2.4). Q.E.D. Proposition 2 that If A is irreducibly diagonally dominant, then, for any as [0,1) and any Q such 0 < w i < 2/(l+li+ui), i = 1,...,n, (2.6) where strict inequality holds for at least one row i' such that li, + u i, < 1, the matrix M(a,9l) is block irreducibly contractive. Proof: It is straightforward to show (by using (2.6) and repeating the proof of Proposition 1) that j, Imijl < 1, for all i, and strict inequality holds for i = i'. Now consider any Sc { 1,...,n} such that mij = 0 for all is S, jo S (if no such S exists, then M(a,92) is irreducible and we are done). We will show that there exists an se S such that Xj Imsjl < 1. [since the choice of S is arbitray, this shows M(a,Q2) to be block irreducibly contractive.] If i'e S, we can simply set s to i'. Otherwise, since A is irreducible, there must exist some se S and some to S such that ast 0 and mst = 0. We distinguish between two cases: (a) t < s and (b) t > s. In case (a), we have from (2.2a) that (aws (-k<s ask mkt) (a - l)w s ast. Hence, from (2.2a)-(2.2c) we obtain that,jt Imsjl< aw- Ws(,jt Yk<slask Imkjl) + (1- a)ws(j<sjtlasjl)+ I1-wsl W + s( j>slasjl ) < aws.l s + (1 - a)ws(l s - lastl) + I1-Wsl + Ws-U s = ws(ls + Us) + 11 - wsl - (1 - a)ws lastl

6 < 1 - (1 - a)w s lastl, where the second inequality follows from the fact Xj Imkjl < 1 for all k and the last inequality follows from the observation (cf. (2.6)) that ws(l s + u s ) + 11 - wsl < 1. In case (b), an analogous argument using (2.2c) instead of (2.2a) yields Xjl ImsjI < 1 - w s la s tl. Since as [0,1), we have j Imsjl < 1 in either case. Q.E.D. We note that the conditions on a given by Proposition 2 is different from that given by Theorem 3 in [Jam73] where it is assumed thate either a = 0 or as (1/2,1]. Does Proposition 2 still hold for a = 1? One suspects that it does, but more analysis is required. [Proposition 2 can be shown to hold for a = 1 if mii 0 O for all i, which in turn can be shown to hold if w i < l/(l+li) for all i (see argument below).] Proposition 3 such that If A is irreducible and diagonally dominant, then, for any as [0,1] and any Q 0 < w i < l/(l+ali), i= l,...,n, (2.7) the matrix M(a,Q) either is block irreducibly contractive or is block irreducibly nonexpansive with diagonal entries having positive real parts. Proof: First note that 1/(l+ali) < 2/(l+li+ui); hence (2.7) implies that (2.6) holds. Then using (2.6) and an argument identical to that in the proof of Proposition 2, we obtain that M(a,Q) is nonexpansive and, if M(a,Q) is not irreducible, then it is block irreducibly contractive. Hence it only remains to show that the diagonal entries of M(a,Q) have positive real parts. We have from (2.2b) that

mii + awi (k<i aik mki) = 1-w i. Hence Re(mii) > 1 - w i - aw i (Xk<i laikllmkil) 1 - wi - aw i (Xkd laikl) = 1 - wi(1 + ali) > 0, where the second inequality follows from the fact that Imkil < 1 for all k, and the third inequality follows from (2.7). Q.E.D. [In practice, we can first choose any 0 < w i < 2/(l+li+ui), for all i; check if M(a,Q) is block irreducibly contractive; and if not, decrease each w i to be below 1/(1 + lij).]

References 8 [Bau78] Baudet, G.M., "Asynchronous Iterative Methods for Multiprocessors," J. ACM, 15 (1978), 226-244. [Ber83] Bertsekas, D.P., "Distributed Asynchronous Computation of Fixed Points," Mathematical Programming, 27 (1983), 107-120. [BeT89] Bertsekas, D.P. and Tsitsiklis, J.N., Parallel and Distributed Computation: Numerical Methods, Prentice-Hall, Englewood Cliffs, NJ (1989). [ChM69] Chazan, D. and Miranker, M., "Chaotic Relaxation," Linear Algebra and its Applications, 2 (1969), 199-222. [Jam73] James, K. R., "Convergence of Matrix Iteration Subject to Diagonal Dominance," SIAM J. Numer. Anal., 10 (1973), 478-484. [JaR74] James, K.R. and Riha, W., "Convergence Criteria for Successive Overrelaxation," SIAM J. Numer. Anal., 12 (1975), 137-143. [Kar84] Karmarkar, N., "A New Polynomial-Time Algorithm for Linear Programming," Combinatorica, 4 (1984), 373-395. [LuD86] Lubachevsky, B. and Mitra, D., "A Chaotic Asynchronous Algorithm for Computing the Fixed Point of a Nonnegative Matrix of Unit Spectral Radius," J. ACM, 33 (1986), 130-150. [OrR70] Ortega, J.M. and Rheinboldt, W.C., Iterative Solution of Nonlinear Equations in Several Variables, Academic Press, New York, NY (1970). [Var62] Varga, R.S., Matrix Iterative Analysis, Prentice-Hall, Englewood Cliffs, NJ (1962). [Var76] Varga, R.S., "On Recurring Theorems on Diagonal Dominance," Linear Algebra and Its Applications, 13 (1976), 1-9. [TBT88] Tseng, P., Bertsekas, D.P., and Tsitsiklis, J.N., "Partially Asynchronous, Parallel Algorithms for Network Flow and Other Problems," LIDS Report P-1832, M. I. T., Cambridge, MA 02139 (November 1988). [Tsi84] Tsitsiklis, J.N., "Problems in Decentralized Decision Making and Computation," Ph.D.

Thesis, Dept. of EECS, M.I.T., Cambridge, MA (1984). 9 [You71] Young, D.M., Iterative Solution of Large Linear Systems, Academic Press, New York, NY (1971).