Computers and Mathematics with Applications. Convergence analysis of the preconditioned Gauss Seidel method for H-matrices

Similar documents
The upper Jacobi and upper Gauss Seidel type iterative methods for preconditioned linear systems

Modified Gauss Seidel type methods and Jacobi type methods for Z-matrices

A NEW EFFECTIVE PRECONDITIONED METHOD FOR L-MATRICES

On optimal improvements of classical iterative schemes for Z-matrices

Comparison results between Jacobi and other iterative methods

CONVERGENCE OF MULTISPLITTING METHOD FOR A SYMMETRIC POSITIVE DEFINITE MATRIX

Some Preconditioning Techniques for Linear Systems

Preconditioning Strategy to Solve Fuzzy Linear Systems (FLS)

Improving the Modified Gauss-Seidel Method for Z-Matrices

Applied Mathematics Letters. Comparison theorems for a subclass of proper splittings of matrices

CAAM 454/554: Stationary Iterative Methods

Generalizations of M-matrices which may not have a nonnegative inverse

Scientific Computing WS 2018/2019. Lecture 9. Jürgen Fuhrmann Lecture 9 Slide 1

Some bounds for the spectral radius of the Hadamard product of matrices

Improving AOR Method for a Class of Two-by-Two Linear Systems

9. Iterative Methods for Large Linear Systems

A generalization of the Gauss-Seidel iteration method for solving absolute value equations

Jordan Journal of Mathematics and Statistics (JJMS) 5(3), 2012, pp A NEW ITERATIVE METHOD FOR SOLVING LINEAR SYSTEMS OF EQUATIONS

Generalized AOR Method for Solving System of Linear Equations. Davod Khojasteh Salkuyeh. Department of Mathematics, University of Mohaghegh Ardabili,

Up to this point, our main theoretical tools for finding eigenvalues without using det{a λi} = 0 have been the trace and determinant formulas

A Method for Constructing Diagonally Dominant Preconditioners based on Jacobi Rotations

Partitioned Methods for Multifield Problems

Key words. Index, Drazin inverse, group inverse, Moore-Penrose inverse, index splitting, proper splitting, comparison theorem, monotone iteration.

Review of matrices. Let m, n IN. A rectangle of numbers written like A =

Abed Elhashash and Daniel B. Szyld. Report August 2007

A note on the unique solution of linear complementarity problem

A Refinement of Gauss-Seidel Method for Solving. of Linear System of Equations

Next topics: Solving systems of linear equations

JACOBI S ITERATION METHOD

Properties for the Perron complement of three known subclasses of H-matrices

BOUNDS OF MODULUS OF EIGENVALUES BASED ON STEIN EQUATION

CHAPTER 5. Basic Iterative Methods

Iterative Methods. Splitting Methods

Solving Linear Systems

Jae Heon Yun and Yu Du Han

ON A HOMOTOPY BASED METHOD FOR SOLVING SYSTEMS OF LINEAR EQUATIONS

Lecture 18 Classical Iterative Methods

Algebraic Multigrid Preconditioners for Computing Stationary Distributions of Markov Processes

Here is an example of a block diagonal matrix with Jordan Blocks on the diagonal: J

ETNA Kent State University

Algebra C Numerical Linear Algebra Sample Exam Problems

CLASSICAL ITERATIVE METHODS

EQUIVALENCE OF CONDITIONS FOR CONVERGENCE OF ITERATIVE METHODS FOR SINGULAR LINEAR SYSTEMS

An Extrapolated Gauss-Seidel Iteration

Iterative Solution methods

9.1 Preconditioned Krylov Subspace Methods

arxiv: v1 [math.ra] 11 Aug 2014

Applied Mathematics Letters

Performance Comparison of Relaxation Methods with Singular and Nonsingular Preconditioners for Singular Saddle Point Problems

Classical iterative methods for linear systems

Yimin Wei a,b,,1, Xiezhang Li c,2, Fanbin Bu d, Fuzhen Zhang e. Abstract

Journal of Computational and Applied Mathematics. Optimization of the parameterized Uzawa preconditioners for saddle point matrices

Linear Algebra Section 2.6 : LU Decomposition Section 2.7 : Permutations and transposes Wednesday, February 13th Math 301 Week #4

c 1995 Society for Industrial and Applied Mathematics Vol. 37, No. 1, pp , March

A LINEAR SYSTEMS OF EQUATIONS. By : Dewi Rachmatin

Ann. Funct. Anal. 5 (2014), no. 2, A nnals of F unctional A nalysis ISSN: (electronic) URL:

Lecture Note 7: Iterative methods for solving linear systems. Xiaoqun Zhang Shanghai Jiao Tong University

An Even Order Symmetric B Tensor is Positive Definite

MATRIX SPLITTING PRINCIPLES

Computational Methods. Systems of Linear Equations

The Solution of Linear Systems AX = B

For δa E, this motivates the definition of the Bauer-Skeel condition number ([2], [3], [14], [15])

The semi-convergence of GSI method for singular saddle point problems

ELA

Discrete mathematical models in the analysis of splitting iterative methods for linear systems

1. Introduction. We consider restricted Schwarz methods for the solution of linear systems of the form

ETNA Kent State University

THE RELATION BETWEEN THE QR AND LR ALGORITHMS

Chapter 7 Iterative Techniques in Matrix Algebra

Splitting Iteration Methods for Positive Definite Linear Systems

Math 5630: Iterative Methods for Systems of Equations Hung Phan, UMass Lowell March 22, 2018

6. Iterative Methods for Linear Systems. The stepwise approach to the solution...

H-matrix theory vs. eigenvalue localization

An inexact parallel splitting augmented Lagrangian method for large system of linear equations

10.2 ITERATIVE METHODS FOR SOLVING LINEAR SYSTEMS. The Jacobi Method

A new approach to solve fuzzy system of linear equations by Homotopy perturbation method

COURSE Iterative methods for solving linear systems

Z-Pencils. November 20, Abstract

Math 1080: Numerical Linear Algebra Chapter 4, Iterative Methods

Some bounds for the spectral radius of the Hadamard product of matrices

R, 1 i 1,i 2,...,i m n.

Introduction. Math 1080: Numerical Linear Algebra Chapter 4, Iterative Methods. Example: First Order Richardson. Strategy

Second Order Iterative Techniques for Boundary Value Problems and Fredholm Integral Equations

Background. Background. C. T. Kelley NC State University tim C. T. Kelley Background NCSU, Spring / 58

Determinants of Partition Matrices

The convergence of stationary iterations with indefinite splitting

Math 471 (Numerical methods) Chapter 3 (second half). System of equations

DEN: Linear algebra numerical view (GEM: Gauss elimination method for reducing a full rank matrix to upper-triangular

Iterative Methods for Solving A x = b

Fast Verified Solutions of Sparse Linear Systems with H-matrices

Tensor Complementarity Problem and Semi-positive Tensors

SEMI-CONVERGENCE ANALYSIS OF THE INEXACT UZAWA METHOD FOR SINGULAR SADDLE POINT PROBLEMS

We first repeat some well known facts about condition numbers for normwise and componentwise perturbations. Consider the matrix

ETNA Kent State University

Solving Linear Systems

Scientific Computing WS 2018/2019. Lecture 10. Jürgen Fuhrmann Lecture 10 Slide 1

NP-hardness of the stable matrix in unit interval family problem in discrete time

Lab 1: Iterative Methods for Solving Linear Systems

Preliminary/Qualifying Exam in Numerical Analysis (Math 502a) Spring 2012

The estimation of eigenvalues of sum, difference, and tensor product of matrices over quaternion division algebra

Transcription:

Computers Mathematics with Applications 56 (2008) 2048 2053 Contents lists available at ScienceDirect Computers Mathematics with Applications journal homepage: wwwelseviercom/locate/camwa Convergence analysis of the preconditioned Gauss Seidel method for H-matrices Qingbing Liu a,b, Guoliang Chen a,, Jing Cai a,c a Department of Mathematics, East China Normal University, Shanghai 200062, PR China b Department of Mathematics, Zhejiang Wanli University, Ningbo 315100, PR China c School of Science, Huzhou Teachers College, Huzhou 313000, PR China a r t i c l e i n f o a b s t r a c t Article history: Received 8 August 2011 Received in revised form 25 February 2012 Accepted 5 March 2012 Keywords: H-matrix Preconditioner Preconditioned iterative method Gauss Seidel method H-splitting In 1997, Kohno et al [Toshiyuki Kohno, Hisashi Kotakemori, Hiroshi Niki, Improving the modified Gauss Seidel method for Z-matrices, Linear Algebra Appl 267 (1997) 113 123] proved that the convergence rate of the preconditioned Gauss Seidel method for irreducibly diagonally dominant Z-matrices with a preconditioner I + S α is superior to that of the basic iterative method In this paper, we present a new preconditioner I + K β which is different from the preconditioner given by Kohno et al [Toshiyuki Kohno, Hisashi Kotakemori, Hiroshi Niki, Improving the modified Gauss Seidel method for Z-matrices, Linear Algebra Appl 267 (1997) 113 123] prove the convergence theory about two preconditioned iterative methods when the coefficient matrix is an H-matrix Meanwhile, two novel sufficient conditions for guaranteeing the convergence of the preconditioned iterative methods are given Crown Copyright 2008-Published by Elsevier Ltd All rights reserved 1 Introduction We consider the following linear system Ax b, where A is a complex n n matrix, x b are n-dimensional vectors For any splitting, A M N with the nonsingular matrix M, the basic iterative method for solving the linear system (1) is as follows: x i+1 M 1 Nx i + M 1 b i 0, 1, 2, Some techniques of preconditioning which improve the rate of convergence of these iterative methods have been developed Let us consider a preconditioned system of (1) PAx Pb, where P is a nonsingular matrix The corresponding basic iterative method is given in general by x i+1 M 1 P N P x i + M 1 P Pb i 0, 1, 2,, where PA M P N P is a splitting of PA (1) (2) This project was granted financial support by the Shanghai Science Technology Committee Shanghai Priority Academic Discipline Foundation (no 062112065) Corresponding author E-mail addresses: lqb_2008@hotmailcom (Q Liu), glchen@mathecnueducn (G Chen)

2049 In 1997, Kohno et al [1] proposed a general method for improving the preconditioned Gauss Seidel method with the preconditioned matrix P I + S α, if A is a nonsingular diagonally dominant Z-matrix with some conditions, where 0 α 1 a 1,2 0 0 0 0 α 2 a 2,3 0 S α 0 0 0 α n1 a n1,n 0 0 0 0 They showed numerically that the preconditioned Gauss Seidel method is superior to the original iterative method if the parameters α i 0 (i 1, 2,, n 1) are chosen appropriately Many other researchers have considered left preconditioners applied to linear system (1) that made the associated Jacobi Gauss Seidel methods converge faster than the original ones Such modifications or improvements based on prechosen preconditioners were considered by Milaszewicz [2] who based his ideas on previous ones (see, eg, [3]), by Gunawardena et al [4], very recently by Li Sun [5] who extended the class of matrices considered in [1] by other researchers (see, eg, [6 12]), many results for more general preconditioned iterative methods were obtained In this paper, besides the above preconditioned method, we will consider the following preconditioned linear system A β x b β, where A β (I + K β )A b β (I + K β )b with 0 0 0 0 β 1 a 2,1 0 0 0 K β 0 β 2 a 3,2 0 0, 0 0 β n1 a n,n1 0 where β i 0 (i 1, 2,, n 1) Our work gives the convergence analysis of the above two preconditioned Gauss Seidel methods for the case when a coefficient matrix A is an H-matrix obtains two sufficient conditions for guaranteeing the convergence of two preconditioned iterative methods (3) 2 Preliminaries Without loss of generality, let the matrix A of the linear system (1) be A I L U, where I is an identity matrix, L U are strictly lower upper triangular matrices obtained from A, respectively We assume a i,i+1 0, considering the preconditioner P I + S α, then we have whenever A α (I + S α )A I L S α L (U S α + S α U) b α (I + Sα)b, α i a i,i+1 a i+1,i 1 for i 1, 2,, n 1, then (I L S α L) 1 exists Hence it is possible to define the Gauss Seidel iteration matrix for A α, namely T α (I L S α L) 1 (U S α + S α U) (4) Similarly, if a i,i1 0, considering the preconditioner P I + K β, then we have A β (I + K β )A I L + K β K β L (U + K β U) b β (I + K β )b, define the Gauss Seidel iteration matrix for A β, namely T β (I L + K β K β L) 1 (U + K β U) We first recall the following: A real vector x (x 1, x 2,, x n ) T is called nonnegative(positive) denoted by x 0 (x > 0), if x i 0 (x i > 0) for all i Similarly, a real matrix A (a i,j ) is called nonnegative denoted by A 0 (A > 0) if a i,j 0 (a i,j > 0) for all i, j, the absolute value of A is denoted by A ( a i,j ) Definition 21 ([13]) A real matrix A is called an M-matrix if A si B, B 0 s > ρ(b), where ρ(b) denotes the spectral radius of B (5)

2050 Definition 22 ([13]) A complex matrix A (a i,j ) is an H-matrix, if its comparison matrix A (ā i,j ) is an M-matrix, where ā i,j is ā i,i a i,i, ā i,j a i,j, i j Definition 23 ([14]) The splitting A M N is called an H-splitting if M N is an M-matrix Lemma 21 ([14]) Let A M N be a splitting If it is an H-splitting, then A M are H-matrices ρ(m 1 N) ρ( M 1 N ) < 1 Lemma 22 ([15]) Let A have nonpositive off-diagonal entries Then a real matrix A is an M-matrix if only if there exists some positive vector u (u 1,, u n ) T > 0 such that Au > 0 3 Convergence results Theorem 31 Let A be an H-matrix with unit diagonal elements, A α (I + S α )A M α N α, M α I L S α L N α US α +S α U Let u (u 1,, u n ) T be a positive vector such that A u > 0 Assume that a i,i+1 0 for i 1, 2,, n1, u i i1 a i,j u j n a i,j u j + a i,i+1 u i+1 α i, a i,i+1 n a i+1,j u j then α i > 1 for i 1, 2,, n 1 for 0 α i < α i, the splitting A α M α N α is an H-splitting ρ(mα 1 N α) < 1 so that the iteration (2) converges to the solution of (1) Proof By assumption, let a positive vector u > 0 satisfy A u > 0, from the definition of A, we have u i a i,j u j > 0 for i 1, 2,, n 1 Therefore, we have u i a i,j u j u i a i,j u j + a i,i+1 u i+1 a i,i+1 a i,j u j + a i,i+1 u i+1 Observe that u i n u i a i,j u j a i+1,j u j a i+1,j u j for i 1, 2,, n 1 +1 a i,j u j > 0 u i+1 n a i+1,j u j > 0, then we have +1 a i,j u j + a i,i+1 u i+1 a i,i+1 a i+1,j u j > 0, u i a i,j u j a i,j u j + a i,i+1 u i+1 > a i,i+1 a i+1,j u j > 0 for i 1, 2,, n 1 This implies u i i1 a i,j u j n a i,j u j + a i,i+1 u i+1 α i > 1 for i 1, 2,, n 1 a i,i+1 n a i+1,j u j Hence, α i > 1 for i 1, 2,, n 1

2051 In order to prove that ρ(mα 1 N α) < 1, we first show that M α N α is an M-matrix Let [( M α N α )u] i be the ith element in the vector ( M α N α )u for i 1, 2,, n 1 Then we have [( M α N α )u] i 1 α i a i,i+1 a i+1,i u i a i,j α i a i,i+1 a i+1,j u j a i,j α i a i,i+1 a i+1,j u j u i α i a i,i+1 a i+1,i u i a i,j u j α i a i,i+1 a i+1,j u j [( M α N α )u] n u n a i,j u j α i a i,i+1 a i+1,j u j 1 α i a i,i+1 u i+1, (6) a n,j u j > 0 (7) j n If 0 α i 1 for i 1, 2,, n 1, then we have [( M α N α )u] i u i α i a i,i+1 a i+1,i u i a i,j u j α i a i,i+1 a i+1,j u j Since u i n u i a i,j u j α i a i,i+1 a i+1,j u j (1 α i ) a i,i+1 u i+1 a i,j u j + α i a i,i+1 u i+1 α i a i,i+1 u i a i,j u j + α i a i,i+1 u i+1 a i,j u j > 0 u i+1 n a i+1,j u j > 0, we have +1 a i+1,j u j +1 a i+1,j u j (8) [( M α N α )u] i > 0 for i 1, 2,, n 1 (9) +1 If 1 < α i < α i for i 1, 2,, n 1, from (6) the definition of α i, then we have [( M α N α )u] i u i α i a i,i+1 a i+1,i u i a i,j u j α i a i,i+1 a i+1,j u j a i,j u j α i a i,i+1 a i+1,j u j (α i 1) a i,i+1 u i+1 u i a i,j u j > 0 Therefore, from (7) to (10), we have ( M α N α )u > 0 for 0 α i < α i a i,j u j + a i,i+1 u i+1 α i a i,i+1 a i+1,j u j By Lemma 22, M α N α is an M-matrix for 0 α i < α i (i 1, 2,, n 1) From Definition 23, A α M α N α is an H-splitting for 0 α i < α i (i 1, 2,, n 1) Hence, according to Lemma 21, we have that A α M α are H-matrices ρ(mα 1 N α) ρ( M α 1 N α ) < 1 for 0 α i < α i (i 1, 2,, n 1) Theorem 32 Let A be an H-matrix with unit diagonal elements, A β (I + K β )A M β N β, M β I L + K β K β L N β U + K β U Let v (v 1,, v n ) T be a positive vector such that A v > 0 Assume that a i,i1 0 for i 2,, n, v i i2 a i,j v j n a i,j v j + a i,i1 v i1 β i, a i,i1 n a i1,j v j (10)

2052 then β i > 1 for i 2,, n for 0 β i < β i, the splitting A β M β N β is an H-splitting ρ(m 1 β N β) < 1 so that the iteration (2) converges to the solution of (1) Proof From the conditions of the theorem, we know there exists a positive vector v > 0 satisfying A u > 0, then we have v i a i,j v j > 0 for i 2,, n Therefore, we have v i a i,j v j v i Since v i n a i,j v j + a i,i1 v i1 a i,i1 a i,j v j + a i,i1 v i1 v i a i,j v j a i1,j v j a i1,j v j for i 2,, n 1 a i,j v j > 0 v i1 n a i1,j v j > 0, we have 1 a i,j v j + a i,i1 v i1 a i,i1 So it is obvious to obtain v i a i,j v j a i,j v j + a i,i1 v i1 > a i,i1 This implies β i v i i2 a i,j v j n a i,j v j + a i,i1 v i1 a i1,j v j > 0 a i1,j v j > 0 for i 2,, n > 1 for i 2,, n a i,i1 n a i1,j v j Namely, β i > 1 for i 2,, n In order to prove that ρ(m 1 β N β) < 1, we first show that M β N β is an M-matrix, let [( M β N β )v] i be the ith element in the vector ( M β N β )v for i 2,, n Then we have [( M β N β )v] i v i β i1 a i,i1 a i1,i v i a i,j β i1 a i,i1 a i1,j v j a i,j β i1 a i,i1 a i1,j v j v i β i1 a i,i1 a i1,i v i a i,j v j β i1 a i,i1 a i1,j v j a i,j v j β i1 a i,i1 a i1,j v j 1 β i1 a i,i1 v i1, (11) [( M β N β )v] 1 v 1 a 1,j v j > 0 (12) j2 If 0 β i1 1 for i 2,, n, then we have [( M β N β )v] i v i β i1 a i,i1 a i1,i v i a i,j v j β i1 a i,i1 a i1,j v j a i,j v j β i1 a i,i1 a i1,j v j (1 β i1 ) a i,i1 v i1

Observe that v i n v i a i,j v j + β i1 a i,i1 v i1 β i1 a i,i1 v i a i,j v j + β i1 a i,i1 v i1 a i1,j v j 1 2053 a i1,j v j (13) 1 a i,j v j > 0 v i1 n a i1,j v j > 0, then we have 1 [( M β N β )v] i > 0 for i 2, 3,, n (14) If 1 < β i1 < β i1 for i 2,, n, from (11) the definition of β i1, then we have [( M β N β )v] i v i β i1 a i,i1 a i1,i v i a i,j v j β i1 a i,i1 a i1,j v j a i,j v j β i1 a i,i1 a i1,j v j (β i1 1) a i,i1 v i1 v i a i,j v j > 0 Therefore, from (12) to (15), we have a i,j v j + a i,i1 v i1 β i1 a i,i1 ( M β N β )v > 0 for 0 β i1 < β i1 a i1,j v j By Lemma 22, M β N β is an M- matrix for 0 β i1 < β i1 (i 2,, n) From Definition 23, A β M β N β is an H-splitting for 0 β i1 < β i1 (i 2,, n) Hence, from Lemma 21, we have that A β M β are H-matrices ρ(m 1 β N β) ρ( M β 1 N β ) < 1 for 0 β i1 < β i1 (i 2,, n) Remark 31 From Theorems 31 32, we observe that the two preconditioned iterative methods converge to the solution of the linear system (1) if the coefficient matrix A which is an H-matrix satisfies the conditions of the corresponding theorems Hence the two theorems provide sufficient conditions for guaranteeing the convergence of the preconditioned iterative methods Acknowledgements The authors wish to express their thanks to Professor EY Rodin the two referees for helpful suggestions comments References [1] Toshiyuki Kohno, Hisashi Kotakemori, Hiroshi Niki, Improving the modified Gauss Seidel method for Z-matrices, Linear Algebra Appl 267 (1997) 113 123 [2] JP Milaszewicz, Improving Jacobi Gauss Seidel iterations, Linear Algebra Appl 93 (1987) 161 170 [3] JP Milaszewicz, On modified Jacobi linear operators, Linear Algebra Appl 51 (1983) 127 136 [4] A Gunawardena, SK Jain, L Snyder, Modified iterative methods for consistent linear systems, Linear Algebra Appl 149/156 (1991) 123 143 [5] W Li, WW Sun, Modified Gauss Seidel type methods Jacobi type methods for Z-matrix, Linear Algebra Appl 317 (2000) 227 240 [6] A Hadjidimos, D Noutsos, M Tzoumas, More on modifications improvements of classical iterative schemes for M-matrices, Linear Algebra Appl 364 (2003) 253 279 [7] D Noutsos, M Tzoumas, On optimal improvements of classical iterative schemes for Z-matrix, J Comput Appl Math 188 (2006) 89 106 [8] M Neumann, RJ Plemmons, Convergence of parallel multisplitting iterative methods for M-matrix, Linear Algebra Appl 88/89 (1987) 559 573 [9] W Li, A note on the preconditioned Gauss Seidel(GS) method for linear systems, J Comput Appl Math 182 (2005) 81 90 [10] XZ Wang, TZ Huang, YD Fu, Comparison results on preconditioned SOR-type method for Z-matrix linear systems, J Comput Appl Math 206 (2007) 726 732 [11] MJ Wu, L Wang, YZ Song, Preconditioned AOR iterative method for linear systems, Appl Numer Math 57 (2007) 572 685 [12] LY Sun, Some extensions of the improved modified Gauss Seidel iterative method for H-matrix, Numer Linear Algebr 13 (2006) 869 876 [13] RS Varga, Matrix Iterative Analysis, Prentice-Hall, Englewood Cliffs, NJ, 1962, pp 22 23 [14] A Frommer, DB Szyld, H-splitting two-stage iterative methods, Numer Math 63 (1992) 345 356 [15] KY Fan, Topological proofs for certain theorems on matrices with non-negative elements, Monatsh Math 62 (1958) 219 237 (15)