BLOCK ILU PRECONDITIONED ITERATIVE METHODS FOR REDUCED LINEAR SYSTEMS

Similar documents
On the Preconditioning of the Block Tridiagonal Linear System of Equations

Preface to the Second Edition. Preface to the First Edition

9.1 Preconditioned Krylov Subspace Methods

Incomplete LU Preconditioning and Error Compensation Strategies for Sparse Matrices

AMS Mathematics Subject Classification : 65F10,65F50. Key words and phrases: ILUS factorization, preconditioning, Schur complement, 1.

ITERATIVE METHODS FOR SPARSE LINEAR SYSTEMS

OUTLINE ffl CFD: elliptic pde's! Ax = b ffl Basic iterative methods ffl Krylov subspace methods ffl Preconditioning techniques: Iterative methods ILU

Fine-grained Parallel Incomplete LU Factorization

A Domain Decomposition Based Jacobi-Davidson Algorithm for Quantum Dot Simulation

A SPARSE APPROXIMATE INVERSE PRECONDITIONER FOR NONSYMMETRIC LINEAR SYSTEMS

Jae Heon Yun and Yu Du Han

An advanced ILU preconditioner for the incompressible Navier-Stokes equations

DELFT UNIVERSITY OF TECHNOLOGY

A Block Compression Algorithm for Computing Preconditioners

Solving Large Nonlinear Sparse Systems

Preconditioning Techniques Analysis for CG Method

A Method for Constructing Diagonally Dominant Preconditioners based on Jacobi Rotations

Preconditioners for the incompressible Navier Stokes equations

The solution of the discretized incompressible Navier-Stokes equations with iterative methods

A Numerical Study of Some Parallel Algebraic Preconditioners

Fast Iterative Solution of Saddle Point Problems

OUTLINE 1. Introduction 1.1 Notation 1.2 Special matrices 2. Gaussian Elimination 2.1 Vector and matrix norms 2.2 Finite precision arithmetic 2.3 Fact

A Robust Preconditioned Iterative Method for the Navier-Stokes Equations with High Reynolds Numbers

Department of Computer Science, University of Illinois at Urbana-Champaign

Robust solution of Poisson-like problems with aggregation-based AMG

ON THE GENERALIZED DETERIORATED POSITIVE SEMI-DEFINITE AND SKEW-HERMITIAN SPLITTING PRECONDITIONER *

Multilevel low-rank approximation preconditioners Yousef Saad Department of Computer Science and Engineering University of Minnesota

AMS526: Numerical Analysis I (Numerical Linear Algebra for Computational and Data Sciences)

Solving Symmetric Indefinite Systems with Symmetric Positive Definite Preconditioners

Generalized AOR Method for Solving System of Linear Equations. Davod Khojasteh Salkuyeh. Department of Mathematics, University of Mohaghegh Ardabili,

Preconditioning Techniques for Large Linear Systems Part III: General-Purpose Algebraic Preconditioners

Many preconditioners have been developed with specic applications in mind and as such their eciency is somewhat limited to those applications. The mai

Multigrid absolute value preconditioning

Chapter 7 Iterative Techniques in Matrix Algebra

Optimal Interface Conditions for an Arbitrary Decomposition into Subdomains

ON ORTHOGONAL REDUCTION TO HESSENBERG FORM WITH SMALL BANDWIDTH

A SHORT NOTE COMPARING MULTIGRID AND DOMAIN DECOMPOSITION FOR PROTEIN MODELING EQUATIONS

The flexible incomplete LU preconditioner for large nonsymmetric linear systems. Takatoshi Nakamura Takashi Nodera

Incomplete Cholesky preconditioners that exploit the low-rank property

Lecture 18 Classical Iterative Methods

Optimal Left and Right Additive Schwarz Preconditioning for Minimal Residual Methods with Euclidean and Energy Norms

Lecture 9 Approximations of Laplace s Equation, Finite Element Method. Mathématiques appliquées (MATH0504-1) B. Dewals, C.

THE solution to electromagnetic wave interaction with material

Multigrid and Domain Decomposition Methods for Electrostatics Problems

Numerische Mathematik

AMS526: Numerical Analysis I (Numerical Linear Algebra) Lecture 23: GMRES and Other Krylov Subspace Methods; Preconditioning

A robust algebraic multilevel preconditioner for non-symmetric M-matrices

DELFT UNIVERSITY OF TECHNOLOGY

A Novel Approach for Solving the Power Flow Equations

Cholesky factorisations of linear systems coming from a finite difference method applied to singularly perturbed problems

Contents. Preface... xi. Introduction...

Key words. conjugate gradients, normwise backward error, incremental norm estimation.

FINE-GRAINED PARALLEL INCOMPLETE LU FACTORIZATION

Mathematics and Computer Science

c 2015 Society for Industrial and Applied Mathematics

An Efficient Low Memory Implicit DG Algorithm for Time Dependent Problems

20. A Dual-Primal FETI Method for solving Stokes/Navier-Stokes Equations

Scientific Computing WS 2018/2019. Lecture 9. Jürgen Fuhrmann Lecture 9 Slide 1

Solving PDEs with Multigrid Methods p.1

C. Vuik 1 R. Nabben 2 J.M. Tang 1

CONVERGENCE BOUNDS FOR PRECONDITIONED GMRES USING ELEMENT-BY-ELEMENT ESTIMATES OF THE FIELD OF VALUES

Scientific Computing with Case Studies SIAM Press, Lecture Notes for Unit VII Sparse Matrix

Parallel Iterative Methods for Sparse Linear Systems. H. Martin Bücker Lehrstuhl für Hochleistungsrechnen

ETNA Kent State University

XIAO-CHUAN CAI AND MAKSYMILIAN DRYJA. strongly elliptic equations discretized by the nite element methods.

Optimal V-cycle Algebraic Multilevel Preconditioning

On the influence of eigenvalues on Bi-CG residual norms

The amount of work to construct each new guess from the previous one should be a small multiple of the number of nonzeros in A.

Iterative Methods and Multigrid

On the performance of the Algebraic Optimized Schwarz Methods with applications

Iterative Methods and High-Order Difference Schemes for 2D Elliptic Problems with Mixed Derivative

Fine-Grained Parallel Algorithms for Incomplete Factorization Preconditioning

5.1 Banded Storage. u = temperature. The five-point difference operator. uh (x, y + h) 2u h (x, y)+u h (x, y h) uh (x + h, y) 2u h (x, y)+u h (x h, y)

J.I. Aliaga 1 M. Bollhöfer 2 A.F. Martín 1 E.S. Quintana-Ortí 1. March, 2009

Optimal multilevel preconditioning of strongly anisotropic problems.part II: non-conforming FEM. p. 1/36

Algorithm for Sparse Approximate Inverse Preconditioners in the Conjugate Gradient Method

Linear Solvers. Andrew Hazel

Multilevel Preconditioning of Graph-Laplacians: Polynomial Approximation of the Pivot Blocks Inverses

Efficient Augmented Lagrangian-type Preconditioning for the Oseen Problem using Grad-Div Stabilization

Numerical behavior of inexact linear solvers

Solving PDEs with CUDA Jonathan Cohen

Domain Decomposition Preconditioners for Spectral Nédélec Elements in Two and Three Dimensions

Iterative Methods for Linear Systems

THE solution of the absolute value equation (AVE) of

Arnoldi Methods in SLEPc

Efficient Solvers for the Navier Stokes Equations in Rotation Form

Residual iterative schemes for largescale linear systems

Iterative Methods for Linear Systems of Equations

ON THE GLOBAL KRYLOV SUBSPACE METHODS FOR SOLVING GENERAL COUPLED MATRIX EQUATIONS

ADDITIVE SCHWARZ FOR SCHUR COMPLEMENT 305 the parallel implementation of both preconditioners on distributed memory platforms, and compare their perfo

M.A. Botchev. September 5, 2014

The Deflation Accelerated Schwarz Method for CFD

ON A GENERAL CLASS OF PRECONDITIONERS FOR NONSYMMETRIC GENERALIZED SADDLE POINT PROBLEMS

A High-Performance Parallel Hybrid Method for Large Sparse Linear Systems

The Conjugate Gradient Method

Local Mesh Refinement with the PCD Method

Preconditioning for Nonsymmetry and Time-dependence

WHEN studying distributed simulations of power systems,

AN INTRODUCTION TO DOMAIN DECOMPOSITION METHODS. Gérard MEURANT CEA

Solving Sparse Linear Systems: Iterative methods

Transcription:

CANADIAN APPLIED MATHEMATICS QUARTERLY Volume 12, Number 3, Fall 2004 BLOCK ILU PRECONDITIONED ITERATIVE METHODS FOR REDUCED LINEAR SYSTEMS N GUESSOUS AND O SOUHAR ABSTRACT This paper deals with the iterative solution of a large sparse matrix when this matrix is in block-partitioned form arising from discrete elliptic Partial Differential Equations (PDEs) We introduce a block red-black coloring technique to reorder the unknowns, which is based on the standard red-black coloring We propose some preconditioning techniques which combine the block incomplete LU factorization and the Schur complement technique, where the reduced system is preconditioned using different preconditioners Some theoretical results to control the residual norm are presented 1 Introduction In this paper, we address the problem of developing preconditioners for the solution of large linear systems (11) Au = b, where A is an n n real sparse matrix The above system is assumed to have the following block form ( ( ) ( B F x f (12) = E C) y g) This block structure arises from the standard and block Red-Black colorings described in Sections 2 and 3, or from the domain decomposition method which is based on decomposing the physical domain Ω into a number of subdomains Ω i, 1 i k, for each of which an independent incomplete factorization can be computed and applied in parallel [7 9, 25, 31] The main idea is to obtain more parallelism at the subdomain level rather than at the grid point level Another approach to increase AMS subject classification: 65F10, 65N06, 65N22 Keywords: Sparse linear systems, ordering methods, block incomplete LU factorization, Schur complement, Krylov subspace Copyright c Applied Mathematics Institute, University of Alberta 351

352 N GUESSOUS AND O SOUHAR the degree of parallelism is to use several hyperplane wavefronts to sweep through the grid For example van der Vorst [37] considered starting wavefronts from each of the four corners in a 2D rectangular grid Earlier, Meurant [26] and van der Vorst [38] used a similar idea in which the grid is divided into equal parts (eg, halves or quadrants) and each part is ordered in its own natural ordering A broad class of preconditioners is based on incomplete LU factorizations which were originally developed for M-matrices arising from the discretization of simple elliptic partial differential equations For more details on this class of preconditioners, the interested readers can be refer to [1, 2, 10, 11, 20, 21, 23, 24, 27 31, 34 36, 39] The matrix A can be written as ( ) ( ) B I B (13) A = 1 F, E S I where S = C EB 1 F is the Schur complement matrix The matrix S does not need to be formed explicitly in order to solve the reduced system by an iterative method For example, if a Krylov subspace method without preconditioning is used, then the only operations that are required with the matrix S are matrix-vector operations w = Sv Such operations can be performed as follows (a) Compute v = F v (b) Solve Bz = v (c) Compute w = Cv Ez The above procedure involves only matrix-vector multiplications and one linear system solution (b) with B Note that the linear system with B must be solved exactly, by using a direct solution technique or by an iterative technique with a high level of accuracy It is well known that the ordering of the unknowns can have a significant effect on the convergence of a preconditioned iterative method and on its implementation on a parallel computer [13, 14], especially in the application of the ILU preconditioners Indeed, if we consider a graph containing a central node and whose only edges join this node to the (n 1) remaining ones And consider now the two extreme orderings, the first giving the number 1 to the central node and the second giving it the last number n Let A 1 and A 2 be the adjacency matrices to these

BLOCK ILU PRECONDITIONED ITERATIVE METHODS 353 two orderings respectively, which can be written as A 1 = A 2 =, Let A 1 = L 1 U 1 and A 2 = L 2 U 2 be the factorizations according to Gaussian elimination process, A 1 and A 2 have the same number of non-zero entries In the first ordering the structure (L 1 +U 1 ) has all its coefficients different from zero, whereas with the second ordering (L 2 + U 2 ) it has the same structure as A 2 Furthermore, the factors governed by the factorization ILU(0) of A 2 are exactly L 2 and U 2! This paper is organized as follows: Section 2 gives a brief overview of a standard red-black coloring technique Section 3 describes an extension of the previous section named block red-black coloring Section 4 provides some preconditioners arising from block ILU factorization and some theoretical results to control the residual norm Section 5 gives some numerical experiments and a few concluding remarks 2 Standard red-black coloring The problem addressed by multicoloring is to determine a coloring of the nodes of the adjacency graph of a matrix such that any two adjacent nodes have different colors For simple two-dimensional finite difference grid (a five-point operator), we can easily separate the grid points into two sets, red and black, so that the nodes in one set are adjacent only to nodes in the other set, as shown in Figure 1 for a 12 8 grid (with white nodes correspond to red nodes) Assume that the unknowns are labeled by listing the red unknowns first

354 N GUESSOUS AND O SOUHAR together, followed by the black ones, we obtain a system of the form (21) ( ) ( ) D1 F x = E D 2 y ( f g), where D 1 and D 2 are diagonal matrices One way to exploit the red-black ordering is to use the standard SSOR or ILU (0) preconditioners for solving (21) The preconditioning operations are highly parallel For example, the linear system that arises from the forward solving in SSOR will have the form (22) ( D1 E ) ( ) x = D 2 y ( f g) Figure 1 A standard red-black coloring of 12 8 grid This system can be solved by performing the following sequence of operations: Solve D 1 x = f Compute g = g Ex Solve D 2 y = g This consists of two diagonal scalings (operations 1 and 3) and a sparse matrix-vector product (operation 2) The situation is identical with the ILU (0) preconditioning A simple look at the structure of the ILU factors reveals that many more elements are dropped with the red-black ordering than with the natural ordering The result is that the number of

BLOCK ILU PRECONDITIONED ITERATIVE METHODS 355 iterations to achieve convergence can be much higher with the red-black ordering than with the natural ordering [30] The second method that has been used in connection with the redblack ordering solves the reduced system which involves only the black unknowns Eliminating the red unknowns from (21) results in the reduced system: (23) (D 2 ED 1 1 F ) y = g ED 1 1 f Note that this new system is again a sparse linear system with about half as many unknowns In addition, it has been observed that for easy problems, the reduced system can often be solved efficiently with only diagonal preconditioning The computation of the reduced system is a highly parallel and inexpensive process We can also use a multi-level ordering of the grid points For more details on this technique and on the order of the condition number see the references of Brand and Heinmann [4] and [12] where they proposed a repeated red-black (RRB) ordering 3 Block red-black coloring We now consider an extension of the red-black coloring which consists of transforming the system (11) into the following block form (31) ( ( ) ( B F x f = E C) y g) Here B and C are block diagonal matrices, where each block is of size 2 A block of size 2 will be found by coupling a node j with its neighbor (j+1) with the same color (Red) followed by two nodes that are different color (Black) Assume again that the unknowns are labeled by listing the red unknowns first together, followed by the black ones This block two-by-two colors is illustrated in Figure 2 for a 12 8 grid (with white nodes correspond to red nodes)

356 N GUESSOUS AND O SOUHAR Figure 2 A block red-black coloring of 12 8 grid This leads to a 2 2 submatrix of the form (32) ( ) bj,j b j,j+1, b j+1,j b j+1,j+1 ( ) cj,j c j,j+1 c j+1,j c j+1,j+1 This recursive block version technique Repeated Block Red-Black ordering (RBRB) can be applied as in [4, 12] This technique starts with all red nodes indicated by, followed by half of the remaining black nodes corresponding to alternating horizontal grid line indicated by It can be easily checked that the remaining black nodes indicated by form a regular grid again with a double mesh spacing as the original one The process can then be repeated recursively until a sufficiently coarse level is reached on which the usual natural ordering can be applied Figure 3 shows an example of RBRB ordering Figure 3 A RBRB ordering, stopping after 1 coarse level (indicated by )

BLOCK ILU PRECONDITIONED ITERATIVE METHODS 357 4 Preconditioning techniques and preconditioned iterations Some results to control the preconditioned residual norm are given in this section First a preconditioner to solve the linear system of the block form is developed: ( ( ) ( B F x f (41) =, E C) y g) where B and C are square matrices Furthermore, we assume that B is nonsingular The block lower-upper (LU) triangular factorization of A can be written as ( ) ( ) ( ) B F B I B (42) = 1 F, E C E S I where S is the Schur complement matrix (43) S = C EB 1 F To compute the solution (x, y) T of the linear system (41), we therefore need to perform the following steps: (a) Compute g = g EB 1 f (b) Solve Sy = g (c) Solve Bx = (f F y) If the cost for computing the exact inverse of B is prohibitive, then we first approximate B 1 by a sparse matrix Y using sparse approximate inverse techniques; see [3, 18, 19], to approach the Schur complement S with a sparse matrix The matrix S = C EY F is sparse And solving (b) with it is more economical since S is sparser than S Note that, we can also use an approximate factorization of S: (44) S = S + R es = L SU S + R es which define a preconditioner for the Schur complement matrix, S = L SU S, and here, R es is the error matrix 41 Left preconditioning ILU In the following we suppose that the inverse of B is computed exactly Then the preconditioner can be written as: ( ) ( ) B I B (45) M = 1 E S F = L I M U M,

358 N GUESSOUS AND O SOUHAR with S being some approximate to the Schur complement S For example, in the form of an approximate LU factorization of S When C is a block diagonal matrix and is not singular, we can choose S = C, since the only inverse that occurs is that of a block diagonal matrix with small blocks, or S = I, ie, no preconditioning of S We can easily show that the preconditioned system has the particular form ( ) ( ) ( ) M 1 I B A = 1 F I I B 1 F (46) I S 1 S I = U 1 M DU M Then the quality of the preconditioner M depends on that of S, and the concentration of the spectrum 1 of M 1 A depends only on that of S 1 S, since σ(m 1 A) = {1} σ( S 1 S) Consider now a GMRES iteration to solve the preconditioned system (47) M 1 Au = M 1 b Given an initial guess u 0 = (x 0, y 0 ) T, we denote the preconditioned initial residual by r 0 = M 1 (b Au 0 ) = ( z0 s 0 ) It is useful to recall that the GMRES algorithm finds a vector u in the affine subspace u 0 + K m, where K m is a Krylov subspace: K m = K m (M 1 A, r 0 ) = Span { r 0, (M 1 A)r 0,, (M 1 A) m 1 r 0 } The GMRES minimizes M 1 (b Au) 2 for an arbitrary u in u 0 + K m u = u 0 + w, w K m Then, the preconditioned residual takes the form: r = M 1 (b Au) = M 1 (b Au 0 Aw) = r 0 M 1 Aw 1 σ(x) denotes the spectrum of a matrix X

BLOCK ILU PRECONDITIONED ITERATIVE METHODS 359 Theorem 41 Assume that the reduced system S 1 Sy = S 1 (g EB 1 f) is solved with GMRES starting with an arbitrary initial guess y 0 and let s m = S 1 (g Sy m ) be the preconditioned residual obtained at the m th step Then the preconditioned residual vector r m obtained at the m th step of GMRES for solving the block system (41) preconditioned by M of (46) and with an initial guess u 0 = (x 0, y 0 ) T satisfies Bx 0 = f F y 0, satisfies the inequality r m 2 (1 + B 1 F 2 ) s m 2 In particular, if s m = 0, then r m = 0 Proof Let s m = p m ( S 1 S)s 0 be the preconditioned residual obtained at the m th step of the GMRES algorithm starting with an arbitrary initial guess y 0 to solve the preconditioned reduced system (48) S 1 Sy = S 1 g, where g = g EB 1 f and p m is the m th residual polynomial, which minimizes p( S 1 S)s 0 2 among all polynomials p of degree m satisfying the constraint p(0) = 1 Let m p m (t) = α i t i The preconditioned residual r m obtained at the m th step of GMRES algorithm for solving the preconditioned unreduced system minimizes p(m 1 A)r 0 2 over all polynomials p of degree m which are consistent, ie, p(0) = 1, and satisfies i=0 where r m 2 p m (M 1 A)r 0 2 ( ) ( ) U 1 pm (I) z0 2 M p m ( S 1 U S) M s 0 ( ( ) U 1 M pm (I) δ 2 p m ( S)) S 1, s 0 2 (49) δ = z 0 + B 1 F s 0

360 N GUESSOUS AND O SOUHAR Due to the particular structure of B, that is a block diagonal matrix, and if we suppose that its blocks have a small size then, we have to solve the system by a direct solution technique On these conditions we have to solve (410) Bx 0 = f F y 0 exactly and we have δ = 0 Then (411) r m 2 U 1 M 2 p m ( S 1 S)s 2 0, that is, r m 2 U 1 M 2 s m 2 (1 + B 1 F 2 ) s m 2 Remark 41 In general, systems involving B may be solved in many ways, depending on their difficulty and what we know about B If B is known to be well conditioned, then triangular solves with incomplete LU factors may be sufficient For more difficult B matrices, the incomplete factors may be used as a preconditioner for an inner iterative process for B Suppose now that an iterative method is used to solve (410) Let x 0 the exact solution and x 0 0 an initial guess, then at kth step, the error is e k 0 = x 0 x k 0 We have Also, we get ( ) ( ) f Bx k 0 F y 0 Be k 0 Mr 0 = b Au 0 = g Ex k 0 Cy = 0 g Mr 0 = ( From the first component we have since B is not singular then Bz 0 + F s 0 Ez 0 + EB 1 F s 0 + Ss 0 B(z 0 + B 1 F s 0 ) = Be k 0, (412) δ = e k 0, and we have (413) r m 2 ( U 1 pm 2 (1) 2 e k 0 2 2 + s m 2 1/2 2) M )

BLOCK ILU PRECONDITIONED ITERATIVE METHODS 361 42 Split preconditioning Let A = L M U M + R A, where R A is the error matrix and ( S) ( ) B I B L M =, U E M = 1 F, I Then ( ) (414) L 1 1 I M AUM = S 1, S and the split preconditioning is obtained by solving: L 1 1 M AUM (415) v = L 1 M b U M u = v The Krylov subspace associated with this system is: K m (L 1 M where 1 AUA, r 0) = Span {r 0, (L 1 r 0 = L 1 M b (L 1 M 1 M AU AU 1 M )r 0,, (L 1 M M ) U Mu 0 = ( z0 s 0 ), AU 1 M )m 1 r 0 }, where u 0 = (x 0, y 0 ) T is an initial guess to solve (41), and v 0 = U A u 0 Theorem 42 Assume that the reduced system (48) is solved with GM- RES starting with an arbitrary initial guess y 0 and let s m = S 1 (g Sy m ) be the preconditioned residual obtained at the m th step Then the preconditioned residual vector r m+1 obtained at the (m+1) th step of GM- RES for solving the block system (415) with an arbitrary initial guess u 0 = (x 0, y 0 ) T satisfies the inequality (416) r m+1 2 I S 1 S 2 s m 2 In particular, if s m = 0, then r m+1 = 0 Proof Let p m be the m th residual polynomial, which minimizes p( S 1 S)s 0 2 among all polynomials p of degree m satisfying the constraint p(0) = 1, then the residual vector of the m th GMRES approximation associated with the preconditioned reduced system (48) is the form s m = p m ( S 1 S)s 0 We define the polynomial q m+1 (t) = (1 t)p m (t) The rest is similar to the proof of Theorem 41

362 N GUESSOUS AND O SOUHAR Definition 41 A matrix B R n n is said to be a Z-matrix if all of its off-diagonal entries (if any) are nonpositive: b i,j 0, for all i j Remark 42 Assume that B is a block diagonal matrix, where each block is of size 2 Furthermore, we suppose that( B is a Z-matrix ) and bj,j b strictly diagonally dominant Then each block j,j+1 is b j+1,j b j+1,j+1 nonsingular and its inverse can be written (417) ( ) 1 bj,j b j,j+1 = 1 b j+1,j b j+1,j+1 det j = ( ) bj+1,j+1 b j,j+1 b j+1,j b j,j ( ) βj,j β j,j+1, β j+1,j β j+1,j+1 where det j = b j,j b j+1,j+1 b j+1,j b j,j+1 Since B is strictly diagonally dominant then there exists ε j > 0 such that det j > ε j, then one easily has (418) 0 β i,j < γ ε B for each 1 i, j m, where ε B = min{ε j } and γ = max 1 i,j n { a i,j } In other words, By Theorem 41, we deduce (419) r m 2 ( m m B 1 2 B 1 F = 2m γ ε B j=1 i=1 β 2 i,j ) 1 2 ( 1 + 2m γ ) F 2 s m 2 ε B 43 Schur complement preconditioning Assume that C has the same order m and the same structure (block diagonal matrix of size 2) and satisfies the same properties of B Then, we can choose S = C as a preconditioner for the Schur complement matrix S In this case and

BLOCK ILU PRECONDITIONED ITERATIVE METHODS 363 from Theorem 42, one easily has r m+1 2 I C 1 S 2 s m 2 C 1 EB 1 F 2 s m 2 C 1 2 B 1 2 E 2 F 2 s m 2 2m γ γ E 2 F 2 s m 2 ε C ε B Then, we get ( γ ) 2 (420) r m+1 2 n E 2 F 2 s m 2, ε where ε = min{ε B, ε C } Now take S = L SU S + R S to be an approximate factorization of S The other version in which the Schur complement matrix S is allowed to be preconditioned from the left and the right by L S and U S, respectively Consider the reduced system L 1 1 SU S S (421) w = L 1 S g U Sy = w, where g = g EB 1 f The system (421) is solved by GMRES with an arbitrary initial guess y 0 (U Sy 0 = w 0 ) Let (422) s m = L 1 S g (L 1 S SU 1 S ) U Sy m Then the residual vector obtained at the (m + 1) th step of GMRES for solving the system (42) with left preconditioning L A and right preconditioning U A and with an initial guess u 0 = (x 0, y 0 ) T satisfies the inequality (423) r m+1 2 I L 1 S SU 1 S 2 s m 2 The proof of this result starts with the equality ( ) 1 ( ) B I B (424) A 1 1 ( F I = E L S U S The rest is similar to that of the previous result L 1 1 S SUS )

364 N GUESSOUS AND O SOUHAR 5 Numerical experiments and conclusion We consider the five point upwind finite difference scheme of the Laplace and convectiondiffusion equations [40] The advantage of this scheme is that it can produce stable discretizations, while with the standard central finite difference approximation and for small mesh sizes, sperious oscillations appear in the discrete solution of the convection-diffusion equation [27] We use the standard and block red-black colorings described in sections 2 and 3 with uniform meshsize h in both x and y directions In these cases, we compute the Schur complement matrix S explicitly In our tests, we present the number of preconditioned GMRES(20) iterations [33] with four different grid sizes In Tables 4 and 7, solu shows the CPU time in seconds for the solution phase; prec shows the CPU time in seconds taken in constructing the preconditioners; spar shows the sparsity ratio which is the ratio between the number of nonzero elements of the preconditioner in its block ILU factorization and that of the original matrix The numerical experiments were conducted on a Pentium III CPU at 550 MHZ with 64Mb RAM The right hand side was generated by assuming that the exact solution is a vector of all ones and the initial guess was a vector of some random numbers The computations were stopped when the 2-norm of the residual was reduced by a factor of 10 8 or when 200 iterations were reached, indicated by (-) We denote M S as the preconditioner defined by (45), where the solution of the reduced system (51) Sy = g EB 1 f, is performed using backward and forward solution steps by appling ILUT (10 4, 10) factorization [29] And by M I and M C the preconditioners where the reduced system (51) is unpreconditioned and preconditioned by C 1, respectively The first set of test problems is a Laplace equation with Dirichlet boundary conditions (52) u = rhs, which is defined on the unit square The second set of test problem is a convection-diffusion equation with Dirichlet boundary conditions, (53) ν u + v u = rhs,

BLOCK ILU PRECONDITIONED ITERATIVE METHODS 365 Standard red-black Block red-black h 1 M I M C M S M I M C M S 17 23 22 5 37 26 6 33 69 69 8 77 60 9 65 140 148 14 197 152 12 105 - - 19 - - 16 TABLE 1: Iterations to convergence for the Laplace problem by different preconditioners which is defined on the unit square v is a convective flow (velocity vector) and the viscosity parameter ν governs the ratio between convection and diffusion and define the inverse of Reynolds number Re = 1/ν Looking at the results for the Laplace problem (Table 1), it turns out that the application of the preconditioner M S offers a small number of iterations to achieve the convergence than M I and M C, for both standard and block red-black orderings Furthermore, M I appears to be better with the standard red-black colorings than with the block version, that is not the case for the preconditioners M C and M S One can easily see that the number of iterations increases strongly with h 1 for both preconditioners M I and M C and weakly for M S Tables 2 and 3, 5 and 6 show the number of iterations for the convectiondiffusion problem in two cases discretized by using standard and block colorings, preconditioned by three preconditioners, Case 2 was proposed by Ruge and Stuben [28] to test their algebraic multigrid method, this case has some semi-recirculation effect because the first of its convection coefficients vanishes at y = 05, while the first case has neither stagnation point nor semi-recirculation effect We observe that for the preconditioner M I, the number is less with the standard ordering than with the block version, except when (10 6 ν 10 2 and for h 1 =33 and 65) in the first case, and when (ν = 10 2 for all mesh-sizes) in the second case For the preconditioner M C and for the second case, the block version requires less number of iterations than the standard version, but for the first case we cannot deduce which ordering is better For the preconditioner M S, we observe a moderate deterioration for 10 3 ν 10 1 for both cases and both orderings We show that with the standard coloring we obtain a number of iterations less than with the block version for both cases, except for h 1 =65 and h 1 =105 when ν = 1 in the first case and 10 2 ν 1 in the second one

366 N GUESSOUS AND O SOUHAR Case 1: v = ( ) cos x sin y h 1 17 33 65 105 ν M I M C M S M I M C M S M I M C M S M I M C M S 1 29 27 5 71 70 8 148 147 14 - - 20 10 1 32 30 5 77 72 7 157 153 12 - - 18 10 2 29 18 3 118 97 5-177 8 - - 11 10 3 29 16 2 122 120 4-169 6 - - 8 10 4 29 16 3 122 121 3-187 4 - - 5 10 5 29 16 2 122 122 3-172 3 - - 4 10 6 29 16 2 122 122 2-178 3 - - 3 TABLE 2: Iterations to convergence for the convection-diffusion problem (Case 1) by different preconditioners using Standard version h 1 17 33 65 105 ν M I M C M S M I M C M S M I M C M S M I M C M S 1 37 26 6 77 60 9 196 151 12 - - 16 10 1 38 24 7 88 67 10 147 157 15 - - 18 10 2 39 20 5 114 59 9 176 168 15 - - 28 10 3 39 18 4 120 77 6 198-9 - - 11 10 4 38 18 4 122 60 5 197-6 - - 8 10 5 38 18 4 118 60 4 197-5 - - 6 10 6 38 18 3 120 60 4 197-4 - - 5 TABLE 3: Iterations to convergence for the convection-diffusion problem (Case 1) by different preconditioners using Block version From Tables 4 and 7, it is seen that with the standard red-black coloring M S requires less CPU time for the processing phase and less space memory than with the block version We also see that the CPU time for the solution phase depends on the number of iterations required to achieve the convergence

BLOCK ILU PRECONDITIONED ITERATIVE METHODS 367 Standard red-black coloring Block red-black coloring h 1 65 105 65 105 ν solu prec spar solu prec spar solu prec spar solu prec spar 1 027 021 290 110 039 290 021 022 307 100 039 308 10 1 021 019 290 095 039 290 032 021 307 119 039 308 10 2 017 013 287 065 038 289 033 020 304 189 038 306 10 3 011 011 281 039 036 286 016 020 297 066 037 300 10 4 008 007 255 027 033 273 015 016 278 039 033 288 10 5 006 004 195 022 022 210 011 014 235 037 028 247 10 6 005 004 154 011 021 161 006 011 195 023 022 205 TABLE 4: Performance of the preconditioner M S for solving the convection-diffusion problem (Case 1) Case 2: v = ( ) (2y 1)(1 x 2 ) 2xy(y 1) h 1 17 33 65 105 ν M I M C M S M I M C M S M I M C M S M I M C M S 1 29 30 5 78 78 8 155 155 14 - - 20 10 1 41 36 5 94 88 8-186 15 - - 22 10 2 60 53 4 160 133 7 - - 12 - - 20 10 3 77 62 3 167 160 5 - - 7 - - 10 10 4 82 61 3 198 173 4 - - 6 - - 8 10 5 83 60 3 197 174 4 - - 5 - - 6 10 6 83 60 3-176 4 - - 5 - - 6 TABLE 5: Iterations to convergence for the convection-diffusion problem (Case 2) preconditioned by different preconditioners using Standard version 6 Conclusion We have presented a technique to reorder the nodes of a graph adjacency to a sparse matrix, this thecnique is based on point red-black coloring, according to this ordering we have also presented a block ILU preconditioner with different approximations of the Schur complement matrix for solving sparse matrices We have shown both from a theoretical and a practical point of view that the efficiency of the global preconditioner M for the block original system depends strongly on how well S approximates the Schur complement matrix S From the

368 N GUESSOUS AND O SOUHAR h 1 17 33 65 105 ν M I M C M S M I M C M S M I M C M S M I M C M S 1 36 27 6 76 60 9 194 153 12 - - 16 10 1 41 30 6 98 60 9 163 158 13 - - 16 10 2 55 32 6 136 92 9 - - 12 - - 18 10 3 96 31 5 182 115 7-178 10 - - 13 10 4 99 19 5-116 6 - - 8 - - 10 10 5 98 21 8-120 5 - - 7 - - 10 10 6 116 37 12-120 14 - - 7 - - 9 TABLE 6: Iterations to convergence for the convection-diffusion problem (Case 2) preconditioned by different preconditioners using Block version Standard red-black coloring Block red-black coloring h 1 65 105 65 105 ν solu prec spar solu prec spar solu prec spar solu prec spar 1 024 021 290 110 038 290 022 024 307 103 042 308 10 1 026 021 290 115 038 290 027 021 307 106 045 308 10 2 022 021 289 112 038 289 019 021 305 116 040 306 10 3 012 013 279 060 034 285 016 019 292 101 038 297 10 4 011 010 256 044 033 263 011 015 263 063 034 274 10 5 011 008 214 038 027 224 006 013 231 060 032 238 10 6 010 008 180 034 022 191 006 010 201 055 024 213 TABLE 7: Performance of the preconditioner M S for solving the convection-diffusion problem (Case 2) numerical experiments that we have conducted in this paper (Laplace and convection-diffusion equations), we have learned that the preconditioner M S is much better than M I and M C where the convergence is not reached for small ν and large problems (with small meshsize h) and leads to a considerably faster convergence As can be seen by comparing Tables, we can solve more problems with the standard red-black ordering and M S preconditioner than with the block version, but the use of large blocks may be more efficient for some parallel architecture Acknowledgment The authors would like to thank the anonymous referee for their accurate remarks in improving the quality of this paper and for directing them to several of the references

BLOCK ILU PRECONDITIONED ITERATIVE METHODS 369 REFERENCES 1 O Axelsson, Iterative Solution Methods, University Press, Cambridge, 1994 2 R Barrett, M Berry, T Chan, J Demmel, J Donato, J Dongarra, V Eijkhout, R Pozo, C Romine and H A van der Vorst, Templates for the Solution of Linear Systems: Building Blocks for Iterative Methods, SIAM, Philadelphia, PA, 1994 3 M Benzi and M Tuma, A sparse approximate inverse preconditioner for nonsymmetric linear systems, SIAM J Sci Comput 19 (1998), 968 994 4 C Brand and Z E Heinemann, A new iterative solution technique for reservoir simulation equations on locally refined grids, SPE, (18410), 1989 5 B L Buzbee, F W Dorr, J A George and G H Golub, The direct solution of the discrete poisson equation on irregular regions, SIAM, J Num Anal 8 (1971), 722 736 6 B L Buzbee, G H Golub and C W Nielson, On direct methods for solving poisson s equations, SIAM, J Num Anal 7 (1970), 627 656 7 T F Chan, Analysis of preconditioners for domain decomposition, SIAM J Numer 24 (2) (1987), 382 390 8 T F Chan and D Goovaerts, A note on the efficiency of domain decomposed incomplete factorization, SIAM J Sci Stat Comput 11 (4) (1990), 794 803 9 T F Chan and H A van der Vorst, Approximate and incomplete factorization, Technical Report 871, Department of Mathematics, University of Utrecht, 1994 10 E Chow and Y Saad, Approximate inverse preconditioner via sparse-sparse iterations, SIAM J Comput 19 (3) (1998), 995 1023 11 E Chow and Y Saad, Approximate inverse techniques for block-partitioned matrices, SIAM J Sci Comput 18 (1997), 1657 1675 12 P Ciarlet, Jr, Repeated red-black ordering: a new approach, Numer Alg (1994) 13 E F D Azevedo, P A Forsyth and Wei-Pai Tang, Ordering methods preconditioned conjugate gradient methods applied to unstructured grid problems, SIAM J Matrix Anal Appl 13 (3) (1992), 944 961 14 I S Duff and G A Meurant, The effect of ordering on preconditioned conjugate gradients, BIT 29 (1989), 635 657 15 H C Elman, A stability analysis of incomplete LU factorization, Math Comp 47 (175) (1986), 191 217 16 H C Elman and G H Golub, Iterative methods for cyclically reduced nonself-adjoint linear systems, Math Comp 54 (1990), 671 700 17 H C Elman and G H Golub, Iterative methods for cyclically reduced nonself-adjoint linear systems II, Math Comp 56 (1991), 215 242 18 N I M Gould and J A Scott, Sparse approximate-inverse preconditioners using normalization techniques, SIAM J Sci Comput 19 (1998), 605 625 19 M G Grote and T Huckle, Parrallel preconditioning with sparse approximate inverses, SIAM J Sci Comput 18 (1997), 838 853 20 N Guessous and O Souhar, Multilevel block ILU preconditioner for sparse nonsymmetric M-matrices, J Comput Appl Math 162 (1) (2004), 231 246 21 N Guessous and O Souhar, Recursive two-level ILU preconditioner for nonsymmetric M-matrices, J Appl Math Comput 16 (1) (2004), 19 35 22 M T Jones and P E Plassman, A parallel graph coloring heuristic, SIAM J Sci Comput 14 (1993), 654 669 23 L YU Kolotilina and A Yu Yermin, Factorized sparse approximate inverse preconditioning, SIAM J Matrix-Anal 14 (1993), 45 58

370 N GUESSOUS AND O SOUHAR 24 T A Manteuffel, An incomplete factorization technique for positive definite linear systems, Math of Comput 32 (1980), 473 497 25 G Meurant, Domain decomposition methods for partial differential equations on parallel computers, Int J Supercomputing Appls 2 (1988), 5 12 26 G Meurant, The block preconditioned conjugate gradient method on vector computers, BIT 24 (1984), 623 633 27 Y Notay, A robust algebraic multilevel preconditioner for nonsymmetric M- matrix, Tech Rep GANMN 99-01, Université Libre de Bruxelles, Brussels, Belgium, 1999 28 J W Ruge and K Stuben, Efficient solution of finite difference and finite element equations, in: Multigrid Methods for Integral and Differential Equations (D J Paddon, H Holstein, eds), Clarendon Press, Oxford, 169 212, 1985 29 Y Saad, ILUT: A dual threshold incomplete LU preconditioner, Numer Linear Algebra Appl 1(4) (1994), 387 402 30 Y Saad, ILUM: A multi-elimination ILU preconditioner for general sparse matrices, SIAM J Sci Comput 17(4) (1996), 830 847 31 Y Saad, Iterative Methods for Sparse Linear Systems, PWS Publishing, New York, NY, 1996 32 Y Saad, SPARSKIT: A Basic Tool Kit for Sparse Matrix Computations, Technical Report CSRD TR 1029, University of Illinois at Urbana-Champaign, IL, 1990 33 Y Saad and M H Schultz, GMRES: A generalized minimal residual algorithm for solving nonsymetric linear systems, This Journal 7 (1986), 856 869 34 Y Saad and J Zhang, BILUM: Block versions of multi-elimination and multilevel ILU preconditioner for general linear sparse systems, SIAM J Sci Comput 20(6) (1999), 2103 2121 35 Y Saad and J Zhang, BILUTM: A domain based multilevel block ILUT preconditioner for general sparse matrices, SIAM J Matrix Anal Appl 21(1) (1999), 279 299 36 Y Saad and J Zhang, Diagonal threshold techniques in robust multi-level ILU preconditioner for general linear sparse systems, Numer Linear Algebra Appl 6 (1999), 257 280 37 H A van der Vorst, High performance preconditioning, SIAM J Sci Stat Comput 10 (1989), 1174 1185 38 H A van der Vorst, Large tridiagonal and block tridiagonal linear systems on vector and parallel computers, Parallel Computing 5 (1987), 45 54 39 J Zhang, A sparse Approximate Inverse Technique for Parallel Preconditioning of General Sparse Matrices, Technical report 281-98, Department of Computer science, University of Kentucky, Lexington, KY, 1998 40 J Zhang, Preconditioned iterative methods and finite difference schemes for convection-diffusion, Appl Math and Comput 109 (2000), 11 30 Ecole Normale supérieure, Bensouda BP 5206 Fès, Morocco E-mail address: guessous najib@hotmailcom Département de Mathématiques et Informatique Faculté des Sciences Dhar-Mahraz BP 1796 Atlas Fès, Morocco E-mail address: drsouhar@hotmailcom