AMS Mathematics Subject Classification : 65F10,65F50. Key words and phrases: ILUS factorization, preconditioning, Schur complement, 1.

Size: px
Start display at page:

Download "AMS Mathematics Subject Classification : 65F10,65F50. Key words and phrases: ILUS factorization, preconditioning, Schur complement, 1."

Transcription

1 J. Appl. Math. & Computing Vol. 15(2004), No. 1, pp BILUS: A BLOCK VERSION OF ILUS FACTORIZATION DAVOD KHOJASTEH SALKUYEH AND FAEZEH TOUTOUNIAN Abstract. ILUS factorization has many desirable properties such as its amenability to the syline format, the ease with which stability may be monitored, and the possibility of constructing a preconditioner with symmetric structure. In this paper we introduce a new preconditioning technique for general sparse linear systems based on the ILUS factorization strategy. The resulting preconditioner has the same properties as the ILUS preconditioner. Some theoretical properties of the new preconditioner are discussed and numerical experiments on test matrices from the Harwell- Boeing collection are tested. Our results indicate that the new preconditioner is cheaper to construct than the ILUS preconditioner. AMS Mathematics Subject Classification : 65F10,65F50. Key words and phrases: ILUS factorization, preconditioning, Schur complement, syline format. 1. Introduction Central to many scientific and engineering problems is the solution of large sparse linear system of equations of the form Ax = b, (1) where A is a matrix of dimension N and usually nonsymmetric and unstructured. It is now accepted that, for solving very large sparse linear systems, iterative methods are becoming the method of choice, due to their more favorable memory and computational costs, comparing to the direct solution methods based on Gaussian elimination. One drawbac of many iterative methods is their lac of robustness, i.e., an iterative method may not yield to an acceptable solution for a given problem. A common strategy to enhance the robustness of iterative methods is to exploit preconditioning techniques. However, most Received May 8, Revised October 8, c 2004 Korean Society for Computational & Applied Mathematics and Korean SIGCAM. 299

2 300 Davod Khojasteh Saluyeh and Faezeh Toutounian robust preconditioners are derived from certain type of incomplete LU factorizations of the coefficient matrix. It can be observed experimentally that ILU factorization may produce L and U factors such that the norm (LU) 1 is very large. The long recurrences associated with solving with these factors are unstable [3,6,7], producing solutions with extremely large components. A sign of this severely poor preconditioning is the erratic behavior of the iterative method, for example, divergence of the iterations due to large numerical errors. One possible remedy for determining in advance whether or not a factorization will fail due instability is to estimate a norm of (LU) 1 in some way. In [6], E. Chow and Y. Saad have proposed the ILUS factorization which has many desirable properties, such as its amenability to the Syline Format, the ease with which stability may be monitored, and the possibility of constructing a preconditioner with symmetric structure. In this paper we show that how to use the ILUS preconditioner strategy in bloc form and to construct BILUS factorization which has the same properties as the ILUS preconditioner. Our results indicate that the new preconditioner is cheaper to construct than the ILUS preconditioner. This paper is organized as follows. In section 2, we give a brief description of ILUS preconditioner. In section 3, we introduce a bloc version of incomplete LU preconditioner in sparse syline format and describe some of its theoretical properties. In section 4, BILUS factorization will be applied to the bloc-tridiagonal matrix and we observe, in this case, the BILUS Algorithm is a variant of a general bloc ILU factorization. In section 5, we consider the use of preconditioning on test matrices from the Harwell-Boeing collection. Some concluding remars are given in section ILUS factorization In this section, we briefly review the ILUS factorization proposed in [6]. In ILUS factorization, the sequence of matrices ( ) A v A +1 =, (2) w α +1 where A n = A, is considered. If A is nonsingular and its LDU factorization A = L D U, (3) is already available, then the LDU factorization of A +1 is as following ( ) ( ) ( ) L 0 D 0 U z A +1 =, (4) y 1 0 d in which z = D 1 L 1 v, (5) y = w U 1 D 1, (6) d +1 = α +1 y D z. (7)

3 A bloc version of ILUS factorization 301 Hence, each row and column of the factorization can be obtained by solving two unit lower triangular systems and computing a scaled dot product. Since a sparse approximate solution is often required in the preconditioning, the ILU factorization based on this approach will consist of two approximate sparse linear system solutions and a sparse dot product. There are a number of ways to compute the sparse approximations required in (5) and (6), for example see [5, 8, 11, 12, 13, 14]. One technique of approximation proposed in [6] is to use the truncated Neumann series z = D 1 L 1 v = D 1 (I + E + E E p )v (8) in which E = I L and p is a small natural number. Note that the matrices E are never formed and that the series is evaluated with Horner s rule and that the vector E j v should be computed in sparse-sparse mode. If the number of nonzero elements in z exceeds the fill-in tolerance lfil, then some of the fill-in elements must be dropped according to some strategy. As mentioned in [6], one advantage of the ILUS factorization strategy is that it can estimate L 1 and U 1 easily and determine the stability of the L and U factors. When instability has been detected, e.g., when a norm estimate exceeds some stable limit norm, the ILUS factorization code exits and indicates that the solver should switch to another preconditioner, or restart ILUS with more allowed fill-in, or attempt to use the other strategies which have been described in [6]. The ILUS Algorithm can be summarized as follows. Algorithm 1. ILUS 1. Set D 1 = a 11, L 1 = U 1 = 1 2. For = 1,..., n 1 Do: 3. Compute a sparse z D 1 L 1 v 4. Compute a sparse y w U 1 D 1 5. Compute d +1 := α +1 y D z 6. Form L +1, D +1 and U +1 via (4) 7. Estimate L and U+1 and exit if either exceeds some limit 8. EndDo. 3. BILUS method Let us consider the sequence of matrices ( ) A V A +1 =, = 1,, l 1, (9) W +1

4 302 Davod Khojasteh Saluyeh and Faezeh Toutounian where A l = A, V IR s m, W IR m s and +1 IR m m in which A IR s s and m n. If A is nonsingular and its LDU factorization A = L D U, (10) is already available, then the LDU factorization of A +1 is as following ( ) ( ) ( ) L 0 D 0 U Z A +1 =, (11) Y L 0 D 0 U in which and Y = W U 1 D 1, (12) Z = D 1 L 1 V, (13) L D U = +1 Y D Z R. (14) So, we can obtain the matrices Y and Z by solving two unit lower systems with multiple right hand sides W T and V, respectively. The matrices L, D, and U can be obtained by computing the LDU factorization of R. Hence, by this way, we can obtain the LDU factorization of matrix A if the LDU factorization of all intermediate R exist. For maing an incomplete LDU factorization, which we call BILUS (Bloc version of ILUS factorization), we can solve the systems (12) and (13) approximately. We can also use an incomplete (or exact) LDU factorization algorithm for (14). For example, we can use the Algorithm 1 for this purpose. One of the advantages of BILUS over ILUS is that the computation of the rows of Y and columns of Z can be done in parallel and this can save much of CPU time. Moreover, as the ILUS factorization, the BILUS factorization strategy is able to estimate L 1 and U 1 easily and to determine the stability of L and U factors. If for the lower triangular factor, we use as in [6], the infinity norm bound L 1 e, where e is a vector of all ones, the solution and norm of L 1 +1e may be updated easily from L 1 e. For the upper triangular factor, as in [6], we can use the infinity norm for its transpose which can be estimated easily. As ILUS, when instability has been detected, an appropriate strategy should be used. Numerical results show that the new preconditioner is cheaper to construct than the ILUS preconditioner. Furthermore, the new technique insures convergence rates of the preconditioned iteration which are comparable with those obtained with ILUS preconditioners. A setch of BILUS Algorithm can be written as follows. Algorithm 2. BILUS 1. Set A 1 = the submatrix of A consisting of the first m 1 rows and columns, and compute an approximate (or exact) LDU factorization of A 1, i.e., A 1 L 1 D 1 U 1 (or A 1 = L 1 D 1 U 1 ) 2. For = 1,..., l 1 Do:

5 A bloc version of ILUS factorization Compute a sparse Y W U 1 4. Compute a sparse Z D 1 L 1 D 1 V 5. Set R = +1 Y D Z 6. Compute an incomplete (or exact) LDU factorization of R, i.e., R L D U (or R = L D U ) 7. Form L +1, D +1 and U +1 via (11) 8. Estimate L 1 some limit 9. EndDo. +1 and U 1 +1 and exit if either exceeds 3.1. Analysis In this section we first show that the exact bloc version of LDU factorization as described above exists if A is a SPD matrix or M-matrix. Then, we derive the conditions which guarantee that the matrices R will be nonsingular and the BILUS Algorithm with pivoting will not brea down. Definition 1. A IR N N is an M-matrix if a ij 0 for all i j, A is nonsingular and A 1 0. The following lemmas which have been proved in [1], will be useful for later applications. Lemma 1. Let A be an M-matrix that is partitioned in bloc matrix form, A = (A ij ), where A ii are square matrices. Then the matrices A ii on the diagonal of A are M-matrices. Lemma 2. Let A be an M-matrix that is partitioned in two-by-two bloc form, i.e., ( ) B E A =, (15) F C where B is square matrix. Then the Schur complement exists and itself is an M-matrix. Now we state and prove the following lemma. S = C F B 1 E (16)

6 304 Davod Khojasteh Saluyeh and Faezeh Toutounian Lemma 3. Let A be a SPD matrix (M-matrix ) and all steps of the Algorithm 2 are done exactly. Then, the BILUS Algorithm will not brea down. Proof. By substituting Z and Y in relation (14), we have R = +1 W A 1 V. So, R is the Schur complement of A +1 and by lemma 2 is a SPD matrix (Mmatrix), because A +1 is a leading principal submatrix of A and a SPD matrix (M-matrix). Therefore the LDU factorization of R exists. Now, we consider the existence of the BILUS factorization. As we now when R is nonsingular, it is possible to obtain the LDU factorization of P R for some permutation matrix P. So, it is enough to find some conditions which imply the invertibility of R. Let  = L D U be the incomplete LDU factorization of A which is generated by Algorithm 2. By P and Q, we denote the residuals obtained at steps 3 and 4 of the Algorithm 2, respectively, i.e., P = W Y D U, Q = V L D Z. Hence, from equation (11) we see that the Algorithm 2 (at step ) generates an incomplete LDU factorization of intermediate matrix ( ) ( ) L D Ω +1 U L D Z  = V Q, Y D U +1 W P +1 where Y and Z are obtained at steps 3 and 4 of the Algorithm 2, respectively. Furthermore, we note that R is the Schur complement of Ω +1. So, the invertibility of Ω +1 results the invertibility of R. By defining the matrix E +1 A +1 Ω +1 = ( A  Q P 0 We observe that if A +1 is nonsingular and E +1 < A for some matrix norm., then Ω +1 is nonsingular, too (see [10], page 218). So, if the steps 3 and 4 of the Algorithm 2 are done with enough accuracy and the norm of A  be sufficiently small then R is nonsingular. When R is nonsingular, it is possible to obtain an incomplete or exact LDU factorization of P R for some permutation matrix P. Hence, an incomplete bloc LDU factorization of P +1 A +1 can be obtained for permutation matrix ( ) I 0 P +1 =, 0 P since, P R = P +1 ( P Y )D Z. ).

7 A bloc version of ILUS factorization 305 Here, we note that the dimension of the identity matrix I is the same as the dimension of A. In order to have a nonsingular matrix P +1 Â +1 with enough accuracy, it is better to perform the exact LDU factorization of P R. This computation will not be expensive since the order of matrix R is small (m N). Finally, as we observe, the nonsingularity of the matrices A, = 1,..., l, and the computation with enough accuracy guarantee that the BILUS Algorithm with pivoting will not brea down. 4. BILUS in a special case Consider the bloc-tridiagonal matrix bloced in the form G 1 E 2 F 2 G 2 E 3 A = (17) F l 1 G l 1 E l F l G l Let G be bloc-diagonal matrix consisting of the diagonal blocs G i, L the bloc strictly-lower triangular matrix consisting of the sub-diagonal blocs F i, and U the bloc strictly-upper triangular matrix consisting of the super-diagonal blocs E i. Then, A is of the form A = L + G + U. First, we investigate the exact LDU factorization of A. Consider the sequence of matrices G 1 E 2 F 2 G 2 E 3 A +1 = , = 1,..., l 1, (18) F G E +1 F +1 G +1 with A 1 = G 1. By letting +1 = G +1, W = (0 0 0 F +1 ), V = (0 0 0 E T +1) T, and a little computation we can rewrite (14) as R = L D U = G +1 W A 1 V = G +1 F +1 (L 1 D 1 U 1 ) 1 E +1. (19)

8 306 Davod Khojasteh Saluyeh and Faezeh Toutounian Therefore, by defining Λ 1 = G 1 = L 0 D 0 U 0 and Λ = L 1 D 1 U 1 for = 2,..., l, we have Λ +1 = G +1 F +1 Λ 1 E +1, for = 1,..., l 1. By the above notations and expressions it can be easily seen where A = (L + Λ)Λ 1 (Λ + U), (20) Λ = diag(λ 1,, Λ l ). Relation (20) shows that, by computing some approximation of matrices Λ, = 1,..., l, it is possible to obtain an incomplete bloc LDU factorization of matrix A. Algorithm 3 gives a sparse approximation matrix X of matrix Λ for = 1,..., l. Algorithm Compute an incomplete LDU factorization of G 1, i.e., G 1 L 0 D 0 U 0 = X 1 2. For = 1,..., l 1 Do: 3. Compute a sparse F +1 F +1 U 1 1D Compute a sparse E +1 D 1 1L 1 1E Set X +1 = G +1 F +1 D 1 E Compute an incomplete LDU factorization of X +1, i.e., X +1 L D U 7. EndDo. As we now [2,4], based on the formula (20) a general incomplete LDU factorization for the bloc-tridiagonal matrix (17) taes the following form. Set Z 1 = G 1 and X 1 = approx 2 (Z 1 ). For = 1,..., l 1, compute and let Z +1 = G +1 F +1 approx 1 (Z 1 )E +1, X +1 = approx 2 (Z +1 ). Then the bloc ILU factorization matrix is defined to be where à = (L + Λ) Λ 1 ( Λ + U), Λ = diag(x 1,..., X l ). The role of approx 1 (.) is to control a prespecified sparsity structure of the approximate Z, and the approx 2 (.) is meant to either control a prescribed sparsity

9 A bloc version of ILUS factorization 307 pattern of X and hence mae them easily factored or if the blocs X 1 are explicitly formed, mae their application to a vector easily computed. The main difference between the BILUS factorization for a bloc-tridiagonal matrix and above ILU factorization scheme can be easily seen if we rewrite the Algorithm 3 in the following form. Set Z 1 = G 1 = L 1 D 1 U 1, X 1 = approx 4 (Z 1 ) = approx 4 (L 1 D 1 U 1 ), and for = 1,..., l 1, compute Z +1 = G +1 approx 3 (F +1 X 1 E +1) X +1 = approx 4 (Z +1 ) = approx 4 (L +1 D +1 U +1 ), In this case, the role of approx 4 (.) is to control the prescribed sparsity patterns of L and U, and the approx 3 (.) is as lie as approx 1 (.), to control a prescribed sparsity pattern of the approximate of Z. As we observed, for a bloc tridiagonal matrix A, the above Algorithm is a variant of BILUS Algorithm. 5. Numerical examples For one of our experiments we consider the equation u + x (ex u) xu = f(x, y), (x, y) Ω = (0, 1) (0, 1) (21) Discretizing (21) on an n x n y grid, by using the second order centered differences for the Laplacian and centered difference for x, gives a linear system of equations of order N = n x n y. In our test, we tae n x = n y = 32, so this yields a matrix of order N = 1024 which is an M-matrix [9]. The boundary conditions are taen so that the exact solution of the system is x = [1,, 1] T. We name this example by F2DA. Also, we use some matrices from Harwell-Boeing collection. These matrices with their properties are shown in Table 1. Table 1 matrix name order symmetric positive definite NOS3 960 yes yes NOS5 468 yes yes GR yes yes FIDAP yes no SHERMAN no no CANITY no no CANITY no no CANITY no no CANITY no no

10 308 Davod Khojasteh Saluyeh and Faezeh Toutounian For these examples, the right hand side of the systems is taen such that the exact solution is x = [1,, 1] T. The split preconditioned conjugate gradient method and left preconditioned GMRES(10) method are used for solving the systems. For all the examples we used the stopping criterion b Ax i For preserving sparsity we used two following strategies[6,13]. 1. The truncated Neumann series described in section 2 for a small p. 2. Dropping entries of z i, columns of Z i, (y j, rows of Y j ) (i.e., replaced by zero) which are less than the relative tolerance τ i (γ j ) obtained by multiplying τ by the original 2-norm of the i-th row (j-th column) of A corresponding to z i (y j ). Numerical results are shown in Tables 2 and 3 without running the steps 3 and 4 of BILUS Algorithm in parallel. For each example, the system (1) was solved for various m. Column 4 represents the number of iterations. Columns 5 and 6 present the CPU time for computing the preconditioner M = LDU and solving the systems with preconditioning, respectively. The total of CPU times are shown in column 7. Note that the case m = 1 corresponds to the ILUS method. Finally, PCG and PGMRES stand for Preconditioned CG [13] and GMRES methods [13,15], respectively. Table 2 PCG CPU time(s) matrix p, τ m iterations precon. solve total NOS3 2, NOS5 2, GR , The columns 4 and 6 of Tables 2 and 3 show that the results of BILUS factorization is as good as ILUS factorization. The column 5 of the Tables show that the CPU time of maing BILUS preconditioner is much smaller than that of maing ILUS preconditioner and decreases more slowly with increasing m. The columns 7 show that the total CPU time of BILUS is always much less than that of ILUS preconditioner and decreases more slowly with increase m. The results for CAVITY8 (Table 3 sign ) show that for m = 1, 6 we have no solution after 500 iterations but for m = 12 we have the solution after 15 iterations. In fact for m = 1, 6 an instability has been detected since L 1 e is equal to , , for m = 1, 6, 12, respectively. Figure 1 shows

11 A bloc version of ILUS factorization 309 the required CPU time for computing three preconditioners as a function of m for matrices SHERMAN4, F2DA and CAVITY05 and it shows that the rate of decrease slows down after m sufficiently increases. So we can conclude that the BILUS Algorithm is a robust technique for constructing a good preconditioner. The column 7 show that the total CPU time of BILUS is always much less than that of ILUS preconditioner and decreases more slowly with increase m. Table 3 PGMRES(10) CPU time(s) matrix p, τ m iterations precon. solve total FIDAP001 2, SHERMAN4 1, F2DA 1, CAVITY05 3, CAVITY06 3, CAVITY07 3, CAVITY08 3, The results for CAVITY8 (Table 3 sign ) show that for m = 1, 6 we have no solution after 500 iterations but for m = 12 we have the solution after 15 iterations. In fact for m = 1, 6 an instability has been detected since L 1 e is equal to , , for m = 1, 6, 12, respectively. Figure 1 shows the required CPU time for computing three preconditioners as a function of m for matrices SHERMAN4, F2DA and CAVITY05 and it shows that the rate of decrease slows down after m sufficiently increases. So we can conclude that the BILUS Algorithm is a robust technique for constructing a good preconditioner.

12 310 Davod Khojasteh Saluyeh and Faezeh Toutounian Figure 1. CPU times for computing three preconditioners as a function of m for matrices SHERMAN4, F2DA and CAVITY05 6. Conclusion We have proposed a new preconditioner for general sparse linear systems based on the ILUS factorization strategy. The constructed preconditioner has the same properties as the ILUS preconditioner. It can estimate L 1 and U 1 easily and to determine the stability of the L and U. One of the advantages of BILUS over ILUS is that the computations can be done in parallel and this can save the CPU time. Our results indicated that the new preconditioner is cheaper to construct than the ILUS preconditioner and, in addition, it retains the efficiency and robustness of ILUS preconditioner. 7. Acnowledgements The authors are grateful to referee for his/her comments which substantially improved the quality of this paper. References [1] O. Axelsson, Iterative solution methods, Cambridge University Press, Cambridge, [2] O. Axelsson and B. Polman, On approximate factorization methods for bloc matrices suitable for vector and parallel processors, Numerical Linear Algebra with applications, Vol. 77(1986), [3] A. M. Bruaset, A. Tveito and R. Winther, On the stability of relaxed incomplete LU factorization, Math. Comp., Vol. 54(1990), [4] T. F. Chan, P. S. Vassilevsi, A framewor for bloc ILU factorizations using bloc-size reduction, Mathematics of Computation, Vol. 64(1995), [5] E. Chow and Y. Saad, Approximate inverse preconditioners via sparse-sparse iterations, SIAM J. Sci, Comput., Vol. 19(1998), , [6] E. Chow and Y. Saad, ILUS: an incomplete LU preconditioner in sparse syline format, International Journal of Numerical Methods in Fluids, Vol. 25(1997), [7] H. C. Elman, A stability analysis of incomplete LU factorization, Math. Comp., Vol. 47(1986), [8] M. J. Grote and T. Hucle, Parallel preconditioning with sparse approximate inverses, SIAM J. Sci. Comput., Vol. 18(1997), [9] A. Kerayechian, D. Khojasteh Saluyeh, On the existence, uniqueness and approximation of a class of elliptic problems, International Journal of Applied Mathematics, Vol. 11(2002), No.1, [10] D. Kincaid, W. Cheney,Numerical Analysis, Broos/Cole Publishing Company, [11] L. Y. Kolotilina and A. Y. Yeremin, Factorized sparse approximate inverse preconditioning I. Theory, SIAM J. Matrix Anal. apll., 14(1993),

13 A bloc version of ILUS factorization 311 [12] Y. Saad, ILUT: a dual thereshold incomplete LU factorization, Numerical Linear Algebra with applications, 1(1994), [13] Y. Saad, Iterative Methods for Sparse linear Systems, PWS press, New Yor, [14] Y. Saad, Preconditioned Krylov subspace methods for CFD applications, in : W. G. Habashi, ed., solution techniques for Large-Scale CFD problems, Wiley, New Yor, [15] Y. Saad and M. H. Schultz, GMRES, A generalized minimal residual algorithm for nonsymmetric linear systems, SIAM J. Sci. Statist. Comput., 7(1986), Davod Khojasteh Saluyeh received his B.Sc from Sharif University of Technology, Tehran, Iran and his M.Sc from Ferdowsi University of Mashhad, Mashhad, Iran. At present he is writing his Ph.D thesis under supervision of professor Faezeh Toutounian at Ferdowsi University of Mashhad. His research interests are mainly iterative methods for sparse linear systems and finite element method. Department of Mathematics, School of Mathematical Sciences, Ferdowsi University of Mashhad, P.O. Box , Mashhad, Iran. hojaste@math.um.ac.ir Faezeh Toutounian received her B.Sc in Mathematics from Ferdowsi University of Mashhad, Iran, two degree of M.Sc in Mathematical statistics and applied computer and her Ph.D in Mathematics from Paris VI University, France. She spent two sabbatical years in 1985 and 1996 at Paris VI University. She is currently a Professor of Mathematics at Ferdowsi University of mashhad. Her research interests are mainly numerical linear algebra, iterative methods and error analysis. Department of Mathematics, School of Mathematical Sciences, Ferdowsi University of Mashhad, P.O. Box , Mashhad, Iran. toutouni@math.um.ac.ir

On the Preconditioning of the Block Tridiagonal Linear System of Equations

On the Preconditioning of the Block Tridiagonal Linear System of Equations On the Preconditioning of the Block Tridiagonal Linear System of Equations Davod Khojasteh Salkuyeh Department of Mathematics, University of Mohaghegh Ardabili, PO Box 179, Ardabil, Iran E-mail: khojaste@umaacir

More information

A SPARSE APPROXIMATE INVERSE PRECONDITIONER FOR NONSYMMETRIC LINEAR SYSTEMS

A SPARSE APPROXIMATE INVERSE PRECONDITIONER FOR NONSYMMETRIC LINEAR SYSTEMS INTERNATIONAL JOURNAL OF NUMERICAL ANALYSIS AND MODELING, SERIES B Volume 5, Number 1-2, Pages 21 30 c 2014 Institute for Scientific Computing and Information A SPARSE APPROXIMATE INVERSE PRECONDITIONER

More information

Jae Heon Yun and Yu Du Han

Jae Heon Yun and Yu Du Han Bull. Korean Math. Soc. 39 (2002), No. 3, pp. 495 509 MODIFIED INCOMPLETE CHOLESKY FACTORIZATION PRECONDITIONERS FOR A SYMMETRIC POSITIVE DEFINITE MATRIX Jae Heon Yun and Yu Du Han Abstract. We propose

More information

Solving Large Nonlinear Sparse Systems

Solving Large Nonlinear Sparse Systems Solving Large Nonlinear Sparse Systems Fred W. Wubs and Jonas Thies Computational Mechanics & Numerical Mathematics University of Groningen, the Netherlands f.w.wubs@rug.nl Centre for Interdisciplinary

More information

Incomplete LU Preconditioning and Error Compensation Strategies for Sparse Matrices

Incomplete LU Preconditioning and Error Compensation Strategies for Sparse Matrices Incomplete LU Preconditioning and Error Compensation Strategies for Sparse Matrices Eun-Joo Lee Department of Computer Science, East Stroudsburg University of Pennsylvania, 327 Science and Technology Center,

More information

A Method for Constructing Diagonally Dominant Preconditioners based on Jacobi Rotations

A Method for Constructing Diagonally Dominant Preconditioners based on Jacobi Rotations A Method for Constructing Diagonally Dominant Preconditioners based on Jacobi Rotations Jin Yun Yuan Plamen Y. Yalamov Abstract A method is presented to make a given matrix strictly diagonally dominant

More information

9.1 Preconditioned Krylov Subspace Methods

9.1 Preconditioned Krylov Subspace Methods Chapter 9 PRECONDITIONING 9.1 Preconditioned Krylov Subspace Methods 9.2 Preconditioned Conjugate Gradient 9.3 Preconditioned Generalized Minimal Residual 9.4 Relaxation Method Preconditioners 9.5 Incomplete

More information

Preface to the Second Edition. Preface to the First Edition

Preface to the Second Edition. Preface to the First Edition n page v Preface to the Second Edition Preface to the First Edition xiii xvii 1 Background in Linear Algebra 1 1.1 Matrices................................. 1 1.2 Square Matrices and Eigenvalues....................

More information

ITERATIVE METHODS FOR SPARSE LINEAR SYSTEMS

ITERATIVE METHODS FOR SPARSE LINEAR SYSTEMS ITERATIVE METHODS FOR SPARSE LINEAR SYSTEMS YOUSEF SAAD University of Minnesota PWS PUBLISHING COMPANY I(T)P An International Thomson Publishing Company BOSTON ALBANY BONN CINCINNATI DETROIT LONDON MADRID

More information

Jordan Journal of Mathematics and Statistics (JJMS) 5(3), 2012, pp A NEW ITERATIVE METHOD FOR SOLVING LINEAR SYSTEMS OF EQUATIONS

Jordan Journal of Mathematics and Statistics (JJMS) 5(3), 2012, pp A NEW ITERATIVE METHOD FOR SOLVING LINEAR SYSTEMS OF EQUATIONS Jordan Journal of Mathematics and Statistics JJMS) 53), 2012, pp.169-184 A NEW ITERATIVE METHOD FOR SOLVING LINEAR SYSTEMS OF EQUATIONS ADEL H. AL-RABTAH Abstract. The Jacobi and Gauss-Seidel iterative

More information

AMS526: Numerical Analysis I (Numerical Linear Algebra) Lecture 23: GMRES and Other Krylov Subspace Methods; Preconditioning

AMS526: Numerical Analysis I (Numerical Linear Algebra) Lecture 23: GMRES and Other Krylov Subspace Methods; Preconditioning AMS526: Numerical Analysis I (Numerical Linear Algebra) Lecture 23: GMRES and Other Krylov Subspace Methods; Preconditioning Xiangmin Jiao SUNY Stony Brook Xiangmin Jiao Numerical Analysis I 1 / 18 Outline

More information

Preconditioning Techniques for Large Linear Systems Part III: General-Purpose Algebraic Preconditioners

Preconditioning Techniques for Large Linear Systems Part III: General-Purpose Algebraic Preconditioners Preconditioning Techniques for Large Linear Systems Part III: General-Purpose Algebraic Preconditioners Michele Benzi Department of Mathematics and Computer Science Emory University Atlanta, Georgia, USA

More information

INCREMENTAL INCOMPLETE LU FACTORIZATIONS WITH APPLICATIONS TO TIME-DEPENDENT PDES

INCREMENTAL INCOMPLETE LU FACTORIZATIONS WITH APPLICATIONS TO TIME-DEPENDENT PDES INCREMENTAL INCOMPLETE LU FACTORIZATIONS WITH APPLICATIONS TO TIME-DEPENDENT PDES C. CALGARO, J. P. CHEHAB, AND Y. SAAD Abstract. This paper addresses the problem of computing preconditioners for solving

More information

A generalization of the Gauss-Seidel iteration method for solving absolute value equations

A generalization of the Gauss-Seidel iteration method for solving absolute value equations A generalization of the Gauss-Seidel iteration method for solving absolute value equations Vahid Edalatpour, Davod Hezari and Davod Khojasteh Salkuyeh Faculty of Mathematical Sciences, University of Guilan,

More information

The solution of the discretized incompressible Navier-Stokes equations with iterative methods

The solution of the discretized incompressible Navier-Stokes equations with iterative methods The solution of the discretized incompressible Navier-Stokes equations with iterative methods Report 93-54 C. Vuik Technische Universiteit Delft Delft University of Technology Faculteit der Technische

More information

Optimal Iterate of the Power and Inverse Iteration Methods

Optimal Iterate of the Power and Inverse Iteration Methods Optimal Iterate of the Power and Inverse Iteration Methods Davod Khojasteh Salkuyeh and Faezeh Toutounian Department of Mathematics, University of Mohaghegh Ardabili, P.O. Box. 79, Ardabil, Iran E-mail:

More information

Linear Solvers. Andrew Hazel

Linear Solvers. Andrew Hazel Linear Solvers Andrew Hazel Introduction Thus far we have talked about the formulation and discretisation of physical problems...... and stopped when we got to a discrete linear system of equations. Introduction

More information

Multilevel Preconditioning of Graph-Laplacians: Polynomial Approximation of the Pivot Blocks Inverses

Multilevel Preconditioning of Graph-Laplacians: Polynomial Approximation of the Pivot Blocks Inverses Multilevel Preconditioning of Graph-Laplacians: Polynomial Approximation of the Pivot Blocks Inverses P. Boyanova 1, I. Georgiev 34, S. Margenov, L. Zikatanov 5 1 Uppsala University, Box 337, 751 05 Uppsala,

More information

A Robust Preconditioned Iterative Method for the Navier-Stokes Equations with High Reynolds Numbers

A Robust Preconditioned Iterative Method for the Navier-Stokes Equations with High Reynolds Numbers Applied and Computational Mathematics 2017; 6(4): 202-207 http://www.sciencepublishinggroup.com/j/acm doi: 10.11648/j.acm.20170604.18 ISSN: 2328-5605 (Print); ISSN: 2328-5613 (Online) A Robust Preconditioned

More information

Generalized AOR Method for Solving System of Linear Equations. Davod Khojasteh Salkuyeh. Department of Mathematics, University of Mohaghegh Ardabili,

Generalized AOR Method for Solving System of Linear Equations. Davod Khojasteh Salkuyeh. Department of Mathematics, University of Mohaghegh Ardabili, Australian Journal of Basic and Applied Sciences, 5(3): 35-358, 20 ISSN 99-878 Generalized AOR Method for Solving Syste of Linear Equations Davod Khojasteh Salkuyeh Departent of Matheatics, University

More information

Numerical Methods I Non-Square and Sparse Linear Systems

Numerical Methods I Non-Square and Sparse Linear Systems Numerical Methods I Non-Square and Sparse Linear Systems Aleksandar Donev Courant Institute, NYU 1 donev@courant.nyu.edu 1 MATH-GA 2011.003 / CSCI-GA 2945.003, Fall 2014 September 25th, 2014 A. Donev (Courant

More information

Preconditioning Techniques Analysis for CG Method

Preconditioning Techniques Analysis for CG Method Preconditioning Techniques Analysis for CG Method Huaguang Song Department of Computer Science University of California, Davis hso@ucdavis.edu Abstract Matrix computation issue for solve linear system

More information

ON THE GLOBAL KRYLOV SUBSPACE METHODS FOR SOLVING GENERAL COUPLED MATRIX EQUATIONS

ON THE GLOBAL KRYLOV SUBSPACE METHODS FOR SOLVING GENERAL COUPLED MATRIX EQUATIONS ON THE GLOBAL KRYLOV SUBSPACE METHODS FOR SOLVING GENERAL COUPLED MATRIX EQUATIONS Fatemeh Panjeh Ali Beik and Davod Khojasteh Salkuyeh, Department of Mathematics, Vali-e-Asr University of Rafsanjan, Rafsanjan,

More information

The flexible incomplete LU preconditioner for large nonsymmetric linear systems. Takatoshi Nakamura Takashi Nodera

The flexible incomplete LU preconditioner for large nonsymmetric linear systems. Takatoshi Nakamura Takashi Nodera Research Report KSTS/RR-15/006 The flexible incomplete LU preconditioner for large nonsymmetric linear systems by Takatoshi Nakamura Takashi Nodera Takatoshi Nakamura School of Fundamental Science and

More information

Incomplete Cholesky preconditioners that exploit the low-rank property

Incomplete Cholesky preconditioners that exploit the low-rank property anapov@ulb.ac.be ; http://homepages.ulb.ac.be/ anapov/ 1 / 35 Incomplete Cholesky preconditioners that exploit the low-rank property (theory and practice) Artem Napov Service de Métrologie Nucléaire, Université

More information

A robust multilevel approximate inverse preconditioner for symmetric positive definite matrices

A robust multilevel approximate inverse preconditioner for symmetric positive definite matrices DICEA DEPARTMENT OF CIVIL, ENVIRONMENTAL AND ARCHITECTURAL ENGINEERING PhD SCHOOL CIVIL AND ENVIRONMENTAL ENGINEERING SCIENCES XXX CYCLE A robust multilevel approximate inverse preconditioner for symmetric

More information

4.8 Arnoldi Iteration, Krylov Subspaces and GMRES

4.8 Arnoldi Iteration, Krylov Subspaces and GMRES 48 Arnoldi Iteration, Krylov Subspaces and GMRES We start with the problem of using a similarity transformation to convert an n n matrix A to upper Hessenberg form H, ie, A = QHQ, (30) with an appropriate

More information

M.A. Botchev. September 5, 2014

M.A. Botchev. September 5, 2014 Rome-Moscow school of Matrix Methods and Applied Linear Algebra 2014 A short introduction to Krylov subspaces for linear systems, matrix functions and inexact Newton methods. Plan and exercises. M.A. Botchev

More information

Chapter 7 Iterative Techniques in Matrix Algebra

Chapter 7 Iterative Techniques in Matrix Algebra Chapter 7 Iterative Techniques in Matrix Algebra Per-Olof Persson persson@berkeley.edu Department of Mathematics University of California, Berkeley Math 128B Numerical Analysis Vector Norms Definition

More information

Iterative Methods for Solving A x = b

Iterative Methods for Solving A x = b Iterative Methods for Solving A x = b A good (free) online source for iterative methods for solving A x = b is given in the description of a set of iterative solvers called templates found at netlib: http

More information

Parallelization of Multilevel Preconditioners Constructed from Inverse-Based ILUs on Shared-Memory Multiprocessors

Parallelization of Multilevel Preconditioners Constructed from Inverse-Based ILUs on Shared-Memory Multiprocessors Parallelization of Multilevel Preconditioners Constructed from Inverse-Based ILUs on Shared-Memory Multiprocessors J.I. Aliaga 1 M. Bollhöfer 2 A.F. Martín 1 E.S. Quintana-Ortí 1 1 Deparment of Computer

More information

Multilevel low-rank approximation preconditioners Yousef Saad Department of Computer Science and Engineering University of Minnesota

Multilevel low-rank approximation preconditioners Yousef Saad Department of Computer Science and Engineering University of Minnesota Multilevel low-rank approximation preconditioners Yousef Saad Department of Computer Science and Engineering University of Minnesota SIAM CSE Boston - March 1, 2013 First: Joint work with Ruipeng Li Work

More information

Scientific Computing

Scientific Computing Scientific Computing Direct solution methods Martin van Gijzen Delft University of Technology October 3, 2018 1 Program October 3 Matrix norms LU decomposition Basic algorithm Cost Stability Pivoting Pivoting

More information

OUTLINE ffl CFD: elliptic pde's! Ax = b ffl Basic iterative methods ffl Krylov subspace methods ffl Preconditioning techniques: Iterative methods ILU

OUTLINE ffl CFD: elliptic pde's! Ax = b ffl Basic iterative methods ffl Krylov subspace methods ffl Preconditioning techniques: Iterative methods ILU Preconditioning Techniques for Solving Large Sparse Linear Systems Arnold Reusken Institut für Geometrie und Praktische Mathematik RWTH-Aachen OUTLINE ffl CFD: elliptic pde's! Ax = b ffl Basic iterative

More information

AN ITERATIVE METHOD TO SOLVE SYMMETRIC POSITIVE DEFINITE MATRIX EQUATIONS

AN ITERATIVE METHOD TO SOLVE SYMMETRIC POSITIVE DEFINITE MATRIX EQUATIONS AN ITERATIVE METHOD TO SOLVE SYMMETRIC POSITIVE DEFINITE MATRIX EQUATIONS DAVOD KHOJASTEH SALKUYEH and FATEMEH PANJEH ALI BEIK Communicated by the former editorial board Let A : R m n R m n be a symmetric

More information

ANONSINGULAR tridiagonal linear system of the form

ANONSINGULAR tridiagonal linear system of the form Generalized Diagonal Pivoting Methods for Tridiagonal Systems without Interchanges Jennifer B. Erway, Roummel F. Marcia, and Joseph A. Tyson Abstract It has been shown that a nonsingular symmetric tridiagonal

More information

In order to solve the linear system KL M N when K is nonsymmetric, we can solve the equivalent system

In order to solve the linear system KL M N when K is nonsymmetric, we can solve the equivalent system !"#$% "&!#' (%)!#" *# %)%(! #! %)!#" +, %"!"#$ %*&%! $#&*! *# %)%! -. -/ 0 -. 12 "**3! * $!#%+,!2!#% 44" #% &#33 # 4"!#" "%! "5"#!!#6 -. - #% " 7% "3#!#3! - + 87&2! * $!#% 44" ) 3( $! # % %#!!#%+ 9332!

More information

Iterative Methods for Sparse Linear Systems

Iterative Methods for Sparse Linear Systems Iterative Methods for Sparse Linear Systems Luca Bergamaschi e-mail: berga@dmsa.unipd.it - http://www.dmsa.unipd.it/ berga Department of Mathematical Methods and Models for Scientific Applications University

More information

BLOCK ILU PRECONDITIONED ITERATIVE METHODS FOR REDUCED LINEAR SYSTEMS

BLOCK ILU PRECONDITIONED ITERATIVE METHODS FOR REDUCED LINEAR SYSTEMS CANADIAN APPLIED MATHEMATICS QUARTERLY Volume 12, Number 3, Fall 2004 BLOCK ILU PRECONDITIONED ITERATIVE METHODS FOR REDUCED LINEAR SYSTEMS N GUESSOUS AND O SOUHAR ABSTRACT This paper deals with the iterative

More information

Linear Algebra Section 2.6 : LU Decomposition Section 2.7 : Permutations and transposes Wednesday, February 13th Math 301 Week #4

Linear Algebra Section 2.6 : LU Decomposition Section 2.7 : Permutations and transposes Wednesday, February 13th Math 301 Week #4 Linear Algebra Section. : LU Decomposition Section. : Permutations and transposes Wednesday, February 1th Math 01 Week # 1 The LU Decomposition We learned last time that we can factor a invertible matrix

More information

Robust Preconditioned Conjugate Gradient for the GPU and Parallel Implementations

Robust Preconditioned Conjugate Gradient for the GPU and Parallel Implementations Robust Preconditioned Conjugate Gradient for the GPU and Parallel Implementations Rohit Gupta, Martin van Gijzen, Kees Vuik GPU Technology Conference 2012, San Jose CA. GPU Technology Conference 2012,

More information

The amount of work to construct each new guess from the previous one should be a small multiple of the number of nonzeros in A.

The amount of work to construct each new guess from the previous one should be a small multiple of the number of nonzeros in A. AMSC/CMSC 661 Scientific Computing II Spring 2005 Solution of Sparse Linear Systems Part 2: Iterative methods Dianne P. O Leary c 2005 Solving Sparse Linear Systems: Iterative methods The plan: Iterative

More information

FEM and sparse linear system solving

FEM and sparse linear system solving FEM & sparse linear system solving, Lecture 9, Nov 19, 2017 1/36 Lecture 9, Nov 17, 2017: Krylov space methods http://people.inf.ethz.ch/arbenz/fem17 Peter Arbenz Computer Science Department, ETH Zürich

More information

A Block Compression Algorithm for Computing Preconditioners

A Block Compression Algorithm for Computing Preconditioners A Block Compression Algorithm for Computing Preconditioners J Cerdán, J Marín, and J Mas Abstract To implement efficiently algorithms for the solution of large systems of linear equations in modern computer

More information

Scientific Computing WS 2018/2019. Lecture 9. Jürgen Fuhrmann Lecture 9 Slide 1

Scientific Computing WS 2018/2019. Lecture 9. Jürgen Fuhrmann Lecture 9 Slide 1 Scientific Computing WS 2018/2019 Lecture 9 Jürgen Fuhrmann juergen.fuhrmann@wias-berlin.de Lecture 9 Slide 1 Lecture 9 Slide 2 Simple iteration with preconditioning Idea: Aû = b iterative scheme û = û

More information

ON A SPLITTING PRECONDITIONER FOR SADDLE POINT PROBLEMS

ON A SPLITTING PRECONDITIONER FOR SADDLE POINT PROBLEMS J. Appl. Math. & Informatics Vol. 36(208, No. 5-6, pp. 459-474 https://doi.org/0.437/jami.208.459 ON A SPLITTING PRECONDITIONER FOR SADDLE POINT PROBLEMS DAVOD KHOJASTEH SALKUYEH, MARYAM ABDOLMALEKI, SAEED

More information

1 Multiply Eq. E i by λ 0: (λe i ) (E i ) 2 Multiply Eq. E j by λ and add to Eq. E i : (E i + λe j ) (E i )

1 Multiply Eq. E i by λ 0: (λe i ) (E i ) 2 Multiply Eq. E j by λ and add to Eq. E i : (E i + λe j ) (E i ) Direct Methods for Linear Systems Chapter Direct Methods for Solving Linear Systems Per-Olof Persson persson@berkeleyedu Department of Mathematics University of California, Berkeley Math 18A Numerical

More information

ENHANCING PERFORMANCE AND ROBUSTNESS OF ILU PRECONDITIONERS BY BLOCKING AND SELECTIVE TRANSPOSITION

ENHANCING PERFORMANCE AND ROBUSTNESS OF ILU PRECONDITIONERS BY BLOCKING AND SELECTIVE TRANSPOSITION SIAM J. SCI. COMPUT. Vol. 39, No., pp. A303 A332 c 207 Society for Industrial and Applied Mathematics ENHANCING PERFORMANCE AND ROBUSTNESS OF ILU PRECONDITIONERS BY BLOCKING AND SELECTIVE TRANSPOSITION

More information

Iterative Methods. Splitting Methods

Iterative Methods. Splitting Methods Iterative Methods Splitting Methods 1 Direct Methods Solving Ax = b using direct methods. Gaussian elimination (using LU decomposition) Variants of LU, including Crout and Doolittle Other decomposition

More information

ON THE GENERALIZED DETERIORATED POSITIVE SEMI-DEFINITE AND SKEW-HERMITIAN SPLITTING PRECONDITIONER *

ON THE GENERALIZED DETERIORATED POSITIVE SEMI-DEFINITE AND SKEW-HERMITIAN SPLITTING PRECONDITIONER * Journal of Computational Mathematics Vol.xx, No.x, 2x, 6. http://www.global-sci.org/jcm doi:?? ON THE GENERALIZED DETERIORATED POSITIVE SEMI-DEFINITE AND SKEW-HERMITIAN SPLITTING PRECONDITIONER * Davod

More information

Lecture 18 Classical Iterative Methods

Lecture 18 Classical Iterative Methods Lecture 18 Classical Iterative Methods MIT 18.335J / 6.337J Introduction to Numerical Methods Per-Olof Persson November 14, 2006 1 Iterative Methods for Linear Systems Direct methods for solving Ax = b,

More information

Solving PDEs with CUDA Jonathan Cohen

Solving PDEs with CUDA Jonathan Cohen Solving PDEs with CUDA Jonathan Cohen jocohen@nvidia.com NVIDIA Research PDEs (Partial Differential Equations) Big topic Some common strategies Focus on one type of PDE in this talk Poisson Equation Linear

More information

PROJECTED GMRES AND ITS VARIANTS

PROJECTED GMRES AND ITS VARIANTS PROJECTED GMRES AND ITS VARIANTS Reinaldo Astudillo Brígida Molina rastudillo@kuaimare.ciens.ucv.ve bmolina@kuaimare.ciens.ucv.ve Centro de Cálculo Científico y Tecnológico (CCCT), Facultad de Ciencias,

More information

A short course on: Preconditioned Krylov subspace methods. Yousef Saad University of Minnesota Dept. of Computer Science and Engineering

A short course on: Preconditioned Krylov subspace methods. Yousef Saad University of Minnesota Dept. of Computer Science and Engineering A short course on: Preconditioned Krylov subspace methods Yousef Saad University of Minnesota Dept. of Computer Science and Engineering Universite du Littoral, Jan 19-3, 25 Outline Part 1 Introd., discretization

More information

Efficient smoothers for all-at-once multigrid methods for Poisson and Stokes control problems

Efficient smoothers for all-at-once multigrid methods for Poisson and Stokes control problems Efficient smoothers for all-at-once multigrid methods for Poisson and Stoes control problems Stefan Taacs stefan.taacs@numa.uni-linz.ac.at, WWW home page: http://www.numa.uni-linz.ac.at/~stefant/j3362/

More information

Fine-grained Parallel Incomplete LU Factorization

Fine-grained Parallel Incomplete LU Factorization Fine-grained Parallel Incomplete LU Factorization Edmond Chow School of Computational Science and Engineering Georgia Institute of Technology Sparse Days Meeting at CERFACS June 5-6, 2014 Contribution

More information

AMS526: Numerical Analysis I (Numerical Linear Algebra for Computational and Data Sciences)

AMS526: Numerical Analysis I (Numerical Linear Algebra for Computational and Data Sciences) AMS526: Numerical Analysis I (Numerical Linear Algebra for Computational and Data Sciences) Lecture 19: Computing the SVD; Sparse Linear Systems Xiangmin Jiao Stony Brook University Xiangmin Jiao Numerical

More information

Structure preserving preconditioner for the incompressible Navier-Stokes equations

Structure preserving preconditioner for the incompressible Navier-Stokes equations Structure preserving preconditioner for the incompressible Navier-Stokes equations Fred W. Wubs and Jonas Thies Computational Mechanics & Numerical Mathematics University of Groningen, the Netherlands

More information

Stabilization and Acceleration of Algebraic Multigrid Method

Stabilization and Acceleration of Algebraic Multigrid Method Stabilization and Acceleration of Algebraic Multigrid Method Recursive Projection Algorithm A. Jemcov J.P. Maruszewski Fluent Inc. October 24, 2006 Outline 1 Need for Algorithm Stabilization and Acceleration

More information

Index. book 2009/5/27 page 121. (Page numbers set in bold type indicate the definition of an entry.)

Index. book 2009/5/27 page 121. (Page numbers set in bold type indicate the definition of an entry.) page 121 Index (Page numbers set in bold type indicate the definition of an entry.) A absolute error...26 componentwise...31 in subtraction...27 normwise...31 angle in least squares problem...98,99 approximation

More information

CONVERGENCE OF MULTISPLITTING METHOD FOR A SYMMETRIC POSITIVE DEFINITE MATRIX

CONVERGENCE OF MULTISPLITTING METHOD FOR A SYMMETRIC POSITIVE DEFINITE MATRIX J. Appl. Math. & Computing Vol. 182005 No. 1-2 pp. 59-72 CONVERGENCE OF MULTISPLITTING METHOD FOR A SYMMETRIC POSITIVE DEFINITE MATRIX JAE HEON YUN SEYOUNG OH AND EUN HEUI KIM Abstract. We study convergence

More information

Lab 1: Iterative Methods for Solving Linear Systems

Lab 1: Iterative Methods for Solving Linear Systems Lab 1: Iterative Methods for Solving Linear Systems January 22, 2017 Introduction Many real world applications require the solution to very large and sparse linear systems where direct methods such as

More information

Nested splitting CG-like iterative method for solving the continuous Sylvester equation and preconditioning

Nested splitting CG-like iterative method for solving the continuous Sylvester equation and preconditioning Adv Comput Math DOI 10.1007/s10444-013-9330-3 Nested splitting CG-like iterative method for solving the continuous Sylvester equation and preconditioning Mohammad Khorsand Zak Faezeh Toutounian Received:

More information

The Conjugate Gradient Method

The Conjugate Gradient Method The Conjugate Gradient Method Classical Iterations We have a problem, We assume that the matrix comes from a discretization of a PDE. The best and most popular model problem is, The matrix will be as large

More information

Multigrid absolute value preconditioning

Multigrid absolute value preconditioning Multigrid absolute value preconditioning Eugene Vecharynski 1 Andrew Knyazev 2 (speaker) 1 Department of Computer Science and Engineering University of Minnesota 2 Department of Mathematical and Statistical

More information

Parallel Iterative Methods for Sparse Linear Systems. H. Martin Bücker Lehrstuhl für Hochleistungsrechnen

Parallel Iterative Methods for Sparse Linear Systems. H. Martin Bücker Lehrstuhl für Hochleistungsrechnen Parallel Iterative Methods for Sparse Linear Systems Lehrstuhl für Hochleistungsrechnen www.sc.rwth-aachen.de RWTH Aachen Large and Sparse Small and Dense Outline Problem with Direct Methods Iterative

More information

Incomplete factorization preconditioners and their updates with applications - I 1,2

Incomplete factorization preconditioners and their updates with applications - I 1,2 Incomplete factorization preconditioners and their updates with applications - I 1,2 Daniele Bertaccini, Fabio Durastante Moscow August 24, 216 Notes of the course: Incomplete factorization preconditioners

More information

Algebraic Multigrid as Solvers and as Preconditioner

Algebraic Multigrid as Solvers and as Preconditioner Ò Algebraic Multigrid as Solvers and as Preconditioner Domenico Lahaye domenico.lahaye@cs.kuleuven.ac.be http://www.cs.kuleuven.ac.be/ domenico/ Department of Computer Science Katholieke Universiteit Leuven

More information

Section 5.6. LU and LDU Factorizations

Section 5.6. LU and LDU Factorizations 5.6. LU and LDU Factorizations Section 5.6. LU and LDU Factorizations Note. We largely follow Fraleigh and Beauregard s approach to this topic from Linear Algebra, 3rd Edition, Addison-Wesley (995). See

More information

6.4 Krylov Subspaces and Conjugate Gradients

6.4 Krylov Subspaces and Conjugate Gradients 6.4 Krylov Subspaces and Conjugate Gradients Our original equation is Ax = b. The preconditioned equation is P Ax = P b. When we write P, we never intend that an inverse will be explicitly computed. P

More information

Preliminary/Qualifying Exam in Numerical Analysis (Math 502a) Spring 2012

Preliminary/Qualifying Exam in Numerical Analysis (Math 502a) Spring 2012 Instructions Preliminary/Qualifying Exam in Numerical Analysis (Math 502a) Spring 2012 The exam consists of four problems, each having multiple parts. You should attempt to solve all four problems. 1.

More information

Solving Symmetric Indefinite Systems with Symmetric Positive Definite Preconditioners

Solving Symmetric Indefinite Systems with Symmetric Positive Definite Preconditioners Solving Symmetric Indefinite Systems with Symmetric Positive Definite Preconditioners Eugene Vecharynski 1 Andrew Knyazev 2 1 Department of Computer Science and Engineering University of Minnesota 2 Department

More information

9. Iterative Methods for Large Linear Systems

9. Iterative Methods for Large Linear Systems EE507 - Computational Techniques for EE Jitkomut Songsiri 9. Iterative Methods for Large Linear Systems introduction splitting method Jacobi method Gauss-Seidel method successive overrelaxation (SOR) 9-1

More information

Mathematics and Computer Science

Mathematics and Computer Science Technical Report TR-2007-002 Block preconditioning for saddle point systems with indefinite (1,1) block by Michele Benzi, Jia Liu Mathematics and Computer Science EMORY UNIVERSITY International Journal

More information

An Assessment of Incomplete-LU Preconditioners for Nonsymmetric Linear Systems 1

An Assessment of Incomplete-LU Preconditioners for Nonsymmetric Linear Systems 1 Informatica 17 page xxx yyy 1 An Assessment of Incomplete-LU Preconditioners for Nonsymmetric Linear Systems 1 John R. Gilbert Xerox Palo Alto Research Center 3333 Coyote Hill Road Palo Alto, CA 930, USA

More information

Scientific Computing with Case Studies SIAM Press, Lecture Notes for Unit VII Sparse Matrix

Scientific Computing with Case Studies SIAM Press, Lecture Notes for Unit VII Sparse Matrix Scientific Computing with Case Studies SIAM Press, 2009 http://www.cs.umd.edu/users/oleary/sccswebpage Lecture Notes for Unit VII Sparse Matrix Computations Part 1: Direct Methods Dianne P. O Leary c 2008

More information

ON A GENERAL CLASS OF PRECONDITIONERS FOR NONSYMMETRIC GENERALIZED SADDLE POINT PROBLEMS

ON A GENERAL CLASS OF PRECONDITIONERS FOR NONSYMMETRIC GENERALIZED SADDLE POINT PROBLEMS U..B. Sci. Bull., Series A, Vol. 78, Iss. 4, 06 ISSN 3-707 ON A GENERAL CLASS OF RECONDIIONERS FOR NONSYMMERIC GENERALIZED SADDLE OIN ROBLE Fatemeh anjeh Ali BEIK his paper deals with applying a class

More information

A parameter tuning technique of a weighted Jacobi-type preconditioner and its application to supernova simulations

A parameter tuning technique of a weighted Jacobi-type preconditioner and its application to supernova simulations A parameter tuning technique of a weighted Jacobi-type preconditioner and its application to supernova simulations Akira IMAKURA Center for Computational Sciences, University of Tsukuba Joint work with

More information

Numerical Methods - Numerical Linear Algebra

Numerical Methods - Numerical Linear Algebra Numerical Methods - Numerical Linear Algebra Y. K. Goh Universiti Tunku Abdul Rahman 2013 Y. K. Goh (UTAR) Numerical Methods - Numerical Linear Algebra I 2013 1 / 62 Outline 1 Motivation 2 Solving Linear

More information

Linear algebra issues in Interior Point methods for bound-constrained least-squares problems

Linear algebra issues in Interior Point methods for bound-constrained least-squares problems Linear algebra issues in Interior Point methods for bound-constrained least-squares problems Stefania Bellavia Dipartimento di Energetica S. Stecco Università degli Studi di Firenze Joint work with Jacek

More information

Iterative methods for Linear System

Iterative methods for Linear System Iterative methods for Linear System JASS 2009 Student: Rishi Patil Advisor: Prof. Thomas Huckle Outline Basics: Matrices and their properties Eigenvalues, Condition Number Iterative Methods Direct and

More information

FINE-GRAINED PARALLEL INCOMPLETE LU FACTORIZATION

FINE-GRAINED PARALLEL INCOMPLETE LU FACTORIZATION FINE-GRAINED PARALLEL INCOMPLETE LU FACTORIZATION EDMOND CHOW AND AFTAB PATEL Abstract. This paper presents a new fine-grained parallel algorithm for computing an incomplete LU factorization. All nonzeros

More information

Computational Methods. Systems of Linear Equations

Computational Methods. Systems of Linear Equations Computational Methods Systems of Linear Equations Manfred Huber 2010 1 Systems of Equations Often a system model contains multiple variables (parameters) and contains multiple equations Multiple equations

More information

Course Notes: Week 1

Course Notes: Week 1 Course Notes: Week 1 Math 270C: Applied Numerical Linear Algebra 1 Lecture 1: Introduction (3/28/11) We will focus on iterative methods for solving linear systems of equations (and some discussion of eigenvalues

More information

Finding Rightmost Eigenvalues of Large, Sparse, Nonsymmetric Parameterized Eigenvalue Problems

Finding Rightmost Eigenvalues of Large, Sparse, Nonsymmetric Parameterized Eigenvalue Problems Finding Rightmost Eigenvalues of Large, Sparse, Nonsymmetric Parameterized Eigenvalue Problems AMSC 663-664 Final Report Minghao Wu AMSC Program mwu@math.umd.edu Dr. Howard Elman Department of Computer

More information

Parallel Algorithms for Solution of Large Sparse Linear Systems with Applications

Parallel Algorithms for Solution of Large Sparse Linear Systems with Applications Parallel Algorithms for Solution of Large Sparse Linear Systems with Applications Murat Manguoğlu Department of Computer Engineering Middle East Technical University, Ankara, Turkey Prace workshop: HPC

More information

A Review of Preconditioning Techniques for Steady Incompressible Flow

A Review of Preconditioning Techniques for Steady Incompressible Flow Zeist 2009 p. 1/43 A Review of Preconditioning Techniques for Steady Incompressible Flow David Silvester School of Mathematics University of Manchester Zeist 2009 p. 2/43 PDEs Review : 1984 2005 Update

More information

THE solution of the absolute value equation (AVE) of

THE solution of the absolute value equation (AVE) of The nonlinear HSS-like iterative method for absolute value equations Mu-Zheng Zhu Member, IAENG, and Ya-E Qi arxiv:1403.7013v4 [math.na] 2 Jan 2018 Abstract Salkuyeh proposed the Picard-HSS iteration method

More information

Fine-Grained Parallel Algorithms for Incomplete Factorization Preconditioning

Fine-Grained Parallel Algorithms for Incomplete Factorization Preconditioning Fine-Grained Parallel Algorithms for Incomplete Factorization Preconditioning Edmond Chow School of Computational Science and Engineering Georgia Institute of Technology, USA SPPEXA Symposium TU München,

More information

Exploiting hyper-sparsity when computing preconditioners for conjugate gradients in interior point methods

Exploiting hyper-sparsity when computing preconditioners for conjugate gradients in interior point methods Exploiting hyper-sparsity when computing preconditioners for conjugate gradients in interior point methods Julian Hall, Ghussoun Al-Jeiroudi and Jacek Gondzio School of Mathematics University of Edinburgh

More information

Further experiences with GMRESR

Further experiences with GMRESR Further experiences with GMRESR Report 92-2 C. Vui Technische Universiteit Delft Delft University of Technology Faculteit der Technische Wisunde en Informatica Faculty of Technical Mathematics and Informatics

More information

MATH 590: Meshfree Methods

MATH 590: Meshfree Methods MATH 590: Meshfree Methods Chapter 34: Improving the Condition Number of the Interpolation Matrix Greg Fasshauer Department of Applied Mathematics Illinois Institute of Technology Fall 2010 fasshauer@iit.edu

More information

Today s class. Linear Algebraic Equations LU Decomposition. Numerical Methods, Fall 2011 Lecture 8. Prof. Jinbo Bi CSE, UConn

Today s class. Linear Algebraic Equations LU Decomposition. Numerical Methods, Fall 2011 Lecture 8. Prof. Jinbo Bi CSE, UConn Today s class Linear Algebraic Equations LU Decomposition 1 Linear Algebraic Equations Gaussian Elimination works well for solving linear systems of the form: AX = B What if you have to solve the linear

More information

J.I. Aliaga 1 M. Bollhöfer 2 A.F. Martín 1 E.S. Quintana-Ortí 1. March, 2009

J.I. Aliaga 1 M. Bollhöfer 2 A.F. Martín 1 E.S. Quintana-Ortí 1. March, 2009 Parallel Preconditioning of Linear Systems based on ILUPACK for Multithreaded Architectures J.I. Aliaga M. Bollhöfer 2 A.F. Martín E.S. Quintana-Ortí Deparment of Computer Science and Engineering, Univ.

More information

ELA THE MINIMUM-NORM LEAST-SQUARES SOLUTION OF A LINEAR SYSTEM AND SYMMETRIC RANK-ONE UPDATES

ELA THE MINIMUM-NORM LEAST-SQUARES SOLUTION OF A LINEAR SYSTEM AND SYMMETRIC RANK-ONE UPDATES Volume 22, pp. 480-489, May 20 THE MINIMUM-NORM LEAST-SQUARES SOLUTION OF A LINEAR SYSTEM AND SYMMETRIC RANK-ONE UPDATES XUZHOU CHEN AND JUN JI Abstract. In this paper, we study the Moore-Penrose inverse

More information

Key words. conjugate gradients, normwise backward error, incremental norm estimation.

Key words. conjugate gradients, normwise backward error, incremental norm estimation. Proceedings of ALGORITMY 2016 pp. 323 332 ON ERROR ESTIMATION IN THE CONJUGATE GRADIENT METHOD: NORMWISE BACKWARD ERROR PETR TICHÝ Abstract. Using an idea of Duff and Vömel [BIT, 42 (2002), pp. 300 322

More information

CME 302: NUMERICAL LINEAR ALGEBRA FALL 2005/06 LECTURE 6

CME 302: NUMERICAL LINEAR ALGEBRA FALL 2005/06 LECTURE 6 CME 302: NUMERICAL LINEAR ALGEBRA FALL 2005/06 LECTURE 6 GENE H GOLUB Issues with Floating-point Arithmetic We conclude our discussion of floating-point arithmetic by highlighting two issues that frequently

More information

Lecture 11: CMSC 878R/AMSC698R. Iterative Methods An introduction. Outline. Inverse, LU decomposition, Cholesky, SVD, etc.

Lecture 11: CMSC 878R/AMSC698R. Iterative Methods An introduction. Outline. Inverse, LU decomposition, Cholesky, SVD, etc. Lecture 11: CMSC 878R/AMSC698R Iterative Methods An introduction Outline Direct Solution of Linear Systems Inverse, LU decomposition, Cholesky, SVD, etc. Iterative methods for linear systems Why? Matrix

More information

CONVERGENCE BOUNDS FOR PRECONDITIONED GMRES USING ELEMENT-BY-ELEMENT ESTIMATES OF THE FIELD OF VALUES

CONVERGENCE BOUNDS FOR PRECONDITIONED GMRES USING ELEMENT-BY-ELEMENT ESTIMATES OF THE FIELD OF VALUES European Conference on Computational Fluid Dynamics ECCOMAS CFD 2006 P. Wesseling, E. Oñate and J. Périaux (Eds) c TU Delft, The Netherlands, 2006 CONVERGENCE BOUNDS FOR PRECONDITIONED GMRES USING ELEMENT-BY-ELEMENT

More information

ITERATIVE METHODS BASED ON KRYLOV SUBSPACES

ITERATIVE METHODS BASED ON KRYLOV SUBSPACES ITERATIVE METHODS BASED ON KRYLOV SUBSPACES LONG CHEN We shall present iterative methods for solving linear algebraic equation Au = b based on Krylov subspaces We derive conjugate gradient (CG) method

More information