SOR as a Preconditioner. A Dissertation. Presented to. University of Virginia. In Partial Fulllment. of the Requirements for the Degree

Size: px
Start display at page:

Download "SOR as a Preconditioner. A Dissertation. Presented to. University of Virginia. In Partial Fulllment. of the Requirements for the Degree"

Transcription

1 SOR as a Preconditioner A Dissertation Presented to The Faculty of the School of Engineering and Applied Science University of Virginia In Partial Fulllment of the Reuirements for the Degree Doctor of Philosophy (Computer Science) by Michael A. DeLong May 1997

2 Abstract Preconditioning is usually necessary for CG-type iterative algorithms for the solution of large sparse nonsymmetric linear systems. However, many good preconditioners have only marginal intrinsic parallelism { ILU and SSOR in the natural ordering are essentially seuential algorithms. Reordering can yield very good parallelism, but can cause (severe) degradation in the rate of convergence. We consider multi-step SOR in the red-black ordering as a preconditioner. It can have very good parallel properties, but it does not cause degradation in the overall rate of convergence, as does ILU or SSOR. We present one-processor results using multi-step SOR to precondition GMRES which substantiate this claim. There has been some confusion in the literature as to whether this approach should be eective. We deal with these objections, and discuss the relationship between the spectral properties of the preconditioned system and the convergence of GMRES. We also present results from the Intel Paragon and the IBM SP2 showing that as a preconditioner, multistep red-black SOR can give good scaled speedup.

3 Contents 1 Introduction The Conjugate Gradient (CG) Method Nonsymmetric Problems Summary of our results Organization of this thesis Linear Algebra, Test Problems, and Methods Basic linear algebra Inner products Orthogonality and conjugacy Norms Bases Special matrices The Jordan canonical form Three test problems Problem Problem Problem The red-black ordering Two important properties Consistently ordered matrices Property A The GMRES algorithm Arnoldi's method The GMRES algorithm Restarting GMRES Breakdown and stagnation Preconditioning GMRES i

4 CONTENTS ii Left preconditioning Right preconditioning The SOR iteration The Gauss-Seidel and SOR iterations SOR in the red-black ordering Optimal values for! SOR as a preconditioner The Bi-CGSTAB algorithm Background The Bi-Conjugate Gradient (BCG) algorithm The Conjugate Gradients Suared (CGS) algorithm The Bi-CGSTAB algorithm Preconditioning the CGS and Bi-CGSTAB algorithms Summary Serial Results Implementation The compressed sparse row (CSR) format Reordering Gauss-Seidel vs. ILU(0) as a preconditioner The eect of! The eect of Results for Problems 2 and Experiments using Bi-CGSTAB Summary and conclusions Theory Convergence of CG Convergence of GMRES Residual norm bounds \Superlinear" convergence Form of the preconditioned matrix ^A Convergence of GMRES for ^A The Gauss-Seidel case:! = ( ^A) lying in a circle:!! opt The \suboptimal"! case: 1 <! <! opt Tighter bounds Summary

5 CONTENTS iii 5 Parallel Implementation Issues Hardware Intel Paragon IBM SP Fortran Message Passing Interface (MPI) Minimizing paging on the Paragon Data structures Parallel data distribution Speedup and scaled speedup Speedup Scaled Speedup Summary Parallel Results Problem Sizes Choosing GMRES parameters Estimated and actual work Experiments on the Intel Paragon Experiments on the IBM SP Summary Summary Summary and conclusions Open problems and future directions

6 Chapter 1 Introduction In this dissertation we address the problem of nding parallel preconditioners for Conjugate Gradient type methods for solving large, sparse, nonsymmetric linear systems. Our goal is to show that the SOR iteration provides an eective preconditioner for an important class of problems, and is highly parallel when used in the red-black ordering. In this chapter we briey overview the problem of using iterative methods to solve symmetric systems of linear euations, and use it as an introduction to the more dicult problem of solving sparse nonsymmetric systems of linear euations. We then outline the rest of the dissertation. Given an n n system of linear euations Ax = b (1) where A is large and sparse, such as arises from the nite dierence or nite element discretization of elliptic partial dierential euations, direct methods such as Gaussian Elimination have storage and computational reuirements that make them prohibitively expensive when (1) arises from the discretization of three-dimensional partial dierential euations. One alternative is to use iterative methods, which generally have more modest storage reuirements and may also reuire less time to produce a solution of comparable accuracy, depending on the method and the problem. They usually also have better parallel properties. One such iterative method is the Conjugate Gradient (CG) method. 1.1 The Conjugate Gradient (CG) Method If A is real and symmetric (A T = A) and positive denite (x T Ax > 0; for any nonzero vector x), then (1) can be solved using the Conjugate Gradient (CG) method [HS52], shown in Figure 1. In this gure, and henceforth, we use Greek letters for scalars (e.g. and ), upper-case letters for 1

7 CHAPTER 1. INTRODUCTION 2 1. Choose x 0. Set p 0 = r 0 = b Ax For k = 1; 2; : : : until convergence do: (a) k = r T k r k =p T k Ap k (b) x k+1 = x k k p k (c) r k+1 = r k + k Ap k (d) If r T r k+1 k+1 < then stop. (e) k = r T k+1 r k+1 =r T k r k (f) p k+1 = r k+1 + k p k Figure 1: The Conjugate Gradient algorithm matrices (A), and bold lower-case letters for vectors, including the iterates (x k ), residuals (r k ), and direction vectors (p k ). It is well-known (see e.g. [GO93]) that the CG algorithm has three desirable properties: It is guaranteed to converge in at most n iterations. If A has m n distinct eigenvalues, it is guaranteed to converge in at most m iterations. If x is the exact solution to (1) and x k is the approximation to x generated by the CG method at iteration k, then the error x x k is minimized in some norm at each iteration k. The iterates are related by short recurrences. Rutishauser (see e.g. [Ort88], Appendix 3) has shown that they are related by a three-term recurrence. The nite termination property is only of theoretical interest, since for large n convergence is generally needed in far fewer than n iterations. As a result, it is usually desirable to accelerate the rate of convergence of the CG algorithm by preconditioning, which transforms the original system to a new system: Ax = b ) SAS T (S T x) = Sb (2) having the same solution. The term preconditioning is generally credited to Turing [Tur48], and while the idea of preconditioning the CG algorithm was rst presented by Hestenes [Hes56], the term and the idea were not brought together until the papers by Evans [Eva73] and Axelsson [Axe74]. It is not necessary to form S or SAS T explicitly; instead, the preconditioning is usually incorporated into the CG algorithm. This is shown in Figure 2, where M = (S T S) 1. Since the CG algorithm reuires that the matrix A be symmetric and positive-denite, the preconditioned matrix ^A = SAS T must also be symmetric and positive-denite. Thus we have three (not entirely distinct) criteria for a good preconditioner for the CG method: 1. The CG algorithm must converge faster for the preconditioned problem (2) than for the original system. Otherwise there is no gain in performance.

8 CHAPTER 1. INTRODUCTION 3 1. Choose x 0 and compute r 0 = b Ax 0. Solve Mz 0 = r 0 : Set p 0 = z For k = 1; 2; : : : until convergence do: (a) k = r T k p k =p T k Ap k (b) x k+1 = x k k p k (c) r k+1 = r k + k Ap k (d) Solve Mz k+1 = r k+1 (e) k = z T k+1 r k+1 =z T k r k (f) p k+1 = z k+1 + k p k Figure 2: The Preconditioned Conjugate Gradient Algorithm 2. The auxiliary system Mz k+1 = r k+1 must be easy to solve. This is primarily a performance consideration { if the solution of the auxiliary system is expensive, then the preconditioner must be of very high uality to oset the expense. 3. The preconditioned matrix ^A must be symmetric positive denite. If ^A is not symmetric positive denite the CG algorithm may produce nonpositive inner products and break down. One popular preconditioning method is the Incomplete Cholesky (IC) factorization [Mv77] (Figure 3). This is an incomplete factorization based on the Cholesky factorization for symmetric l 11 = p a 11 For i = 2 to n For j = 1 to i 1 If a ij = 0 then l ij = 0 else l ij = l ii = r a ii P j 1 a ij l k=1 ikl jk =l jj P i 1 k=1 l2 ik Figure 3: Incomplete Cholesky Factorization positive-denite matrices, a variant of Gaussian Elimination. The IC factorization computes M = LL T ; where A = M R; (3) and R 6= 0 is referred to as the remainder matrix. By discarding ll-in elements as the factorization progresses, this factorization also retains the sparsity of the original matrix. The form given in Figure 3 is the no-ll form, but allowing limited ll-in makes R approach 0 in some sense, and improves the uality of the preconditioner.

9 CHAPTER 1. INTRODUCTION 4 Another approach is to use several steps of another iterative method as a subsidiary iteration. This basic idea is implicit in the paper by Concus, Golub, and O'Leary [CGO76], where they consider CG and another iterative method, Symmetric Successive Over-Relaxation (SSOR), working together as a \generalized CG method." They cite three previous papers ([Axe74], [YHS75], [Ehr75]), all of which consider CG as an accelerator for some other iterative method, in particular, the SSOR method. Their formulation is euivalent to preconditioning CG with a single SSOR step. More generally, subsidiary iteration preconditioning eectively forms z = (I + H H m 1 )P 1 r; (4) where H is the iteration matrix of the subsidiary iteration, so M 1 = (I + H H m 1 )P 1. Dubois, Greenbaum, and Rodrigue [DGR79] used a Jacobi iteration as a preconditioner. Johnson, Micchelli, and Paul [JMP83] use least-suares and min-max polynomials in conjunction with the Jacobi iteration, and give a framework for more general subsidiary iteration preconditioning with their discussion of inner and outer iterations. Adams [Ada83], [Ada85] used m-step SSOR preconditioning and also discussed the more general case of a polynomial SSOR preconditioner. Adams [Ada83] briey discusses preconditioning CG with SOR, and gives an example showing that the SOR iteration is not a valid preconditioner for CG: the loss of symmetry of the preconditioned system leads to a loss of convergence. One problem that arises when using IC or SSOR on a parallel computer is that they reuire the solution of large sparse triangular systems of euations, making them dicult to carry out in parallel. It is possible to extract some parallelism in the triangular solve using wavefront orderings ([AG88], [Gre86], [Sal90]), and this approach is somewhat eective for vector computers. But Stotland [Sto93] has shown that the wavefront algorithm computation is completely swamped by communication on distributed-memory multicomputers such as the Intel ipsc/860. A dierent approach is to reorder the euations. An important classical reordering for systems arising from partial dierential euations is the red-black ordering [You71]. It decouples the underlying euations into two groups, and eectively replaces the triangular solve with two half-sized matrix-vector multiplies. However, it has been shown ([AG88], [DM89], [DFT90], [HO90], [Ort91]) for both IC and SSOR preconditionings that the red-black ordering may cause severe degradation in the rate of convergence of the Preconditioned Conjugate Gradient algorithm as compared with the row-wise natural ordering. Poole and Ortega [PO87] gave results showing this to be the case for orderings with more than two colors as well. There are other parallel preconditioners for the CG method that are not IC, SSOR, nor any of the polynomial preconditioners mentioned earlier. A survey of preconditioners for the CG method may be found in [BBC + 94] and [Ort88].

10 CHAPTER 1. INTRODUCTION Nonsymmetric Problems If A is nonsymmetric (A T 6= A) the CG algorithm is not valid: its minimization property is lost, and its nite termination property may be lost as well (see e.g. [Ada83], p. 127). Instead, several methods have been developed that are CG-like in that They are optimal in some sense, or Their iterates or residuals are related by short recurrences. We will discuss several of these methods in Chapter 2. The crucial properties of preconditioners for nonsymmetric CG-type methods are not as wellunderstood as they are for the CG method: there is no need to maintain symmetry; furthermore, eigenvalue information alone for the preconditioned system may not be sucient (see e.g. [GS94a], [GPS96]). In addition, characteristics of a problem that may be good for one nonsymmetric CG-type method may be bad for another [NRT92]. It is dicult to know a priori what will make a good preconditioner for using a given solver to solve a particular problem or class of problems. As a result, nding the \best" preconditioner-solver combination for a given problem is often a process of trial and error [BBC + 94]. One popular preconditoner for nonsymmetric systems uses Incomplete LU (ILU) factorization [Mv77]. Like IC, it is a variant of Gaussian Elimination, or LU factorization, in which A is factored as A = LU + R (5) where U is upper triangular, L is lower triangular, and R 6= 0 is the residual matrix. The preconditioning matrix M is just M = LU: (6) The preconditioned residual z = M 1 r is then obtained by solving the triangular systems Ly = r and U z = y. There are many ways to do ILU preconditioning, distinguished by the characteristics of the residual matrix R. No-ll, or ILU(0), preconditioning (Figure 4) retains the sparsity pattern of A. As is the case with IC preconditioning, it is possible to improve the uality of the preconditioner by adding some kind of diagonal compensation (the so-called modied factorizations) [Gus78], or by allowing ll elements, according to their level, prole, or magnitude. A discussion of ll techniues can be found in [Saa95], Chapter 10. A more concise discussion can be found in the tech report [CSW96]. It is also possible to use wavefront and multicolor orderings to parallelize ILU preconditioning, and, like IC, multicolor orderings can lead to a severe degradation in the overall rate of convergence of the iterative method. We will give a few experimental results to this eect in Chapter 3.

11 CHAPTER 1. INTRODUCTION 6 For k = 1 to n For i = k + 1 to n If (a ki = 0) then l ik = 0 else l ik = a ik =a kk For j = k + 1 to n If (a ij 6= 0) then a ij = a ij l ik a kj Figure 4: No-ll ILU Factorization Since it is not necessary to maintain the symmetry of the coecient matrix, it is possible to consider the SOR iteration itself as a preconditioner for nonsymmetric systems, even though it is not useful with CG. However, in [BBC + 94], the authors claim that for nonsymmetric problems \[SOR and Gauss-Seidel] are never used as preconditioners, for a rather technical reason." They explain that preconditioning with SOR maps the eigenvalues of the preconditioned matrix onto a circle in the complex plane, and that polynomial acceleration of SOR yields no improvement over simple SOR with optimal!, where! is the SOR relaxation parameter. Both of these claims are in fact true, but as we will show they do not speak to the uestion of whether SOR can be used eectively as a preconditioner. Another negative aspect of SOR as a preconditioner appears in a paper by Shadid and Tuminaro [ST94], in which they compare a variety of parallel preconditioners. One of these is one-step Gauss- Seidel, which does not perform well as compared to their other precondtioners. However, as we will show, it is critical to use multiple Gauss-Seidel steps to obtain useful preconditioning. In fact, Saad [Saa94] briey considers multiple Gauss-Seidel steps, but does not use a over-relaxation factor. 1.3 Summary of our results In this dissertation we will show that For an important class of problems, SOR is an eective preconditioner, despite the statement to the contrary in [BBC + 94]. Unlike SSOR, SOR does not suer serious degradation in the rate of convergence in the redblack ordering. It is critical to use multiple steps of Gauss-Seidel. This is why the use of a single Gauss-Seidel step in [ST94] was ineective.

12 CHAPTER 1. INTRODUCTION 7 At least a factor of two improvement over Gauss-Seidel may be achieved by using SOR with a nominal value for the relaxation parameter!. We will also give results showing the eect of the variation of! on the rate of convergence. SOR as a preconditioner can be highly parallel in the red-black ordering. In fact, for our implementation on the Paragon, our data partitioning and distribution yields superlinear scaled speedup. This suggests that multi-step red-black SOR may be an eective preconditioner for nonsymmetric CG-type iterative methods. 1.4 Organization of this thesis In Chapter 2 we introduce three partial dierential euations that lead to nonsymmetric test matrices, review some basic linear algebra, and discuss GMRES, a CG-type method for nonsymmetric problems that has a minimization property, and the BCG family of methods, which have short recurrences. We also review the SOR method, and show how SOR may be used to precondition GMRES and the BCG methods. In Chapter 3 we show that the matrices that arise when the partial dierential euations are discretized using nite dierences can be put into the red-black ordering, and show that for these problems, in the red-black ordering, the SOR iteration iteration is highly parallel. We also give experimental results on a serial machine showing the degradation in the rate of convergence of ILU(0)-preconditioned GMRES, and we show that for the same problem SOR-preconditioned GM- RES does not suer degradation. We then compare SOR-GMRES to stand-alone SOR and to unpreconditioned GMRES, and conclude that SOR can be used eectively as a preconditioner for GMRES. In Chapter 4 we give some mathematical justication for these experimental results by showing the eect of the preconditioner on the spectrum of the preconditioned matrix and discuss the rate of convergence in that context. In Chapter 5 we discuss some of the details of our parallel implementation, including the data partitioning and distribution, and in Chapter 6 we give parallel experimental results from the Intel Paragon, some of which show superlinear scaled speedup. We then explain the reasons for this, and show that the data distribution we have chosen may be the preferred way to proceed on this machine. We also give experimental results from the IBM SP2. In Chapter 7 we discuss some conclusions, future directions, and open problems.

13 Chapter 2 Linear Algebra, Test Problems, and Methods The Generalized Minimal Residual (GMRES) [SS86] method is an iterative method that can be used to solve nonsymmetric systems of linear euations. As is the case for the Conjugate Gradient (CG) method, the convergence of GMRES can be accelerated through the use of an eective preconditioner. Another iterative method, SOR, can be used as a preconditioner, and under some conditions it can have good parallel properties. We begin this chapter by reviewing some basic linear algebra. We then introduce three test problems, and show that in the discretization scheme we have chosen, the matrices that arise are consistently ordered with Property A, two properties that are crucial to our discussion of SOR. We then describe the GMRES method and the SOR iteration, and show how SOR can be used to precondition GMRES. We will use the resulting method, SOR(k)-GMRES(m), applied to these three test problems, for experiments in later chapters. We also discuss the Bi-Conjugate Gradient (BCG) family of methods, including the Stabilized Bi-Conjugate Gradient (Bi-CGSTAB) method, another nonsymmetric iterative method we will use in the next chapter. 2.1 Basic linear algebra In this section we review some basic linear algebra that we will use in the remainder of this dissertation. As in Chapter 1 we use the following conventions: Greek letters for scalars (e.g. and ), upper-case letters for matrices (A and M), and bold lower-case letters for vectors (x and y). We also use these denitions: x is a column vector and x T is a row vector. For both matrices and vectors T denotes the transpose. 8

14 CHAPTER 2. LINEAR ALGEBRA, TEST PROBLEMS, AND METHODS 9 A = [a 1 ; a 2 ; : : :; a n ] is a matrix with columns a 1 ; a 2 ; : : :; a n. A T is the transpose of A, and will have rows a T 1 ; a T 2 ; : : :; a T n. The elements a ii ; i = 1; 2; : : :n, are the main diagonal. The diagonals above it are referred to as the as the rst superdiagonal, the second superdiagonal, and so on. Analgously the diagonals below the main diagonal are referred to as the subdiagonals Inner products Let C n be the space of complex vectors of length n. Then an inner product is a complex-valued function of two vector variables, denoted ( ; ), that satises 1. (x; x) 0 for all x 2 C n, and (x; x) = 0 only if x = (x; y) = (y; x) for all real x; y 2 C n. (x; y) = (y; x) if any entry of x or y is complex. 3. (x; y) = (x; y) for all x; y 2 C n and complex scalars. 4. (x + y; z) = (x; z) + (y; z), for all x; y; z 2 C n. For the space R n of real vectors, the usual Euclidean inner product is just the dot product (x; y) = x T y = nx i=1 x i y i : (7) If A is a symmetric positive-denite matrix the product (x; y) A x T Ay also denes an inner product, as it satises the denition above. It is referred to as the A inner product Orthogonality and conjugacy If (x; y) = 0 for some inner product, the vectors x and y are said to be orthogonal or conjugate with respect to that inner product. A seuence of vectors x i ; i = 1; 2; : : :k is said to be orthogonal if (x i ; x j ) = 0 for i 6= j, and orthonormal if they are orthogonal and (x i ; x i ) = 1 for i = 1; 2; : : :k. If a seuence of vectors is orthogonal with respect to the A inner product, they are said to be A- conjugate. Lastly, if there are two seuences of vectors x i ; i = 1; 2; : : :k and y j ; j = 1; 2; : : :k such that each of the x i is orthogonal to all of the y j and vice versa, then the two seuences of vectors are said to be biconjugate. It can be shown (see e.g. [GO93], Chapter 9) that the residuals r i generated by the CG method are orthogonal, while the direction vectors p i are A-conjugate Norms A norm is a real-valued function over R n denoted by k k, having these three properties:

15 CHAPTER 2. LINEAR ALGEBRA, TEST PROBLEMS, AND METHODS kxk 0, and kxk = 0 only if x = kxk = jjkxk for all x 2 R n and complex scalars. 3. kx + yk kxk + kyk for all x; y 2 R n. For any inner product there is a related norm: kxk = p (x; x): (8) This leads directly to the Euclidean norm and the A-norm, where A is symmetric and positivedenite: kxk 2 = vu u t X n x 2 i i=1! ; kxk A = p x T Ax: (9) There are many other norms we will not mention here. The only other norm we will consider is the innity norm: which is also called the max norm. Any vector norm gives rise to a matrix norm by the denition For the innity norm this leads to kxk 1 = max 1in jx ij; (10) kaxk kak = max x6=0 kxk = max kaxk: (11) kxk=1 kak 1 = max nx 1in j=1 If i ; 1 i n are the eigenvalues of A, then the spectral radius of A is ja ij j: (12) (A) = max 1in j ij (13) It can be shown (see e.g. [Ort90]) that the Euclidean norm of A is kak 2 = (A T A); (14) and if A is symmetric this reduces to kak 2 = (A). For a matrix A and some norm, A has a condition number relative to that norm: (A) = kak ka 1 k (15) where A 1 is the inverse of A. In a sense, a condition number measures the diculty of solving a linear system with coecient matrix A: the identity matrix has a condition number of (I) = 1, and matrices that have condition numbers for a given norm that are close to one are close to the identity matrix I in that norm. In fact, we will show in a later chapter that condition numbers enter explicitly into the estimates of the rate of convergence for both CG and GMRES.

16 CHAPTER 2. LINEAR ALGEBRA, TEST PROBLEMS, AND METHODS Bases A vector v is said to be a linear combination of some set of p n vectors fx 1 ; x 2 ; : : : ; x p g if there are scalars 1 ; 2 ; : : :; p such that v = px i=1 i x i : (16) The vectors x i ; i = 1; : : :; p are linearly independent if no x i can be written as a linear combination of the other x j ; j 6= i. The set of all linear combinations of the x i, denoted spanfx 1 ; x 2 ; : : : ; x p g, forms a subspace of R n of dimension p. The x i span the subspace, and are called its basis Special matrices There are several special types of matrices that will be of interest in our discussion. A triangular matrix has non-zero entries on its main diagonal and either above it or below it, but not both. A matrix which is triangular except for non-zero entries on the rst diagonal above or below the main diagonal is a Hessenberg matrix. A real matrix Q such that Q T Q = I, where I is the identity matrix, is an orthogonal matrix. Its columns are orthonormal. A special kind of orthogonal matrix having entries that are either ones or zeros eectively reorders the euations represented by the matrix, and for this reason it is called a permutation matrix The Jordan canonical form The eigenvalues of A are known collectively as the spectrum of A, and are denoted: (A) = f 1 ; 2 ; : : :; n g : (17) Associated with each distinct eigenvalue i there is a vector p i such that Ap i = i p i : (18) The p i are called the eigenvectors of A. Evey matrix A can be written in its Jordan Canonical Form: A = P JP 1 (19) where P is a nonsingular matrix and J is diagonal or nearly so. If A has n linearly independent eigenvectors, then P = [p 1 ; p 2 ; : : : ; p n ] is the matrix of eigenvectors of A and J is a diagonal matrix having j ii = i. Symmetric matrices are guaranteed to have a full set of linearly independent eigenvectors (see e.g. [Ort87]) but nonsymmetric matrices are not. If A has fewer than n linearly independent eigenvectors, then it has some number of generalized eigenvectors that do not satisfy (18). In this case J is not diagonal: in addition to having j ii = i will also have some number of

17 CHAPTER 2. LINEAR ALGEBRA, TEST PROBLEMS, AND METHODS 12 ones on the rst superdiagonal. Associated with the Jordan canonical form is the Jordan condition number, which is the the condition number of P, measured in the Euclidean norm: j(a) = kp k 2 kp 1 k 2 : (20) When A is symmetric and positive denite, j(a) = 1, but when A is nonsymmetric its eigenvectors can fail to be orthogonal, so that j(a) > 1. In fact, j(a) can be arbitrarily large. 2.2 Three test problems Our three test problems are all partial dierential euations that, when discretized, yield large, sparse nonsymmetric systems of linear euations Problem 1 Our rst problem is a two-dimensional convection-diusion euation: u + u x + u y = f; (21) where u is the Laplacian operator applied to u (u = u xx + u yy ), and the rst-derivative terms u x and u y are referred to as the convection terms. For simplicity we consider only the unit suare [0; 1] [0; 1] as the domain, and we use uniform grid spacing h = 1=(N + 1), where N is the number of gridpoints in a given direction, in both the x and y directions. Then the grid points are and the grid points interior to the domain are (x i ; y j ) = (ih; jh); i; j = 0; 1; : : :N; N + 1; (22) (x i ; y j ) = (ih; jh); i; j = 1; : : :N: (23) We use Dirichlet conditions (u = 0) on the boundary. To discretize we use standard second-order nite dierences to approximate u xx and u yy at every interior node (x i ; y j ): u xx (x i ; y j ) u yy (x i ; y j ) so that the Laplacian u at (x i ; y j ) is : = h 1 2 (u(x i+1 ; y j ) 2u(x i ; y j ) + u(x i 1 ; y j )); : = h 1 2 (u(x i ; y j+1 ) 2u(x i ; y j ) + u(x i ; y j 1 )); u(x i ; y j ) : = 1 h 2 (4u(x i; y j ) u(x i+1 ; y j ) u(x i 1 ; y j ) u(x i ; y j+1 ) u(x i ; y j 1 )): (25) We use centered dierences to approximate the rst-derivative terms u x and u y at (x i ; y j ): u x (x i ; y j ) u y (x i ; y j ) : = 1 : = 1 2h (u(x i+1; y j ) u(x i 1 ; y j )); 2h (u(x i; y j+1 ) u(x i ; y j 1 )): (24) (26)

18 CHAPTER 2. LINEAR ALGEBRA, TEST PROBLEMS, AND METHODS 13 We substitute (25) and (26) into (21) and collect like terms. Now we order the gridpoints in the row-wise \natural" ordering: u 1 ; : : : ; u N 2 = u(x 1 ; y 1 ); : : :; u(x N ; y 1 ); u(x 1 ; y 2 ); : : :; u(x N ; y 2 );. : : :. u(x 1 ; y N ); : : :; u(x N ; y N ): (27) Figure 5(a) shows the interior grid points, labeled in the natural ordering, for N = 5. As there are (a) Grid d e n w d e n w d e n w d e n w d n s d e n s w d e n s w d e n s w d e n s w d n s d e n s w d e n s w d e n s w d e n s w d n s d e n s w d e n s w d e n s w d e n s w d n s d e s w d e s w d e s w d e s w d (b) Assembled matrix Figure 5: A 5 5 grid on the unit suare in the natural ordering, and the related matrix. n = N 2 interior gridpoints, (21) is approximated by an n n linear system in which a typical euation is of the form Au = b (28) d i u i + e i u i+1 + w i u i 1 + n i u i+n + s i u i N = h 2 f i ; (29) where d i = 4; e i = 1 + h 2 ; w h i = 1 2 ; n i = 1 + h 2 ; s h i = 1 2 : (30) The assembled matrix A has a structure analogous to the one shown in Figure 5(b) for the case N = 5. This matrix A is structurally symmetric, but it is not numerically symmetric unless = = 0. For testing purposes we choose b for this problem so that the exact solution to (28) is known (u is a vector of ones).

19 CHAPTER 2. LINEAR ALGEBRA, TEST PROBLEMS, AND METHODS Problem 2 Our next problem is taken from [van81], and was also used in [ST94]: u x u xx + (1 + y 2 )(u y u yy ) = f (31) over the unit suare with Dirichlet boundary conditions and right-hand side chosen such that the exact solution is u(x; y) = e x+y + x 2 (1 x) 2 ln(1 + y 2 ) (32) It is discretized in the same way as (21) and its typical discrete euation has the form (29) with d i = 4 + 2y 2 ; e i = h; w i = Problem 3 Like Problem 2, our third problem is also used in [ST94]: 1 2 h; n i = (1 + y 2 )e i ; s i = (1 + y 2 )w i (33) u + (x 2 u x + y 2 u y ) u = f (34) over the unit suare with Dirichlet boundary conditions. We use the same discretization scheme as in Problems 1 and 2, so that d i = 4 h 2 ; e i = x2 h; w i = x2 h; n i = y2 h; s i = y2 h; (35) and again we choose the right-hand side so the exact solution to the discrete problem is known (u is a vector of ones). 2.3 The red-black ordering Because we have chosen the ve-point discretization scheme (25) and Dirichlet boundary conditions, all three problems yield matrices with the structure shown in Figure 5(b) when the matrix is assembled using the natural ordering. If instead we label the gridpoints in the red-black ordering [You71], shown in Figure 6(a) for a 5 5 grid, and assemble the matrix by ordering rst all the red points, then all the black points, the resulting matrix has the structure shown in Figure 6(b). Matrices in the natural and red-black orderings are related by A nat = P A rb P T (36) where P is a permutation matrix. The matrix in the red-black ordering, A rb (Figure 6(b)), has a block structure: " # D R E A rb = (37) F D B where D R and D B are diagonal matrices. We will exploit this special structure in a later section in our discussion of preconditioning with SOR.

20 CHAPTER 2. LINEAR ALGEBRA, TEST PROBLEMS, AND METHODS 15 r11 b11 r12 b12 r13 b 8 r 9 b 9 r10 b10 r 6 b 6 r 7 b 7 r 8 b 3 r 4 b 4 r 5 b 5 r 1 b 1 r 2 b 2 r 3 (a) grid d e n d w e n d w n d s w e n d s w e n d s e n d s w e n d s w n d s w e n d s w e n d s e d s w e d s w w e n d w e n d s e n d s w e n d s w n d s w e n d s w e n d s e n d s w e n d s w n d s w e d s w e d (b) matrix Figure 6: 5 5 grid on the unit suare in the red-black ordering, and the related matrix. 2.4 Two important properties The matrices arising from all three of the test problems discussed in this chapter have two properties which will be very important when we discuss SOR, namely that they are consistently ordered and have what Young calls Property A [You71]. In this section we give the denitions of these two properties and show that the matrices arising from Probems 1, 2, and 3 are consistently ordered with Property A Consistently ordered matrices Young [You71] gives the following denitions: Given a matrix A = (a i;j ), the integers i and j are associated with respect to A if a i;j 6= 0 or a j;i 6= 0. The matrix A of order N is consistently ordered if for some t there exist subsets S 1 ; S 2 ; : : :; S t of W = f1; 2; : : :; Ng such that S t k=1 S k = W and such that i and j are associated, then j 2 S k+1 if j > i and j 2 S k 1 if j < i, where i 2 S k. If the grid of interest is m m, then the ve-point nite-dierence stencil connects any gridpoint i to the gridpoints i + 1; i 1; i + m; i m. Thus, if W = f1; : : :; m 2 g and the subsets S i of W are S 1 = f1g; S 2 = f2; m + 1g; S 3 = f3; m + 2; 2m + 1g; : : :; S 2m 2 = fm 2 m; m 2 1g; S 2m 1 = fm 2 g; (38) corresponding to diagonal lines of gridpoints, then the assembled matrix A is consistently ordered. For a 55 suare grid, the subsets are shown Figure 7, where S 1 = f1g, S 2 = f2; 6g, S 3 = f3; 7; 11g,

21 CHAPTER 2. LINEAR ALGEBRA, TEST PROBLEMS, AND METHODS 16 through S 8 = f20; 24g and S 9 = f25g. For example, gridpoint 9 will be associated with gridpoints Figure 7: Diagonal subsets for a 5 5 grid 4, 8, 10, and 14 by the nite-dierence approximations (25) and (26). It is in S 5, while gridpoints 4 and 8 are in S 4, and gridpoints 10 and 14 are in S 6. Following this example the matrices arising from Problems 1-3 are consistently ordered Property A Ortega [Ort90] gives the following denition for Property A: The matrix A is 2-cyclic or has Property A if there is a permutation matrix P such that " # P AP T D 1 C 1 = (39) C 2 D 2 where D 1 and D 2 are diagonal. From our discussion of the block structure of A rb (see (37)), it is clear that the matrices arising from Problems 1-3 have Property A. 2.5 The GMRES algorithm Recall from Chapter 1 that the CG algorithm has three desirable properties: In exact arithmetic, it converges in at most n iterations if A is n n. It minimizes the error x x k in some norm at every iteration. The iterates are related by short recurrences.

22 CHAPTER 2. LINEAR ALGEBRA, TEST PROBLEMS, AND METHODS 17 In this section we develop the GMRES algorithm [SS86] from Arnoldi's method, and show that it is also guaranteed to converge in at most n iterations, and that it minimizes the residual r k = b Ax k in some sense at every iteration. In a later section we will discuss the BCG family of methods, which if they avoid breakdown are guaranteed to converge in no more than n iterations, and produce iterates that are related by short recurrences Arnoldi's method In exact arithmetic, Arnoldi's method generates a basis for the Krylov subspace K m = K m (A; v 1 ) = span v 1 ; Av 1 ; A 2 v 1 ; : : :; A m 1 v 1 (40) where m n. It can be implemented using a Gram-Schmidt process (Figure 8). The v j vectors it 1. Choose an initial vector v 1 with kv 1 k 2 = 1 2. For j = 1; 2; : : :; m (a) h i;j = (Av j ; v i ); i = 1; 2; : : :; j: (b) ^v j+1 = Av j P j i=1 h i;jv i ; (c) h j+1;j = k^v j+1 k 2 (d) v j+1 = (1=h j+1;j )^v j+1 produces are orthonormal: Figure 8: Arnoldi's method using Gram-Schmidt orthogonalization and they span K m. By construction h j+1;j v j+1 = Av j jx i=1 v T i v j = ( 1 if i = j 0 otherwise h i;j v i ; or Av j = j+1 X If we dene the matrix V m = [v 1 ; : : :; v m ], then this becomes i=1 (41) h i;j v i ; j = 1; 2; : : :; m: (42) AV m = V m H m + ^v m+1 e T m ; (43) where e T m = (0; 0; : : :; 0; 1) and H m = (h i;j ). H m is upper Hessenberg (upper triangular except for the rst subdiagonal). The last term, ^v m+1 e T m, is a vector that is zero except for the last element, which is just the last element of the \unorthogonalized" Arnoldi vector, ^v m+1 = h m+1;m v m+1, the other elements being forced to be zero by the orthogonality of the v i. We rewrite (43) as " # H AV m = V m+1hm ; where Hm m = : (44) k^v m+1 k 2

23 CHAPTER 2. LINEAR ALGEBRA, TEST PROBLEMS, AND METHODS 18 That is, Hm is an (m + 1) m matrix consisting of H m augmented by a row that is zero except for the (m + 1; m) entry. If we multiply (44) by Vm T and make use of the orthogonality of the v j, we obtain Vm T AV m = H m : (45) In practice the vectors generated using the Gram-Schmidt process can lose orthogonality: because of rounding error and cancellation there may be vectors v i and v j that are far from orthogonal. This can be mitigated by implementing Arnoldi's method using a modied Gram-Schmidt process [Ste73] (Figure 9) which rearranges the computations in steps 2(a) and 2(b) of Figure 8. The two methods 1. Choose an initial vector v 1 with kv 1 k 2 = 1 2. For j = 1; 2; : : :; m (a) ^v j+1 = Av j (b) For i = 1; 2; : : :j h i;j = (^v j+1 ; v i ) ^v j+1 = ^v j+1 h i;j v i (c) h j+1;j = k^v j+1 k 2 (d) v j+1 = (1=h j+1;j )^v j+1 Figure 9: Arnoldi's method using modied Gram-Schmidt orthogonalization are mathematically euivalent. However, the parallel properties of the modied Gram-Schmidt process are less desirable: since each of the inner products (^v j+1 ; v i ) in Figure 9 reuires all the processors to synchronize, the j inner products reuire j synchronizations, while the unmodied Gram-Schmidt method allows its j inner products to be done all at once, eectively saving j 1 synchronizations The GMRES algorithm If x 0 is an initial guess for the solution to Ax = b, the initial residual is r 0 = b Ax 0 : (46) The GMRES algorithm (Figure 10) uses Arnoldi's method to generate a basis for the Krylov space K m (A; r 0 ). It then forms the iterate x m = x 0 + mx i=1 y i v i (47) as a linear combination of the Arnoldi vectors, where the coecients y i are chosen so the iterate x m minimizes the l 2 norm of the residual (kr m k 2 ) over the space K m (A; r 0 ). Thus, the GMRES algorithm has two desirable properties:

24 CHAPTER 2. LINEAR ALGEBRA, TEST PROBLEMS, AND METHODS 19 In exact arithmetic, it is guaranteed to converge in at most n iterations. The Euclidean norms of the residuals are non-increasing: kr k+1 k 2 kr k k 2 ; for k Choose x 0. Compute r 0 = b Ax 0, = kr 0 k 2, v 1 = (1=)r For j = 1; 2; : : :; m; : : : until converged do: (a) ^v j+1 = Av j : (b) h i;j = (^v j+1 ; v i ); i = 1; 2; : : :; j: (c) ^v j+1 = ^v j+1 P j i=1 h i;jv i ; (d) h j+1;j = k^v j+1 k 2 (e) Check the convergence criterion: if kr k k 2 < then GMRES has converged. (f) v j+1 = (1=h j+1;j )^v j Form the approximate solution: x m = x 0 + V m y m, where y m minimizes ke 1 Hm yk 2 ; y 2 R m : Figure 10: The GMRES algorithm We now show how to nd the minimizer y m = (y 1 ; y 2 ; : : :; y m ) T and how to monitor convergence. Our presentation follows that given in [SS86]. A more detailed discussion can be found in [Saa95]. Let = kr 0 k 2 and let v 1 = (1=)r 0 be the starting vector for Arnoldi's method. If V m = [v 1 ; : : :; v m ], then (47) can be written x m = x 0 + V m y, where y is an m-vector. Dene J(y) kb Axk 2 = kb A(x 0 + V m y)k 2 : (48) We want to choose y so that V m y minimizes the residual norm J(y) over K m (A; r 0 ). Note that b A(x 0 + V m y) = (b Ax 0 ) AV m y (49) = r 0 AV m y; (50) and recall that r 0 = v 1. Since V m+1 e 1 = v 1, from the relationship (44) r 0 AV m y = v 1 V m+1hm y (51) = V m+1 (e 1 Hm y): (52) Because all the v i are orthonormal, V T V m+1 m+1 = I and kv m+1 k 2 = kk 2 for any vector. Thus (48) through (52) give J(y) = kb Axk 2 = kv m+1 (e 1 Hm y)k 2 = ke 1 Hm yk 2 : (53) Therefore, the minimizer y m will minimize the l 2 norm of e 1 Hm y. This can be found using QR reduction: Q mhm = R m ; (54)

25 CHAPTER 2. LINEAR ALGEBRA, TEST PROBLEMS, AND METHODS 20 factoring the Hessenberg matrix Hm into Q T m, an (m + 1) (m + 1) orthogonal matrix, and R m, an (m + 1) m upper-triangular matrix whose last row is zero. This decomposition yields J(y) = ke 1 Hm yk 2 = kq m e 1 Hm y k 2 = kg m R m yk 2 : (55) The last line can be rewritten J(y) = " # " g g m+1 R 0 #" ym 0 # 2 (56) where R is an upper triangular mm matrix, and g m+1 is the (m+1)st entry of g m. The minimizer y m is therefore the solution of the nonsingular m m triangular system Ry m = g: (57) Substituting back into (56) gives J(y m ) = jg m+1 j: (58) We only need the minimizer y m every m iterations, but we need the residual norm kr k k 2 at every iteration for the convergence test. By factoring the Hessenberg matrix H k at every iteration we can form g k = Q k (e 1 ) inexpensively, and from the above construction we have kr k k 2 = jg k+1 j; (59) the absolute value of the last entry of g k Restarting GMRES As GMRES progresses, both its memory and computational reuirements increase. At iteration j it needs to store the new vector ^v j+1 and it performs j inner products and j \axpys" (vector operations that add a vector to a scalar times a vector: z = x + y) to orthogonalize ^v j+1 against all previous v i. If ^v j+1 is an n-vector it is easy to see that GMRES has memory reuirements that grow like jn and computational reuirements that grow like j 2 n. A modication of the algorithm that is less expensive per iteration restarts every m steps, where m is small relative to n. After step m the restarted algorithm, denoted GMRES(m), updates the iterate x m and the residual r m, discards the vectors v 1 ; : : :; v m, and restarts with x 0 = x m and r 0 = r m. Since GMRES does not compute the residual r k at each step it will need to be computed at restart. This can be done in either of two ways: Compute r m = b Ax m, or Use (52): r m = V m+1 (e 1 Hm y m ).

26 CHAPTER 2. LINEAR ALGEBRA, TEST PROBLEMS, AND METHODS 21 These two expressions are mathematically euivalent, but if m is small the second expression can be substantially cheaper, and for this reason it is the form we use for our experiments. However, as m grows large the second expression can give an r m that is less accurate, due to a loss of orthogonality of the columns of V m, and the rst expression can be more desirable. GMRES(m) is not guaranteed to converge in n steps, but will still converge if m is chosen large enough. Saad and Schultz [SS86] give conditions under which GMRES(m) will converge. Joubert [Jou94] has introduced a strategy for choosing the restart freuency adaptively. At each iteration the method produces two estimates of the residual norm for the next iteration: one with restarting, one without, and chooses to restart if restarting is expected to be more ecient. The eectiveness of this strategy is determined by the accuracy of the residual norm estimates, and by the correctness of the underlying assumption that a single-step prediction can nd the strategy that is most ecient overall. We do not use Joubert's strategy Breakdown and stagnation Many iterative methods for solving nonsymmetric linear systems can break down, and fail to produce a new iterate. Brown [Bro91] contrasts the breakdown behavior of GMRES with that of an earlier method developed by Saad [Saa81], the Full Orthogonalization Method (FOM). FOM uses Arnoldi's method to generate a basis for the Krylov subspace K m (A; r 0 ), and constructs the iterate x m = x 0 + V m y, but instead of minimizing the l 2 norm of the residual it solves H m y = e 1 (60) where H m is the m m Hessenberg matrix in (44). In exact arithmetic, FOM can break down in one of two ways: h k+1;k = 0. Then AV k = V k H k, the vectors v 1 ; : : :; v k span K m (A; r 0 ), so the current iterate x k solves Ax = b exactly. This is called lucky breakdown [SS86] or happy breakdown [Bro91]. It rarely occurs. H k is singular. When H k is singular FOM cannot generate the new iterate x m. In the rst case both FOM and GMRES generate the exact solution and stop. Brown [Bro91] shows that when FOM generates a singular H k, GMRES will under the same circumstances generate the same iterate again: x k = x k 1 ; but does not break down. This behavior can persist for several iterations, and it is called stagnation. In exact arithmetic GMRES can recover from stagnation and produce a seuence of iterates that converge to the solution. However, when GMRES is restarted this recovery can be lost { if GMRES(m) stagnates for a full restart cycle, then convergence is lost. Saad and Schultz [SS86] give the example " # " # " # A = ; b = ; x 0 = ; (61)

27 CHAPTER 2. LINEAR ALGEBRA, TEST PROBLEMS, AND METHODS 22 where full GMRES converges in two steps, but the restarted method GMRES(1) never does. In machine arithmetic it is possible for GMRES to produce a seuence of nearly singular H k. Because of rounding error this can cause fatal stagnation: GMRES can produce a seuence of iterates that change very little, and fail to converge to the solution A 1 b. 2.6 Preconditioning GMRES Since n, the order of A, is usually very large, the nite convergence property of CG or GMRES is of only theoretical interest. In practice, convergence is determined by some convergence criterion such as the norm of the residual being suitably small (see Figure 10). In many cases, the rate of convergence can be increased by preconditioning. For CG, this means transforming A so as to maintain symmetry and positive-deniteness: ^A = SAS T : (62) Here S is a nonsingular matrix, chosen so that ^A has fewer distinct eigenvalues than A, or has a smaller condition number. For a nonsymmetric matrix, (62) is replaced by ^A = S 1 AS 2 ; (63) where S 1 and S 2 are nonsingular. If S 1 = I; the identity matrix, this is called right preconditioning, and if S 2 = I, it is called left preconditioning. We will consider preconditioning on only one side, so that the preconditioning matrix is either M = S 1 1 or M = S 1 2. The preconditioned system is not formed explicitly, but rather a solution of an auxiliary system involving the matrix M is incorporated into each iteration. In principle, we could choose M = A, which would give M 1 A = I, and GMRES would converge in a single step. However this is not practical, since applying the preconditioner would be as dicult as solving the original problem Ax = b. Instead we think of M as approximating A in some sense. For CG, M must be chosen so that ^A is symmetric positive denite, but for a nonsymmetric matrix the only constraint is that M be nonsingular, and the auxiliary system be \easy" to solve Left preconditioning GMRES is preconditioned on the left by solving Mz = y with z = ^v j+1 and y = Av j, so that step 2(a) in Figure 10 becomes Solve M ^v j+1 = Av j : It also reuires that the initial residual r 0 be preconditioned. The resulting algorithm is given in Figure 11. The modied residuals are ^r j = M 1 b M 1 Ax j = M 1 (b Ax j ) = M 1 r j ; (64)

SOLVING SPARSE LINEAR SYSTEMS OF EQUATIONS. Chao Yang Computational Research Division Lawrence Berkeley National Laboratory Berkeley, CA, USA

SOLVING SPARSE LINEAR SYSTEMS OF EQUATIONS. Chao Yang Computational Research Division Lawrence Berkeley National Laboratory Berkeley, CA, USA 1 SOLVING SPARSE LINEAR SYSTEMS OF EQUATIONS Chao Yang Computational Research Division Lawrence Berkeley National Laboratory Berkeley, CA, USA 2 OUTLINE Sparse matrix storage format Basic factorization

More information

Chapter 7 Iterative Techniques in Matrix Algebra

Chapter 7 Iterative Techniques in Matrix Algebra Chapter 7 Iterative Techniques in Matrix Algebra Per-Olof Persson persson@berkeley.edu Department of Mathematics University of California, Berkeley Math 128B Numerical Analysis Vector Norms Definition

More information

The amount of work to construct each new guess from the previous one should be a small multiple of the number of nonzeros in A.

The amount of work to construct each new guess from the previous one should be a small multiple of the number of nonzeros in A. AMSC/CMSC 661 Scientific Computing II Spring 2005 Solution of Sparse Linear Systems Part 2: Iterative methods Dianne P. O Leary c 2005 Solving Sparse Linear Systems: Iterative methods The plan: Iterative

More information

Preface to the Second Edition. Preface to the First Edition

Preface to the Second Edition. Preface to the First Edition n page v Preface to the Second Edition Preface to the First Edition xiii xvii 1 Background in Linear Algebra 1 1.1 Matrices................................. 1 1.2 Square Matrices and Eigenvalues....................

More information

Jordan Journal of Mathematics and Statistics (JJMS) 5(3), 2012, pp A NEW ITERATIVE METHOD FOR SOLVING LINEAR SYSTEMS OF EQUATIONS

Jordan Journal of Mathematics and Statistics (JJMS) 5(3), 2012, pp A NEW ITERATIVE METHOD FOR SOLVING LINEAR SYSTEMS OF EQUATIONS Jordan Journal of Mathematics and Statistics JJMS) 53), 2012, pp.169-184 A NEW ITERATIVE METHOD FOR SOLVING LINEAR SYSTEMS OF EQUATIONS ADEL H. AL-RABTAH Abstract. The Jacobi and Gauss-Seidel iterative

More information

AMS526: Numerical Analysis I (Numerical Linear Algebra) Lecture 23: GMRES and Other Krylov Subspace Methods; Preconditioning

AMS526: Numerical Analysis I (Numerical Linear Algebra) Lecture 23: GMRES and Other Krylov Subspace Methods; Preconditioning AMS526: Numerical Analysis I (Numerical Linear Algebra) Lecture 23: GMRES and Other Krylov Subspace Methods; Preconditioning Xiangmin Jiao SUNY Stony Brook Xiangmin Jiao Numerical Analysis I 1 / 18 Outline

More information

Iterative methods for Linear System

Iterative methods for Linear System Iterative methods for Linear System JASS 2009 Student: Rishi Patil Advisor: Prof. Thomas Huckle Outline Basics: Matrices and their properties Eigenvalues, Condition Number Iterative Methods Direct and

More information

9.1 Preconditioned Krylov Subspace Methods

9.1 Preconditioned Krylov Subspace Methods Chapter 9 PRECONDITIONING 9.1 Preconditioned Krylov Subspace Methods 9.2 Preconditioned Conjugate Gradient 9.3 Preconditioned Generalized Minimal Residual 9.4 Relaxation Method Preconditioners 9.5 Incomplete

More information

Algorithms that use the Arnoldi Basis

Algorithms that use the Arnoldi Basis AMSC 600 /CMSC 760 Advanced Linear Numerical Analysis Fall 2007 Arnoldi Methods Dianne P. O Leary c 2006, 2007 Algorithms that use the Arnoldi Basis Reference: Chapter 6 of Saad The Arnoldi Basis How to

More information

Iterative Methods for Sparse Linear Systems

Iterative Methods for Sparse Linear Systems Iterative Methods for Sparse Linear Systems Luca Bergamaschi e-mail: berga@dmsa.unipd.it - http://www.dmsa.unipd.it/ berga Department of Mathematical Methods and Models for Scientific Applications University

More information

Linear Solvers. Andrew Hazel

Linear Solvers. Andrew Hazel Linear Solvers Andrew Hazel Introduction Thus far we have talked about the formulation and discretisation of physical problems...... and stopped when we got to a discrete linear system of equations. Introduction

More information

Preconditioned Conjugate Gradient-Like Methods for. Nonsymmetric Linear Systems 1. Ulrike Meier Yang 2. July 19, 1994

Preconditioned Conjugate Gradient-Like Methods for. Nonsymmetric Linear Systems 1. Ulrike Meier Yang 2. July 19, 1994 Preconditioned Conjugate Gradient-Like Methods for Nonsymmetric Linear Systems Ulrike Meier Yang 2 July 9, 994 This research was supported by the U.S. Department of Energy under Grant No. DE-FG2-85ER25.

More information

Course Notes: Week 1

Course Notes: Week 1 Course Notes: Week 1 Math 270C: Applied Numerical Linear Algebra 1 Lecture 1: Introduction (3/28/11) We will focus on iterative methods for solving linear systems of equations (and some discussion of eigenvalues

More information

Conjugate gradient method. Descent method. Conjugate search direction. Conjugate Gradient Algorithm (294)

Conjugate gradient method. Descent method. Conjugate search direction. Conjugate Gradient Algorithm (294) Conjugate gradient method Descent method Hestenes, Stiefel 1952 For A N N SPD In exact arithmetic, solves in N steps In real arithmetic No guaranteed stopping Often converges in many fewer than N steps

More information

FEM and sparse linear system solving

FEM and sparse linear system solving FEM & sparse linear system solving, Lecture 9, Nov 19, 2017 1/36 Lecture 9, Nov 17, 2017: Krylov space methods http://people.inf.ethz.ch/arbenz/fem17 Peter Arbenz Computer Science Department, ETH Zürich

More information

ITERATIVE METHODS FOR SPARSE LINEAR SYSTEMS

ITERATIVE METHODS FOR SPARSE LINEAR SYSTEMS ITERATIVE METHODS FOR SPARSE LINEAR SYSTEMS YOUSEF SAAD University of Minnesota PWS PUBLISHING COMPANY I(T)P An International Thomson Publishing Company BOSTON ALBANY BONN CINCINNATI DETROIT LONDON MADRID

More information

6.4 Krylov Subspaces and Conjugate Gradients

6.4 Krylov Subspaces and Conjugate Gradients 6.4 Krylov Subspaces and Conjugate Gradients Our original equation is Ax = b. The preconditioned equation is P Ax = P b. When we write P, we never intend that an inverse will be explicitly computed. P

More information

Chapter 7. Iterative methods for large sparse linear systems. 7.1 Sparse matrix algebra. Large sparse matrices

Chapter 7. Iterative methods for large sparse linear systems. 7.1 Sparse matrix algebra. Large sparse matrices Chapter 7 Iterative methods for large sparse linear systems In this chapter we revisit the problem of solving linear systems of equations, but now in the context of large sparse systems. The price to pay

More information

Topics. The CG Algorithm Algorithmic Options CG s Two Main Convergence Theorems

Topics. The CG Algorithm Algorithmic Options CG s Two Main Convergence Theorems Topics The CG Algorithm Algorithmic Options CG s Two Main Convergence Theorems What about non-spd systems? Methods requiring small history Methods requiring large history Summary of solvers 1 / 52 Conjugate

More information

M.A. Botchev. September 5, 2014

M.A. Botchev. September 5, 2014 Rome-Moscow school of Matrix Methods and Applied Linear Algebra 2014 A short introduction to Krylov subspaces for linear systems, matrix functions and inexact Newton methods. Plan and exercises. M.A. Botchev

More information

Institute for Advanced Computer Studies. Department of Computer Science. Iterative methods for solving Ax = b. GMRES/FOM versus QMR/BiCG

Institute for Advanced Computer Studies. Department of Computer Science. Iterative methods for solving Ax = b. GMRES/FOM versus QMR/BiCG University of Maryland Institute for Advanced Computer Studies Department of Computer Science College Park TR{96{2 TR{3587 Iterative methods for solving Ax = b GMRES/FOM versus QMR/BiCG Jane K. Cullum

More information

Iterative methods for Linear System of Equations. Joint Advanced Student School (JASS-2009)

Iterative methods for Linear System of Equations. Joint Advanced Student School (JASS-2009) Iterative methods for Linear System of Equations Joint Advanced Student School (JASS-2009) Course #2: Numerical Simulation - from Models to Software Introduction In numerical simulation, Partial Differential

More information

OUTLINE ffl CFD: elliptic pde's! Ax = b ffl Basic iterative methods ffl Krylov subspace methods ffl Preconditioning techniques: Iterative methods ILU

OUTLINE ffl CFD: elliptic pde's! Ax = b ffl Basic iterative methods ffl Krylov subspace methods ffl Preconditioning techniques: Iterative methods ILU Preconditioning Techniques for Solving Large Sparse Linear Systems Arnold Reusken Institut für Geometrie und Praktische Mathematik RWTH-Aachen OUTLINE ffl CFD: elliptic pde's! Ax = b ffl Basic iterative

More information

Reduced Synchronization Overhead on. December 3, Abstract. The standard formulation of the conjugate gradient algorithm involves

Reduced Synchronization Overhead on. December 3, Abstract. The standard formulation of the conjugate gradient algorithm involves Lapack Working Note 56 Conjugate Gradient Algorithms with Reduced Synchronization Overhead on Distributed Memory Multiprocessors E. F. D'Azevedo y, V.L. Eijkhout z, C. H. Romine y December 3, 1999 Abstract

More information

Krylov Subspace Methods that Are Based on the Minimization of the Residual

Krylov Subspace Methods that Are Based on the Minimization of the Residual Chapter 5 Krylov Subspace Methods that Are Based on the Minimization of the Residual Remark 51 Goal he goal of these methods consists in determining x k x 0 +K k r 0,A such that the corresponding Euclidean

More information

Preliminary/Qualifying Exam in Numerical Analysis (Math 502a) Spring 2012

Preliminary/Qualifying Exam in Numerical Analysis (Math 502a) Spring 2012 Instructions Preliminary/Qualifying Exam in Numerical Analysis (Math 502a) Spring 2012 The exam consists of four problems, each having multiple parts. You should attempt to solve all four problems. 1.

More information

Lab 1: Iterative Methods for Solving Linear Systems

Lab 1: Iterative Methods for Solving Linear Systems Lab 1: Iterative Methods for Solving Linear Systems January 22, 2017 Introduction Many real world applications require the solution to very large and sparse linear systems where direct methods such as

More information

Krylov Space Methods. Nonstationary sounds good. Radu Trîmbiţaş ( Babeş-Bolyai University) Krylov Space Methods 1 / 17

Krylov Space Methods. Nonstationary sounds good. Radu Trîmbiţaş ( Babeş-Bolyai University) Krylov Space Methods 1 / 17 Krylov Space Methods Nonstationary sounds good Radu Trîmbiţaş Babeş-Bolyai University Radu Trîmbiţaş ( Babeş-Bolyai University) Krylov Space Methods 1 / 17 Introduction These methods are used both to solve

More information

Scientific Computing WS 2018/2019. Lecture 9. Jürgen Fuhrmann Lecture 9 Slide 1

Scientific Computing WS 2018/2019. Lecture 9. Jürgen Fuhrmann Lecture 9 Slide 1 Scientific Computing WS 2018/2019 Lecture 9 Jürgen Fuhrmann juergen.fuhrmann@wias-berlin.de Lecture 9 Slide 1 Lecture 9 Slide 2 Simple iteration with preconditioning Idea: Aû = b iterative scheme û = û

More information

The Lanczos and conjugate gradient algorithms

The Lanczos and conjugate gradient algorithms The Lanczos and conjugate gradient algorithms Gérard MEURANT October, 2008 1 The Lanczos algorithm 2 The Lanczos algorithm in finite precision 3 The nonsymmetric Lanczos algorithm 4 The Golub Kahan bidiagonalization

More information

Iterative Methods and Multigrid

Iterative Methods and Multigrid Iterative Methods and Multigrid Part 3: Preconditioning 2 Eric de Sturler Preconditioning The general idea behind preconditioning is that convergence of some method for the linear system Ax = b can be

More information

Iterative Methods for Solving A x = b

Iterative Methods for Solving A x = b Iterative Methods for Solving A x = b A good (free) online source for iterative methods for solving A x = b is given in the description of a set of iterative solvers called templates found at netlib: http

More information

Lecture 18 Classical Iterative Methods

Lecture 18 Classical Iterative Methods Lecture 18 Classical Iterative Methods MIT 18.335J / 6.337J Introduction to Numerical Methods Per-Olof Persson November 14, 2006 1 Iterative Methods for Linear Systems Direct methods for solving Ax = b,

More information

In order to solve the linear system KL M N when K is nonsymmetric, we can solve the equivalent system

In order to solve the linear system KL M N when K is nonsymmetric, we can solve the equivalent system !"#$% "&!#' (%)!#" *# %)%(! #! %)!#" +, %"!"#$ %*&%! $#&*! *# %)%! -. -/ 0 -. 12 "**3! * $!#%+,!2!#% 44" #% &#33 # 4"!#" "%! "5"#!!#6 -. - #% " 7% "3#!#3! - + 87&2! * $!#% 44" ) 3( $! # % %#!!#%+ 9332!

More information

Numerical Methods - Numerical Linear Algebra

Numerical Methods - Numerical Linear Algebra Numerical Methods - Numerical Linear Algebra Y. K. Goh Universiti Tunku Abdul Rahman 2013 Y. K. Goh (UTAR) Numerical Methods - Numerical Linear Algebra I 2013 1 / 62 Outline 1 Motivation 2 Solving Linear

More information

FINE-GRAINED PARALLEL INCOMPLETE LU FACTORIZATION

FINE-GRAINED PARALLEL INCOMPLETE LU FACTORIZATION FINE-GRAINED PARALLEL INCOMPLETE LU FACTORIZATION EDMOND CHOW AND AFTAB PATEL Abstract. This paper presents a new fine-grained parallel algorithm for computing an incomplete LU factorization. All nonzeros

More information

Solving Sparse Linear Systems: Iterative methods

Solving Sparse Linear Systems: Iterative methods Scientific Computing with Case Studies SIAM Press, 2009 http://www.cs.umd.edu/users/oleary/sccs Lecture Notes for Unit VII Sparse Matrix Computations Part 2: Iterative Methods Dianne P. O Leary c 2008,2010

More information

Solving Sparse Linear Systems: Iterative methods

Solving Sparse Linear Systems: Iterative methods Scientific Computing with Case Studies SIAM Press, 2009 http://www.cs.umd.edu/users/oleary/sccswebpage Lecture Notes for Unit VII Sparse Matrix Computations Part 2: Iterative Methods Dianne P. O Leary

More information

Numerical Methods I Non-Square and Sparse Linear Systems

Numerical Methods I Non-Square and Sparse Linear Systems Numerical Methods I Non-Square and Sparse Linear Systems Aleksandar Donev Courant Institute, NYU 1 donev@courant.nyu.edu 1 MATH-GA 2011.003 / CSCI-GA 2945.003, Fall 2014 September 25th, 2014 A. Donev (Courant

More information

Applied Mathematics 205. Unit V: Eigenvalue Problems. Lecturer: Dr. David Knezevic

Applied Mathematics 205. Unit V: Eigenvalue Problems. Lecturer: Dr. David Knezevic Applied Mathematics 205 Unit V: Eigenvalue Problems Lecturer: Dr. David Knezevic Unit V: Eigenvalue Problems Chapter V.4: Krylov Subspace Methods 2 / 51 Krylov Subspace Methods In this chapter we give

More information

The rate of convergence of the GMRES method

The rate of convergence of the GMRES method The rate of convergence of the GMRES method Report 90-77 C. Vuik Technische Universiteit Delft Delft University of Technology Faculteit der Technische Wiskunde en Informatica Faculty of Technical Mathematics

More information

Scientific Computing with Case Studies SIAM Press, Lecture Notes for Unit VII Sparse Matrix

Scientific Computing with Case Studies SIAM Press, Lecture Notes for Unit VII Sparse Matrix Scientific Computing with Case Studies SIAM Press, 2009 http://www.cs.umd.edu/users/oleary/sccswebpage Lecture Notes for Unit VII Sparse Matrix Computations Part 1: Direct Methods Dianne P. O Leary c 2008

More information

Lecture Note 7: Iterative methods for solving linear systems. Xiaoqun Zhang Shanghai Jiao Tong University

Lecture Note 7: Iterative methods for solving linear systems. Xiaoqun Zhang Shanghai Jiao Tong University Lecture Note 7: Iterative methods for solving linear systems Xiaoqun Zhang Shanghai Jiao Tong University Last updated: December 24, 2014 1.1 Review on linear algebra Norms of vectors and matrices vector

More information

4.8 Arnoldi Iteration, Krylov Subspaces and GMRES

4.8 Arnoldi Iteration, Krylov Subspaces and GMRES 48 Arnoldi Iteration, Krylov Subspaces and GMRES We start with the problem of using a similarity transformation to convert an n n matrix A to upper Hessenberg form H, ie, A = QHQ, (30) with an appropriate

More information

Numerical Methods in Matrix Computations

Numerical Methods in Matrix Computations Ake Bjorck Numerical Methods in Matrix Computations Springer Contents 1 Direct Methods for Linear Systems 1 1.1 Elements of Matrix Theory 1 1.1.1 Matrix Algebra 2 1.1.2 Vector Spaces 6 1.1.3 Submatrices

More information

FINE-GRAINED PARALLEL INCOMPLETE LU FACTORIZATION

FINE-GRAINED PARALLEL INCOMPLETE LU FACTORIZATION FINE-GRAINED PARALLEL INCOMPLETE LU FACTORIZATION EDMOND CHOW AND AFTAB PATEL Abstract. This paper presents a new fine-grained parallel algorithm for computing an incomplete LU factorization. All nonzeros

More information

CME342 Parallel Methods in Numerical Analysis. Matrix Computation: Iterative Methods II. Sparse Matrix-vector Multiplication.

CME342 Parallel Methods in Numerical Analysis. Matrix Computation: Iterative Methods II. Sparse Matrix-vector Multiplication. CME342 Parallel Methods in Numerical Analysis Matrix Computation: Iterative Methods II Outline: CG & its parallelization. Sparse Matrix-vector Multiplication. 1 Basic iterative methods: Ax = b r = b Ax

More information

AMS526: Numerical Analysis I (Numerical Linear Algebra for Computational and Data Sciences)

AMS526: Numerical Analysis I (Numerical Linear Algebra for Computational and Data Sciences) AMS526: Numerical Analysis I (Numerical Linear Algebra for Computational and Data Sciences) Lecture 19: Computing the SVD; Sparse Linear Systems Xiangmin Jiao Stony Brook University Xiangmin Jiao Numerical

More information

Linear Algebra (part 1) : Vector Spaces (by Evan Dummit, 2017, v. 1.07) 1.1 The Formal Denition of a Vector Space

Linear Algebra (part 1) : Vector Spaces (by Evan Dummit, 2017, v. 1.07) 1.1 The Formal Denition of a Vector Space Linear Algebra (part 1) : Vector Spaces (by Evan Dummit, 2017, v. 1.07) Contents 1 Vector Spaces 1 1.1 The Formal Denition of a Vector Space.................................. 1 1.2 Subspaces...................................................

More information

Fine-grained Parallel Incomplete LU Factorization

Fine-grained Parallel Incomplete LU Factorization Fine-grained Parallel Incomplete LU Factorization Edmond Chow School of Computational Science and Engineering Georgia Institute of Technology Sparse Days Meeting at CERFACS June 5-6, 2014 Contribution

More information

IDR(s) as a projection method

IDR(s) as a projection method Delft University of Technology Faculty of Electrical Engineering, Mathematics and Computer Science Delft Institute of Applied Mathematics IDR(s) as a projection method A thesis submitted to the Delft Institute

More information

MAA507, Power method, QR-method and sparse matrix representation.

MAA507, Power method, QR-method and sparse matrix representation. ,, and representation. February 11, 2014 Lecture 7: Overview, Today we will look at:.. If time: A look at representation and fill in. Why do we need numerical s? I think everyone have seen how time consuming

More information

Lecture 11: CMSC 878R/AMSC698R. Iterative Methods An introduction. Outline. Inverse, LU decomposition, Cholesky, SVD, etc.

Lecture 11: CMSC 878R/AMSC698R. Iterative Methods An introduction. Outline. Inverse, LU decomposition, Cholesky, SVD, etc. Lecture 11: CMSC 878R/AMSC698R Iterative Methods An introduction Outline Direct Solution of Linear Systems Inverse, LU decomposition, Cholesky, SVD, etc. Iterative methods for linear systems Why? Matrix

More information

Summary of Iterative Methods for Non-symmetric Linear Equations That Are Related to the Conjugate Gradient (CG) Method

Summary of Iterative Methods for Non-symmetric Linear Equations That Are Related to the Conjugate Gradient (CG) Method Summary of Iterative Methods for Non-symmetric Linear Equations That Are Related to the Conjugate Gradient (CG) Method Leslie Foster 11-5-2012 We will discuss the FOM (full orthogonalization method), CG,

More information

Contents. Preface... xi. Introduction...

Contents. Preface... xi. Introduction... Contents Preface... xi Introduction... xv Chapter 1. Computer Architectures... 1 1.1. Different types of parallelism... 1 1.1.1. Overlap, concurrency and parallelism... 1 1.1.2. Temporal and spatial parallelism

More information

Iterative Methods. Splitting Methods

Iterative Methods. Splitting Methods Iterative Methods Splitting Methods 1 Direct Methods Solving Ax = b using direct methods. Gaussian elimination (using LU decomposition) Variants of LU, including Crout and Doolittle Other decomposition

More information

c 2015 Society for Industrial and Applied Mathematics

c 2015 Society for Industrial and Applied Mathematics SIAM J. SCI. COMPUT. Vol. 37, No. 2, pp. C169 C193 c 2015 Society for Industrial and Applied Mathematics FINE-GRAINED PARALLEL INCOMPLETE LU FACTORIZATION EDMOND CHOW AND AFTAB PATEL Abstract. This paper

More information

4.6 Iterative Solvers for Linear Systems

4.6 Iterative Solvers for Linear Systems 4.6 Iterative Solvers for Linear Systems Why use iterative methods? Virtually all direct methods for solving Ax = b require O(n 3 ) floating point operations. In practical applications the matrix A often

More information

ITERATIVE METHODS BASED ON KRYLOV SUBSPACES

ITERATIVE METHODS BASED ON KRYLOV SUBSPACES ITERATIVE METHODS BASED ON KRYLOV SUBSPACES LONG CHEN We shall present iterative methods for solving linear algebraic equation Au = b based on Krylov subspaces We derive conjugate gradient (CG) method

More information

On the influence of eigenvalues on Bi-CG residual norms

On the influence of eigenvalues on Bi-CG residual norms On the influence of eigenvalues on Bi-CG residual norms Jurjen Duintjer Tebbens Institute of Computer Science Academy of Sciences of the Czech Republic duintjertebbens@cs.cas.cz Gérard Meurant 30, rue

More information

Last Time. Social Network Graphs Betweenness. Graph Laplacian. Girvan-Newman Algorithm. Spectral Bisection

Last Time. Social Network Graphs Betweenness. Graph Laplacian. Girvan-Newman Algorithm. Spectral Bisection Eigenvalue Problems Last Time Social Network Graphs Betweenness Girvan-Newman Algorithm Graph Laplacian Spectral Bisection λ 2, w 2 Today Small deviation into eigenvalue problems Formulation Standard eigenvalue

More information

5 and A,1 = B = is obtained by interchanging the rst two rows of A. Write down the inverse of B.

5 and A,1 = B = is obtained by interchanging the rst two rows of A. Write down the inverse of B. EE { QUESTION LIST EE KUMAR Spring (we will use the abbreviation QL to refer to problems on this list the list includes questions from prior midterm and nal exams) VECTORS AND MATRICES. Pages - of the

More information

The Conjugate Gradient Method

The Conjugate Gradient Method The Conjugate Gradient Method Classical Iterations We have a problem, We assume that the matrix comes from a discretization of a PDE. The best and most popular model problem is, The matrix will be as large

More information

A Method for Constructing Diagonally Dominant Preconditioners based on Jacobi Rotations

A Method for Constructing Diagonally Dominant Preconditioners based on Jacobi Rotations A Method for Constructing Diagonally Dominant Preconditioners based on Jacobi Rotations Jin Yun Yuan Plamen Y. Yalamov Abstract A method is presented to make a given matrix strictly diagonally dominant

More information

Parallel Numerics, WT 2016/ Iterative Methods for Sparse Linear Systems of Equations. page 1 of 1

Parallel Numerics, WT 2016/ Iterative Methods for Sparse Linear Systems of Equations. page 1 of 1 Parallel Numerics, WT 2016/2017 5 Iterative Methods for Sparse Linear Systems of Equations page 1 of 1 Contents 1 Introduction 1.1 Computer Science Aspects 1.2 Numerical Problems 1.3 Graphs 1.4 Loop Manipulations

More information

Linear Algebra. Brigitte Bidégaray-Fesquet. MSIAM, September Univ. Grenoble Alpes, Laboratoire Jean Kuntzmann, Grenoble.

Linear Algebra. Brigitte Bidégaray-Fesquet. MSIAM, September Univ. Grenoble Alpes, Laboratoire Jean Kuntzmann, Grenoble. Brigitte Bidégaray-Fesquet Univ. Grenoble Alpes, Laboratoire Jean Kuntzmann, Grenoble MSIAM, 23 24 September 215 Overview 1 Elementary operations Gram Schmidt orthonormalization Matrix norm Conditioning

More information

AMS526: Numerical Analysis I (Numerical Linear Algebra for Computational and Data Sciences)

AMS526: Numerical Analysis I (Numerical Linear Algebra for Computational and Data Sciences) AMS526: Numerical Analysis I (Numerical Linear Algebra for Computational and Data Sciences) Lecture 1: Course Overview; Matrix Multiplication Xiangmin Jiao Stony Brook University Xiangmin Jiao Numerical

More information

Conjugate Gradient (CG) Method

Conjugate Gradient (CG) Method Conjugate Gradient (CG) Method by K. Ozawa 1 Introduction In the series of this lecture, I will introduce the conjugate gradient method, which solves efficiently large scale sparse linear simultaneous

More information

Lecture 8: Fast Linear Solvers (Part 7)

Lecture 8: Fast Linear Solvers (Part 7) Lecture 8: Fast Linear Solvers (Part 7) 1 Modified Gram-Schmidt Process with Reorthogonalization Test Reorthogonalization If Av k 2 + δ v k+1 2 = Av k 2 to working precision. δ = 10 3 2 Householder Arnoldi

More information

1 Vectors. Notes for Bindel, Spring 2017 Numerical Analysis (CS 4220)

1 Vectors. Notes for Bindel, Spring 2017 Numerical Analysis (CS 4220) Notes for 2017-01-30 Most of mathematics is best learned by doing. Linear algebra is no exception. You have had a previous class in which you learned the basics of linear algebra, and you will have plenty

More information

Iterative Methods for Linear Systems of Equations

Iterative Methods for Linear Systems of Equations Iterative Methods for Linear Systems of Equations Projection methods (3) ITMAN PhD-course DTU 20-10-08 till 24-10-08 Martin van Gijzen 1 Delft University of Technology Overview day 4 Bi-Lanczos method

More information

A MULTIGRID ALGORITHM FOR. Richard E. Ewing and Jian Shen. Institute for Scientic Computation. Texas A&M University. College Station, Texas SUMMARY

A MULTIGRID ALGORITHM FOR. Richard E. Ewing and Jian Shen. Institute for Scientic Computation. Texas A&M University. College Station, Texas SUMMARY A MULTIGRID ALGORITHM FOR THE CELL-CENTERED FINITE DIFFERENCE SCHEME Richard E. Ewing and Jian Shen Institute for Scientic Computation Texas A&M University College Station, Texas SUMMARY In this article,

More information

The solution of the discretized incompressible Navier-Stokes equations with iterative methods

The solution of the discretized incompressible Navier-Stokes equations with iterative methods The solution of the discretized incompressible Navier-Stokes equations with iterative methods Report 93-54 C. Vuik Technische Universiteit Delft Delft University of Technology Faculteit der Technische

More information

Stabilization and Acceleration of Algebraic Multigrid Method

Stabilization and Acceleration of Algebraic Multigrid Method Stabilization and Acceleration of Algebraic Multigrid Method Recursive Projection Algorithm A. Jemcov J.P. Maruszewski Fluent Inc. October 24, 2006 Outline 1 Need for Algorithm Stabilization and Acceleration

More information

Computational Linear Algebra

Computational Linear Algebra Computational Linear Algebra PD Dr. rer. nat. habil. Ralf Peter Mundani Computation in Engineering / BGU Scientific Computing in Computer Science / INF Winter Term 2017/18 Part 2: Direct Methods PD Dr.

More information

Solving Large Nonlinear Sparse Systems

Solving Large Nonlinear Sparse Systems Solving Large Nonlinear Sparse Systems Fred W. Wubs and Jonas Thies Computational Mechanics & Numerical Mathematics University of Groningen, the Netherlands f.w.wubs@rug.nl Centre for Interdisciplinary

More information

A DISSERTATION. Extensions of the Conjugate Residual Method. by Tomohiro Sogabe. Presented to

A DISSERTATION. Extensions of the Conjugate Residual Method. by Tomohiro Sogabe. Presented to A DISSERTATION Extensions of the Conjugate Residual Method ( ) by Tomohiro Sogabe Presented to Department of Applied Physics, The University of Tokyo Contents 1 Introduction 1 2 Krylov subspace methods

More information

AMS526: Numerical Analysis I (Numerical Linear Algebra)

AMS526: Numerical Analysis I (Numerical Linear Algebra) AMS526: Numerical Analysis I (Numerical Linear Algebra) Lecture 23: GMRES and Other Krylov Subspace Methods Xiangmin Jiao SUNY Stony Brook Xiangmin Jiao Numerical Analysis I 1 / 9 Minimizing Residual CG

More information

DS-GA 1002 Lecture notes 0 Fall Linear Algebra. These notes provide a review of basic concepts in linear algebra.

DS-GA 1002 Lecture notes 0 Fall Linear Algebra. These notes provide a review of basic concepts in linear algebra. DS-GA 1002 Lecture notes 0 Fall 2016 Linear Algebra These notes provide a review of basic concepts in linear algebra. 1 Vector spaces You are no doubt familiar with vectors in R 2 or R 3, i.e. [ ] 1.1

More information

Solving Ax = b, an overview. Program

Solving Ax = b, an overview. Program Numerical Linear Algebra Improving iterative solvers: preconditioning, deflation, numerical software and parallelisation Gerard Sleijpen and Martin van Gijzen November 29, 27 Solving Ax = b, an overview

More information

ON ORTHOGONAL REDUCTION TO HESSENBERG FORM WITH SMALL BANDWIDTH

ON ORTHOGONAL REDUCTION TO HESSENBERG FORM WITH SMALL BANDWIDTH ON ORTHOGONAL REDUCTION TO HESSENBERG FORM WITH SMALL BANDWIDTH V. FABER, J. LIESEN, AND P. TICHÝ Abstract. Numerous algorithms in numerical linear algebra are based on the reduction of a given matrix

More information

A short course on: Preconditioned Krylov subspace methods. Yousef Saad University of Minnesota Dept. of Computer Science and Engineering

A short course on: Preconditioned Krylov subspace methods. Yousef Saad University of Minnesota Dept. of Computer Science and Engineering A short course on: Preconditioned Krylov subspace methods Yousef Saad University of Minnesota Dept. of Computer Science and Engineering Universite du Littoral, Jan 19-3, 25 Outline Part 1 Introd., discretization

More information

Lecture # 20 The Preconditioned Conjugate Gradient Method

Lecture # 20 The Preconditioned Conjugate Gradient Method Lecture # 20 The Preconditioned Conjugate Gradient Method We wish to solve Ax = b (1) A R n n is symmetric and positive definite (SPD). We then of n are being VERY LARGE, say, n = 10 6 or n = 10 7. Usually,

More information

Lecture 17: Iterative Methods and Sparse Linear Algebra

Lecture 17: Iterative Methods and Sparse Linear Algebra Lecture 17: Iterative Methods and Sparse Linear Algebra David Bindel 25 Mar 2014 Logistics HW 3 extended to Wednesday after break HW 4 should come out Monday after break Still need project description

More information

A JACOBI-DAVIDSON ITERATION METHOD FOR LINEAR EIGENVALUE PROBLEMS. GERARD L.G. SLEIJPEN y AND HENK A. VAN DER VORST y

A JACOBI-DAVIDSON ITERATION METHOD FOR LINEAR EIGENVALUE PROBLEMS. GERARD L.G. SLEIJPEN y AND HENK A. VAN DER VORST y A JACOBI-DAVIDSON ITERATION METHOD FOR LINEAR EIGENVALUE PROBLEMS GERARD L.G. SLEIJPEN y AND HENK A. VAN DER VORST y Abstract. In this paper we propose a new method for the iterative computation of a few

More information

On the Preconditioning of the Block Tridiagonal Linear System of Equations

On the Preconditioning of the Block Tridiagonal Linear System of Equations On the Preconditioning of the Block Tridiagonal Linear System of Equations Davod Khojasteh Salkuyeh Department of Mathematics, University of Mohaghegh Ardabili, PO Box 179, Ardabil, Iran E-mail: khojaste@umaacir

More information

total work bounds to approximately solve the linear systems are O(n); if = O(1): For class (4) we are within a small increasing factor of these bounds

total work bounds to approximately solve the linear systems are O(n); if = O(1): For class (4) we are within a small increasing factor of these bounds Ecient Approximate Solution of Sparse Linear Systems John H. Reif z Abstract. We consider the problem of approximate solution ex of a linear system Ax = b over the reals, such that kaex? bk kbk; for a

More information

LINEAR SYSTEMS (11) Intensive Computation

LINEAR SYSTEMS (11) Intensive Computation LINEAR SYSTEMS () Intensive Computation 27-8 prof. Annalisa Massini Viviana Arrigoni EXACT METHODS:. GAUSSIAN ELIMINATION. 2. CHOLESKY DECOMPOSITION. ITERATIVE METHODS:. JACOBI. 2. GAUSS-SEIDEL 2 CHOLESKY

More information

Computational Methods. Systems of Linear Equations

Computational Methods. Systems of Linear Equations Computational Methods Systems of Linear Equations Manfred Huber 2010 1 Systems of Equations Often a system model contains multiple variables (parameters) and contains multiple equations Multiple equations

More information

Bindel, Fall 2016 Matrix Computations (CS 6210) Notes for

Bindel, Fall 2016 Matrix Computations (CS 6210) Notes for 1 Iteration basics Notes for 2016-11-07 An iterative solver for Ax = b is produces a sequence of approximations x (k) x. We always stop after finitely many steps, based on some convergence criterion, e.g.

More information

Notes on PCG for Sparse Linear Systems

Notes on PCG for Sparse Linear Systems Notes on PCG for Sparse Linear Systems Luca Bergamaschi Department of Civil Environmental and Architectural Engineering University of Padova e-mail luca.bergamaschi@unipd.it webpage www.dmsa.unipd.it/

More information

EXAMPLES OF CLASSICAL ITERATIVE METHODS

EXAMPLES OF CLASSICAL ITERATIVE METHODS EXAMPLES OF CLASSICAL ITERATIVE METHODS In these lecture notes we revisit a few classical fixpoint iterations for the solution of the linear systems of equations. We focus on the algebraic and algorithmic

More information

Linear Algebra Massoud Malek

Linear Algebra Massoud Malek CSUEB Linear Algebra Massoud Malek Inner Product and Normed Space In all that follows, the n n identity matrix is denoted by I n, the n n zero matrix by Z n, and the zero vector by θ n An inner product

More information

1 Extrapolation: A Hint of Things to Come

1 Extrapolation: A Hint of Things to Come Notes for 2017-03-24 1 Extrapolation: A Hint of Things to Come Stationary iterations are simple. Methods like Jacobi or Gauss-Seidel are easy to program, and it s (relatively) easy to analyze their convergence.

More information

Further experiences with GMRESR

Further experiences with GMRESR Further experiences with GMRESR Report 92-2 C. Vui Technische Universiteit Delft Delft University of Technology Faculteit der Technische Wisunde en Informatica Faculty of Technical Mathematics and Informatics

More information

Review Questions REVIEW QUESTIONS 71

Review Questions REVIEW QUESTIONS 71 REVIEW QUESTIONS 71 MATLAB, is [42]. For a comprehensive treatment of error analysis and perturbation theory for linear systems and many other problems in linear algebra, see [126, 241]. An overview of

More information

Fundamentals of Engineering Analysis (650163)

Fundamentals of Engineering Analysis (650163) Philadelphia University Faculty of Engineering Communications and Electronics Engineering Fundamentals of Engineering Analysis (6563) Part Dr. Omar R Daoud Matrices: Introduction DEFINITION A matrix is

More information

Numerical Analysis Lecture Notes

Numerical Analysis Lecture Notes Numerical Analysis Lecture Notes Peter J Olver 8 Numerical Computation of Eigenvalues In this part, we discuss some practical methods for computing eigenvalues and eigenvectors of matrices Needless to

More information

DELFT UNIVERSITY OF TECHNOLOGY

DELFT UNIVERSITY OF TECHNOLOGY DELFT UNIVERSITY OF TECHNOLOGY REPORT 06-05 Solution of the incompressible Navier Stokes equations with preconditioned Krylov subspace methods M. ur Rehman, C. Vuik G. Segal ISSN 1389-6520 Reports of the

More information

Laboratoire d'informatique Fondamentale de Lille

Laboratoire d'informatique Fondamentale de Lille Laboratoire d'informatique Fondamentale de Lille Publication AS-181 Modied Krylov acceleration for parallel environments C. Le Calvez & Y. Saad February 1998 c LIFL USTL UNIVERSITE DES SCIENCES ET TECHNOLOGIES

More information