1. Introduction. We address the problem of solving a set of linear equations

Size: px
Start display at page:

Download "1. Introduction. We address the problem of solving a set of linear equations"

Transcription

1 SIAM J. SCI. COMPUT. Vol. 0, No. 0, pp c 20XX Society for Industrial and Applied Mathematics THE ITERATIVE SOLVER RISOLV WITH APPLICATION TO THE EXTERIOR HELMHOLTZ PROBLEM HILLEL TAL-EZER AND E. TURKEL Abstract. The innermost computational kernel of many large-scale scientific applications is often a large set of linear equations of the form Ax = b which typically consumes a significant portion of the overall computational time required by the simulation. The traditional approach for solving this problem is to use direct methods. This approach is often preferred in industry because direct solvers are robust and effective for moderate size problems. However, direct methods can consume a huge amount of memory, and CPU time, in large-scale cases. In these cases, iterative techniques are the only viable alternative. Unfortunately, iterative methods lack the robustness of direct methods. The situation is especially difficult when the matrix is nonsymmetric. A lot of research has been devoted to trying to develop a robust iterative algorithm for nonsymmetric systems. The present paper describes a new robust and efficient algorithm aimed at solving iteratively nonsymmetric linear systems. It is based on looking for an approximation to the optimal polynomial P m(z) which satisfies P m(z) =min Q Πm Q(z), z D, whereπ m is the set of all polynomials of degree m which satisfies Q m(0) = 1 and D is a domain in the complex plane which includes all the eigenvalues of A. The resulting algorithm is an efficient one, especially in the case where we have a set of linear systems which share the same matrix A. We present several applications, including the exterior Helmholtz problem, which leads to a large indefinite, nonsymmetric, and complex system. Key words. Risolv, Krylov, Helmholtz exterior, preconditioning AMS subject classifications. 15A15, 15A09, 15A23 DOI / X 1. Introduction. We address the problem of solving a set of linear equations (1.1) Ax = b, where A is a large, general N N matrix and b is an N 1 vector. In real-life problems, preconditioning is mandatory. In this paper we do not address this issue. When A is a symmetric, positive definite matrix, conjugate gradient [13] is shown to be the optimal algorithm. Similarly, there are optimal algorithms for the indefinite, symmetric case (e.g., MINRES, SYMMLQ [23]). The situation is significantly more complicated in the general, nonsymmetric case. A popular algorithm for this type of problem is the well-known GMRES algorithm [25], which is optimal in the following sense. Let x m 1 be the solution vector after applying m 1 matrix-vector multiplications: (1.2) x m 1 = x 0 + Q m 1 (A)r 0, where x 0, r 0 are the initial guess and residual, respectively. Hence, (1.3) r m 1 = b Ax m 1 = b A(x 0 + Q m 1 r 0 )=(I AQ m 1 (A)) r 0 = P m (A)r 0, Received by the editors May 19, 2008; accepted for publication (in revised form) August 5, 2009; published electronically DATE. School of Computer Science, Academic College of Tel-Aviv Yaffo, Tel Aviv 64044, Israel (hillel@ mta.ac.il). School of Mathematical Sciences, Tel Aviv University, Tel Aviv 69978, Israel (turkel@post.tau. ac.il). 1

2 2 HILLEL TAL-EZER AND E. TURKEL and it satisfies (1.4) r m 1 2 = min T Π m T (A)r 0 2. Π m is the set of polynomials, T,ofdegreem with T (0) = 1. The major disadvantage of the GMRES algorithm is that, in order to get the optimal solution, one has to store m vectors. For real-life problems, the algorithm poses stringent demands on computer memory and CPU time. In order to overcome this obstacle, the standard approach is to implement the so-called restarted GMRES. In this approach, one gives up optimality and restarts the algorithm after k (k m) matrix-vector multiplications. This tactic can result in a highly efficient algorithm, but, as frequently has been observed, the rate of convergence can deteriorate or even complete stagnation can occur. In the last 20 years, many studies have aimed to improve upon this drawback, for example, [22, 34, 6, 35, 20, 21]. Another family of methods aimed at solving nonsymmetric linear systems is based on the Lanczos biconjugate algorithm. These methods include QMR [9], TFQMR [10], and BICGSTAB [34]. The main advantage of these algorithms is the elimination of the need to store many vectors. On the other hand, since these methods do not incorporate minimization of the residual, erratic behavior of the error can be observed. Recently, methods which combine the biconjugate and GMRES approaches have been developed [27, 29]. A third approach is based on a polynomial approximation in the complex plane. Since, in a general polynomial approach, the residual vector satisfies (1.3), we look for P m (z) such that (1.5) P m (z) = min Q Π m Q(z), z D. D is a domain in the complex plane which includes all the eigenvalues of A. When D is a real interval, then P m is the scaled Chebyshev polynomial. Manteuffel [15, 16] had developed an algorithm which considers the situation when D is an ellipse in the complex plane. This algorithm had been generalized in [17]. In [28], the authors treat general domains in the complex plane where P m is expanded in Faber polynomials. The Risolv algorithm, presented in this paper, belongs to the third family of algorithms described above. It is based on efficiently computing the roots of P m (z) such that P m (z) is close to optimal. 2. Risolv. Although Risolv belongs to the family of methods based on polynomial approximation in the complex plane, it can also be considered as a slight modification of restarted GMRES. Applying GMRES(s), at the j restart, we have (2.1) x j+1 = x j + V j y j, where V j is an N s matrix whose columns constitute an orthonormal basis of the Krylov space (2.2) K s = Span{r j,ar j,...,a s 1 r j }, and y j is chosen such that r j+1 2 is minimized. It can be shown that r j+1 satisfies s ( (2.3) r j+1 = I 1 ) A r j, z ij i=1

3 RISOLV HELMHOLTZ EQUATION 3 where z 1j,...,z sj are harmonic Ritz values (the eigenvalues of the modified Hessenberg matrix) [11]. Hence, after n restarts we have n s ( (2.4) r m = I 1 ) A r 0 (m = ns). z ij j=1 i=1 The difference between Risolv and restarted GMRES is in the choice of the vector y j. For Risolv, y j is chosen so that z 1j,...,z sj are uniformly (not equally) distributed in D. This can be achieved by employing the Leja algorithm [24] as follows. Let H be the Hessenberg matrix that satisfies (2.5) AV = V H = VH+ h s+1,s v s+1 e H s, and Ĥ is the modified Hessenberg defined as (2.6) Ĥ = H + h 2 s+1,sfe H s, where f = H H e s and e s =[0,...,1] H. At each restart we have in our possession a set of 2s matrices whose eigenvalues can be considered as approximated points in the domain D. These matrices are H i, Ĥi, 1 i s. Since each matrix is of dimension i i, all together we have a set of s 2 points. From this set we choose, using the Leja algorithm, a set of s points uniformly distributed in D (taking into account the already chosen sj points). Hence, after n restarts, the residual is the same as (2.4), but now the set of points (2.7) {z ij } 1 i s,1 j n is almost uniformly distributed in D. By (2.3), we have (2.8) r j+1 = Vy j, (2.9) x j+1 = x j + Vμ j, where y j and μ j are vectors of length s +1ands, respectively. These vectors can be computed by the following algorithm. NewVectors. μ =[0,...,0] H,y=[1, 0,...,0] H, for i =1:s w = 1 z i ȳ (ȳ i = y i, 1 i s) μ = μ + w y = y Hw end Based on the above algorithm the Risolv algorithm is described as follows. Risolv algorithm. r 0 = b Ax 0,z=emptyset, tol for j = 0 : until satisfied [ V, H] = Arnoldi(r j ) z new =Leja( H,z) z = z z new [μ j,y j ] = NewVectors( H,z new ) x j+1 = x j + Vμ j r j+1 = Vy j if r j+1 2 tol b 2, break end

4 4 HILLEL TAL-EZER AND E. TURKEL Fig. 1. Eigenvalue distribution of matrix A, (3.1). Risolv is especially efficient for solving sets of linear systems of the form (2.10) Ax i = b i, 1 i k. After solving first system, we have a set of points which is almost uniformly distributed, and Risolv becomes essentially a simple Richardson algorithm: (2.11) x k+1 = x k + 1 w k (b Ax k ), where w k is a point in this set. This stage is free of inner products. If needed, a few additional matrix-vector multiplications will reduce the error to the desired accuracy. 3. Numerical examples. In this section we consider a variety of numerical problems and compare the results shown by Risolv and GMRES. All the computations described below were done in MATLAB. Example 1. We consider a modified Grcar matrix (3.1) A = and the right-hand-side vector b =[1,...,1] T. The exact eigenvalues of A, ascomputed by MATLAB, are shown in Figure 1. We solved this system both with restarted GMRES and Risolv. The size of the Krylov space is 20 and the desired accuracy is ItrequiredGMRESandRISOLV 3226 and 1195 matrix-vector multiplications, respectively, to reach an approximate solution of the desired accuracy, The set z ij (2.3) for the two algorithms is given in Figures 2(a) and 2(b). The modified Grcar matrix is a highly nonnormal matrix. Its nonnormality is in-

5 RISOLV HELMHOLTZ EQUATION 5 (a) GMRES (b) RISOLV Fig. 2. z ij for matrix (3.1). Fig. 3. Eigenvalue distribution of matrix B, (3.2). creased by adding upper diagonals. Adding upper diagonals to (3.1) yields (3.2) B = The exact eigenvalues of (3.2) are shown in Figure 3. The set {z ij } for (3.2) for GMRES and RISOLV is given in Figures 4(a) and 4(b). Due to the fact that now the pseudospectra of the matrix almost surround the origin, solving the linear system iteratively becomes a much more difficult task. While

6 6 HILLEL TAL-EZER AND E. TURKEL (a) GMRES (b) RISOLV Fig. 4. z ij for matrix (3.2). Table 1 Matrices from Matrix-Market. Matrix Risolv GMRES mat-vecs mat-vecs ADD MEMPLUS ORSIRR ORSIRR ORSREG RAEFSKY SAYLR GMRES stagnated completely, it took 1400 matrix-vector multiplications for RISOLV to reach an accuracy of Example 2. We next present several linear systems with matrices taken from Matrix-Market [18]. In Table 1 we compare the number of matrix-vector multiplies required for RISOLV and GMRES to compute solutions with a relative accuracy of 10 4 with right-hand-side vector [1,...,1] T. Example 3. We now demonstrate the high efficiency of RISOLV when several right-hand sides are given. Hence, we consider the following: (3.3) Ax i = b i, 1 i s. In this experiment we solve 10 linear systems. A is the ORSIRR1 matrix and b i are random vectors. The results are shown in Table 2. We see that the number of matrix multiplies is almost the same for each right-hand side. However, the number of inner products reduces substantially for the second right-hand side and becomes negligible for further solutions of the system with different right-hand sides. Example 4. Finally, this example is for a convection-diffusion problem taken from SPARSKIT2 of Y. Saad. The grid is shown in Figure 5. In this case N(size of the matrix) = and NNZ (number of nonzeros) = We used ILUT as the preconditioner. A drop tolerance of 0.1 was good enough for RISOLV yielding

7 RISOLV HELMHOLTZ EQUATION 7 Table 2 Operation count for problem with multiple right-hand sides. No. system Mat-vecs Inn-prod Fig. 5. Grid for convection-diffusion problem. nonzeros in the ILU matrices. A drop tolerance of was needed for GMRES resulting in nonzeros in the ILU matrices Exterior Helmholtz problem. We now consider the Helmholtz equation for scattering exterior to a body. This leads to a difficult indefinite, nonsymmetric and complex equation (3.4) Δu + k 2 u =0. Since we consider the exterior of a body it is necessary to truncate the infinite domain and introduce an artificial outer surface. On this surface we introduce boundary conditions that reduce reflections back into the physical domain. With this boundary condition the solution is complex. The resultant discretized system is not positive definite and is not self-adjoint. Several recent surveys discuss the difficulties in numerically solving the Helmholtz equation exterior to a body; see, for example, [12, 30, 31, 33]. We consider two dimensional scattering about a general boundary. The total field is given by u tot = u scat + u inc,withu inc given by a plane wave. Then [5] (3.5) Δu + k 2 u = 0 exterior to D on D, u = u scat + e ik x d (soft body) u or n = n (uscat + e ik x d ) (hard body), ( u scat ) (x) lim r iku scat (x) = 0 (Sommerfeld radiation condition). r r

8 8 HILLEL TAL-EZER AND E. TURKEL We discretize the equation in the interior of D by a standard linear finite element method in polar coordinates. To approximate the Sommerfeld radiation condition we truncate the domain by introducing an artificial elliptical surface. On this artificial surface we impose the absorbing boundary condition of Kriegsmann, Taflove, and Umashankar [14] and Medvinsky, Turkel, and Hetmaniuk [19], which is an extension of BGT [3]. This has been modified to be in divergence-free form [1, 19, 33]: (3.6) u κu = iku n 2 κ2 u 8(ik κ) s ( 1 2(ik κ) We shall consider various Krylov subspace iterations to solve the resultant linear matrix problem. It is well known that Krylov methods converge slowly unless the system is preconditioned. Two types of preconditioners have been developed [33]. In the first kind, the preconditioner utilizes algebraic properties of the matrix. A typical example is ILU. A different class of preconditioners is based on properties of the differential system, i.e., an operator-based preconditioner. An example of this is basing the preconditioner on the solution of the Laplace equation in Bayliss, Goldstein, and Turkel [4]. This works well for low wavenumbers but becomes less efficient as k increases. Erlangga, Vuik, and Oosterlee [7] and Erlangga [8] extended this idea by considering a Helmholtz equation with a complex wavenumber as the preconditioner. One still has to solve a system the same size as the original system. So, the modified Helmholtz equation needs to be also solved by an inner iterative method. We shall also solve the inner preconditioned problem by a Krylov space technique. A complex k introduces dissipation which makes the preconditioned system easier to solve than the original system with a real k. Furthermore, we have a reasonable initial guess for the inner iteration. Hence, we consider a preconditioner which is an approximate solution to (3.7) Δv + k 2 precv =0, k 2 prec =(1+iα)k 2. As a model case we consider scattering about a hard ellipse (natural boundary condition at the scatterer) with an additional tower to resemble a submarine which has an aspect ratio of 10 while the outer surface is an ellipse with a semimajor axis of 1.2 and a semiminor axis of 0.6. A typical grid is shown in Figure 6. The absorbing boundary condition (3.6) is imposed on the outer artificial ellipse. The outer boundary fits snugly about the scatterer. Because of the high aspect ratio of the scatterer and the close position of the outer boundary this is a difficult problem for a solver. For all the iterative solvers we reduce the residual by seven orders of magnitude. We have several layers of iterations. The outer loop is based on solving the original scattering problem with a Krylov method. This is preconditioned by the approximate solution of a modified Helmholtz problem with a complex wavenumber. The initial guess of this outer iteration is based on an ILU(0) decomposition of the modified Helmholtz equation. The modified Helmholtz problem is again solved by a Krylov method. For some of the methods we shall also investigate the use of ILU(0) of the modified Helmholtz equation as the preconditioner. This is not a good preconditioner but has the advantage of not being a variable preconditioner. We will investigate the optimal value of the imaginary portion of k and check the sensitivity to this value. In [32] we presented results using BICGSTABwithanILU(0)preconditioning to solve the preconditioned problem (3.7). For the more difficult cases considered here this did not always converge. Furthermore, the convergence is very sensitive to the mesh. Instead, we first consider the Helmholtz equation solved by BICGSTAB u s ).

9 RISOLV HELMHOLTZ EQUATION 9 Fig. 6. Submarine and grid. but with the modified Helmholtz equation solved by 20 iterations of GMRES with kprec 2 =(1+.3i)k 2. We use ILU(0) as the preconditioning for the inner problem. Again, the number of iterations is very sensitive to the details of the wavenumber and corresponding mesh size. For k =40anda mesh it did not converge. Similarly for the finer meshes this did not converge. This lack of robustness makes this approach less feasible. In summary, when the Helmholtz equation is solved by BICGSTAB we found the iterative method very sensitive whether we used BICGSTAB or GMRES to solve the preconditioned modified Helmholtz equation. Instead we use a restarted GMRES method for the original Helmholtz equation. The inner loop is solved by GMRES with kprec 2 =(1+.5i)k2. GMRES essentially builds a polynomial, P m (z), in the matrix. With each restart we get a different polynomial. Thus, we have a variable preconditioner, and so we use flexible GMRES (FGMRES) to solve the Helmholtz equation rather than GMRES. A maximum of 20 iterations is allowed in the solution of the modified Helmholtz equation with the FGMRES restarted after 10 iterations. In Table 3 we present the results. Interestingly, the number of iterations required to reduce the residual by seven orders of magnitude decreases as the wavenumber increases. For fixed k as we refine the mesh we require more outer iterations than when the modified Helmholtz equation was solved exactly. So solving the inner problem by GMRES reduces the effectiveness of the preconditioner. Also, for a fixed wavenumber the number of iterations required increases with the size of the mesh. Nevertheless, it is more efficient and more robust than the BICGSTAB solver. Finally, we consider Risolv for the Helmholtz equation with a preconditioner based on the modified Helmholtz equation (3.7). Since Risolv, like GMRES, does not allow for a variable preconditioner we instead construct the system AP 1 v = b, Pu = v, wherea corresponds to the original Helmholtz equation (3.5) and P refers to the ILU(0) factorization of the modified Helmholtz equation (3.7). Since we are not solving the full modified Helmholtz equation this is a poorer preconditioner than we used for FGMRES. The convergence rate is fairly insensitive to the imaginary part of k prec. This is despite that the eigenvalue distributions of the various AP 1 differ considerably. One of the advantages of Risolv is that it automatically outputs a set of points distributed in the vicinity of the boundaries of the spectra (or pseudospectra in the nonnormal case) of AP 1. In Figure 7(a) we display this set when k prec = k.

10 10 HILLEL TAL-EZER AND E. TURKEL Table 3 Convergence rate FGMRES; modified problem solved by GMRES. k Mesh # outer iterations # total iterations Time (a) k prec = k (b) k 2 prec =(1+i)k 2 Fig. 7. Effect of k prec on eigenvalues, k =40, grid, ILU (0). In this case the points are almost all real. In Figure 7(b) we plot the set of points when k 2 prec =(1+i)k 2 and see the difference in the distribution. If we choose P 1 as the ILU(0) decomposition of the original equation (3.4), then we save storage and the convergence is affected only mildly. However, we lose robustness since the ILU(0) of the original matrix may not exist. In Table 4 we present the results. We see that the computer times are slightly larger than with FGMRES. Thus, when computing a single case it is preferable to solve the Helmholtz equation and the preconditioner by FGMRES. We again stress that with Risolv we are using a poorer preconditioner than with FGMRES. As described before, a significant advantage of Risolv is its efficiency when subsequent problems are solved with different right-hand sides. This would occur if one is solving a scattering problem with incident waves at various angles. Alternatively, one can consider the Helmholtz equation with various wavenumbers k. Avery, Farhat, and

11 RISOLV HELMHOLTZ EQUATION 11 Table 4 Convergence rate for Risolv solving modified problem by ILU (0). k k prec k Mesh # matrix vector multiplies Time 1 1+i i i i i i i i i i i i i i i i NC NC 40 1+i i Table 5 Comparison of FGMRES with Risolv for multiple right-hand sides. k Scheme k Grid Preconditioner prec k First time Later times FGMRES mod Helm 1+.5i FGMRES ILU(0) 1+.5i Risolv ILU(0) Risolv ILU(0) 1+.2i Risolv ILU(0) 1+.5i Risolv ILU(0) 1+i Risolv ILU(0) Risolv ILU(0) 1+.6i Risolv ILU(0) 1+i Risolv ILU(0) 1i NC NC Risolv ILU(0) 1+.6i Risolv ILU(0) 1+i Reese [2] demonstrate how to reformulate the problem using a Padé expansion. This requires the solution of the Helmholtz equation with a single frequency but multiple right-hand sides. So the CPU time information is misleading since Risolv spends time on computing optimal points in the domain of eigenvalues. These computations do not depend on N (the size of the problem) but on k (the dimension of the Krylov space). Hence, for large N, this part of Risolv is negligible. In Table 5 we compare FGMRES with Risolv when we compute several right-hand sides. In particular we chose incident plane waves at angles of 0, 10, 20, 30. The later times column reflects an average over the later three angles. It contains both differences in convergence rates for various angles of incidence and the effect of the algorithm in computing multiple right-hand sides. For FGMRES we see that using the full modified Helmholtz equation is faster than using than the simplified precon-

12 12 HILLEL TAL-EZER AND E. TURKEL ditioner of the ILU(0) decomposition of the modified Helmholtz equation. For Risolv we cannot use a variable preconditioner, and so we choose only the preconditioner as the ILU(0) decomposition. Nevertheless, depending on k prec Risolv is as efficient or slightly worse than FGMRES for the first angle. For subsequent angles Risolv is almost twice as fast. Note that for this case k prec = k is optimal and we do not need to compute and store the matrix for the modified Helmholtz equation. However, in general the ILU(0) decomposition of the original Helmholtz equation may not exist. For example, for k =40andk prec = k Risolv did not converge because the ILU(0) preconditioner did not exist. This occurred because the mesh was too coarse to resolve this wavenumber. Increasing the imaginary component to k prec =(1+.5i)k the ILU(0) decomposition exists but is a poor preconditioner as seen from the eigenvalue distribution (not shown). Further increasing it to k prec =(1+i)k gives an improved preconditioner and Risolv converges fast. Hence, as seen from the table, for robustness one might wish to choose kprec k =1+i. If the waves are resolved, then k prec = k should be sufficient. 4. Conclusion. We have constructed a new iterative solver, Risolv, based on polynomial approximation in the complex plane. We have shown the advantages of this algorithm for a range of problems. In addition to several matrices from standard libraries we have considered scattering about a submarine-like object using the Helmholtz equation and a suitable preconditioner. This problem involves complex numbers, is nonpositive, and is nonself-adjoint. Risolv is particularly effective when multiple solutions are required with different right-hand sides. A by-product of the algorithm is that it automatically outputs a set of points distributed in the vicinity of the boundaries of the spectra (or pseudospectra in the nonnormal case) of AP 1. REFERENCES [1] X. Antoine, Fast approximate computation of a time-harmonic scattered field using the onsurface radiation condition method, IMA J. Appl. Math., 66 (2001), pp [2] A. Avery, C. Farhat, and G. Reese, Fast frequency sweep computations using a multi-point Padé-based reconstruction method and an efficient iterative solver, Internat. J. Numer. Methods Engrg., 69 (2007), pp [3] A. Bayliss, M. Gunzburger, and E. Turkel, Boundary conditions for the numerical solution of elliptic equations in exterior regions, SIAM J. Appl. Math., 42 (1982) pp [4] A. Bayliss, C. I. Goldstein, and E. Turkel, An iterative method for the Helmholtz equation, J. Comput. Phys., 49 (1983), pp [5] D. Colton and R. Kress, Inverse Acoustic and Electromagnetic Scattering Theory, 2nded., Springer-Verlag, Berlin, [6] E. De Sturler and D. R. Fokkema, Nested Krylov Methods and Preserving the Orthogonality, Tech. Rep. Preprint 796, Utrecht University, Utrecht, The Netherlands, [7] Y. A. Erlangga, C. Vuik, and C. W. Oosterlee, On a class of preconditioners for the Helmholtz equation, Appl. Numer. Math., 50 (2004), pp [8] Y. A. Erlangga, Advances in iterative methods and preconditioners for the Helmholtz equation, Arch. Comput. Methods Engrg., 15 (2008), pp [9] R. W. Freund and N. Nachtigal, QMR: A quasi-minimal residual method for non-hermitian linear systems, Numer. Math., 60 (1991), pp [10] R. W. Freund, A transpose-free quasi-minimum residual algorithm for non-hermitian linear systems, SIAM J. Sci. Comput., 14 (1993), pp [11] S. Goossens and D. Roose, Ritz and harmonic Ritz values and the convergence of FOM and GMRES, Numer. Linear Algebra Appl., 6 (1999), pp [12] I. Harari, A survey of finite element methods for time-harmonic acoustics, Comput. Methods Appl. Mech. Engrg., 195 (2006), pp [13] M. R. Hestenes and E. Stiefel, Methods of conjugate gradients for solving linear systems, J. Research Nat. Bur. Standards, 49 (1954), pp

13 RISOLV HELMHOLTZ EQUATION 13 [14] G. A. Kriegsmann, A. Taflove, and K. R. Umashankar, A new formulation of electromagnetic scattering using on surface radiation condition approach, IEEE Trans. Antennas and Propagation, 35 (1987), pp [15] T. A. Manteuffel, The Tchebychev iteration for nonsymmetric linear systems, Numer. Math., 28 (1977), pp [16] T. A. Manteuffel, Adaptive procedure for estimation of parameters for the nonsymmetric Tchebychev iteration, Numer. Math., 31 (1978), pp [17] T. A. Manteuffel and G. Starke, On hybrid iterative methods for nonsymmetric systems of linear equations, Numer. Math., 73 (1996), pp [18] Matrix-Market, A Collection of 500 Sparse Matrices from a Variety of Applications, math.nist.gov/matrixmarket/. [19] M. Medvinsky, E. Turkel, and U. Hetmaniuk, Local absorbing boundary conditions for elliptical shaped boundaries, J. Comput. Phys., 227 (2008), pp [20] R. B. Morgan, Implicitly restarted GMRES and Arnoldi methods for nonsymmetric systems of equations, SIAM J. Matrix Anal. Appl., 21 (2000), pp [21] R. B. Morgan, A restarted GMRES method augmented with eigenvectors, SIAMJ.Matrix Anal. Appl., 16 (1995), pp [22] N. M. Nachtigal, L. Reichel, and L. N. Trefethen, A hybrid GMRES algorithm for nonsymmetric linear systems, SIAM J. Matrix Anal. Appl., 13 (1992), pp [23] C. C. Paige and M. A. Saunders, Solution of sparse indefinite systems of linear equations, SIAM J. Numer. Anal., 12 (1975), pp [24] L. Reichel, Newton interpolation at Leja points, BIT, 30 (1990), pp [25] Y. Saad and M. H. Schultz, GMRES: A generalized minimal residual algorithm for solving nonsymmetric linear systems, SIAM J. Sci. Stat. Comput., 7 (1986), pp [26] Y. Saad, Iterative Methods for Sparse Linear Systems, 2nd ed., SIAM, Philadelphia, [27] G. L. G. Sleijpen, H. A. van Der Vorst, and D. R. Fokkema, BiCGSTAB(l) and other hybrid Bi-CG methods, Numer. Algorithms, 7 (1994), pp [28] G. Starke and R. S. Varga, A hybrid Arnoldi Faber iterative method for nonsymmetric systems of linear equations, Numer. Math., 64 (1993), pp [29] P. Sonneveld and M. B. Van Gijzen, IDR(s): A family of simple and fast algorithms for solving large nonsymmetric systems of linear equations, SIAM J. Sci. Comput., 31 (2008), pp [30] L. L. Thompson, A review of finite-element methods for time-harmonic acoustics, J. Acoust. Soc. Am., 119 (2006), pp [31] E. Turkel, Numerical difficulties solving time harmonic equations, in Multiscale Computational Methods in Chemistry and Physics, A. Brandt, J. Bernholc, and K. Binder, eds., IOS Press, Amsterdam, Ohmsha, Tokyo, 2001, pp [32] E. Turkel and Y. Erlangga, Preconditioning a finite element solver of the exterior Helmholtz equation, in Proceedings of ECCOMAS CFD 2006, P. Wesseling, E. Omate, and J. Priaux, eds., Egmond aan Zee, The Netherlands, [33] E. Turkel, Boundary conditions and iterative schemes for the Helmholtz equation in unbounded regions, in Computational Methods for Acoustics Problems, F. Magoules, ed., Saxe Coburg, Stirling, UK, [34] H. A. van der Vorst, Bi-CGSTAB: A fast and smoothly convergent variant of Bi-CG for the solution of nonsymmetric linear systems, SIAM J. Sci. Statist. Comput., 13 (1992), pp [35] C. Vuik, A comparison of some GMRES-like methods, in Iterative Methods in Linear Algebra, North Holland, Amsterdam, 1992, pp

Key words. linear equations, polynomial preconditioning, nonsymmetric Lanczos, BiCGStab, IDR

Key words. linear equations, polynomial preconditioning, nonsymmetric Lanczos, BiCGStab, IDR POLYNOMIAL PRECONDITIONED BICGSTAB AND IDR JENNIFER A. LOE AND RONALD B. MORGAN Abstract. Polynomial preconditioning is applied to the nonsymmetric Lanczos methods BiCGStab and IDR for solving large nonsymmetric

More information

DELFT UNIVERSITY OF TECHNOLOGY

DELFT UNIVERSITY OF TECHNOLOGY DELFT UNIVERSITY OF TECHNOLOGY REPORT 16-02 The Induced Dimension Reduction method applied to convection-diffusion-reaction problems R. Astudillo and M. B. van Gijzen ISSN 1389-6520 Reports of the Delft

More information

Iterative Methods for Linear Systems of Equations

Iterative Methods for Linear Systems of Equations Iterative Methods for Linear Systems of Equations Projection methods (3) ITMAN PhD-course DTU 20-10-08 till 24-10-08 Martin van Gijzen 1 Delft University of Technology Overview day 4 Bi-Lanczos method

More information

A PRECONDITIONER FOR THE HELMHOLTZ EQUATION WITH PERFECTLY MATCHED LAYER

A PRECONDITIONER FOR THE HELMHOLTZ EQUATION WITH PERFECTLY MATCHED LAYER European Conference on Computational Fluid Dynamics ECCOMAS CFD 2006 P. Wesseling, E. Oñate and J. Périaux (Eds) c TU Delft, The Netherlands, 2006 A PRECONDITIONER FOR THE HELMHOLTZ EQUATION WITH PERFECTLY

More information

IDR(s) A family of simple and fast algorithms for solving large nonsymmetric systems of linear equations

IDR(s) A family of simple and fast algorithms for solving large nonsymmetric systems of linear equations IDR(s) A family of simple and fast algorithms for solving large nonsymmetric systems of linear equations Harrachov 2007 Peter Sonneveld en Martin van Gijzen August 24, 2007 1 Delft University of Technology

More information

Preface to the Second Edition. Preface to the First Edition

Preface to the Second Edition. Preface to the First Edition n page v Preface to the Second Edition Preface to the First Edition xiii xvii 1 Background in Linear Algebra 1 1.1 Matrices................................. 1 1.2 Square Matrices and Eigenvalues....................

More information

The amount of work to construct each new guess from the previous one should be a small multiple of the number of nonzeros in A.

The amount of work to construct each new guess from the previous one should be a small multiple of the number of nonzeros in A. AMSC/CMSC 661 Scientific Computing II Spring 2005 Solution of Sparse Linear Systems Part 2: Iterative methods Dianne P. O Leary c 2005 Solving Sparse Linear Systems: Iterative methods The plan: Iterative

More information

AMS526: Numerical Analysis I (Numerical Linear Algebra) Lecture 23: GMRES and Other Krylov Subspace Methods; Preconditioning

AMS526: Numerical Analysis I (Numerical Linear Algebra) Lecture 23: GMRES and Other Krylov Subspace Methods; Preconditioning AMS526: Numerical Analysis I (Numerical Linear Algebra) Lecture 23: GMRES and Other Krylov Subspace Methods; Preconditioning Xiangmin Jiao SUNY Stony Brook Xiangmin Jiao Numerical Analysis I 1 / 18 Outline

More information

RESIDUAL SMOOTHING AND PEAK/PLATEAU BEHAVIOR IN KRYLOV SUBSPACE METHODS

RESIDUAL SMOOTHING AND PEAK/PLATEAU BEHAVIOR IN KRYLOV SUBSPACE METHODS RESIDUAL SMOOTHING AND PEAK/PLATEAU BEHAVIOR IN KRYLOV SUBSPACE METHODS HOMER F. WALKER Abstract. Recent results on residual smoothing are reviewed, and it is observed that certain of these are equivalent

More information

Incomplete LU Preconditioning and Error Compensation Strategies for Sparse Matrices

Incomplete LU Preconditioning and Error Compensation Strategies for Sparse Matrices Incomplete LU Preconditioning and Error Compensation Strategies for Sparse Matrices Eun-Joo Lee Department of Computer Science, East Stroudsburg University of Pennsylvania, 327 Science and Technology Center,

More information

The solution of the discretized incompressible Navier-Stokes equations with iterative methods

The solution of the discretized incompressible Navier-Stokes equations with iterative methods The solution of the discretized incompressible Navier-Stokes equations with iterative methods Report 93-54 C. Vuik Technische Universiteit Delft Delft University of Technology Faculteit der Technische

More information

Key words. linear equations, eigenvalues, polynomial preconditioning, GMRES, GMRES-DR, deflation, QCD

Key words. linear equations, eigenvalues, polynomial preconditioning, GMRES, GMRES-DR, deflation, QCD 1 POLYNOMIAL PRECONDITIONED GMRES AND GMRES-DR QUAN LIU, RONALD B. MORGAN, AND WALTER WILCOX Abstract. We look at solving large nonsymmetric systems of linear equations using polynomial preconditioned

More information

Solving Sparse Linear Systems: Iterative methods

Solving Sparse Linear Systems: Iterative methods Scientific Computing with Case Studies SIAM Press, 2009 http://www.cs.umd.edu/users/oleary/sccs Lecture Notes for Unit VII Sparse Matrix Computations Part 2: Iterative Methods Dianne P. O Leary c 2008,2010

More information

Solving Sparse Linear Systems: Iterative methods

Solving Sparse Linear Systems: Iterative methods Scientific Computing with Case Studies SIAM Press, 2009 http://www.cs.umd.edu/users/oleary/sccswebpage Lecture Notes for Unit VII Sparse Matrix Computations Part 2: Iterative Methods Dianne P. O Leary

More information

Fast iterative solvers

Fast iterative solvers Utrecht, 15 november 2017 Fast iterative solvers Gerard Sleijpen Department of Mathematics http://www.staff.science.uu.nl/ sleij101/ Review. To simplify, take x 0 = 0, assume b 2 = 1. Solving Ax = b for

More information

Conjugate gradient method. Descent method. Conjugate search direction. Conjugate Gradient Algorithm (294)

Conjugate gradient method. Descent method. Conjugate search direction. Conjugate Gradient Algorithm (294) Conjugate gradient method Descent method Hestenes, Stiefel 1952 For A N N SPD In exact arithmetic, solves in N steps In real arithmetic No guaranteed stopping Often converges in many fewer than N steps

More information

ITERATIVE METHODS FOR SPARSE LINEAR SYSTEMS

ITERATIVE METHODS FOR SPARSE LINEAR SYSTEMS ITERATIVE METHODS FOR SPARSE LINEAR SYSTEMS YOUSEF SAAD University of Minnesota PWS PUBLISHING COMPANY I(T)P An International Thomson Publishing Company BOSTON ALBANY BONN CINCINNATI DETROIT LONDON MADRID

More information

A decade of fast and robust Helmholtz solvers

A decade of fast and robust Helmholtz solvers A decade of fast and robust Helmholtz solvers Werkgemeenschap Scientific Computing Spring meeting Kees Vuik May 11th, 212 1 Delft University of Technology Contents Introduction Preconditioning (22-28)

More information

Topics. The CG Algorithm Algorithmic Options CG s Two Main Convergence Theorems

Topics. The CG Algorithm Algorithmic Options CG s Two Main Convergence Theorems Topics The CG Algorithm Algorithmic Options CG s Two Main Convergence Theorems What about non-spd systems? Methods requiring small history Methods requiring large history Summary of solvers 1 / 52 Conjugate

More information

Summary of Iterative Methods for Non-symmetric Linear Equations That Are Related to the Conjugate Gradient (CG) Method

Summary of Iterative Methods for Non-symmetric Linear Equations That Are Related to the Conjugate Gradient (CG) Method Summary of Iterative Methods for Non-symmetric Linear Equations That Are Related to the Conjugate Gradient (CG) Method Leslie Foster 11-5-2012 We will discuss the FOM (full orthogonalization method), CG,

More information

TWO-GRID DEFLATED KRYLOV METHODS FOR LINEAR EQUATIONS

TWO-GRID DEFLATED KRYLOV METHODS FOR LINEAR EQUATIONS 1 TWO-GRID DEFLATED KRYLOV METHODS FOR LINEAR EQUATIONS RONALD B. MORGAN AND ZHAO YANG Abstract. An approach is given for solving large linear systems that combines Krylov methods with use of two different

More information

Nested Krylov methods for shifted linear systems

Nested Krylov methods for shifted linear systems Nested Krylov methods for shifted linear systems M. Baumann, and M. B. van Gizen Email: M.M.Baumann@tudelft.nl Delft Institute of Applied Mathematics Delft University of Technology Delft, The Netherlands

More information

Comparison of Fixed Point Methods and Krylov Subspace Methods Solving Convection-Diffusion Equations

Comparison of Fixed Point Methods and Krylov Subspace Methods Solving Convection-Diffusion Equations American Journal of Computational Mathematics, 5, 5, 3-6 Published Online June 5 in SciRes. http://www.scirp.org/journal/ajcm http://dx.doi.org/.436/ajcm.5.5 Comparison of Fixed Point Methods and Krylov

More information

Spectral analysis of complex shifted-laplace preconditioners for the Helmholtz equation

Spectral analysis of complex shifted-laplace preconditioners for the Helmholtz equation Spectral analysis of complex shifted-laplace preconditioners for the Helmholtz equation C. Vuik, Y.A. Erlangga, M.B. van Gijzen, and C.W. Oosterlee Delft Institute of Applied Mathematics c.vuik@tudelft.nl

More information

Solving Symmetric Indefinite Systems with Symmetric Positive Definite Preconditioners

Solving Symmetric Indefinite Systems with Symmetric Positive Definite Preconditioners Solving Symmetric Indefinite Systems with Symmetric Positive Definite Preconditioners Eugene Vecharynski 1 Andrew Knyazev 2 1 Department of Computer Science and Engineering University of Minnesota 2 Department

More information

Performance Evaluation of GPBiCGSafe Method without Reverse-Ordered Recurrence for Realistic Problems

Performance Evaluation of GPBiCGSafe Method without Reverse-Ordered Recurrence for Realistic Problems Performance Evaluation of GPBiCGSafe Method without Reverse-Ordered Recurrence for Realistic Problems Seiji Fujino, Takashi Sekimoto Abstract GPBiCG method is an attractive iterative method for the solution

More information

Iterative methods for Linear System of Equations. Joint Advanced Student School (JASS-2009)

Iterative methods for Linear System of Equations. Joint Advanced Student School (JASS-2009) Iterative methods for Linear System of Equations Joint Advanced Student School (JASS-2009) Course #2: Numerical Simulation - from Models to Software Introduction In numerical simulation, Partial Differential

More information

Iterative Methods for Sparse Linear Systems

Iterative Methods for Sparse Linear Systems Iterative Methods for Sparse Linear Systems Luca Bergamaschi e-mail: berga@dmsa.unipd.it - http://www.dmsa.unipd.it/ berga Department of Mathematical Methods and Models for Scientific Applications University

More information

A Domain Decomposition Based Jacobi-Davidson Algorithm for Quantum Dot Simulation

A Domain Decomposition Based Jacobi-Davidson Algorithm for Quantum Dot Simulation A Domain Decomposition Based Jacobi-Davidson Algorithm for Quantum Dot Simulation Tao Zhao 1, Feng-Nan Hwang 2 and Xiao-Chuan Cai 3 Abstract In this paper, we develop an overlapping domain decomposition

More information

M.A. Botchev. September 5, 2014

M.A. Botchev. September 5, 2014 Rome-Moscow school of Matrix Methods and Applied Linear Algebra 2014 A short introduction to Krylov subspaces for linear systems, matrix functions and inexact Newton methods. Plan and exercises. M.A. Botchev

More information

FEM and sparse linear system solving

FEM and sparse linear system solving FEM & sparse linear system solving, Lecture 9, Nov 19, 2017 1/36 Lecture 9, Nov 17, 2017: Krylov space methods http://people.inf.ethz.ch/arbenz/fem17 Peter Arbenz Computer Science Department, ETH Zürich

More information

DELFT UNIVERSITY OF TECHNOLOGY

DELFT UNIVERSITY OF TECHNOLOGY DELFT UNIVERSITY OF TECHNOLOGY REPORT 14-1 Nested Krylov methods for shifted linear systems M. Baumann and M. B. van Gizen ISSN 1389-652 Reports of the Delft Institute of Applied Mathematics Delft 214

More information

Applied Mathematics 205. Unit V: Eigenvalue Problems. Lecturer: Dr. David Knezevic

Applied Mathematics 205. Unit V: Eigenvalue Problems. Lecturer: Dr. David Knezevic Applied Mathematics 205 Unit V: Eigenvalue Problems Lecturer: Dr. David Knezevic Unit V: Eigenvalue Problems Chapter V.4: Krylov Subspace Methods 2 / 51 Krylov Subspace Methods In this chapter we give

More information

A DISSERTATION. Extensions of the Conjugate Residual Method. by Tomohiro Sogabe. Presented to

A DISSERTATION. Extensions of the Conjugate Residual Method. by Tomohiro Sogabe. Presented to A DISSERTATION Extensions of the Conjugate Residual Method ( ) by Tomohiro Sogabe Presented to Department of Applied Physics, The University of Tokyo Contents 1 Introduction 1 2 Krylov subspace methods

More information

WHEN studying distributed simulations of power systems,

WHEN studying distributed simulations of power systems, 1096 IEEE TRANSACTIONS ON POWER SYSTEMS, VOL 21, NO 3, AUGUST 2006 A Jacobian-Free Newton-GMRES(m) Method with Adaptive Preconditioner and Its Application for Power Flow Calculations Ying Chen and Chen

More information

On complex shifted Laplace preconditioners for the vector Helmholtz equation

On complex shifted Laplace preconditioners for the vector Helmholtz equation On complex shifted Laplace preconditioners for the vector Helmholtz equation C. Vuik, Y.A. Erlangga, M.B. van Gijzen, C.W. Oosterlee, D. van der Heul Delft Institute of Applied Mathematics c.vuik@tudelft.nl

More information

Preconditioned inverse iteration and shift-invert Arnoldi method

Preconditioned inverse iteration and shift-invert Arnoldi method Preconditioned inverse iteration and shift-invert Arnoldi method Melina Freitag Department of Mathematical Sciences University of Bath CSC Seminar Max-Planck-Institute for Dynamics of Complex Technical

More information

AMS526: Numerical Analysis I (Numerical Linear Algebra)

AMS526: Numerical Analysis I (Numerical Linear Algebra) AMS526: Numerical Analysis I (Numerical Linear Algebra) Lecture 23: GMRES and Other Krylov Subspace Methods Xiangmin Jiao SUNY Stony Brook Xiangmin Jiao Numerical Analysis I 1 / 9 Minimizing Residual CG

More information

CONVERGENCE BOUNDS FOR PRECONDITIONED GMRES USING ELEMENT-BY-ELEMENT ESTIMATES OF THE FIELD OF VALUES

CONVERGENCE BOUNDS FOR PRECONDITIONED GMRES USING ELEMENT-BY-ELEMENT ESTIMATES OF THE FIELD OF VALUES European Conference on Computational Fluid Dynamics ECCOMAS CFD 2006 P. Wesseling, E. Oñate and J. Périaux (Eds) c TU Delft, The Netherlands, 2006 CONVERGENCE BOUNDS FOR PRECONDITIONED GMRES USING ELEMENT-BY-ELEMENT

More information

DELFT UNIVERSITY OF TECHNOLOGY

DELFT UNIVERSITY OF TECHNOLOGY DELFT UNIVERSITY OF TECHNOLOGY REPORT 06-05 Solution of the incompressible Navier Stokes equations with preconditioned Krylov subspace methods M. ur Rehman, C. Vuik G. Segal ISSN 1389-6520 Reports of the

More information

6.4 Krylov Subspaces and Conjugate Gradients

6.4 Krylov Subspaces and Conjugate Gradients 6.4 Krylov Subspaces and Conjugate Gradients Our original equation is Ax = b. The preconditioned equation is P Ax = P b. When we write P, we never intend that an inverse will be explicitly computed. P

More information

A SPARSE APPROXIMATE INVERSE PRECONDITIONER FOR NONSYMMETRIC LINEAR SYSTEMS

A SPARSE APPROXIMATE INVERSE PRECONDITIONER FOR NONSYMMETRIC LINEAR SYSTEMS INTERNATIONAL JOURNAL OF NUMERICAL ANALYSIS AND MODELING, SERIES B Volume 5, Number 1-2, Pages 21 30 c 2014 Institute for Scientific Computing and Information A SPARSE APPROXIMATE INVERSE PRECONDITIONER

More information

ON A ROBUST ITERATIVE METHOD FOR HETEROGENEOUS HELMHOLTZ PROBLEMS FOR GEOPHYSICS APPLICATIONS

ON A ROBUST ITERATIVE METHOD FOR HETEROGENEOUS HELMHOLTZ PROBLEMS FOR GEOPHYSICS APPLICATIONS INTERNATIONAL JOURNAL OF NUMERICAL ANALYSIS AND MODELING Volume 2, Supp, Pages 197 28 c 25 Institute for Scientific Computing and Information ON A ROBUST ITERATIVE METHOD FOR HETEROGENEOUS HELMHOLTZ PROBLEMS

More information

Multigrid absolute value preconditioning

Multigrid absolute value preconditioning Multigrid absolute value preconditioning Eugene Vecharynski 1 Andrew Knyazev 2 (speaker) 1 Department of Computer Science and Engineering University of Minnesota 2 Department of Mathematical and Statistical

More information

Iterative methods for Linear System

Iterative methods for Linear System Iterative methods for Linear System JASS 2009 Student: Rishi Patil Advisor: Prof. Thomas Huckle Outline Basics: Matrices and their properties Eigenvalues, Condition Number Iterative Methods Direct and

More information

Electromagnetic wave propagation. ELEC 041-Modeling and design of electromagnetic systems

Electromagnetic wave propagation. ELEC 041-Modeling and design of electromagnetic systems Electromagnetic wave propagation ELEC 041-Modeling and design of electromagnetic systems EM wave propagation In general, open problems with a computation domain extending (in theory) to infinity not bounded

More information

DELFT UNIVERSITY OF TECHNOLOGY

DELFT UNIVERSITY OF TECHNOLOGY DELFT UNIVERSITY OF TECHNOLOGY REPORT 10-16 An elegant IDR(s) variant that efficiently exploits bi-orthogonality properties Martin B. van Gijzen and Peter Sonneveld ISSN 1389-6520 Reports of the Department

More information

PROJECTED GMRES AND ITS VARIANTS

PROJECTED GMRES AND ITS VARIANTS PROJECTED GMRES AND ITS VARIANTS Reinaldo Astudillo Brígida Molina rastudillo@kuaimare.ciens.ucv.ve bmolina@kuaimare.ciens.ucv.ve Centro de Cálculo Científico y Tecnológico (CCCT), Facultad de Ciencias,

More information

Variants of BiCGSafe method using shadow three-term recurrence

Variants of BiCGSafe method using shadow three-term recurrence Variants of BiCGSafe method using shadow three-term recurrence Seiji Fujino, Takashi Sekimoto Research Institute for Information Technology, Kyushu University Ryoyu Systems Co. Ltd. (Kyushu University)

More information

On the influence of eigenvalues on Bi-CG residual norms

On the influence of eigenvalues on Bi-CG residual norms On the influence of eigenvalues on Bi-CG residual norms Jurjen Duintjer Tebbens Institute of Computer Science Academy of Sciences of the Czech Republic duintjertebbens@cs.cas.cz Gérard Meurant 30, rue

More information

Chebyshev semi-iteration in Preconditioning

Chebyshev semi-iteration in Preconditioning Report no. 08/14 Chebyshev semi-iteration in Preconditioning Andrew J. Wathen Oxford University Computing Laboratory Tyrone Rees Oxford University Computing Laboratory Dedicated to Victor Pereyra on his

More information

The flexible incomplete LU preconditioner for large nonsymmetric linear systems. Takatoshi Nakamura Takashi Nodera

The flexible incomplete LU preconditioner for large nonsymmetric linear systems. Takatoshi Nakamura Takashi Nodera Research Report KSTS/RR-15/006 The flexible incomplete LU preconditioner for large nonsymmetric linear systems by Takatoshi Nakamura Takashi Nodera Takatoshi Nakamura School of Fundamental Science and

More information

ON ORTHOGONAL REDUCTION TO HESSENBERG FORM WITH SMALL BANDWIDTH

ON ORTHOGONAL REDUCTION TO HESSENBERG FORM WITH SMALL BANDWIDTH ON ORTHOGONAL REDUCTION TO HESSENBERG FORM WITH SMALL BANDWIDTH V. FABER, J. LIESEN, AND P. TICHÝ Abstract. Numerous algorithms in numerical linear algebra are based on the reduction of a given matrix

More information

A Robust Preconditioned Iterative Method for the Navier-Stokes Equations with High Reynolds Numbers

A Robust Preconditioned Iterative Method for the Navier-Stokes Equations with High Reynolds Numbers Applied and Computational Mathematics 2017; 6(4): 202-207 http://www.sciencepublishinggroup.com/j/acm doi: 10.11648/j.acm.20170604.18 ISSN: 2328-5605 (Print); ISSN: 2328-5613 (Online) A Robust Preconditioned

More information

KRYLOV SUBSPACE ITERATION

KRYLOV SUBSPACE ITERATION KRYLOV SUBSPACE ITERATION Presented by: Nab Raj Roshyara Master and Ph.D. Student Supervisors: Prof. Dr. Peter Benner, Department of Mathematics, TU Chemnitz and Dipl.-Geophys. Thomas Günther 1. Februar

More information

Preconditioned Conjugate Gradient-Like Methods for. Nonsymmetric Linear Systems 1. Ulrike Meier Yang 2. July 19, 1994

Preconditioned Conjugate Gradient-Like Methods for. Nonsymmetric Linear Systems 1. Ulrike Meier Yang 2. July 19, 1994 Preconditioned Conjugate Gradient-Like Methods for Nonsymmetric Linear Systems Ulrike Meier Yang 2 July 9, 994 This research was supported by the U.S. Department of Energy under Grant No. DE-FG2-85ER25.

More information

IDR(s) Master s thesis Goushani Kisoensingh. Supervisor: Gerard L.G. Sleijpen Department of Mathematics Universiteit Utrecht

IDR(s) Master s thesis Goushani Kisoensingh. Supervisor: Gerard L.G. Sleijpen Department of Mathematics Universiteit Utrecht IDR(s) Master s thesis Goushani Kisoensingh Supervisor: Gerard L.G. Sleijpen Department of Mathematics Universiteit Utrecht Contents 1 Introduction 2 2 The background of Bi-CGSTAB 3 3 IDR(s) 4 3.1 IDR.............................................

More information

RANA03-02 January Jacobi-Davidson methods and preconditioning with applications in pole-zero analysis

RANA03-02 January Jacobi-Davidson methods and preconditioning with applications in pole-zero analysis RANA03-02 January 2003 Jacobi-Davidson methods and preconditioning with applications in pole-zero analysis by J.Rommes, H.A. van der Vorst, EJ.W. ter Maten Reports on Applied and Numerical Analysis Department

More information

Krylov Space Solvers

Krylov Space Solvers Seminar for Applied Mathematics ETH Zurich International Symposium on Frontiers of Computational Science Nagoya, 12/13 Dec. 2005 Sparse Matrices Large sparse linear systems of equations or large sparse

More information

Arnoldi Methods in SLEPc

Arnoldi Methods in SLEPc Scalable Library for Eigenvalue Problem Computations SLEPc Technical Report STR-4 Available at http://slepc.upv.es Arnoldi Methods in SLEPc V. Hernández J. E. Román A. Tomás V. Vidal Last update: October,

More information

Further experiences with GMRESR

Further experiences with GMRESR Further experiences with GMRESR Report 92-2 C. Vui Technische Universiteit Delft Delft University of Technology Faculteit der Technische Wisunde en Informatica Faculty of Technical Mathematics and Informatics

More information

ETNA Kent State University

ETNA Kent State University Electronic ransactions on Numerical Analysis. Volume 2, pp. 57-75, March 1994. Copyright 1994,. ISSN 1068-9613. ENA MINIMIZAION PROPERIES AND SHOR RECURRENCES FOR KRYLOV SUBSPACE MEHODS RÜDIGER WEISS Dedicated

More information

1. Introduction. Krylov subspace methods are used extensively for the iterative solution of linear systems of equations Ax = b.

1. Introduction. Krylov subspace methods are used extensively for the iterative solution of linear systems of equations Ax = b. SIAM J. SCI. COMPUT. Vol. 31, No. 2, pp. 1035 1062 c 2008 Society for Industrial and Applied Mathematics IDR(s): A FAMILY OF SIMPLE AND FAST ALGORITHMS FOR SOLVING LARGE NONSYMMETRIC SYSTEMS OF LINEAR

More information

Jae Heon Yun and Yu Du Han

Jae Heon Yun and Yu Du Han Bull. Korean Math. Soc. 39 (2002), No. 3, pp. 495 509 MODIFIED INCOMPLETE CHOLESKY FACTORIZATION PRECONDITIONERS FOR A SYMMETRIC POSITIVE DEFINITE MATRIX Jae Heon Yun and Yu Du Han Abstract. We propose

More information

Since the early 1800s, researchers have

Since the early 1800s, researchers have the Top T HEME ARTICLE KRYLOV SUBSPACE ITERATION This survey article reviews the history and current importance of Krylov subspace iteration algorithms. 1521-9615/00/$10.00 2000 IEEE HENK A. VAN DER VORST

More information

IDR(s) as a projection method

IDR(s) as a projection method Delft University of Technology Faculty of Electrical Engineering, Mathematics and Computer Science Delft Institute of Applied Mathematics IDR(s) as a projection method A thesis submitted to the Delft Institute

More information

A Block Compression Algorithm for Computing Preconditioners

A Block Compression Algorithm for Computing Preconditioners A Block Compression Algorithm for Computing Preconditioners J Cerdán, J Marín, and J Mas Abstract To implement efficiently algorithms for the solution of large systems of linear equations in modern computer

More information

Contents. Preface... xi. Introduction...

Contents. Preface... xi. Introduction... Contents Preface... xi Introduction... xv Chapter 1. Computer Architectures... 1 1.1. Different types of parallelism... 1 1.1.1. Overlap, concurrency and parallelism... 1 1.1.2. Temporal and spatial parallelism

More information

A Novel Approach for Solving the Power Flow Equations

A Novel Approach for Solving the Power Flow Equations Vol.1, Issue.2, pp-364-370 ISSN: 2249-6645 A Novel Approach for Solving the Power Flow Equations Anwesh Chowdary 1, Dr. G. MadhusudhanaRao 2 1 Dept.of.EEE, KL University-Guntur, Andhra Pradesh, INDIA 2

More information

Numerische Mathematik

Numerische Mathematik Numer. Math. (1999) 83: 231 257 Numerische Mathematik c Springer-Verlag 1999 Eigenanalysis of some preconditioned Helmholtz problems Howard C. Elman, Dianne P. O Leary Computer Science Department and Institute

More information

Valeria Simoncini and Daniel B. Szyld. Report October 2009

Valeria Simoncini and Daniel B. Szyld. Report October 2009 Interpreting IDR as a Petrov-Galerkin method Valeria Simoncini and Daniel B. Szyld Report 09-10-22 October 2009 This report is available in the World Wide Web at http://www.math.temple.edu/~szyld INTERPRETING

More information

Fast iterative solvers

Fast iterative solvers Solving Ax = b, an overview Utrecht, 15 november 2017 Fast iterative solvers A = A no Good precond. yes flex. precond. yes GCR yes A > 0 yes CG no no GMRES no ill cond. yes SYMMLQ str indef no large im

More information

GMRESR: A family of nested GMRES methods

GMRESR: A family of nested GMRES methods GMRESR: A family of nested GMRES methods Report 91-80 H.A. van der Vorst C. Vui Technische Universiteit Delft Delft University of Technology Faculteit der Technische Wisunde en Informatica Faculty of Technical

More information

Solution of eigenvalue problems. Subspace iteration, The symmetric Lanczos algorithm. Harmonic Ritz values, Jacobi-Davidson s method

Solution of eigenvalue problems. Subspace iteration, The symmetric Lanczos algorithm. Harmonic Ritz values, Jacobi-Davidson s method Solution of eigenvalue problems Introduction motivation Projection methods for eigenvalue problems Subspace iteration, The symmetric Lanczos algorithm Nonsymmetric Lanczos procedure; Implicit restarts

More information

An advanced ILU preconditioner for the incompressible Navier-Stokes equations

An advanced ILU preconditioner for the incompressible Navier-Stokes equations An advanced ILU preconditioner for the incompressible Navier-Stokes equations M. ur Rehman C. Vuik A. Segal Delft Institute of Applied Mathematics, TU delft The Netherlands Computational Methods with Applications,

More information

Multilevel low-rank approximation preconditioners Yousef Saad Department of Computer Science and Engineering University of Minnesota

Multilevel low-rank approximation preconditioners Yousef Saad Department of Computer Science and Engineering University of Minnesota Multilevel low-rank approximation preconditioners Yousef Saad Department of Computer Science and Engineering University of Minnesota SIAM CSE Boston - March 1, 2013 First: Joint work with Ruipeng Li Work

More information

The Lanczos and conjugate gradient algorithms

The Lanczos and conjugate gradient algorithms The Lanczos and conjugate gradient algorithms Gérard MEURANT October, 2008 1 The Lanczos algorithm 2 The Lanczos algorithm in finite precision 3 The nonsymmetric Lanczos algorithm 4 The Golub Kahan bidiagonalization

More information

A short course on: Preconditioned Krylov subspace methods. Yousef Saad University of Minnesota Dept. of Computer Science and Engineering

A short course on: Preconditioned Krylov subspace methods. Yousef Saad University of Minnesota Dept. of Computer Science and Engineering A short course on: Preconditioned Krylov subspace methods Yousef Saad University of Minnesota Dept. of Computer Science and Engineering Universite du Littoral, Jan 19-3, 25 Outline Part 1 Introd., discretization

More information

Harmonic Projection Methods for Large Non-symmetric Eigenvalue Problems

Harmonic Projection Methods for Large Non-symmetric Eigenvalue Problems NUMERICAL LINEAR ALGEBRA WITH APPLICATIONS Numer. Linear Algebra Appl., 5, 33 55 (1998) Harmonic Projection Methods for Large Non-symmetric Eigenvalue Problems Ronald B. Morgan 1 and Min Zeng 2 1 Department

More information

Iterative Solvers in the Finite Element Solution of Transient Heat Conduction

Iterative Solvers in the Finite Element Solution of Transient Heat Conduction Iterative Solvers in the Finite Element Solution of Transient Heat Conduction Mile R. Vuji~i} PhD student Steve G.R. Brown Senior Lecturer Materials Research Centre School of Engineering University of

More information

DELFT UNIVERSITY OF TECHNOLOGY

DELFT UNIVERSITY OF TECHNOLOGY DELFT UNIVERSITY OF TECHNOLOGY REPORT 15-05 Induced Dimension Reduction method for solving linear matrix equations R. Astudillo and M. B. van Gijzen ISSN 1389-6520 Reports of the Delft Institute of Applied

More information

DELFT UNIVERSITY OF TECHNOLOGY

DELFT UNIVERSITY OF TECHNOLOGY DELFT UNIVERSITY OF TECHNOLOGY REPORT 11-14 On the convergence of inexact Newton methods R. Idema, D.J.P. Lahaye, and C. Vuik ISSN 1389-6520 Reports of the Department of Applied Mathematical Analysis Delft

More information

Linear Solvers. Andrew Hazel

Linear Solvers. Andrew Hazel Linear Solvers Andrew Hazel Introduction Thus far we have talked about the formulation and discretisation of physical problems...... and stopped when we got to a discrete linear system of equations. Introduction

More information

The quadratic eigenvalue problem (QEP) is to find scalars λ and nonzero vectors u satisfying

The quadratic eigenvalue problem (QEP) is to find scalars λ and nonzero vectors u satisfying I.2 Quadratic Eigenvalue Problems 1 Introduction The quadratic eigenvalue problem QEP is to find scalars λ and nonzero vectors u satisfying where Qλx = 0, 1.1 Qλ = λ 2 M + λd + K, M, D and K are given

More information

Preconditioning Techniques for Large Linear Systems Part III: General-Purpose Algebraic Preconditioners

Preconditioning Techniques for Large Linear Systems Part III: General-Purpose Algebraic Preconditioners Preconditioning Techniques for Large Linear Systems Part III: General-Purpose Algebraic Preconditioners Michele Benzi Department of Mathematics and Computer Science Emory University Atlanta, Georgia, USA

More information

On the Preconditioning of the Block Tridiagonal Linear System of Equations

On the Preconditioning of the Block Tridiagonal Linear System of Equations On the Preconditioning of the Block Tridiagonal Linear System of Equations Davod Khojasteh Salkuyeh Department of Mathematics, University of Mohaghegh Ardabili, PO Box 179, Ardabil, Iran E-mail: khojaste@umaacir

More information

ITERATIVE METHODS BASED ON KRYLOV SUBSPACES

ITERATIVE METHODS BASED ON KRYLOV SUBSPACES ITERATIVE METHODS BASED ON KRYLOV SUBSPACES LONG CHEN We shall present iterative methods for solving linear algebraic equation Au = b based on Krylov subspaces We derive conjugate gradient (CG) method

More information

Domain Decomposition Algorithms for an Indefinite Hypersingular Integral Equation in Three Dimensions

Domain Decomposition Algorithms for an Indefinite Hypersingular Integral Equation in Three Dimensions Domain Decomposition Algorithms for an Indefinite Hypersingular Integral Equation in Three Dimensions Ernst P. Stephan 1, Matthias Maischak 2, and Thanh Tran 3 1 Institut für Angewandte Mathematik, Leibniz

More information

Iterative methods for symmetric eigenvalue problems

Iterative methods for symmetric eigenvalue problems s Iterative s for symmetric eigenvalue problems, PhD McMaster University School of Computational Engineering and Science February 11, 2008 s 1 The power and its variants Inverse power Rayleigh quotient

More information

HARMONIC RAYLEIGH RITZ EXTRACTION FOR THE MULTIPARAMETER EIGENVALUE PROBLEM

HARMONIC RAYLEIGH RITZ EXTRACTION FOR THE MULTIPARAMETER EIGENVALUE PROBLEM HARMONIC RAYLEIGH RITZ EXTRACTION FOR THE MULTIPARAMETER EIGENVALUE PROBLEM MICHIEL E. HOCHSTENBACH AND BOR PLESTENJAK Abstract. We study harmonic and refined extraction methods for the multiparameter

More information

Convergence property of IDR variant methods in the integral equation analysis of electromagnetic scattering problems

Convergence property of IDR variant methods in the integral equation analysis of electromagnetic scattering problems LETTER IEICE Electronics Express, Vol.11, No.9, 1 8 Convergence property of IDR variant methods in the integral equation analysis of electromagnetic scattering problems Hidetoshi Chiba a), Toru Fukasawa,

More information

Solving Ax = b, an overview. Program

Solving Ax = b, an overview. Program Numerical Linear Algebra Improving iterative solvers: preconditioning, deflation, numerical software and parallelisation Gerard Sleijpen and Martin van Gijzen November 29, 27 Solving Ax = b, an overview

More information

On the choice of abstract projection vectors for second level preconditioners

On the choice of abstract projection vectors for second level preconditioners On the choice of abstract projection vectors for second level preconditioners C. Vuik 1, J.M. Tang 1, and R. Nabben 2 1 Delft University of Technology 2 Technische Universität Berlin Institut für Mathematik

More information

Lecture 11: CMSC 878R/AMSC698R. Iterative Methods An introduction. Outline. Inverse, LU decomposition, Cholesky, SVD, etc.

Lecture 11: CMSC 878R/AMSC698R. Iterative Methods An introduction. Outline. Inverse, LU decomposition, Cholesky, SVD, etc. Lecture 11: CMSC 878R/AMSC698R Iterative Methods An introduction Outline Direct Solution of Linear Systems Inverse, LU decomposition, Cholesky, SVD, etc. Iterative methods for linear systems Why? Matrix

More information

BiCGCR2: A new extension of conjugate residual method for solving non-hermitian linear systems

BiCGCR2: A new extension of conjugate residual method for solving non-hermitian linear systems BiCGCR2: A new extension of conjugate residual method for solving non-hermitian linear systems Xian-Ming Gu 1,2, Ting-Zhu Huang 1, Bruno Carpentieri 3, 1. School of Mathematical Sciences, University of

More information

Preconditioning Techniques Analysis for CG Method

Preconditioning Techniques Analysis for CG Method Preconditioning Techniques Analysis for CG Method Huaguang Song Department of Computer Science University of California, Davis hso@ucdavis.edu Abstract Matrix computation issue for solve linear system

More information

On prescribing Ritz values and GMRES residual norms generated by Arnoldi processes

On prescribing Ritz values and GMRES residual norms generated by Arnoldi processes On prescribing Ritz values and GMRES residual norms generated by Arnoldi processes Jurjen Duintjer Tebbens Institute of Computer Science Academy of Sciences of the Czech Republic joint work with Gérard

More information

Solving Large Nonlinear Sparse Systems

Solving Large Nonlinear Sparse Systems Solving Large Nonlinear Sparse Systems Fred W. Wubs and Jonas Thies Computational Mechanics & Numerical Mathematics University of Groningen, the Netherlands f.w.wubs@rug.nl Centre for Interdisciplinary

More information

9.1 Preconditioned Krylov Subspace Methods

9.1 Preconditioned Krylov Subspace Methods Chapter 9 PRECONDITIONING 9.1 Preconditioned Krylov Subspace Methods 9.2 Preconditioned Conjugate Gradient 9.3 Preconditioned Generalized Minimal Residual 9.4 Relaxation Method Preconditioners 9.5 Incomplete

More information

Recent computational developments in Krylov Subspace Methods for linear systems. Valeria Simoncini and Daniel B. Szyld

Recent computational developments in Krylov Subspace Methods for linear systems. Valeria Simoncini and Daniel B. Szyld Recent computational developments in Krylov Subspace Methods for linear systems Valeria Simoncini and Daniel B. Szyld A later version appeared in : Numerical Linear Algebra w/appl., 2007, v. 14(1), pp.1-59.

More information