Splitting of Expanded Tridiagonal Matrices. Seongjai Kim. Abstract. The article addresses a regular splitting of tridiagonal matrices.

Size: px
Start display at page:

Download "Splitting of Expanded Tridiagonal Matrices. Seongjai Kim. Abstract. The article addresses a regular splitting of tridiagonal matrices."

Transcription

1 Splitting of Expanded Tridiagonal Matrices ga = B? R for Which (B?1 R) = 0 Seongai Kim Abstract The article addresses a regular splitting of tridiagonal matrices. The given tridiagonal matrix A is rst expanded to an equivalent matrix A e and then split as e A = B? R for which B is block-diagonal and every eigenvalue of B?1 R is zero, i.e., (M?1 N ) = 0. The optimal splitting technique is applicable to various algorithms that incorporate one-dimensional solves or their approximations. Examples can be found in the parallelization of alternating direction iterative (ADI) methods and ecient parameter choices for domain decomposition (DD) methods for elliptic and parabolic problems. Numerical results solving the Helmholtz wave equation in two dimensions are presented to demonstrate usefulness and eciency of the splitting technique. Key words. Matrix expansion, regular splitting, ADI, domain decomposition method, Helmholtz wave equation. AMS subect classications. 5F15, 5F10, 5N Introduction Lots of modern computational algorithms incorporate tridiagonal matrix solves of the form Au = b; (1.1) where A IR nn and u; b IR n. Examples in numerical PDE are the alternating direction iterative (ADI) method [,, 1] and parameter choices for multigrid (MG) and domain decomposition (DD) methods [10, 0]. Department of Mathematics, University of Kentucky, Lexington, Kentucky USA skim@ms.uky.edu 1

2 Splitting of Tridiagonal Matrices This article studies a regular splitting of tridiagonal matrices. The given tridiagonal matrix A is rst expanded to an equivalent matrix e A. Then, we consider a regular splitting for e A of the form ea = B? R; (1.) where B is block-diagonal and every eigenvalue of B?1 R is zero, i.e., (B?1 R) = 0: (1.3) Such an optimal splitting technique has not been studied for general tridiagonal matrices in the literature of numerical analysis; as a matter of fact, its usefulness is not yet recognized. The main obect of the article is to introduce the splitting technique (1.)-(1.3) and address its usefulness and eciency. The tridiagonal systems can become expensive to solve in some applications. For example, the ADI method is poorly tting with parallel computers of distributed memory, since it requires tridiagonal solves in all directions. The parallelization eciency is deteriorated when the tridiagonal systems are to be solved across the computer memory boundaries. We may try to solve the tridiagonal systems by an iterative algorithm (as an inner loop) incompletely. Such an idea was tested to be ecient even if -3 red-black (point) Gauss-Seidel iterations were performed for an approximation of the tridiagonal solve [, 5, 15]. Some computational algorithms require ecient choices of algorithm parameters, such as in domain decomposition (DD) methods and multigrid (MG) methods. As an application of (1.)-(1.3), we will present a strategy for parameter choices for a nonoverlapping DD method applied to the Helmholtz problem which is complexvalued, indenite, and non-hermitian. The article is organized as follows. In the next section, we present the numerical strategy of nding the optimal splitting of expanded matrices of tridiagonal matrices. Section 3 includes applications of the optimal splitting to the parallelization of ADI methods, and parameter choices for a DD method for the Helmholtz problem and the heat equation. Numerical results are presented to show eciency and usefulness of the optimal splitting.

3 Splitting of Tridiagonal Matrices 3. Optimal splitting of tridiagonal matrices.1. Matrix expansion and splitting We rst introduce the notion of equivalent matrix expansion of tridiagonal matrices. As an example, we consider the following algebraic system a 11 a 1 u 1 b 1 a 1 a a 3 Au a 3 a 33 a 3 u a 3 a a u 3 b 5 5 = u b 3 5 b: (.1) b 5 a 5 a 55 u 5 b 5 An expansion of the system (.1) having two blocks can be written as a 11 a 1 u 1 b 1 a 1 a a 3 u eaeu a 3 a 33 +' 1?' 1 a 3 b a 3?' 1 a 33 +' 1 a 3 u 3 u a 3 a a = b 3 b u 5 3 b: e (.) b 5 a 5 a 55 u 5 b 5 When (.) admits a unique solution, the corresponding components of the solution eu are clearly the same as those of u. In the case, the matrix expansion e A in (.) is said to be equivalent to A. Now, consider a general tridiagonal matrix A IR nn and its related algebraic system (1.1). Then, we can expand A into A e of M blocks, M 1. We assume that each block has the same dimension m for simplicity; n = (m? 1)M + 1. Dene a tridiagonal matrix of dimension m m by T m (; ; ) = m?1;m? m?1;m?1 m?1;m m;m?1 m;m + Then, the splitting of e A can be formulated as follows: ea = B? R; (.3) 3 5 : where B = diag(d 1 ; ; D M ); R = 0 F 1 E 0 F E M?1 0 F M?1 E M :

4 Splitting of Tridiagonal Matrices Here the matrices are dened by D 1 = T m ( 1 ; 0; ' 1 ); D = T m ( ; ' ;?1 ; ' ;+1 ); = ; ; M? 1; D M = T m ( M ; ' M;M?1 ; 0); where, = 1; ; M, are obtained from A, correspondingly, and E = for appropriate ea ;W and ea ;E ea ;W ' ;? ; F = ' ;+1 ea ;E 0 0 Remark. With the splitting of the expanded matrix, we can solve (1.1) applying the iteration eu n = B?1 (Reu n?1 + e b); n = 1; ; ; (.) for a given eu 0. We know that the iteration converges if and only if (B?1 R) < 1. Remark. If A is an M-matrix, equivalence between A and A e can be proved as suggested by Tang [0], which is closely related to the existence of D?1, = 1; ; M? 1. For an expansion of tridiagonal matrices of constant diagonals for which the parameters ' i; are constant, i.e., ' = ' ;+1 = ' +1; for = 1; ; M? 1, Tang has utilized the analytic formula for D?1 to nd the single parameter ' that minimizes the spectral radius of the matrix B?1 R... The optimal parameters f' ;+1 g Recall the main goal of the section is to nd f' ;+1 g such that the spectral radius of G(:= B?1 R) is zero. It should be noticed that R has only (M? 1) nonzero elements, and therefore the matrix G has only (M? 1) nonzero columns that are related to the rst or last columns of the matrices D?1 )`;k, the (`; k)-element of D?1 t () `;k = (D?1 3 5 ;, = 1; ; M. Let. For a simple expression, we rst consider the case of two blocks: M =. Then, removing zero columns and the corresponding rows from matrix G, we have (G) = f0g [ (G 0 ); where (G) is the spectrum of the matrix G and G 0 = 0 0 ' 1 t (1) m?1;m ea 1;E t (1) m?1;m 0 0 ' 1 t (1) m;m ea 1;E t (1) m;m ea ;W t () 1;1 ' 1 t () 1;1 0 0 ea ;W t () ;1 ' 1 t () ; :

5 Splitting of Tridiagonal Matrices 5 Let P = Then a simple calculation yields where P?1 G 0 P = 1?' 1 =ea ;W ?' 1 =ea 1;E : g ea 1;E t (1) m;m ea ;W t () 1; g (1) g 1 = ea 1;E t m?1;m + ' 1 t (1) () m;m ; g1 = ea ;W t ;1 + ' 1 t ea ;W ea 1;1 () : (.5) 1;E Again, reducing zero columns and the corresponding rows from P?1 G 0 P, we obtain (G) = f0g [ (G 0 ) = f0g [ (G 00 ); 3 5 ; where Now, it is easy to see (G) = 0 if G 00 = " 0 g1 g 1 0 # : g 1 = 0; or g 1 = 0: Now let us generalize the above argument. For example, consider the case: M =. Then, it is not dicult to derive the following: where G 00 = where, for = 1; ; M? 1, (G) = f0g [ (G 00 ); (.) 0 g g g 0 0 g g g g 0 0 g g g 5 0 g?1; = ea ;E t () m?1;m + ' +1; ea +1;W t () (+1) g ;?1 = ea +1;W t ;1 + ' ;+1 ea ;E 3 5 m;m ; t (+1) 1;1 ; (.) : (.8)

6 Splitting of Tridiagonal Matrices What we want to do is to nd ' ;+1 and ' +1; such that g?1; = 0; or g ;?1 = 0; = 1; ; M? 1: In the case, we can see from (.)-(.) that (G) = 0. Set Then, from (.8), we have g?1; = 0; = 1; ; M? 1: ' +1; =?ea +1;W t () m?1;m t () m;m ; = 1; ; M? 1: (.9) The problem is thus reduced to the problem of nding the terms in the right side of (.9). How can we nd them? Since we do not assume any good features of A except nonsingularity, it is impossible to utilize the explicit formulas for the inverse of symmetric tridiagonal matrices of constant diagonals; see [8, 1] and and references therein. A computer-aid method can be designed as follows. Set ' ;+1 = ' +1; ; = 1; ; M? 1; (.10) and assume ' ;?1 is known. (The above setting is not a requirement, but a choice.) Let X be the last column of D?1, i.e., D X = I m ; X = (t () 1;m; t () ;m; ; t () m;m )T : (.11) where I m = (0; ; 0; 1) T. Now, we can nd X explicitly by inverting the tridiagonal matrix D. In fact, D need not be inverted completely: let D be factorized into L U, where the diagonal elements of L are all one. It can be easily veried that L U X = U X = I m and the (m? 1)-th equation of the linear system U X = I m is U ;m?1;m?1 t () m?1;m + U ;m?1;m t () m;m = 0; (.1) where U ;`;k is the (`; k)-th element of U. It follows from (.9) and (.1) that ' +1; = ea +1;W U ;m?1;m U ;m?1;m?1 ; = 1; ; M? 1: (.13) We have assumed that ' ;?1 is known to nd ' +1;. Note that one can begin the procedure from D 1 with ' 1;0 = 0 to nd ' 1; = ' ;1, and then nd ' ;+1 = ' +1;, = ; ; M? 1, recursively. We have shown that (G) = 0 when ' +1; are obtained recursively by using (.13) and setting ' ;+1 = ' +1; (a choice). The procedure works whenever the LU-factorizations of D exist with or without pivoting for = 1; ; M? 1. That is, we can nd parameters ' +1; if all D are nonsingular.

7 Splitting of Tridiagonal Matrices 3. Applications 3.1. Parallelization of the ADI method In this section we discuss the classical ADI method and its parallelization which has motivated the main result presented in the previous section. Let = (a x ; b x ) (a y ; b y ) be a domain in IR with its boundary? Consider the problem (a)?r (aru) = f; x ; (b) au + u = g; x?; (3.1) where a = a(x) > 0 is the diusion coecient, the subscript denotes the outer unit normal from the domain, = (x) 0 is a given function, and f and g are the sources. Assume that the problem admits a unique solution. Let the problem (3.1) be discretized by e.g. the central nite dierence (FD) method on a uniform mesh of (N x? 1) (N y? 1) cells; h x = (b x? a x )=(N x? 1) and h y = (b y? a y )=(N y? 1). Then, its linear system reads Au = (H + V )u = b; (3.) where A IR NN, where N = N x N y. Here the matrices H and V are the discretizations of the dierential operators in the horizontal and vertical directions respectively. Each of them is assumed to be reducible to the direct sum of irreducible Stieltes matrices [1]. A Stieltes matrix is by denition a real symmetric and positive denite matrix with nonpositive o-diagonal entries. (We have assumed that A is symmetric. For the central FD method for the problem of the Neumann or mixed boundary condition, one can symmetrize the algebraic system by multiplying certain numbers to the rows of the matrix and the source vector that correspond to nodal points on the boundary. In this case, the matrix turns out to be the same as what one can obtain from the bi-linear nite element method with the Trapezoidal quadrature rule.) To introduce an ADI method, we rewrite (3.) as a pair of matrix equations (ri + H)u = (ri? V )u + b; (ri + V )u = (ri? H)u + b; (3.3) for some positive number r. The ADI iteration is then dened as [,, 1] (a) (r m+1 I + H)u m+1= = (r m+1 I? V )u m + b; (b) (r m+1 I + V )u m+1 = (r m+1 I? H)u m+1= + b; m 0; (3.) where u 0 is an initial approximation of the solution of (3.), and the r m 's are positive constants called acceleration parameters, which are to be chosen so as to make the convergence of the iteration rapid. For a given ordering of grid points (e.g., row-wise

8 Splitting of Tridiagonal Matrices 8 Figure 1: Numerical grid distribution for parallel computers. The thick line segments indicate computer memory boundaries. ordering), the matrix r m+1 I +H is tridiagonal. One can nd a suitable permutation of the rows and columns which converts r m+1 I + V to a tridiagonal matrix; the systems (3.) are easily solved by using e.g. the Gaussian elimination. ADI methods can be accelerated with multiple parameters of cycle length n c. There are well developed theory to accelerate the convergence of ADI by the choice of dierent r's used cyclically. Among others, the optimal parameters suggested by Wachspress [] for the cases n c = t, t > 0, are easy to nd and ecient. See also [1, Ch. ] and [3] for dierent choices of acceleration parameters and analysis for the asymptotic spectral radius of (3.). On parallel computers, tridiagonal solves across the computer memory boundaries often slow down the ADI method (3.) to be unsatisfactory. Let the data be distributed as in Figure 1 for a parallel computer. The vertical sweep (3.b) can be solved eciently after simple communications of data between adacent processors. The horizontal sweep (3.a) requires parallel tridiagonal solves, which is presumably expensive. We suggest the iteration of the form (.) for solving tridiagonal systems in (3.a) by a nonoverlapping DD technique [, 10, 1]. The matrix expansion presented in x has been motivated from the DD method, which is discussed below. Let y0 be a horizontal grid line passing the grid point (a x ; y 0 ) and? y0 =?\@ y0. Let a;hu h c; u h be the centered (second-order) nite dierence approximations of (au x ) x and au, respectively. Then, the problem (3.a) restricted to y0 can be rewritten in the following form (a)? a;hu h + qu h = w; x y0 ; c; u h + u h = g; x? y0 ; (3.5) where iteration indices are dropped to simplify the presentation and w is the nite dierence approximation of (au m y ) y + r m+1 u m + f: Let M be the number of subdomains/processors; decompose y0 into nonoverlapping M subintervals f y0 ; : = 1; ; Mg ordered from left to right and partition the subintervals equally into (m? 1) elements each. (The total number of grid points

9 Splitting of Tridiagonal Matrices 9 W O E y0 ;? y0 ;k y0 ;k Figure : The subintervals y0 ; and y0 ;k and their interface? y0 ;k. The interface point is denoted by O and its adacent points by W y0 ; and E y0 ;k. in the x-direction is N x = M (m? 1) + 1:) Let? y0 ; =? y0 y0 ;;? y0 ;k y0 ; y0 ;k; k = + 1 or? 1: The restriction of u h onto y0 ; is denoted by u h. Now, we can dene the DD method for (3.5) over f y0 ;g: nd fu h g satisfying (a)? a;h uh + quh = w; x y 0 ;; c; u h + u h = g; x? y0 ;; (c) u h = uh k f;ku h =?@ b;ku h k ; x? y 0 ;k; (3.) f;k u h b;k u h k are respectively the forward and backward dierences for a@u =@ and a@u k =@ k on? y0 ;k. For the case in Figure, they are dened f;k u h (O) = ea u h (E)? uh (O) ;E b;k u h k h (O) = ea u h k(o)? u h k(e) ;E ; x h x where h x is the grid size and ea ;E is an average value of a on the interval [O; E], e.g., ea ;E = a(o) + a(e) : (One may try to use the harmonic average instead of the arithmetic average.) The outer average ea ;W of a at the left end point of y0 ; can be dened in the same way. For example, ea k;w = (a(o) + a(w))= in Figure. Remark. For DD methods for the dierential problem, both the solution fu g and the conormal derivative fa@u =@ g should be continuous on the interfaces. On the other hand, for the discrete problem, fu g and fa@u =@ g cannot be simultaneously continuous unless the solution is linear over the whole domain, a totally uninteresting case. Finite dierence/element methods impose the continuity of the solution and allow the conormal derivatives to be discontinuous on the interfaces of elements. To implement this situation elegantly, we have introduced the forward-backward matching for the conormal derivatives; see the second equation of (3.c). The equations (3.c) can be equivalently written as follows (called the Robin interface boundary f;k u h + uh =?@ b;ku h k + uh k ;? k = 1; (3.)

10 Splitting of Tridiagonal Matrices 10 where is a positive function on the interfaces f? y0 ;kg. Remark. It is easy to check that (3.5) and (3.) are equivalent, i.e., the solution u h of (3.5) restricted onto y0 ; is the same as u h, the solution of (3.) on y0 ;. We will leave it to interested readers. The basic idea of a DD iterative method is to localize the computations to smaller subdomain problems. It is feasible to localize to each y0 ; by evaluating the quantities in (3.a)-(3.b) and (3.) related to each y0 ; at the new iterate level and those related to neighboring subdomains y0 ;k at the old level. The iterative algorithm can be dened as follows: choose an initial guess fu h;0 : = 1; ; M x g, then recursively build the sequences fu h;n : = 1; ; M x g, n 1, by solving (a) (b) f;k u h;n? a;hu c; u h;n + u h;n + qu h;n = w; x y0 ;; + u h;n = g; x? y0 ;; =?@ b;k u h;n?1 k + u h;n?1 k ; x? y0 ;k; (3.8) Its algebraic system can be formulated as in (.) where B and R are dened in (.3) and ' +1; = +1; h? ea +1;W ; ' ;+1 = ;+1 h? ea ;E : (3.9) One can nd the optimal parameters ' +1; using the procedure in x and ' ;+1 = ' ;+1 + ea ;E? ea +1;W from (3.9). Here we have set ;+1 = +1; to satisfy the continuity of the solution on the interfaces. For a test of (3.8) incorporating the optimal splitting technique, we choose y0 = (0; 1), replace (3.8b) by the Dirichlet boundary condition, and set u(x) = sin(x); a(x) = 1 + sin(x)=; (3.10) where w is given correspondingly. The unit interval is partitioned into 58 elements (h = 1=58). Following the argument in [] for n c = 8, we have found the eight cyclic acceleration parameters for the model problem (a(x) 1) as 3.1e+0; 8.98e-1;.13e-1; 5.01e-; 1.18e-;.e-3;.55e-; 1.8e-: For these parameters applied to the model problem in D, a cycle of eight ADI iterations would reduce the error by the factor of.e-3. As suggested in [], these parameters can be eciently incorporated in (3.) with a certain scaling. Note that the average of the diusion coecient in (3.10) is one. When the above parameters are used for the algorithm (3.8) with (3.10), the algorithm would converge fast for the acceleration parameters except two or three smallest ones. In Table 1, we present the relative error E n 1 = ku? u h;n k 1 =kuk 1 ;

11 Splitting of Tridiagonal Matrices 11 r =5.01e- r =1.8e- n M=8 M=1 M=3 M=8 M=1 M=3 1.5e-1.e-1 5.e-1.e-1 8.e-1 9.0e-1 3.9e-5.1e- 1.e- 1.8e-1.e-1.e e- 9.0e-.1e-.9e-.8e- 1.e-1 1.e- 1.e-.3e-5.1e-.0e- 1.0e e- 1.e- 9.e-.8e-3.8e-.5e- Table 1: The error E n 1 for the rst ve iterations of the algorithm (3.8). for the rst ve iterations of (3.8), for r = 5.01e- and r 8 = 1.8e-. Set u h;0 0; M denotes the number of subdomains, i.e., the number of processors for a parallel computation. The discretization error seems 1.e-. As one can see from the table, when 1 or less processors are utilized, -3 iterations of (3.8) are almost the same as the direct tridiagonal solve for large parameters, while -3 iterations improve one-digit accuracy of the solution for the smallest acceleration parameter. Note that the inner loop of nested iterative algorithms can be solved incompletely without signicantly deteriorating the overall convergence rate. More systematic treatments of ADI and their parallel implementation will appear elsewhere [1]. In practice, the iterative line solves (3.8) can be implemented with the red-black ordering of subproblems, which speeds up convergence further. 3.. Parameter choice for domain decomposition methods As another application of the splitting discussed in Section, we consider parameter choices for a nonoverlapping DD method for the Helmholtz wave equation. The Helmholtz problem. Let = (a x ; b x ) (a y ; b y ) and? Consider the following complex-valued indenite problem?u? K u = S; x ; u + iu = 0; x?; (3.11) where S is the wave source and i is the imaginary unit. The coecients K and satisfy K = p? iq ; 0 < p 0 p(x) p 1 < 1; 0 q 0 q(x) q 1 < 1; = r? i i ; r (x) > 0; i (x) 0; for some positive constants p 0, p 1, q 0, and q 1, and are suciently regular that the existence and uniqueness of a solution of (3.11) lying in H 1 () are assured for reasonable S. The coecient is properly chosen such that the second equation of (3.11) represents a rst-order absorbing boundary condition (ABC) that allows normally incident

12 Splitting of Tridiagonal Matrices 1 waves to pass out of transparently. The problem (3.11) models the propagation of time-harmonic waves such as electromagnetic waves, seismic waves, and underwater acoustics. In most applications, the wave number p is given as p =! v = f v ; where f is the frequency,!(= f) is the angular frequency, and v is the wave velocity in the medium under consideration. The wave problem (3.11) is dicult to solve, in particular, when 0 q p: (3.1) In addition to having a complex-valued solution, it is neither Hermitian symmetric nor coercive; as a consequence, most standard iterative methods either fail to converge or converge very slowly. In acoustic wave applications, for example, we often need to simulate waves of 0-50 wavelengths. (The wavelength is =p by denition.) It is known that the second-order FD scheme requires at least to 8 grid points per wavelength for a stability reason [10, 19]. Domain decomposition method: In solving (3.11), it is often the case that the coarse grid problem of DD or MG methods is still large and the coarse grid problem should be solved without introducing another coarser grid correction due to stability reasons. Here we introduce a nonoverlapping DD method not incorporating coarser meshes. The DD method can be utilized as either the main solver or the coarse grid solver of a MG method. Let the domain be partitioned into (N x?1)(n y?1) cells; h x = (b x?a x )=(N x?1), h y = (b y? a y )=(N y? 1), and h = max(h x ; h y ). Let u h be the solution of the FD approximation of (3.11):? h u h? K u h = S; x c u h + iu h = 0; x?; (3.13) where h u h is the centered 5-point dierence approximation of u c u h denotes the centered dierence for u on the boundary?. Let f ; = 1; ; Mg be the subdomains of : = M[ =1 ; \ k = ;; = k: Assume that are also rectangular regions consisting of a group of cells. Let? =? ;? k =? k k :

13 Splitting of Tridiagonal Matrices 13 Let u be the restriction of u on. Then, following the idea discussed in x3.1, we can formulate a nonoverlapping DD method for (3.13) as follows: choose initial guess g, then nd fu h;n g, n 1, satisfying fu f;k u h;n? h u c; u h;n + i u h;n? K u h;n = S; x ; + iu h;n = 0; x? ; =?@ b;k u h;n?1 k + i u h;n?1 k ; x? k ; where = r? i i, r > 0, i 0, is the acceleration parameter. (3.1) Despres [, 3], indebted to Lions [1], analyzed a DD algorithm for (3.11) on a dierential, rather than discrete, level. In his implementation of the algorithm incorporating the mixed nite element method of index zero [18], he chose K for the acceleration parameter, i.e., = K. The choice has been motivated by the ABC; when K is real and constant, = K satises the rst-order ABC. The convergence of algorithm (3.1) was analyzed by the author [10], without introducing the coarse grid correction. Various acceleration techniques are incorporated for (3.1) to converge with a linear dependence on the subdomain size H, but independently on the mesh size h and the wave number K; see [9, 10, 11, 13]. The splitting technique in x is applied to nding the acceleration parameter. When the rectangular domain is decomposed into M x M y subdomains, for example, the one-dimensional problem restricted on a horizontal (resp. vertical) grid line can be viewed as a DD method of M x (resp. M y ) subdomains. Applying the optimal splitting technique, we can nd on each of horizontal and vertical grid lines for which the spectral radius of the iteration matrix for the 1D-restricted problem becomes zero. Such a parameter is called the alternating direction optimal parameter (ADOP). Note that the computation of ADOP requires less than N operations, where N is the number of all grid points, which is negligible compared to the total cost. Numerical experiments. Let = (0; 1). The wave speed v(x; y) is chosen as v 1 (x; y) = 1; v (x; y) = ( + sin(x))(? sin(3y)): (3.15) We set the frequency f = 10, p = f=v =!=v :8=v; q is chosen such that the quality factor Q(:= p =q ) is 100. (The average value of q for v = v is 1..) The source S is selected such that u(x; y) = is the true solution of (3.11). (x) (y)! ; (x) = e i!(x?1) + e?i!x? ; (3.1) The domain is partitioned into (N x? 1) (N x? 1) uniform cells of the edge size h = 1=(N x?1). The cells are grouped into 1 strip-type subdomains of the same size.

14 Splitting of Tridiagonal Matrices 1 v = v 1 v = v 1=h n r n 1 CPU n r n 1 CPU e e e e-3.9 Table : Numerical results for (3.1) with ADOP. The domain is decomposed into 1 strip-type subdomains and the solution incorporates 10 wavelengths. Each subproblem is solved directly by the Gauss-elimination. The error is estimated by the relative maximum norm r n 1, and the iteration is stopped when s n 1 10? : r n 1 = kuh;n? uk L 1 () kuk L 1 () ; s n 1 = kuh;n? u h;n?1 k L 1 () ku h;n k L 1 () ; where u h;n is the approximate solution of the n-th iteration. Zero initial values are assumed: u h;0 0. In Table, we present numerical results for (3.1) incorporating ADOP. The integer n denotes the number of DD iterations and CPU is the user time (in second) on a Gateway Solo, a MHz laptop having 18M memory and a Linux operating system. As one can see from the table, the convergence is quite successful. When the parameter is chosen as in Despres [], the algorithm (3.1) is still running at iteration 1000 for v = v 1 and diverged for v = v. ADOP has been successfully tested also with box-type subdomains. The parameter choice ADOP works at least 15 times better than any other constant parameters, for all numerical tests DD methods for the heat equation When the heat equation is discretized with the backward Euler (BE) or the Crank- Nicolson (CN) method, the problem to be solved at the m-th time level is as follows: (a)?u m + p t um = F m ; x ; (b) u m + u m = g m ; x?; (3.1) where t is the timestep size and p = 1 (BE) or p = (CN). The problem (3.1) is not dicult to solve; its algebraic system is strongly diagonally dominant and most iterative algorithms can nd a good initial guess (u m?1 or (u m?1? u m? )). The problem is much better conditioned than the indenite non-hermitian problem (3.11). When the DD method incorporating ADOP is applied to (3.1), it can be easily expected to converge fast. The DD method turns out to be superior to the PCG- ILU (the conugate gradient method preconditioned by incomplete LU-factorization). Furthermore, it can be parallelized with a maximum eciency, since it requires local

15 Splitting of Tridiagonal Matrices 15 communications only. Details and comparisons with various other algorithms will appear elsewhere [1]. Acknowledgment I sincerely thank Dr. G. Wachspress for his helpful comments and interests. References [1] Y. Cha and S. Kim, Parallelization of ADI methods for elliptic problems. In preparation, 000. [] B. Despres, Domain decomposition method and the Helmholtz problem, in Mathematical and Numerical Aspects of Wave Propagation Phenomena, G. Cohen, L. Halpern, and P. Joly, eds., Philadelphia, 1991, SIAM, pp. {5. [3], Domain decomposition method and the Helmholtz problem (part II), in Mathematical and Numerical Aspects of Wave Propagation, R. Kleinman, ed., Philadelphia, 1993, SIAM, pp. 19{0. [] C. Douglas, S. Malhotra, and M. Schultz, Parallel multigrid with ADIlike smoothers in two dimensions. Preprint, [5], Transpose free alternating direction smoothers for serial and parallel multigrid methods. Preprint, [] J. Douglas, Jr., On the numerical integration methods, J. Soc. Indust. Appl. Math., 3 (1955), pp. {5. @t by implicit [] J. Douglas, Jr. and D. Peaceman, Numerical solution of two-dimensional heat ow problems, American Institute of Chemical Engineering Journal, 1 (1955), pp. 505{51. [8] S. Kim, A parallelizable iterative procedure for the Helmholtz problem, Appl. Numer. Math., 1 (199), pp. 35{9. [9], Parallel multidomain iterative algorithms for the Helmholtz wave equation, Appl. Numer. Math., 1 (1995), pp. 11{9. [10], Domain decomposition iterative procedures for solving scalar waves in the frequency domain, Numer. Math., 9 (1998), pp. 31{59. [11], On the use of rational iterations and domain decomposition methods for solving the Helmholtz problem, Numer. Math., 9 (1998), pp. 59{55.

16 Splitting of Tridiagonal Matrices 1 [1], Numerical methods for parabolic problems. In preparation, 000. [13] S. Kim and M. Lee, Articial damping techniques for scalar waves in the frequency domain, Computers Math. Applic., 31, No. 8 (199), pp. 1{1. [1] P. Lions, On the Schwarz alternating method III: a variant for nonoverlapping subdomains, in Domain Decomposition Methods for Partial Dierential Equations, T. Chan, R. Glowinski, J. Periaux, and O. Widlund, eds., Philadelphia, PA, 1990, SIAM, pp. 0{3. [15] S. Malhotra, C. Douglas, and M. Schultz, Parameter choices for ADIlike methods on parallel computers, Comp. Appl. Math., 1 (1998), pp. 1{3. [1] G. Meurant, A review on the inverse of symmetric tridiagonal and block tridiagonal matrices, SIAM J. Matrix Anal. Appl., 13 (199), pp. 0{8. [1] D. Peaceman and H. Rachford, The numerical solution of parabolic and elliptic equations, J. Soc. Indust. Appl. Math., 3 (1955), pp. 8{1. [18] P.-A. Raviart and J. M. Thomas, A mixed nite element method for second order elliptic problems, in Mathematical Aspects of the Finite Element Method, I. Galligani and E. Magenes, eds., vol. 0 of Lecture Notes in Mathematics, Springer-Verlag, Berlin, New York, 19, pp. 9{315. [19] V. Shaidurov and E. Ogorodnikov, Some numerical methods of solving Helmholtz wave equation, in Mathematical and Numerical Aspects of Wave Propagation Phenomena, G. Cohen, L. Halpern, and P. Joly, eds., Philadelphia, 1991, SIAM, pp. 3{9. [0] W. Tang, Generalized Schwarz splittings, SIAM J. Sci. Stat. Comput., 13 (199), pp. 53{595. [1] R. Varga, Matrix Iterative Analysis, Prentice-Hall, Englewood Clis, NJ, 19. [] E. Wachspress, Optimum alternating-direction-implicit iteration parameters for a model problem, J. Soc. Indust. Appl. Math., 10 (19), pp. 339{350. [3], The ADI model problem, Self-published book, 980 Montego Ct., Windsor, CA 959 USA, 1995.

A MULTIGRID ALGORITHM FOR. Richard E. Ewing and Jian Shen. Institute for Scientic Computation. Texas A&M University. College Station, Texas SUMMARY

A MULTIGRID ALGORITHM FOR. Richard E. Ewing and Jian Shen. Institute for Scientic Computation. Texas A&M University. College Station, Texas SUMMARY A MULTIGRID ALGORITHM FOR THE CELL-CENTERED FINITE DIFFERENCE SCHEME Richard E. Ewing and Jian Shen Institute for Scientic Computation Texas A&M University College Station, Texas SUMMARY In this article,

More information

XIAO-CHUAN CAI AND MAKSYMILIAN DRYJA. strongly elliptic equations discretized by the nite element methods.

XIAO-CHUAN CAI AND MAKSYMILIAN DRYJA. strongly elliptic equations discretized by the nite element methods. Contemporary Mathematics Volume 00, 0000 Domain Decomposition Methods for Monotone Nonlinear Elliptic Problems XIAO-CHUAN CAI AND MAKSYMILIAN DRYJA Abstract. In this paper, we study several overlapping

More information

Optimal Interface Conditions for an Arbitrary Decomposition into Subdomains

Optimal Interface Conditions for an Arbitrary Decomposition into Subdomains Optimal Interface Conditions for an Arbitrary Decomposition into Subdomains Martin J. Gander and Felix Kwok Section de mathématiques, Université de Genève, Geneva CH-1211, Switzerland, Martin.Gander@unige.ch;

More information

ADDITIVE SCHWARZ FOR SCHUR COMPLEMENT 305 the parallel implementation of both preconditioners on distributed memory platforms, and compare their perfo

ADDITIVE SCHWARZ FOR SCHUR COMPLEMENT 305 the parallel implementation of both preconditioners on distributed memory platforms, and compare their perfo 35 Additive Schwarz for the Schur Complement Method Luiz M. Carvalho and Luc Giraud 1 Introduction Domain decomposition methods for solving elliptic boundary problems have been receiving increasing attention

More information

Waveform Relaxation Method with Toeplitz. Operator Splitting. Sigitas Keras. August Department of Applied Mathematics and Theoretical Physics

Waveform Relaxation Method with Toeplitz. Operator Splitting. Sigitas Keras. August Department of Applied Mathematics and Theoretical Physics UNIVERSITY OF CAMBRIDGE Numerical Analysis Reports Waveform Relaxation Method with Toeplitz Operator Splitting Sigitas Keras DAMTP 1995/NA4 August 1995 Department of Applied Mathematics and Theoretical

More information

Kasetsart University Workshop. Multigrid methods: An introduction

Kasetsart University Workshop. Multigrid methods: An introduction Kasetsart University Workshop Multigrid methods: An introduction Dr. Anand Pardhanani Mathematics Department Earlham College Richmond, Indiana USA pardhan@earlham.edu A copy of these slides is available

More information

Domain Decomposition Preconditioners for Spectral Nédélec Elements in Two and Three Dimensions

Domain Decomposition Preconditioners for Spectral Nédélec Elements in Two and Three Dimensions Domain Decomposition Preconditioners for Spectral Nédélec Elements in Two and Three Dimensions Bernhard Hientzsch Courant Institute of Mathematical Sciences, New York University, 51 Mercer Street, New

More information

Multi-Factor Finite Differences

Multi-Factor Finite Differences February 17, 2017 Aims and outline Finite differences for more than one direction The θ-method, explicit, implicit, Crank-Nicolson Iterative solution of discretised equations Alternating directions implicit

More information

t x 0.25

t x 0.25 Journal of ELECTRICAL ENGINEERING, VOL. 52, NO. /s, 2, 48{52 COMPARISON OF BROYDEN AND NEWTON METHODS FOR SOLVING NONLINEAR PARABOLIC EQUATIONS Ivan Cimrak If the time discretization of a nonlinear parabolic

More information

Iterative Methods. Splitting Methods

Iterative Methods. Splitting Methods Iterative Methods Splitting Methods 1 Direct Methods Solving Ax = b using direct methods. Gaussian elimination (using LU decomposition) Variants of LU, including Crout and Doolittle Other decomposition

More information

Numerical Solution Techniques in Mechanical and Aerospace Engineering

Numerical Solution Techniques in Mechanical and Aerospace Engineering Numerical Solution Techniques in Mechanical and Aerospace Engineering Chunlei Liang LECTURE 3 Solvers of linear algebraic equations 3.1. Outline of Lecture Finite-difference method for a 2D elliptic PDE

More information

Numerical Methods I Non-Square and Sparse Linear Systems

Numerical Methods I Non-Square and Sparse Linear Systems Numerical Methods I Non-Square and Sparse Linear Systems Aleksandar Donev Courant Institute, NYU 1 donev@courant.nyu.edu 1 MATH-GA 2011.003 / CSCI-GA 2945.003, Fall 2014 September 25th, 2014 A. Donev (Courant

More information

Scientific Computing WS 2018/2019. Lecture 9. Jürgen Fuhrmann Lecture 9 Slide 1

Scientific Computing WS 2018/2019. Lecture 9. Jürgen Fuhrmann Lecture 9 Slide 1 Scientific Computing WS 2018/2019 Lecture 9 Jürgen Fuhrmann juergen.fuhrmann@wias-berlin.de Lecture 9 Slide 1 Lecture 9 Slide 2 Simple iteration with preconditioning Idea: Aû = b iterative scheme û = û

More information

The Algebraic Multigrid Projection for Eigenvalue Problems; Backrotations and Multigrid Fixed Points. Sorin Costiner and Shlomo Ta'asan

The Algebraic Multigrid Projection for Eigenvalue Problems; Backrotations and Multigrid Fixed Points. Sorin Costiner and Shlomo Ta'asan The Algebraic Multigrid Projection for Eigenvalue Problems; Backrotations and Multigrid Fixed Points Sorin Costiner and Shlomo Ta'asan Department of Applied Mathematics and Computer Science The Weizmann

More information

Linear Algebra Section 2.6 : LU Decomposition Section 2.7 : Permutations and transposes Wednesday, February 13th Math 301 Week #4

Linear Algebra Section 2.6 : LU Decomposition Section 2.7 : Permutations and transposes Wednesday, February 13th Math 301 Week #4 Linear Algebra Section. : LU Decomposition Section. : Permutations and transposes Wednesday, February 1th Math 01 Week # 1 The LU Decomposition We learned last time that we can factor a invertible matrix

More information

2 IVAN YOTOV decomposition solvers and preconditioners are described in Section 3. Computational results are given in Section Multibloc formulat

2 IVAN YOTOV decomposition solvers and preconditioners are described in Section 3. Computational results are given in Section Multibloc formulat INTERFACE SOLVERS AND PRECONDITIONERS OF DOMAIN DECOMPOSITION TYPE FOR MULTIPHASE FLOW IN MULTIBLOCK POROUS MEDIA IVAN YOTOV Abstract. A multibloc mortar approach to modeling multiphase ow in porous media

More information

LINEAR SYSTEMS (11) Intensive Computation

LINEAR SYSTEMS (11) Intensive Computation LINEAR SYSTEMS () Intensive Computation 27-8 prof. Annalisa Massini Viviana Arrigoni EXACT METHODS:. GAUSSIAN ELIMINATION. 2. CHOLESKY DECOMPOSITION. ITERATIVE METHODS:. JACOBI. 2. GAUSS-SEIDEL 2 CHOLESKY

More information

/00 $ $.25 per page

/00 $ $.25 per page Contemporary Mathematics Volume 00, 0000 Domain Decomposition For Linear And Nonlinear Elliptic Problems Via Function Or Space Decomposition UE-CHENG TAI Abstract. In this article, we use a function decomposition

More information

Scientific Computing with Case Studies SIAM Press, Lecture Notes for Unit VII Sparse Matrix

Scientific Computing with Case Studies SIAM Press, Lecture Notes for Unit VII Sparse Matrix Scientific Computing with Case Studies SIAM Press, 2009 http://www.cs.umd.edu/users/oleary/sccswebpage Lecture Notes for Unit VII Sparse Matrix Computations Part 1: Direct Methods Dianne P. O Leary c 2008

More information

Mathematics Research Report No. MRR 003{96, HIGH RESOLUTION POTENTIAL FLOW METHODS IN OIL EXPLORATION Stephen Roberts 1 and Stephan Matthai 2 3rd Febr

Mathematics Research Report No. MRR 003{96, HIGH RESOLUTION POTENTIAL FLOW METHODS IN OIL EXPLORATION Stephen Roberts 1 and Stephan Matthai 2 3rd Febr HIGH RESOLUTION POTENTIAL FLOW METHODS IN OIL EXPLORATION Stephen Roberts and Stephan Matthai Mathematics Research Report No. MRR 003{96, Mathematics Research Report No. MRR 003{96, HIGH RESOLUTION POTENTIAL

More information

A Nonoverlapping Subdomain Algorithm with Lagrange Multipliers and its Object Oriented Implementation for Interface Problems

A Nonoverlapping Subdomain Algorithm with Lagrange Multipliers and its Object Oriented Implementation for Interface Problems Contemporary Mathematics Volume 8, 998 B 0-88-0988--03030- A Nonoverlapping Subdomain Algorithm with Lagrange Multipliers and its Object Oriented Implementation for Interface Problems Daoqi Yang. Introduction

More information

Review of matrices. Let m, n IN. A rectangle of numbers written like A =

Review of matrices. Let m, n IN. A rectangle of numbers written like A = Review of matrices Let m, n IN. A rectangle of numbers written like a 11 a 12... a 1n a 21 a 22... a 2n A =...... a m1 a m2... a mn where each a ij IR is called a matrix with m rows and n columns or an

More information

Solving Large Nonlinear Sparse Systems

Solving Large Nonlinear Sparse Systems Solving Large Nonlinear Sparse Systems Fred W. Wubs and Jonas Thies Computational Mechanics & Numerical Mathematics University of Groningen, the Netherlands f.w.wubs@rug.nl Centre for Interdisciplinary

More information

Today s class. Linear Algebraic Equations LU Decomposition. Numerical Methods, Fall 2011 Lecture 8. Prof. Jinbo Bi CSE, UConn

Today s class. Linear Algebraic Equations LU Decomposition. Numerical Methods, Fall 2011 Lecture 8. Prof. Jinbo Bi CSE, UConn Today s class Linear Algebraic Equations LU Decomposition 1 Linear Algebraic Equations Gaussian Elimination works well for solving linear systems of the form: AX = B What if you have to solve the linear

More information

arxiv: v1 [math.na] 11 Jul 2011

arxiv: v1 [math.na] 11 Jul 2011 Multigrid Preconditioner for Nonconforming Discretization of Elliptic Problems with Jump Coefficients arxiv:07.260v [math.na] Jul 20 Blanca Ayuso De Dios, Michael Holst 2, Yunrong Zhu 2, and Ludmil Zikatanov

More information

Domain Decomposition Algorithms for an Indefinite Hypersingular Integral Equation in Three Dimensions

Domain Decomposition Algorithms for an Indefinite Hypersingular Integral Equation in Three Dimensions Domain Decomposition Algorithms for an Indefinite Hypersingular Integral Equation in Three Dimensions Ernst P. Stephan 1, Matthias Maischak 2, and Thanh Tran 3 1 Institut für Angewandte Mathematik, Leibniz

More information

Preface to the Second Edition. Preface to the First Edition

Preface to the Second Edition. Preface to the First Edition n page v Preface to the Second Edition Preface to the First Edition xiii xvii 1 Background in Linear Algebra 1 1.1 Matrices................................. 1 1.2 Square Matrices and Eigenvalues....................

More information

Department of Mathematics California State University, Los Angeles Master s Degree Comprehensive Examination in. NUMERICAL ANALYSIS Spring 2015

Department of Mathematics California State University, Los Angeles Master s Degree Comprehensive Examination in. NUMERICAL ANALYSIS Spring 2015 Department of Mathematics California State University, Los Angeles Master s Degree Comprehensive Examination in NUMERICAL ANALYSIS Spring 2015 Instructions: Do exactly two problems from Part A AND two

More information

CLASSICAL ITERATIVE METHODS

CLASSICAL ITERATIVE METHODS CLASSICAL ITERATIVE METHODS LONG CHEN In this notes we discuss classic iterative methods on solving the linear operator equation (1) Au = f, posed on a finite dimensional Hilbert space V = R N equipped

More information

OUTLINE 1. Introduction 1.1 Notation 1.2 Special matrices 2. Gaussian Elimination 2.1 Vector and matrix norms 2.2 Finite precision arithmetic 2.3 Fact

OUTLINE 1. Introduction 1.1 Notation 1.2 Special matrices 2. Gaussian Elimination 2.1 Vector and matrix norms 2.2 Finite precision arithmetic 2.3 Fact Computational Linear Algebra Course: (MATH: 6800, CSCI: 6800) Semester: Fall 1998 Instructors: { Joseph E. Flaherty, aherje@cs.rpi.edu { Franklin T. Luk, luk@cs.rpi.edu { Wesley Turner, turnerw@cs.rpi.edu

More information

Convergence Behavior of a Two-Level Optimized Schwarz Preconditioner

Convergence Behavior of a Two-Level Optimized Schwarz Preconditioner Convergence Behavior of a Two-Level Optimized Schwarz Preconditioner Olivier Dubois 1 and Martin J. Gander 2 1 IMA, University of Minnesota, 207 Church St. SE, Minneapolis, MN 55455 dubois@ima.umn.edu

More information

Stabilization and Acceleration of Algebraic Multigrid Method

Stabilization and Acceleration of Algebraic Multigrid Method Stabilization and Acceleration of Algebraic Multigrid Method Recursive Projection Algorithm A. Jemcov J.P. Maruszewski Fluent Inc. October 24, 2006 Outline 1 Need for Algorithm Stabilization and Acceleration

More information

ITERATIVE METHODS FOR NONLINEAR ELLIPTIC EQUATIONS

ITERATIVE METHODS FOR NONLINEAR ELLIPTIC EQUATIONS ITERATIVE METHODS FOR NONLINEAR ELLIPTIC EQUATIONS LONG CHEN In this chapter we discuss iterative methods for solving the finite element discretization of semi-linear elliptic equations of the form: find

More information

Partial Differential Equations

Partial Differential Equations Partial Differential Equations Introduction Deng Li Discretization Methods Chunfang Chen, Danny Thorne, Adam Zornes CS521 Feb.,7, 2006 What do You Stand For? A PDE is a Partial Differential Equation This

More information

Linear Algebra (part 1) : Vector Spaces (by Evan Dummit, 2017, v. 1.07) 1.1 The Formal Denition of a Vector Space

Linear Algebra (part 1) : Vector Spaces (by Evan Dummit, 2017, v. 1.07) 1.1 The Formal Denition of a Vector Space Linear Algebra (part 1) : Vector Spaces (by Evan Dummit, 2017, v. 1.07) Contents 1 Vector Spaces 1 1.1 The Formal Denition of a Vector Space.................................. 1 1.2 Subspaces...................................................

More information

AMS526: Numerical Analysis I (Numerical Linear Algebra)

AMS526: Numerical Analysis I (Numerical Linear Algebra) AMS526: Numerical Analysis I (Numerical Linear Algebra) Lecture 24: Preconditioning and Multigrid Solver Xiangmin Jiao SUNY Stony Brook Xiangmin Jiao Numerical Analysis I 1 / 5 Preconditioning Motivation:

More information

Next topics: Solving systems of linear equations

Next topics: Solving systems of linear equations Next topics: Solving systems of linear equations 1 Gaussian elimination (today) 2 Gaussian elimination with partial pivoting (Week 9) 3 The method of LU-decomposition (Week 10) 4 Iterative techniques:

More information

Solving PDEs with CUDA Jonathan Cohen

Solving PDEs with CUDA Jonathan Cohen Solving PDEs with CUDA Jonathan Cohen jocohen@nvidia.com NVIDIA Research PDEs (Partial Differential Equations) Big topic Some common strategies Focus on one type of PDE in this talk Poisson Equation Linear

More information

AN INTRODUCTION TO DOMAIN DECOMPOSITION METHODS. Gérard MEURANT CEA

AN INTRODUCTION TO DOMAIN DECOMPOSITION METHODS. Gérard MEURANT CEA Marrakech Jan 2003 AN INTRODUCTION TO DOMAIN DECOMPOSITION METHODS Gérard MEURANT CEA Introduction Domain decomposition is a divide and conquer technique Natural framework to introduce parallelism in the

More information

Chapter 1: Systems of linear equations and matrices. Section 1.1: Introduction to systems of linear equations

Chapter 1: Systems of linear equations and matrices. Section 1.1: Introduction to systems of linear equations Chapter 1: Systems of linear equations and matrices Section 1.1: Introduction to systems of linear equations Definition: A linear equation in n variables can be expressed in the form a 1 x 1 + a 2 x 2

More information

Spurious Chaotic Solutions of Dierential. Equations. Sigitas Keras. September Department of Applied Mathematics and Theoretical Physics

Spurious Chaotic Solutions of Dierential. Equations. Sigitas Keras. September Department of Applied Mathematics and Theoretical Physics UNIVERSITY OF CAMBRIDGE Numerical Analysis Reports Spurious Chaotic Solutions of Dierential Equations Sigitas Keras DAMTP 994/NA6 September 994 Department of Applied Mathematics and Theoretical Physics

More information

Chapter 7 Iterative Techniques in Matrix Algebra

Chapter 7 Iterative Techniques in Matrix Algebra Chapter 7 Iterative Techniques in Matrix Algebra Per-Olof Persson persson@berkeley.edu Department of Mathematics University of California, Berkeley Math 128B Numerical Analysis Vector Norms Definition

More information

New constructions of domain decomposition methods for systems of PDEs

New constructions of domain decomposition methods for systems of PDEs New constructions of domain decomposition methods for systems of PDEs Nouvelles constructions de méthodes de décomposition de domaine pour des systèmes d équations aux dérivées partielles V. Dolean?? F.

More information

Scientific Computing: An Introductory Survey

Scientific Computing: An Introductory Survey Scientific Computing: An Introductory Survey Chapter 11 Partial Differential Equations Prof. Michael T. Heath Department of Computer Science University of Illinois at Urbana-Champaign Copyright c 2002.

More information

1e N

1e N Spectral schemes on triangular elements by Wilhelm Heinrichs and Birgit I. Loch Abstract The Poisson problem with homogeneous Dirichlet boundary conditions is considered on a triangle. The mapping between

More information

Preface. 2 Linear Equations and Eigenvalue Problem 22

Preface. 2 Linear Equations and Eigenvalue Problem 22 Contents Preface xv 1 Errors in Computation 1 1.1 Introduction 1 1.2 Floating Point Representation of Number 1 1.3 Binary Numbers 2 1.3.1 Binary number representation in computer 3 1.4 Significant Digits

More information

Numerical Solution of partial differential equations

Numerical Solution of partial differential equations G. D. SMITH Brunei University Numerical Solution of partial differential equations FINITE DIFFERENCE METHODS THIRD EDITION CLARENDON PRESS OXFORD Contents NOTATION 1. INTRODUCTION AND FINITE-DIFFERENCE

More information

Finite difference method for elliptic problems: I

Finite difference method for elliptic problems: I Finite difference method for elliptic problems: I Praveen. C praveen@math.tifrbng.res.in Tata Institute of Fundamental Research Center for Applicable Mathematics Bangalore 560065 http://math.tifrbng.res.in/~praveen

More information

Iterative Methods for Problems in Computational Fluid. Howard C. Elman. University of Maryland. David J. Silvester. University of Manchester

Iterative Methods for Problems in Computational Fluid. Howard C. Elman. University of Maryland. David J. Silvester. University of Manchester Report no. 96/19 Iterative Methods for Problems in Computational Fluid Dynamics Howard C. Elman University of Maryland David J. Silvester University of Manchester Andrew J. Wathen Oxford University We

More information

Contents. Preface... xi. Introduction...

Contents. Preface... xi. Introduction... Contents Preface... xi Introduction... xv Chapter 1. Computer Architectures... 1 1.1. Different types of parallelism... 1 1.1.1. Overlap, concurrency and parallelism... 1 1.1.2. Temporal and spatial parallelism

More information

MATH2210 Notebook 2 Spring 2018

MATH2210 Notebook 2 Spring 2018 MATH2210 Notebook 2 Spring 2018 prepared by Professor Jenny Baglivo c Copyright 2009 2018 by Jenny A. Baglivo. All Rights Reserved. 2 MATH2210 Notebook 2 3 2.1 Matrices and Their Operations................................

More information

Incomplete LU Preconditioning and Error Compensation Strategies for Sparse Matrices

Incomplete LU Preconditioning and Error Compensation Strategies for Sparse Matrices Incomplete LU Preconditioning and Error Compensation Strategies for Sparse Matrices Eun-Joo Lee Department of Computer Science, East Stroudsburg University of Pennsylvania, 327 Science and Technology Center,

More information

Incomplete Block LU Preconditioners on Slightly Overlapping. E. de Sturler. Delft University of Technology. Abstract

Incomplete Block LU Preconditioners on Slightly Overlapping. E. de Sturler. Delft University of Technology. Abstract Incomplete Block LU Preconditioners on Slightly Overlapping Subdomains for a Massively Parallel Computer E. de Sturler Faculty of Technical Mathematics and Informatics Delft University of Technology Mekelweg

More information

Jae Heon Yun and Yu Du Han

Jae Heon Yun and Yu Du Han Bull. Korean Math. Soc. 39 (2002), No. 3, pp. 495 509 MODIFIED INCOMPLETE CHOLESKY FACTORIZATION PRECONDITIONERS FOR A SYMMETRIC POSITIVE DEFINITE MATRIX Jae Heon Yun and Yu Du Han Abstract. We propose

More information

Lecture 18 Classical Iterative Methods

Lecture 18 Classical Iterative Methods Lecture 18 Classical Iterative Methods MIT 18.335J / 6.337J Introduction to Numerical Methods Per-Olof Persson November 14, 2006 1 Iterative Methods for Linear Systems Direct methods for solving Ax = b,

More information

A Domain Decomposition Based Jacobi-Davidson Algorithm for Quantum Dot Simulation

A Domain Decomposition Based Jacobi-Davidson Algorithm for Quantum Dot Simulation A Domain Decomposition Based Jacobi-Davidson Algorithm for Quantum Dot Simulation Tao Zhao 1, Feng-Nan Hwang 2 and Xiao-Chuan Cai 3 Abstract In this paper, we develop an overlapping domain decomposition

More information

Jordan Journal of Mathematics and Statistics (JJMS) 5(3), 2012, pp A NEW ITERATIVE METHOD FOR SOLVING LINEAR SYSTEMS OF EQUATIONS

Jordan Journal of Mathematics and Statistics (JJMS) 5(3), 2012, pp A NEW ITERATIVE METHOD FOR SOLVING LINEAR SYSTEMS OF EQUATIONS Jordan Journal of Mathematics and Statistics JJMS) 53), 2012, pp.169-184 A NEW ITERATIVE METHOD FOR SOLVING LINEAR SYSTEMS OF EQUATIONS ADEL H. AL-RABTAH Abstract. The Jacobi and Gauss-Seidel iterative

More information

JACOBI S ITERATION METHOD

JACOBI S ITERATION METHOD ITERATION METHODS These are methods which compute a sequence of progressively accurate iterates to approximate the solution of Ax = b. We need such methods for solving many large linear systems. Sometimes

More information

Efficient smoothers for all-at-once multigrid methods for Poisson and Stokes control problems

Efficient smoothers for all-at-once multigrid methods for Poisson and Stokes control problems Efficient smoothers for all-at-once multigrid methods for Poisson and Stoes control problems Stefan Taacs stefan.taacs@numa.uni-linz.ac.at, WWW home page: http://www.numa.uni-linz.ac.at/~stefant/j3362/

More information

Multigrid solvers for equations arising in implicit MHD simulations

Multigrid solvers for equations arising in implicit MHD simulations Multigrid solvers for equations arising in implicit MHD simulations smoothing Finest Grid Mark F. Adams Department of Applied Physics & Applied Mathematics Columbia University Ravi Samtaney PPPL Achi Brandt

More information

A NEW EFFECTIVE PRECONDITIONED METHOD FOR L-MATRICES

A NEW EFFECTIVE PRECONDITIONED METHOD FOR L-MATRICES Journal of Mathematical Sciences: Advances and Applications Volume, Number 2, 2008, Pages 3-322 A NEW EFFECTIVE PRECONDITIONED METHOD FOR L-MATRICES Department of Mathematics Taiyuan Normal University

More information

Lecture 9 Approximations of Laplace s Equation, Finite Element Method. Mathématiques appliquées (MATH0504-1) B. Dewals, C.

Lecture 9 Approximations of Laplace s Equation, Finite Element Method. Mathématiques appliquées (MATH0504-1) B. Dewals, C. Lecture 9 Approximations of Laplace s Equation, Finite Element Method Mathématiques appliquées (MATH54-1) B. Dewals, C. Geuzaine V1.2 23/11/218 1 Learning objectives of this lecture Apply the finite difference

More information

x n+1 = x n f(x n) f (x n ), n 0.

x n+1 = x n f(x n) f (x n ), n 0. 1. Nonlinear Equations Given scalar equation, f(x) = 0, (a) Describe I) Newtons Method, II) Secant Method for approximating the solution. (b) State sufficient conditions for Newton and Secant to converge.

More information

Iterative Methods and Multigrid

Iterative Methods and Multigrid Iterative Methods and Multigrid Part 3: Preconditioning 2 Eric de Sturler Preconditioning The general idea behind preconditioning is that convergence of some method for the linear system Ax = b can be

More information

LU Factorization. Marco Chiarandini. DM559 Linear and Integer Programming. Department of Mathematics & Computer Science University of Southern Denmark

LU Factorization. Marco Chiarandini. DM559 Linear and Integer Programming. Department of Mathematics & Computer Science University of Southern Denmark DM559 Linear and Integer Programming LU Factorization Marco Chiarandini Department of Mathematics & Computer Science University of Southern Denmark [Based on slides by Lieven Vandenberghe, UCLA] Outline

More information

Multilevel Preconditioning of Graph-Laplacians: Polynomial Approximation of the Pivot Blocks Inverses

Multilevel Preconditioning of Graph-Laplacians: Polynomial Approximation of the Pivot Blocks Inverses Multilevel Preconditioning of Graph-Laplacians: Polynomial Approximation of the Pivot Blocks Inverses P. Boyanova 1, I. Georgiev 34, S. Margenov, L. Zikatanov 5 1 Uppsala University, Box 337, 751 05 Uppsala,

More information

A Finite Element Method for an Ill-Posed Problem. Martin-Luther-Universitat, Fachbereich Mathematik/Informatik,Postfach 8, D Halle, Abstract

A Finite Element Method for an Ill-Posed Problem. Martin-Luther-Universitat, Fachbereich Mathematik/Informatik,Postfach 8, D Halle, Abstract A Finite Element Method for an Ill-Posed Problem W. Lucht Martin-Luther-Universitat, Fachbereich Mathematik/Informatik,Postfach 8, D-699 Halle, Germany Abstract For an ill-posed problem which has its origin

More information

Algebraic Multigrid as Solvers and as Preconditioner

Algebraic Multigrid as Solvers and as Preconditioner Ò Algebraic Multigrid as Solvers and as Preconditioner Domenico Lahaye domenico.lahaye@cs.kuleuven.ac.be http://www.cs.kuleuven.ac.be/ domenico/ Department of Computer Science Katholieke Universiteit Leuven

More information

Iterative Methods for Solving A x = b

Iterative Methods for Solving A x = b Iterative Methods for Solving A x = b A good (free) online source for iterative methods for solving A x = b is given in the description of a set of iterative solvers called templates found at netlib: http

More information

Algebra C Numerical Linear Algebra Sample Exam Problems

Algebra C Numerical Linear Algebra Sample Exam Problems Algebra C Numerical Linear Algebra Sample Exam Problems Notation. Denote by V a finite-dimensional Hilbert space with inner product (, ) and corresponding norm. The abbreviation SPD is used for symmetric

More information

CHAPTER 6. Direct Methods for Solving Linear Systems

CHAPTER 6. Direct Methods for Solving Linear Systems CHAPTER 6 Direct Methods for Solving Linear Systems. Introduction A direct method for approximating the solution of a system of n linear equations in n unknowns is one that gives the exact solution to

More information

Chapter Two: Numerical Methods for Elliptic PDEs. 1 Finite Difference Methods for Elliptic PDEs

Chapter Two: Numerical Methods for Elliptic PDEs. 1 Finite Difference Methods for Elliptic PDEs Chapter Two: Numerical Methods for Elliptic PDEs Finite Difference Methods for Elliptic PDEs.. Finite difference scheme. We consider a simple example u := subject to Dirichlet boundary conditions ( ) u

More information

INTERGRID OPERATORS FOR THE CELL CENTERED FINITE DIFFERENCE MULTIGRID ALGORITHM ON RECTANGULAR GRIDS. 1. Introduction

INTERGRID OPERATORS FOR THE CELL CENTERED FINITE DIFFERENCE MULTIGRID ALGORITHM ON RECTANGULAR GRIDS. 1. Introduction Trends in Mathematics Information Center for Mathematical Sciences Volume 9 Number 2 December 2006 Pages 0 INTERGRID OPERATORS FOR THE CELL CENTERED FINITE DIFFERENCE MULTIGRID ALGORITHM ON RECTANGULAR

More information

Least-Squares Spectral Collocation with the Overlapping Schwarz Method for the Incompressible Navier Stokes Equations

Least-Squares Spectral Collocation with the Overlapping Schwarz Method for the Incompressible Navier Stokes Equations Least-Squares Spectral Collocation with the Overlapping Schwarz Method for the Incompressible Navier Stokes Equations by Wilhelm Heinrichs Universität Duisburg Essen, Ingenieurmathematik Universitätsstr.

More information

MULTIGRID PRECONDITIONING FOR THE BIHARMONIC DIRICHLET PROBLEM M. R. HANISCH

MULTIGRID PRECONDITIONING FOR THE BIHARMONIC DIRICHLET PROBLEM M. R. HANISCH MULTIGRID PRECONDITIONING FOR THE BIHARMONIC DIRICHLET PROBLEM M. R. HANISCH Abstract. A multigrid preconditioning scheme for solving the Ciarlet-Raviart mixed method equations for the biharmonic Dirichlet

More information

GAUSSIAN ELIMINATION AND LU DECOMPOSITION (SUPPLEMENT FOR MA511)

GAUSSIAN ELIMINATION AND LU DECOMPOSITION (SUPPLEMENT FOR MA511) GAUSSIAN ELIMINATION AND LU DECOMPOSITION (SUPPLEMENT FOR MA511) D. ARAPURA Gaussian elimination is the go to method for all basic linear classes including this one. We go summarize the main ideas. 1.

More information

Preliminary Examination, Numerical Analysis, August 2016

Preliminary Examination, Numerical Analysis, August 2016 Preliminary Examination, Numerical Analysis, August 2016 Instructions: This exam is closed books and notes. The time allowed is three hours and you need to work on any three out of questions 1-4 and any

More information

The Solution of Linear Systems AX = B

The Solution of Linear Systems AX = B Chapter 2 The Solution of Linear Systems AX = B 21 Upper-triangular Linear Systems We will now develop the back-substitution algorithm, which is useful for solving a linear system of equations that has

More information

Convergence of The Multigrid Method With A Wavelet. Abstract. This new coarse grid operator is constructed using the wavelet

Convergence of The Multigrid Method With A Wavelet. Abstract. This new coarse grid operator is constructed using the wavelet Convergence of The Multigrid Method With A Wavelet Coarse Grid Operator Bjorn Engquist Erding Luo y Abstract The convergence of the two-level multigrid method with a new coarse grid operator is studied.

More information

Positive Denite Matrix. Ya Yan Lu 1. Department of Mathematics. City University of Hong Kong. Kowloon, Hong Kong. Abstract

Positive Denite Matrix. Ya Yan Lu 1. Department of Mathematics. City University of Hong Kong. Kowloon, Hong Kong. Abstract Computing the Logarithm of a Symmetric Positive Denite Matrix Ya Yan Lu Department of Mathematics City University of Hong Kong Kowloon, Hong Kong Abstract A numerical method for computing the logarithm

More information

0 Finite Element Method of Environmental Problems. Chapter where c is a constant dependent only on and where k k 0; and k k ; are the L () and H () So

0 Finite Element Method of Environmental Problems. Chapter where c is a constant dependent only on and where k k 0; and k k ; are the L () and H () So Chapter SUBSTRUCTURE PRECONDITIONING FOR POROUS FLOW PROBLEMS R.E. Ewing, Yu. Kuznetsov, R.D. Lazarov, and S. Maliassov. Introduction Let be a convex polyhedral domain in IR, f(x) L () and A(x) be a suciently

More information

Block-tridiagonal matrices

Block-tridiagonal matrices Block-tridiagonal matrices. p.1/31 Block-tridiagonal matrices - where do these arise? - as a result of a particular mesh-point ordering - as a part of a factorization procedure, for example when we compute

More information

6. Iterative Methods for Linear Systems. The stepwise approach to the solution...

6. Iterative Methods for Linear Systems. The stepwise approach to the solution... 6 Iterative Methods for Linear Systems The stepwise approach to the solution Miriam Mehl: 6 Iterative Methods for Linear Systems The stepwise approach to the solution, January 18, 2013 1 61 Large Sparse

More information

the sum of two projections. Finally, in Section 5, we apply the theory of Section 4 to the case of nite element spaces. 2. Additive Algorithms for Par

the sum of two projections. Finally, in Section 5, we apply the theory of Section 4 to the case of nite element spaces. 2. Additive Algorithms for Par ON THE SPECTRA OF SUMS OF ORTHOGONAL PROJECTIONS WITH APPLICATIONS TO PARALLEL COMPUTING PETTER E. BJRSTAD y AND JAN MANDEL z Abstract. Many parallel iterative algorithms for solving symmetric, positive

More information

Fraction-free Row Reduction of Matrices of Skew Polynomials

Fraction-free Row Reduction of Matrices of Skew Polynomials Fraction-free Row Reduction of Matrices of Skew Polynomials Bernhard Beckermann Laboratoire d Analyse Numérique et d Optimisation Université des Sciences et Technologies de Lille France bbecker@ano.univ-lille1.fr

More information

Stability of implicit extrapolation methods. Abstract. Multilevel methods are generally based on a splitting of the solution

Stability of implicit extrapolation methods. Abstract. Multilevel methods are generally based on a splitting of the solution Contemporary Mathematics Volume 00, 0000 Stability of implicit extrapolation methods Abstract. Multilevel methods are generally based on a splitting of the solution space associated with a nested sequence

More information

ECE539 - Advanced Theory of Semiconductors and Semiconductor Devices. Numerical Methods and Simulation / Umberto Ravaioli

ECE539 - Advanced Theory of Semiconductors and Semiconductor Devices. Numerical Methods and Simulation / Umberto Ravaioli ECE539 - Advanced Theory of Semiconductors and Semiconductor Devices 1 General concepts Numerical Methods and Simulation / Umberto Ravaioli Introduction to the Numerical Solution of Partial Differential

More information

A Robust Preconditioned Iterative Method for the Navier-Stokes Equations with High Reynolds Numbers

A Robust Preconditioned Iterative Method for the Navier-Stokes Equations with High Reynolds Numbers Applied and Computational Mathematics 2017; 6(4): 202-207 http://www.sciencepublishinggroup.com/j/acm doi: 10.11648/j.acm.20170604.18 ISSN: 2328-5605 (Print); ISSN: 2328-5613 (Online) A Robust Preconditioned

More information

NUMERICAL METHODS FOR ENGINEERING APPLICATION

NUMERICAL METHODS FOR ENGINEERING APPLICATION NUMERICAL METHODS FOR ENGINEERING APPLICATION Second Edition JOEL H. FERZIGER A Wiley-Interscience Publication JOHN WILEY & SONS, INC. New York / Chichester / Weinheim / Brisbane / Singapore / Toronto

More information

Simple Examples on Rectangular Domains

Simple Examples on Rectangular Domains 84 Chapter 5 Simple Examples on Rectangular Domains In this chapter we consider simple elliptic boundary value problems in rectangular domains in R 2 or R 3 ; our prototype example is the Poisson equation

More information

An Iterative Substructuring Method for Mortar Nonconforming Discretization of a Fourth-Order Elliptic Problem in two dimensions

An Iterative Substructuring Method for Mortar Nonconforming Discretization of a Fourth-Order Elliptic Problem in two dimensions An Iterative Substructuring Method for Mortar Nonconforming Discretization of a Fourth-Order Elliptic Problem in two dimensions Leszek Marcinkowski Department of Mathematics, Warsaw University, Banacha

More information

Numerical Analysis: Solutions of System of. Linear Equation. Natasha S. Sharma, PhD

Numerical Analysis: Solutions of System of. Linear Equation. Natasha S. Sharma, PhD Mathematical Question we are interested in answering numerically How to solve the following linear system for x Ax = b? where A is an n n invertible matrix and b is vector of length n. Notation: x denote

More information

Partitioned Methods for Multifield Problems

Partitioned Methods for Multifield Problems Partitioned Methods for Multifield Problems Joachim Rang, 10.5.2017 10.5.2017 Joachim Rang Partitioned Methods for Multifield Problems Seite 1 Contents Blockform of linear iteration schemes Examples 10.5.2017

More information

1 Multiply Eq. E i by λ 0: (λe i ) (E i ) 2 Multiply Eq. E j by λ and add to Eq. E i : (E i + λe j ) (E i )

1 Multiply Eq. E i by λ 0: (λe i ) (E i ) 2 Multiply Eq. E j by λ and add to Eq. E i : (E i + λe j ) (E i ) Direct Methods for Linear Systems Chapter Direct Methods for Solving Linear Systems Per-Olof Persson persson@berkeleyedu Department of Mathematics University of California, Berkeley Math 18A Numerical

More information

On the performance of the Algebraic Optimized Schwarz Methods with applications

On the performance of the Algebraic Optimized Schwarz Methods with applications On the performance of the Algebraic Optimized Schwarz Methods with applications Report 12-1-18 Department of Mathematics Temple University October 212 This report is available at http://www.math.temple.edu/~szyld

More information

AIMS Exercise Set # 1

AIMS Exercise Set # 1 AIMS Exercise Set #. Determine the form of the single precision floating point arithmetic used in the computers at AIMS. What is the largest number that can be accurately represented? What is the smallest

More information

Geometric Multigrid Methods

Geometric Multigrid Methods Geometric Multigrid Methods Susanne C. Brenner Department of Mathematics and Center for Computation & Technology Louisiana State University IMA Tutorial: Fast Solution Techniques November 28, 2010 Ideas

More information

Linear Algebra and Matrix Inversion

Linear Algebra and Matrix Inversion Jim Lambers MAT 46/56 Spring Semester 29- Lecture 2 Notes These notes correspond to Section 63 in the text Linear Algebra and Matrix Inversion Vector Spaces and Linear Transformations Matrices are much

More information

Solutions Preliminary Examination in Numerical Analysis January, 2017

Solutions Preliminary Examination in Numerical Analysis January, 2017 Solutions Preliminary Examination in Numerical Analysis January, 07 Root Finding The roots are -,0, a) First consider x 0 > Let x n+ = + ε and x n = + δ with δ > 0 The iteration gives 0 < ε δ < 3, which

More information

1. Fast Iterative Solvers of SLE

1. Fast Iterative Solvers of SLE 1. Fast Iterative Solvers of crucial drawback of solvers discussed so far: they become slower if we discretize more accurate! now: look for possible remedies relaxation: explicit application of the multigrid

More information