NAVIER-STOKES OPERATORS HOWARD C. ELMAN UMIACS-TR November Revised July 1996

Size: px
Start display at page:

Download "NAVIER-STOKES OPERATORS HOWARD C. ELMAN UMIACS-TR November Revised July 1996"

Transcription

1 PERTURBATION OF EIGENVALUES OF PRECONDITIONED NAVIER-STOKES OPERATORS HOWARD C. ELMAN Report CS-TR-3559 UMIACS-TR-95- November 995 Revised July 996 Abstract. We study the sensitivity of algebraic eigenvalue problems associated with matrices arising from linearization and discretization of the steady-state Navier-Stokes equations. In particular, for several choices of preconditioners applied to the system of discrete equations, we derive upper bounds on perturbations of eigenvalues as functions of the viscosity and discretization mesh size. The bounds suggest that the sensitivity of the eigenvalues is at worst linear in the inverse of the viscosity and quadratic in the inverse of the mesh size, and that scaling can be used to decrease the sensitivity in some cases. Experimental results supplement these results and conrm the relatively mild dependence on viscosity. They also indicate a dependence on the mesh size of magnitude smaller than the analysis suggests. Department of Computer Science and Institute for Advanced Computer Studies, University of Maryland, College Park, MD 74, elman@cs.umd.edu. This work was supported by the U. S. Army Research Oce under grant DAAL-39-G-6 and the U. S. National Science Foundation under grants ASC and DMS

2 PERTURBATION OF EIGENVALUES OF PRECONDITIONED NAVIER-STOKES OPERATORS HOWARD C. ELMAN Abstract. We study the sensitivity of algebraic eigenvalue problems associated with matrices arising from linearization and discretization of the steady-state Navier-Stokes equations. In particular, for several choices of preconditioners applied to the system of discrete equations, we derive upper bounds on perturbations of eigenvalues as functions of the viscosity and discretization mesh size. The bounds suggest that the sensitivity of the eigenvalues is at worst linear in the inverse of the viscosity and quadratic in the inverse of the mesh size, and that scaling can be used to decrease the sensitivity in some cases. Experimental results supplement these results and conrm the relatively mild dependence on viscosity. They also indicate a dependence on the mesh size of magnitude smaller than the analysis suggests. Key words. eigenvalues, perturbation analysis, Navier-Stokes, preconditioning. AMS(MOS) subject classications. Primary 65F, 65N; secondary 5A6. Introduction. This paper concerns properties of the eigenvalues of matrices arising from the discrete linearized steady-state Navier-Stokes equations. The continuous problem is () u (u grad) u grad p = f in ; together with the incompressibility constraint () div u = in ; subject to suitable boundary conditions where is an open bounded domain in R. These equations constitute a fundamental problem in computational uid dynamics, see e.g., [],[6],[8]. The two-dimensional vector eld u represents the velocity in, p represents pressure, and the scalar is the viscosity, roughly speaking, the ratio of convection to diusion in the system. A methodology for computing the numerical solution is to discretize (){() using nite dierence or nite element methods, and then to solve the resulting nonlinear system by some iterative method. Linearization leads to a set of matrix equations of the form F B T u (3) = f ; B p where u and p now represent discrete versions of velocity and pressure, respectively. We will restrict our attention to the discrete Oseen equations (4) u (w grad) u grad p = f div u = Department of Computer Science and Institute for Advanced Computer Studies, University of Maryland, College Park, MD 74, elman@cs.umd.edu. This work was supported by the U. S. Army Research Oce under grant DAAL-39-G-6 and the U. S. National Science Foundation under grants ASC and DMS

3 where w is given such that div w =. These equations arise from a nonlinear iteration of the form u (m) (u (m ) grad) u (m) grad p (m) = f; div u (m) =, see []. In this case F = A N where A consists of a pair of uncoupled discrete Laplace operators, corresponding to diusion, and N is a skew-symmetric matrix representing convection. We will also assume that the velocity and pressure discretizations are div-stable; see, e.g., [], p. 57, [8], pp., [6]. In matrix notation, this is equivalent to the condition (5) (p; BA B T p) (p; Mp) for all p ; where (; ) denotes the Euclidean inner product, and are constants that are independent of the discretization mesh size h, and for nite elements M is the pressure mass matrix, i.e., the Grammian matrix of basis functions dening the discrete pressure space. For nite dierences on uniform grids, a natural analogue is M = h I. Let L denote the coecient matrix of (3). The following preconditioning matrices were introduced in [3] a block diagonal preconditioner (6) Q D = F M and a block triangular preconditioner (7) Q T = F BT M It was shown in [3] that the eigenvalues of each of the preconditioned matrices A D = LQ D and A T = LQ T are uniformly bounded independent of the mesh size used in the discretization. Numerical experiments also suggested that Krylov subspace iterative methods such as the generalized minimal residual (GMRES) [3] and quasiminimal residual (QMR) methods [5] can be used to solve the preconditioned system with iteration counts independent of the mesh size. We are concerned with the sensitivity of the eigenvalues of the preconditioned Oseen matrix using the two preconditioners (6) and (7). Motivation for studying this lies in the fact that use of either preconditioner in an iteration entails applying the action of the inverse of the matrix of either (6) or (7) to a vector at each step. These in turn require the computation of the action of F, which, if direct methods are used, will dominate the cost. An alternative that was considered in [3] is to approximate the action of F (i.e., compute an approximate solution to systems with coecient matrix F ) using an inner iteration. Unless very stringent stopping criteria are used here, the resulting preconditioned operators can be viewed as perturbations of those of (6) { (7). Thus, we are interested the sensitivity of the eigenvalues to perturbation. If the preconditioned matrix is perturbed by matrix of size, then the perturbations of eigenvalues will depend on and also on parameters associated with the An inequality analogous to (5) also holds, with dierent constants, if M is any matrix spectrally equivalent to the mass matrix; for example, M could be the diagonal matrix consisting of the diagonal of the mass matrix [7]. In the sequel, we will not distinguish among such possibilities for M.

4 underlying problem, specically, the viscosity and mesh size h. In this paper, we examine these dependencies using a combination of analytic and experimental results. The analysis derives from Wilkinson's classical perturbation analysis [8], which shows that if there are no nonlinear elementary divisors, the perturbations are of magnitude O(). The analytic bounds are stated as functions of and h but they also depend on properties of certain matrices associated with the Schur complement BF B T derived from (4). The latter quantities are studied in a series of numerical experiments. The combination of analytic and experimental results indicates that there is an increase in sensitivity to perturbation as the viscosity decreases, with growth roughly linear in =. This eect can be mitigated to some extent by scaling the rst equation of (4) (the momentum equation). The bounds also establish linear dependence on =h with the preconditioner Q T and quadratic dependence with Q D, although the experimental results suggest that perturbations are considerably less sensitive to this parameter. An outline of the paper is as follows. In x, we derive preliminary bounds and relations for several operators associated with the preconditioned matrices. In x3, we derive the analytic perturbation bounds for the block tridiagonal preconditioner, and in x4, we present the analysis for the block diagonal preconditioner. For simplicity, the analysis is done for the case where the coecient matrix of (4) has full rank, although often in practice (and in our experiments) it is rank decient by one because the pressure p is uniquely dened only up to a constant. In x5, we show that the analytic results carry over to this case. In x6, we present the experimental results, and in x7, we show how the analysis applies for the case of inexact computation of the action of F.. Preliminary results. In this section we derive preliminary bounds and relations for several operators associated with the preconditioners (6) { (7). We will assume that the discrete problem (3) arises from a standard nite dierence or loworder nite element scheme on a uniform grid with mesh size h, and that the discrete problem is scaled so that the extreme eigenvalues of the discrete Laplace operators of A are contained in an interval of the form [c h ; c ], where here and below c i denotes a generic constant that is independent of h and. This is a natural scaling for nite elements, and for nite dierences on a uniform grid it corresponds to the ve-point operator with 4 in the diagonal entries and in the o-diagonal entries. With this normalization, BB T =h is also a scaled discrete Laplace operator and its eigenvalues are contained in an interval of the same form. Let the discrete velocity and pressure spaces have dimension n u and n p, respectively. For div-stable discretizations, n u n p, and typically n u is signicantly larger than n p. It will be convenient to use the symbol Q instead of M in the matrices of (6). The preconditioned matrices are then given by (8) A D = F BT B F = Q I G K T for the block diagonal preconditioner, and (9) A T = F BT B F F B T Q = Q I G H For two-dimensional problems, the vector u has two components of grid vectors, and stability considerations often also lead to more grid points for velocity than for pressure [8]. 3

5 for the block triangular preconditioner. The submatrices on the right of (8) { (9) are G = BF ; K T = B T Q ; H = GK T = BF B T Q The identity matrices are of order n u. We recall some results from [3], which give bounds on the eigenvalues of H. Let S = BF B T denote the Schur complement matrix for (3), and let C = B F F T B T ; R = B F F T denote the symmetric and skew-symmetric parts of S, respectively. It is shown in [3] that B T ; () (q; Cq) (q; Qq) ; j(q; Rq)j (q; Qq) ; where and are as in (5) and is the largest eigenvalue of A N, which is also uniformly bounded independent of h [4]. Consequently, Bendixson's theorem [4, p. 48] implies that the eigenvalues of H are contained in the box () " ; # " ; in the complex plane. We rst derive bounds on the singular values of G, which will be used in the perturbation analysisis for the block triangular preconditioner. Theorem.. The largest singular value of G is bounded above by a quantity of magnitude O(=) which is independent of h as h. The smallest singular value of G is bounded below by a quantify of magnitude O(h) which is bounded independent of as. Proof. The singular values of G are the square roots of the eigenvalues of GG T, and the largest and smallest of these eigenvalues are the extrema of (q; GG T q)=(q; q). This Rayleigh quotient can be rewritten as # () (q; GG T q) (q; q) = (F T B T q; F T B T q) (q; q) = (F T B T q; F T B T q) (F T B T q; B T q) (F T B T q; B T q) (q; q) We consider the two terms in the product on the right of () separately. For the rst term, the substitution w = F T B T q gives (F T B T q; F T B T q) (F T B T q; B T q) = (w; w) (w; F T w) = (w; w) (w; F F T w) = (w; w) (w; Aw) Under the assumption on the scaling of the discrete Laplacian operators composing A, it follows that (3) c (F T B T q; F T B T q) c h (F T B T q; B T q) 4

6 The second term of the product in () is (4) (q; BF B T q) (q; q) = (q; Cq) (q; q) (q; Cq) (q; Qq) = (q; Qq) (q; q) It is well known (see [7]) that the pressure mass matrix is spectrally equivalent to h I, so that (5) c h (q; Qq) (q; q) c h Thus, the bounds for the symmetric part in () together with () { (5) imply c h (q; GGT q) c (q; q) The singular values of K will be used to analyze the block diagonal preconditioner. Lemma.. The largest singular value of K is bounded above by a quantity of magnitude O h. The smallest singular value is bounded below by a quantity of magnitude O(). Proof. The largest singular value is kkk = kk T k. Using K T = B T Q and Q = M, we have kk T k kb T k km k But km k = O(h ), and our assumptions on B imply that kb T k = kbb T k = = O(h). The smallest singular value is the inverse of k(kk T ) k =. Then k(kk T ) k kmk k(bb T ) k = c h 4 c h 4 Consider an alternative scaling in problems () and (4) in which the rst equation is multiplied by. This does not change the solutions, but as we will show in xx3{4, it aects the sensitivity of discrete eigenvalues. For (4), scaling gives (6) u (w grad) u grad p = f The new discrete problem is as in (3) except that F, p and f are replaced by ^F = F, p and f, respectively. Let ^Q = Q = M, ^G = B ^F, ^KT = B T ^Q, and ^H = ^G ^KT. The analogues of () and the bounds of Theorem. and Lemma. are given below. The proof follows from the facts that ^G = G, ^K = K and ^H = H. Corollary.3. With the scaling of (6), the eigenvalues of ^H are contained in the box (); the singular values of ^G are bounded above by a quantity of magnitude O() and below by a quantity of magnitude O(h); and the singular values of ^K are bounded above by a quantity of magnitude O and below by a quantity of h magnitude O(). 5

7 In the following, we will not specically identify the matrices associated with this scaling using the \hat"-symbol. Instead, we will use the notation of (8) { (9) to refer generically to both scalings. We will distinguish them as derived from either the \original" formulation (4) or the \scaled" formulation (6) of the Oseen equations. Finally, we will use the notation G;K to denote the secant of the largest principal angle between Range(G T ) and Range(K T ). That is, if Q G and Q K are matrices whose columns represent orthogonal bases of Range(G T ) and Range(K T ), respectively, then (7) G;K = k(q T G Q K) k = min (Q T G Q K) ; where min denotes the smallest singular value (see [7, p. 584]). It is easily shown (e.g., using QR decompositions) that (8) G;K = kk T (GK T ) Gk 3. Analytic bounds for the block triangular preconditioner. It is evident from (9) that the eigenvalues of A T consist of = of multiplicity n u together with the eigenvalues of H. We seek a factorization (9) A T = V T D T V T that provides insight into the sensitivity of these eigenvalues to perturbation. We will look for factors of the form () D T = I n u n p n u n p D nu n p ; V T = V V V V nu n p where the sub-blocks of D T and V T must be determined and the dimensions are as indicated. Let H have Jordan canonical form H = P P. The requirement A T V T = V T D T is satised if () V = I; V = ; G = (I H)V P D ; V = P We distinguish between two cases, (H) and = (H). Suppose rst that H has no eigenvalues equal to. The choice D = in () leads to V = (I H) G. In this case, (9) represents a Jordan form for A T with () V T = I (I H) G P Let A T () = A T E be a perturbation of A T, where E = E E (3) E E ; V T = I P (I H) G P The classical perturbation analysis of Wilkinson based on Gerschgorin theory [8, pp. 7] shows that for every eigenvalue of A T with only linear elementary divisors, perturbations of in (A T ()) are contained in a circle centered at with radius of 6

8 size c where c is independent of. The structure of V T can be used to obtain further insight into the sizes of the perturbations. Let ^E T = V T EV ^E T = ^E (4) ; ^E ^E so that we are concerned with the eigenvalues of (5) V T A T ()V T = D T ^ET = I ^E ^E ^E ^E Here and in the following, the symbol \c" represents a generic constant that is independent of the parameters h, and. Lemma 3.. If 6 (H), then there are n u eigenvalues ^ of A T () (counting multiplicity) satisfying (6) j^ j k ^E k c k ^E k If there are m n p eigenvalues of H with linear elementary divisors, then there are m eigenvalues ^ of D T (), distinct from those of (6), that satisfy (7) j^ j k ^E k c k ^E k Proof. Multiplying the rst block row on the right side of (5) by =, multiplying the rst block column by = and then applying Gerschgorin's theorem leads to the bound ^ ( [ ^E ] ii ) X j6=i [ ^E ] ij = X j [ ^E ] ij ; where is such that the Gerschgorin disks for the rst and second blocks of the scaled matrix are disjoint (see [8, p. 73]). Assertion (6) follows. The argument for (7) is identical, applied to the second block row of (5). Thus, we can restrict our attention to the block diagonal entries of ^E T. Using (){(4) we have (8) ^E = E E (I H) G ; ^E = P (I H) GE P P E P The following result gives bounds on the perturbations of eigenvalues as functions of the viscosity and mesh size h, for a perturbation satisfying kek. 3 Theorem 3.. Assume kek. If 6 (H), then the eigenvalues ^ of A T () that are perturbations of = satisfy (9) j^ j 8 >< > c h O( ) c h O( ) for the original formulation; for the scaled formulation; 3 This assumption is stronger than the inequality je ijj used by Wilkinson; the latter condition follows from our assumption. 7

9 where c = c k(i H) k. For eigenvalues of H with linear elementary divisors, the perturbations ^ (A T ()) satisfy (3) j^ j 8 >< > c h O( ) c h O( ) for the original formulation; for the scaled formulation; where c = c (P ) k(i H) k. Proof. The relations (8) together with standard bounds on matrix l p -norms give k ^E k c h k ^E k c h k(i H) k kgk ; k ^E k c h k ^E k c h (P ) k(i H) k kgk The conclusions follow from the upper bounds on kgk in Theorem.. In the case (H), we use an alternative version of (). By analogy with the analysis above, let V = (I H) y G where (I H) y is the pseudo-inverse. Then V T = I (I H) y G P The similarity (9) then holds for the choice D T = ; V T = I P (I H) y G P I P G where is the orthogonal projection onto the null space of I H T. The perturbation of D T associated with A T () is then V T A T ()V T = D T ^E T = I ^E ^E (3) P G ^E ^E First consider the eigenvalues of H dierent from with linear elementary divisors. Let be a parameter in (; ). Multiplying the second block row of (3) by and the second block column by produces the matrix (3) I ^E ^E P G ^E ^E For any (; ) and small enough, the Gerschgorin disk for a perturbation ^ of (H), 6=, is disjoint from the disks corresponding to perturbations of the eigenvalue. Consequently, if has only linear elementary divisors, then j^ j kp Gk O() For eigenvalues equal to, multiplying the second block row of (3) by = and multiplying the second block column by = produces I ^E = ^E (33) = P G 3= ^E ^E 8

10 Thus, for = with linear or quadratic elementary divisors, the perturbations satisfy j^ j = max k ^E k ; kp Gk O() Bounding the matrix innity norms in (3) and (33) gives the following result. Theorem 3.3. If (H) with linear or quadratic elementary divisors, then the eigenvalues ^ of A T () that are perturbations of = satisfy (34) j^ j = c h max kp k ; kp k kgk O() For (H) dierent from with linear elementary divisors, the perturbations ^ (A T ()) satisfy (35) j^ j c h kp k kgk O() for any (; ). Here kgk = O(=) in the original formulation and kgk = O() for the scaled formulation. Note that for any (A T ) with nonlinear elementary divisors, bounds analogous to (9){(3) and (34){(35) can be obtained in which the dependence on is of the form =r where r is the order of the largest Jordan block for ; see [8, p. 79]. 4. Analytic bounds for the block diagonal preconditioner. For the block diagonal preconditioner, we have results only in the case where the Schur complement matrix H has no nonlinear elementary divisors. As in Theorem 3., the perturbation analysis will be stated in terms of properties of P, a xed matrix of eigenvectors of H. In addition, some of the bounds depend on G;K, the secant of the largest principal angle between Range(G T ) and Range(K T ), as dened in (7). Consider the eigenvalue problem for A D, This leads to the conditions I G K T u p = u p (36) K T p = ( )u; Gu = p on the components of the eigenvectors. One solution corresponds to the eigenvalue = ; the assumption that B and therefore K have full rank then implies that any associated eigenvector satises p =, Gu =. We use two approaches to identify eigenvalues 6= and corresponding eigenvectors.. Substitution of u = K T p into the second equation of (36) gives GK T p = ( ) p That is, any eigenpair (; p) of H = GK T leads to two eigenvalues (37) of A D. The associated eigenvectors are (38) = p 4 ; = p 4 u p ; 9 u p ;

11 where (39) u = K T p; u = K T p. Alternatively, substitution of p = Gu into the rst equation of (36) gives K T Gu = ( ) u K T G has a zero eigenvalue of multiplicity n u n p (the dimension of the null space of G), plus n p nonzero eigenvalues. If (; u) is an eigenpair with 6=, then (37) denes a pair of eigenvalues of A D. In this case, the eigenvectors have the form (4) where (4) p = u ; p u ; p Gu; p = Gu It is straightforward to show that any u or u from (39) is an eigenvector of K T G, and any p or p from (4) is an eigenvector of GK T. Therefore, we will use the symbol \" to refer to the pairs of eigenvalues and eigenvectors of A D, i.e., u will refer to the rst entries of either (38) or (4), p to the second entries, and to the associated eigenvalue. We have established the following result. Theorem 4.. The eigenvectors of A D corresponding to the eigenvalue = have the form u where Gu =. The eigenvectors corresponding to eigenvalues dierent from have the form u, where the associated eigenvalues p satisfy (37), and Hp = p ; K T Gu = u ; p = Gu ; u = K T p Let denote a diagonal matrix with entries f g from (37). If the eigenvalues of H have only linear elementary divisors, then A D V D = V D D D, where B C D D I A n u n p n p ; V D = V U U C A n u ; n p P P n p n u n n p n p n u p n p n p in which the columns of V form an orthogonal basis of the null space of G and (4) P = GU ; U = K T P ( I) We will normalize the matrices U and P as follows n p (43) U = U = U = K T P; P = P ( I); P = P ( I)

12 It is then easily veried that V D = I ( ) ( ) C A P G P P G P C A ; where (44) = V T (I U(GU) G) = V T (I K T (GK T ) G) Now let A D () = A D E be a perturbation of A D where E is as in (3). The eigenvalues of this perturbed matrix are the same as those of (45) V D A D()V D = D D ^ED = I ^E ^E ^E 3 ^E ^E ^E 3 ^E 3 ^E 3 ^E 33 C A ; where ^E D = V D EV D. As in x3, Gerschgorin analysis implies that for small, the eects of perturbation can be bounded using the perturbations on the block diagonal of (45). These are given by (46) ^E = T E V; ^E = ( ) hp G(E U E P ) P (E U E P ) ; ^E 33 = ( ) hp G(E U E P ) P (E U E P ) i i Theorem 4.. If H has no nonlinear elementary divisors, then the eigenvalues ^ of A D () that are perturbations of = satisfy j^ j c h ( G;K) O( ) for both the original and scaled formulations of the discrete Oseen equations. The perturbations of eigenvalues dierent from satisfy (47) j^ j and (48) 8 >< > j^ j c (P ) G;K h c h (P ) G;K 8 >< > c h O( ) O( ) for the original formulation; for the scaled formulation; c (P ) h c O( ) for the original formulation; h (P ) c h O( ) for the scaled formulation Proof. For the perturbations of =, it follows from (8), (44) and (45) that j^ j c k ^E k c h k ^E k c h ( G;K)

13 For the perturbations of eigenvalues dierent from, note that (37) and the bounds () on the eigenvalues of H imply that the norm of each of the diagonal matrices (49) I; I; ( ) ; ; ( I) ; is bounded independent of and h. Gerschgorin analysis gives (5) j^ j c h k ^Ejj k ; where j = corresponds to and j = 3 to. Relations (43) and (46) then imply (5) k ^Ejj k k( ) k kp Gk kp k kk T k k Ik kp k Consider kp Gk. Using the expression for P in (43), we have kp Gk k( I) k kp k kgk ; and combining this with (5) gives k ^E 33 k (P ) k( ) k k( I) k kgk kk T k k Ik Result (48) follows from (5), the boundedness of the matrices of (49), and the bounds on kgk and kk T k in Lemma. and Corollary.3. For kp Gk, rst observe that (4) and (43) imply (P G)(K T P ) = I Let (P G) T = Q R and K T P = Q R be QR factorizations, where R and R are square and nonsingular. Then R T QT Q R = I, and (5) kp Gk = kr k kr k k(q T Q ) k = G;K kr k But kr k is the inverse of the smallest singular value of K T P, so that kr k k P (KK T ) P T k = k k kp k k(kk T ) k = From Lemma., Corollary.3 and the boundedness of k k, the term on the right is bounded by c= kp k for the original formulation and c kp k for the alternative formulation. Result (47) then follows from (5), (5) (with j = ) and (5). Remark 4.. The dierence between (47) and (48) stems from the fact that k( I) k is not independent of, so we cannot bound kp Gk directly. The results of x6 suggest that the perturbations do not behave dierently, but we see no way to avoid including G;K in (47). 5. The rank-decient case. Unless an additional constraint is imposed on the pressure in (){() or (4), the matrix of (3) will be rank decient by one. This is the case for the test problems of x6. We outline here how the analysis above carries over in the rank-decient case.

14 For the block triangular preconditioner, A T V T = V T D T where, for 6 (H), the analogues of the matrices dened in () { () are D T = I I ; V T = ; V y (I H) G P T = I P y (I H) G P y is a square matrix of order n p whose eigenvalues are the nonzero eigenvalues of H, and P spans the associated invariant subspace. In particular, P y replaces P and V y T replaces V T. The analysis of perturbations of the nonzero eigenvalues of A T then carries through verbatim with the inverse of the smallest singular value of P in place of kp k. The case (H) is generalized in a similar manner with (I H) y in place of (I H) For the block diagonal preconditioner, the analogue of the similarity transformation of Theorem 4. is A D V D = V D D D where B D D I n u n p n p n p C A nu np n p ; V D = n p V U U P P n u n p n p Once again, all the analysis of x4 goes through with P referring to the matrix of eigenvectors corresponding to nonzero eigenvalues of H, P y in place of P, and the inverse of the smallest nonzero eigenvalue of KK T in place of k(kk T ) k. 6. Experimental results. We now present the results of numerical experiments that supplement the analysis of xx3{5. Our test problem is a discrete Oseen operator (4) on on = ( ; ) ( ; ), with Dirichlet boundary conditions u = u = on the three xed walls (x =, y =, x = ), and u =, u = on the moving wall (y = ). The coecients of the convection terms describe a circular vortex w = y( x ); w = x( y ) We discretize using bilinear nite elements on a uniform rectangular n n velocity grids of width h = =n, augmented by streamline upwinding [9, p. 85]. To impose div-stability, the pressure discretization uses \macro-elements" of width h; see [8, p. 3]. The hydrostatic pressure is not explicitly specied, so that the matrices (3) are rank-decient by one. Additional details about this problem are given in [3]. All computations were performed in MATLAB on either a Sun SPARC- workstation or a DEC-Alpha 4/75 workstation. We rst show in Figure some sample distributions of eigenvalues of A T and A D, for = = and n = 3. The plot on the left gives an indication of the rectangle enclosing the eigenvalues of H (see ()); the plot on the right represents the result of the mapping 7 p 4 = of (37). Both pictures include the eigenvalue, of multiplicity n u = 78 for A T and n u n p = 89 for A D. We present our results primarily as tabulations of maximum perturbations and other quantities, for various choices of viscosity parameter and grid parameter n. The rows and columns of the tables indicate behavior as either or n gets large (h ). Note that accurate discrete solutions to (4) are obtained only if is not too 3 n p C A n u n p

15 Fig.. Eigenvalues of A T (left) and A D (right), for = = and n = small relative to h. This diculty can be ameliorated to some extent by an appropriate choice of discretization such as the streamline upwinding method used here [],[9, p. 6]. In practical experiments, it is often desired to compute solutions of a xed accuracy for a variety of values of by letting h and simultaneously. In an eort to follow trends in the data, we will consider some combinations of and h that could produce inaccurate solutions. Our main results are in Tables and. Table shows the eects of perturbation of the block tridiagonal preconditioner, for two values of n and various. This data was obtained by computing the eigenvalues of a set of ten perturbed matrices A T () = A T E where = 8 and E is a dense matrix with uniformly distributed random numbers in an interval [ a n ; a n ], where a n = 6=n. (For this choice, kek is approximately constant for all n.) For =, the table presents max j^ j ^ (A T E) where (A T E) is the set of n p eigenvalues that are closest to, and the maximum is over all the perturbations. For 6=, the table shows max j^ j ^ (A T E) where (A T E) denotes the perturbations of (A T ), 6= and 6=. 4 Analogous results for the block diagonal preconditioner are shown in Table. Here we also distinguish between the eigenvalues with positive and negative real parts. Before discussing this data, we present experimental results for three other quantities appearing in the bounds of xx3{4 (P ), k(i H) k, and G;K. These are shown in Tables 3, 4, and 5, respectively. Here (P ) refers to the version used for the 4 These were obtained by sorting (A T ) and (A T E) for each E and comparing the ordered sets. For all tests, similar results were obtained for other values of. 4 /; /;

16 Table Maximum normalized perturbations of eigenvalues of A T. n = 6 = = =3 =5 = Original = Formulation 6= Scaled = Formulation 6= n = 3 Original = Formulation 6= Scaled = Formulation 6= Table Maximum normalized perturbations of eigenvalues of A D. n = 6 = = =3 =5 = Original = Formulation Re() < Re() > Scaled = Formulation Re() < Re() > n = 3 Original = Formulation Re() < Re() > Scaled = Formulation Re() < Re() >

17 Table 3 Condition number of matrix P of eigenvectors of H. = = =3 =5 = n = n = n = e4 Table 4 Norm of (I H). = = =3 =5 = n = n = n = e rank-decient problem as outlined in x5. It was obtained by computing the matrix of eigenvectors P of H, normalizing the columns of P to have unit l -norm, and then computing the ratio of largest to smallest singular values of the submatrix of P corresponding to the nonzero eigenvalues of H. (The normalization ensures that we are within a factor of p n p of the condition number of the optimally scaled version of P [5].) Note that (P ) and G;K exhibit a general tendency to increase as either or n increases, with (P ) being particularly volatile. Now consider the data of Table for the block triangular preconditioner. Several trends are apparent. In the original formulation, the perturbations of both = and 6= are increasing with =. In the scaled formulation, the perturbations of = are insensitive to. This behavior is consistent with the results of Theorem 3... In the scaled formulation, the perturbations of 6= show some growth with = for n = 3, although there is no clear trend. This may derive from growth in the product (P ) k(i H) k. 3. There is little or no increase in perturbation size (and a decrease in some cases) with the change of grid size from n = 6 to n = 3. This contrasts with the bounds from the analysis, which degrade as h. Next, consider Table.. The perturbations of 6= for both the unscaled and scaled problems show some growth with =. The qualitative trends are similar, but the perturbations for the scaled formulations are larger. In contrast, the analysis (Theorem 4.) suggests that the scaled version would be smaller; therefore it appears that the upper bounds of this analysis are not giving a complete indication of dependence on.. The perturbations of = are insensitive to and there is essentially no dierence between the perturbations of the eigenvalues with positive and negative real parts. As noted in Remark 4., we believe the dependence on G;K is an artifact of the proof. 3. The dependence on the mesh size is more pronounced for the triangular preconditioning, although there is no consistent pattern. Thus, the analysis gives upper bounds on perturbation sizes, although it is not possible to completely correlate the analytic bounds and experimental results. This 6

18 Table 5 Secant of largest principal angle between Range(G T ) and Range(K T ) (G;K). = = =3 =5 = n = n = Table 6 Maximum normalized perturbations of eigenvalues of H. = = =3 =5 = n = n = n = is likely due to the lack of a clear pattern in the behavior of (P ) and k(i H) k, together with the use of norm inequalities throughout the analysis that are not necessarily tight. For example, the analysis combines bounds on matrix l -norms derived from the underlying dierential operator with l norms derived from Gerschgorin bounds. We suspect that the factor of =h used to relate these quantities articially inates the dependence of the bounds on the mesh size, although we see no way to avoid introducing this term. In general, both the analysis and experiments indicate a tendency for perturbations of the eigenvalues dierent from to increase with =, but growth appears to be at worst linear and for the block triangular preconditioning it can be reduced by scaling. Finally, Table 6 examines the sensitivity to perturbation of the eigenvalues of the reduced matrix H. The entries are max j^ j ^(HE) The results also indicate that the perturbations increase as the viscosity decreases, and they display some growth as the number of mesh points increase, roughly like that for the block diagonal preconditioner displayed in Table. Note that the dependence on both = and n is much less severe than that of (P ) shown in Table 3. Once again, this stands in contrast with analytic bounds such as those obtainable from the Bauer-Fike theorem [7, p. 34], which suggests perturbations proportional to (P ). 7. Inexact inner iteration. We conclude by examining the eect of inexact computation of the action of F. For brevity, we restrict our attention to the unscaled version of A T ; similar analysis leads to essentially the same conclusions for A D. If the action of F is approximated using an iterative method for each system F w = v, then F can be replaced by F E in (9). The perturbed preconditioned matrix is then A T E where E = F E F EKT BE BEK T =; 7

19 For this analysis it will be useful to consider the complete version of (4), ^E T = F E(I K T (I H) G) P (I (I H) )BE (I K T (I H) G) Suppose the approximate solution ~w satises F EK T P P (I (I H) )BEK T P C A (53) kv F ~wk kvk for some tolerance. Standard inequalities yield the bounds on the relative error kw ~wk kwk kf k kf k ; kw ~wk kwk kf k kek Treating these as approximate equalities gives kek kf k, it follows that k ^E k kf k kf k kk T k k(i H) k kgk ; k ^E k kf k kf k kk T k kp k ; k ^E k kp k ( k(i H) k) kbk kf k k ^E k (P ) ki (I H) k kf k kk T k kk T k k(i H) k kgk ; Using the l -norm, let us consider the dependence of these bounds on the viscosity under the assumption (derived from the experimental results presented above) that the inuence of P and H is not signicant. The bounds on k ^E k, k ^E k, and k ^E k include the product kk T k kgk = O() O(=), which is independent of. Using the coercivity and continuity of the convection-diusion operator, it can be shown that as functions of, kf k = O(); kf k = O(=) (An algebraic proof can be found in [4, Theorem ].) This implies that the bounds on k ^E k and k ^E k grow as = decreases, whereas the bounds on k ^E k and k ^E k are independent of. If is small, then as in x3 we can restrict our attention to the block diagonal entries, which indicates that the perturbations of eigenvalues dierent from in this particular case do not depend on. On the other hand, if is large, then it is not possible to exclude ^E from the Gerschgorin analysis, and the presence of kf k in this bound suggests that perturbations in all eigenvalues grow like =. Consider the implication of these observations on the performance of iterative methods for solving the discrete Oseen equations (3). We demonstrated in [3] that the iteration counts of Krylov subspace methods such as GMRES with the preconditioner Q T (7) are independent of the mesh size, but that there is some deterioration in performance as decreases. Now suppose an inner iteration with stopping criterion (53) is used to approximate the action of F. The analysis given here suggests that if is large, there may be additional degradation of performance for small. In contrast, if is small then only = is sensitive to perturbation, which suggests that (extra) degradation of the inner iteration with decreasing may not be as pronounced. Figure 8

20 Fig.. Iterations of FGMRES with inexact convection-diusion solves. n=6 n= tau= tau=.5 tau=. tau=.d 3 Iterations 8 6 tau=. tau=.d 3 tau= (direct) Iterations 8 6 tau= (direct) /nu /nu shows the results of numerical experiments that corroborate these observations. The gure plots iteration counts of right-preconditioned \exible" GMRES (FGMRES) [] as a function of, for solving the discrete problem (3). The inner iteration for the convection-diusion subproblems F w = v was a line Gauss-Seidel method with stopping criterion (53). 5 The test matrices were those used in x6; the right hand side f consisted of normally distributed random numbers with mean and variance. The outer iteration used a zero initial guess and was stopped when the relative residual in the Euclidean norm was less than or equal to 6. The results with inner iteration are compared with using a direct method for the action of F. They indicate that for the relatively modest tolerance = 3, the inexact inner solves lead to little increase in outer iterations for any. For less stringent, additional outer iterations are required and the number of additional iterations becomes larger as decreases. Remark 7.. It can be shown that for small the perturbations of the eigenvalues of the scaled system behave in the same way as the those for the unscaled system. However, scaling aects the relative weighting given to the two block equations of (3), which in turn may aect the iterative solver. Therefore, we have restricted our attention here to the unscaled system. Acknowledgements. The author acknowledges helpful discussions with Santiago Arteaga, Luca Pavarino and Pete Stewart, and insightful comments from Hans Weinberger. The numerical results with FGMRES used software produced by Tim Kelley []. REFERENCES [] F. Brezzi and M. Fortin. Mixed and Hybrid Finite Element Methods. Springer-Verlag, New York, FGMRES allows the preconditioner to vary during the course of the iteration; it was used here because the number of inner iterations was not generally constant, i.e., the preconditioner is not a xed linear operator. FGMRES is equivalent to GMRES when a direct solve is used for F. 9

21 [] A. Brooks and T. Hughes. Streamline upwind/petrov-galerkin formulations for convection dominated ows with particular emphasis on the incompressible Navier-Stokes equations. Comp. Meth. Appl. Mech. Engng., 399{59, 98. [3] H. Elman and D. Silvester. Fast nonsymmetric iterations and preconditioning for Navier-Stokes equations. SIAM J. Sci. Comput., 733{46, 996. [4] H. C. Elman and M. H. Schultz. Preconditioning by fast direct methods for nonselfadjoint nonseparable elliptic problems. SIAM J. Numer. Anal, 344{57, 986. [5] R. Freund and N. M. Nachtigal. QMR a quasi-minimal residual method for non-hermitian linear systems. Numer. Math., 635{339, 99. [6] R. Glowinski. Numerical Methods for Nonlinear Variational Problems. Springer-Verlag, New York, 984. [7] G. H. Golub and C. F. Van Loan. Matrix Computations. The Johns Hopkins University Press, Baltimore, second edition, 989. [8] M. Gunzburger. Finite Element Methods for Viscous Incompressible Flows. Academic Press, San Diego, 989. [9] C. Johnson. Numerical Solution of Partial Dierential Equations by the Finite Element Method. Cambridge University Press, New York, 987. [] O. A. Karakashian. On a Galerkin-Lagrange multiplier method for the stationary Navier-Stokes equations. SIAM. J. Numer. Anal., 999{93, 98. [] T. Kelley. Iterative Methods for Linear and Nonlinear Equations. SIAM, Philadelphia, 995. [] Y. Saad. A exible inner-outer preconditioned GMRES algorithm. SIAM J. Sci. Comput., 446{469, 993. [3] Y. Saad and M. H. Schultz. GMRES A generalized minimal residual algorithm for solving nonsymmetric linear systems. SIAM J. Sci. Stat. Comput., 7856{869, 986. [4] J. Stoer and R. Bulirsch. Introduction to Numerical Analysis. Springer-Verlag, New York, second edition, 993. [5] A. van der Sluis. Condition numbers and equilibration of matrices. Numer. Math., 44{3, 969. [6] R. Verfurth. A combined conjugate gradient-multigrid algorithm for the numerical solution of the Stokes problem. IMA J. Numer. Anal., 444{455, 984. [7] A. J. Wathen. Realistic eigenvalue bounds for the Galerkin mass matrix. IMA J. Numer. Anal., 7449{457, 987. [8] J. H. Wilkinson. The Algebraic Eigenvalue Problem. Oxford University Press, Oxford, 965.

EQUATIONS WITH LOW VISCOSITY HOWARD C. ELMAN UMIACS-TR November 1996

EQUATIONS WITH LOW VISCOSITY HOWARD C. ELMAN UMIACS-TR November 1996 PRECONDITIONING FOR THE STEADY-STATE NAVIER-STOKES EQUATIONS WITH LOW VISCOSITY HOWARD C. ELMAN Report CS-TR-372 UMIACS-TR-96-82 November 996 Abstract. We introduce a preconditioner for the linearized

More information

UNIVERSITY OF MARYLAND PRECONDITIONERS FOR THE DISCRETE STEADY-STATE NAVIER-STOKES EQUATIONS CS-TR #4164 / UMIACS TR # JULY 2000

UNIVERSITY OF MARYLAND PRECONDITIONERS FOR THE DISCRETE STEADY-STATE NAVIER-STOKES EQUATIONS CS-TR #4164 / UMIACS TR # JULY 2000 UNIVERSITY OF MARYLAND INSTITUTE FOR ADVANCED COMPUTER STUDIES DEPARTMENT OF COMPUTER SCIENCE PERFORMANCE AND ANALYSIS OF SADDLE POINT PRECONDITIONERS FOR THE DISCRETE STEADY-STATE NAVIER-STOKES EQUATIONS

More information

A Review of Preconditioning Techniques for Steady Incompressible Flow

A Review of Preconditioning Techniques for Steady Incompressible Flow Zeist 2009 p. 1/43 A Review of Preconditioning Techniques for Steady Incompressible Flow David Silvester School of Mathematics University of Manchester Zeist 2009 p. 2/43 PDEs Review : 1984 2005 Update

More information

A MULTIGRID ALGORITHM FOR. Richard E. Ewing and Jian Shen. Institute for Scientic Computation. Texas A&M University. College Station, Texas SUMMARY

A MULTIGRID ALGORITHM FOR. Richard E. Ewing and Jian Shen. Institute for Scientic Computation. Texas A&M University. College Station, Texas SUMMARY A MULTIGRID ALGORITHM FOR THE CELL-CENTERED FINITE DIFFERENCE SCHEME Richard E. Ewing and Jian Shen Institute for Scientic Computation Texas A&M University College Station, Texas SUMMARY In this article,

More information

A Robust Preconditioned Iterative Method for the Navier-Stokes Equations with High Reynolds Numbers

A Robust Preconditioned Iterative Method for the Navier-Stokes Equations with High Reynolds Numbers Applied and Computational Mathematics 2017; 6(4): 202-207 http://www.sciencepublishinggroup.com/j/acm doi: 10.11648/j.acm.20170604.18 ISSN: 2328-5605 (Print); ISSN: 2328-5613 (Online) A Robust Preconditioned

More information

Fast solvers for steady incompressible flow

Fast solvers for steady incompressible flow ICFD 25 p.1/21 Fast solvers for steady incompressible flow Andy Wathen Oxford University wathen@comlab.ox.ac.uk http://web.comlab.ox.ac.uk/~wathen/ Joint work with: Howard Elman (University of Maryland,

More information

DELFT UNIVERSITY OF TECHNOLOGY

DELFT UNIVERSITY OF TECHNOLOGY DELFT UNIVERSITY OF TECHNOLOGY REPORT 18-05 Efficient and robust Schur complement approximations in the augmented Lagrangian preconditioner for high Reynolds number laminar flows X. He and C. Vuik ISSN

More information

Mathematics and Computer Science

Mathematics and Computer Science Technical Report TR-2007-002 Block preconditioning for saddle point systems with indefinite (1,1) block by Michele Benzi, Jia Liu Mathematics and Computer Science EMORY UNIVERSITY International Journal

More information

Fast Iterative Solution of Saddle Point Problems

Fast Iterative Solution of Saddle Point Problems Michele Benzi Department of Mathematics and Computer Science Emory University Atlanta, GA Acknowledgments NSF (Computational Mathematics) Maxim Olshanskii (Mech-Math, Moscow State U.) Zhen Wang (PhD student,

More information

Numerische Mathematik

Numerische Mathematik Numer. Math. (2002) 90: 665 688 Digital Object Identifier (DOI) 10.1007/s002110100300 Numerische Mathematik Performance and analysis of saddle point preconditioners for the discrete steady-state Navier-Stokes

More information

Efficient Augmented Lagrangian-type Preconditioning for the Oseen Problem using Grad-Div Stabilization

Efficient Augmented Lagrangian-type Preconditioning for the Oseen Problem using Grad-Div Stabilization Efficient Augmented Lagrangian-type Preconditioning for the Oseen Problem using Grad-Div Stabilization Timo Heister, Texas A&M University 2013-02-28 SIAM CSE 2 Setting Stationary, incompressible flow problems

More information

Institute for Advanced Computer Studies. Department of Computer Science. Iterative methods for solving Ax = b. GMRES/FOM versus QMR/BiCG

Institute for Advanced Computer Studies. Department of Computer Science. Iterative methods for solving Ax = b. GMRES/FOM versus QMR/BiCG University of Maryland Institute for Advanced Computer Studies Department of Computer Science College Park TR{96{2 TR{3587 Iterative methods for solving Ax = b GMRES/FOM versus QMR/BiCG Jane K. Cullum

More information

7.4 The Saddle Point Stokes Problem

7.4 The Saddle Point Stokes Problem 346 CHAPTER 7. APPLIED FOURIER ANALYSIS 7.4 The Saddle Point Stokes Problem So far the matrix C has been diagonal no trouble to invert. This section jumps to a fluid flow problem that is still linear (simpler

More information

Iterative Methods for Problems in Computational Fluid. Howard C. Elman. University of Maryland. David J. Silvester. University of Manchester

Iterative Methods for Problems in Computational Fluid. Howard C. Elman. University of Maryland. David J. Silvester. University of Manchester Report no. 96/19 Iterative Methods for Problems in Computational Fluid Dynamics Howard C. Elman University of Maryland David J. Silvester University of Manchester Andrew J. Wathen Oxford University We

More information

PRECONDITIONING TECHNIQUES FOR NEWTON S METHOD FOR THE INCOMPRESSIBLE NAVIER STOKES EQUATIONS

PRECONDITIONING TECHNIQUES FOR NEWTON S METHOD FOR THE INCOMPRESSIBLE NAVIER STOKES EQUATIONS BIT Numerical Mathematics 43: 961 974, 2003. 2003 Kluwer Academic Publishers. Printed in the Netherlands. 961 PRECONDITIONING TECHNIQUES FOR NEWTON S METHOD FOR THE INCOMPRESSIBLE NAVIER STOKES EQUATIONS

More information

M.A. Botchev. September 5, 2014

M.A. Botchev. September 5, 2014 Rome-Moscow school of Matrix Methods and Applied Linear Algebra 2014 A short introduction to Krylov subspaces for linear systems, matrix functions and inexact Newton methods. Plan and exercises. M.A. Botchev

More information

Efficient Solvers for the Navier Stokes Equations in Rotation Form

Efficient Solvers for the Navier Stokes Equations in Rotation Form Efficient Solvers for the Navier Stokes Equations in Rotation Form Computer Research Institute Seminar Purdue University March 4, 2005 Michele Benzi Emory University Atlanta, GA Thanks to: NSF (MPS/Computational

More information

UNIVERSITY OF MARYLAND NAVIER-STOKES EQUATIONS CS-TR #4073 / UMIACS TR #99-66 OCTOBER 1999

UNIVERSITY OF MARYLAND NAVIER-STOKES EQUATIONS CS-TR #4073 / UMIACS TR #99-66 OCTOBER 1999 UNIVERSITY OF MARYLAND INSTITUTE FOR ADVANCED COMPUTER STUDIES DEPARTMENT OF COMPUTER SCIENCE EFFICIENT PRECONDITIONING OF THE LINEARIZED NAVIER-STOKES EQUATIONS CS-TR #4073 / UMIACS TR #99-66 DAVID SILVESTER

More information

Solving Large Nonlinear Sparse Systems

Solving Large Nonlinear Sparse Systems Solving Large Nonlinear Sparse Systems Fred W. Wubs and Jonas Thies Computational Mechanics & Numerical Mathematics University of Groningen, the Netherlands f.w.wubs@rug.nl Centre for Interdisciplinary

More information

MODIFIED STREAMLINE DIFFUSION SCHEMES FOR CONVECTION-DIFFUSION PROBLEMS. YIN-TZER SHIH AND HOWARD C. ELMAN y

MODIFIED STREAMLINE DIFFUSION SCHEMES FOR CONVECTION-DIFFUSION PROBLEMS. YIN-TZER SHIH AND HOWARD C. ELMAN y MODIFIED STREAMLINE DIFFUSION SCHEMES FOR CONVECTION-DIFFUSION PROBLEMS YIN-TZER SHIH AND HOWARD C. ELMAN y Abstract. We consider the design of robust and accurate nite element approximation methods for

More information

The solution of the discretized incompressible Navier-Stokes equations with iterative methods

The solution of the discretized incompressible Navier-Stokes equations with iterative methods The solution of the discretized incompressible Navier-Stokes equations with iterative methods Report 93-54 C. Vuik Technische Universiteit Delft Delft University of Technology Faculteit der Technische

More information

Preconditioned GMRES Revisited

Preconditioned GMRES Revisited Preconditioned GMRES Revisited Roland Herzog Kirk Soodhalter UBC (visiting) RICAM Linz Preconditioning Conference 2017 Vancouver August 01, 2017 Preconditioned GMRES Revisited Vancouver 1 / 32 Table of

More information

UMIACS-TR July CS-TR 2721 Revised March Perturbation Theory for. Rectangular Matrix Pencils. G. W. Stewart.

UMIACS-TR July CS-TR 2721 Revised March Perturbation Theory for. Rectangular Matrix Pencils. G. W. Stewart. UMIAS-TR-9-5 July 99 S-TR 272 Revised March 993 Perturbation Theory for Rectangular Matrix Pencils G. W. Stewart abstract The theory of eigenvalues and eigenvectors of rectangular matrix pencils is complicated

More information

Numerical Methods in Matrix Computations

Numerical Methods in Matrix Computations Ake Bjorck Numerical Methods in Matrix Computations Springer Contents 1 Direct Methods for Linear Systems 1 1.1 Elements of Matrix Theory 1 1.1.1 Matrix Algebra 2 1.1.2 Vector Spaces 6 1.1.3 Submatrices

More information

ON ORTHOGONAL REDUCTION TO HESSENBERG FORM WITH SMALL BANDWIDTH

ON ORTHOGONAL REDUCTION TO HESSENBERG FORM WITH SMALL BANDWIDTH ON ORTHOGONAL REDUCTION TO HESSENBERG FORM WITH SMALL BANDWIDTH V. FABER, J. LIESEN, AND P. TICHÝ Abstract. Numerous algorithms in numerical linear algebra are based on the reduction of a given matrix

More information

Domain decomposition on different levels of the Jacobi-Davidson method

Domain decomposition on different levels of the Jacobi-Davidson method hapter 5 Domain decomposition on different levels of the Jacobi-Davidson method Abstract Most computational work of Jacobi-Davidson [46], an iterative method suitable for computing solutions of large dimensional

More information

Finding Rightmost Eigenvalues of Large, Sparse, Nonsymmetric Parameterized Eigenvalue Problems

Finding Rightmost Eigenvalues of Large, Sparse, Nonsymmetric Parameterized Eigenvalue Problems Finding Rightmost Eigenvalues of Large, Sparse, Nonsymmetric Parameterized Eigenvalue Problems AMSC 663-664 Final Report Minghao Wu AMSC Program mwu@math.umd.edu Dr. Howard Elman Department of Computer

More information

Lecture Note III: Least-Squares Method

Lecture Note III: Least-Squares Method Lecture Note III: Least-Squares Method Zhiqiang Cai October 4, 004 In this chapter, we shall present least-squares methods for second-order scalar partial differential equations, elastic equations of solids,

More information

ON THE GENERALIZED DETERIORATED POSITIVE SEMI-DEFINITE AND SKEW-HERMITIAN SPLITTING PRECONDITIONER *

ON THE GENERALIZED DETERIORATED POSITIVE SEMI-DEFINITE AND SKEW-HERMITIAN SPLITTING PRECONDITIONER * Journal of Computational Mathematics Vol.xx, No.x, 2x, 6. http://www.global-sci.org/jcm doi:?? ON THE GENERALIZED DETERIORATED POSITIVE SEMI-DEFINITE AND SKEW-HERMITIAN SPLITTING PRECONDITIONER * Davod

More information

OUTLINE 1. Introduction 1.1 Notation 1.2 Special matrices 2. Gaussian Elimination 2.1 Vector and matrix norms 2.2 Finite precision arithmetic 2.3 Fact

OUTLINE 1. Introduction 1.1 Notation 1.2 Special matrices 2. Gaussian Elimination 2.1 Vector and matrix norms 2.2 Finite precision arithmetic 2.3 Fact Computational Linear Algebra Course: (MATH: 6800, CSCI: 6800) Semester: Fall 1998 Instructors: { Joseph E. Flaherty, aherje@cs.rpi.edu { Franklin T. Luk, luk@cs.rpi.edu { Wesley Turner, turnerw@cs.rpi.edu

More information

216 S. Chandrasearan and I.C.F. Isen Our results dier from those of Sun [14] in two asects: we assume that comuted eigenvalues or singular values are

216 S. Chandrasearan and I.C.F. Isen Our results dier from those of Sun [14] in two asects: we assume that comuted eigenvalues or singular values are Numer. Math. 68: 215{223 (1994) Numerische Mathemati c Sringer-Verlag 1994 Electronic Edition Bacward errors for eigenvalue and singular value decomositions? S. Chandrasearan??, I.C.F. Isen??? Deartment

More information

A Method for Constructing Diagonally Dominant Preconditioners based on Jacobi Rotations

A Method for Constructing Diagonally Dominant Preconditioners based on Jacobi Rotations A Method for Constructing Diagonally Dominant Preconditioners based on Jacobi Rotations Jin Yun Yuan Plamen Y. Yalamov Abstract A method is presented to make a given matrix strictly diagonally dominant

More information

Linear Solvers. Andrew Hazel

Linear Solvers. Andrew Hazel Linear Solvers Andrew Hazel Introduction Thus far we have talked about the formulation and discretisation of physical problems...... and stopped when we got to a discrete linear system of equations. Introduction

More information

A Domain Decomposition Based Jacobi-Davidson Algorithm for Quantum Dot Simulation

A Domain Decomposition Based Jacobi-Davidson Algorithm for Quantum Dot Simulation A Domain Decomposition Based Jacobi-Davidson Algorithm for Quantum Dot Simulation Tao Zhao 1, Feng-Nan Hwang 2 and Xiao-Chuan Cai 3 Abstract In this paper, we develop an overlapping domain decomposition

More information

Perturbation results for nearly uncoupled Markov. chains with applications to iterative methods. Jesse L. Barlow. December 9, 1992.

Perturbation results for nearly uncoupled Markov. chains with applications to iterative methods. Jesse L. Barlow. December 9, 1992. Perturbation results for nearly uncoupled Markov chains with applications to iterative methods Jesse L. Barlow December 9, 992 Abstract The standard perturbation theory for linear equations states that

More information

ETNA Kent State University

ETNA Kent State University Electronic Transactions on Numerical Analysis. Volume 36, pp. 39-53, 009-010. Copyright 009,. ISSN 1068-9613. ETNA P-REGULAR SPLITTING ITERATIVE METHODS FOR NON-HERMITIAN POSITIVE DEFINITE LINEAR SYSTEMS

More information

UMIACS-TR July CS-TR 2494 Revised January An Updating Algorithm for. Subspace Tracking. G. W. Stewart. abstract

UMIACS-TR July CS-TR 2494 Revised January An Updating Algorithm for. Subspace Tracking. G. W. Stewart. abstract UMIACS-TR-9-86 July 199 CS-TR 2494 Revised January 1991 An Updating Algorithm for Subspace Tracking G. W. Stewart abstract In certain signal processing applications it is required to compute the null space

More information

Numerical behavior of inexact linear solvers

Numerical behavior of inexact linear solvers Numerical behavior of inexact linear solvers Miro Rozložník joint results with Zhong-zhi Bai and Pavel Jiránek Institute of Computer Science, Czech Academy of Sciences, Prague, Czech Republic The fourth

More information

Institute for Advanced Computer Studies. Department of Computer Science. On the Perturbation of. LU and Cholesky Factors. G. W.

Institute for Advanced Computer Studies. Department of Computer Science. On the Perturbation of. LU and Cholesky Factors. G. W. University of Maryland Institute for Advanced Computer Studies Department of Computer Science College Park TR{95{93 TR{3535 On the Perturbation of LU and Cholesky Factors G. W. Stewart y October, 1995

More information

AMS526: Numerical Analysis I (Numerical Linear Algebra) Lecture 23: GMRES and Other Krylov Subspace Methods; Preconditioning

AMS526: Numerical Analysis I (Numerical Linear Algebra) Lecture 23: GMRES and Other Krylov Subspace Methods; Preconditioning AMS526: Numerical Analysis I (Numerical Linear Algebra) Lecture 23: GMRES and Other Krylov Subspace Methods; Preconditioning Xiangmin Jiao SUNY Stony Brook Xiangmin Jiao Numerical Analysis I 1 / 18 Outline

More information

The Eigenvalue Problem: Perturbation Theory

The Eigenvalue Problem: Perturbation Theory Jim Lambers MAT 610 Summer Session 2009-10 Lecture 13 Notes These notes correspond to Sections 7.2 and 8.1 in the text. The Eigenvalue Problem: Perturbation Theory The Unsymmetric Eigenvalue Problem Just

More information

PROOF OF TWO MATRIX THEOREMS VIA TRIANGULAR FACTORIZATIONS ROY MATHIAS

PROOF OF TWO MATRIX THEOREMS VIA TRIANGULAR FACTORIZATIONS ROY MATHIAS PROOF OF TWO MATRIX THEOREMS VIA TRIANGULAR FACTORIZATIONS ROY MATHIAS Abstract. We present elementary proofs of the Cauchy-Binet Theorem on determinants and of the fact that the eigenvalues of a matrix

More information

Iterative Methods for Solving A x = b

Iterative Methods for Solving A x = b Iterative Methods for Solving A x = b A good (free) online source for iterative methods for solving A x = b is given in the description of a set of iterative solvers called templates found at netlib: http

More information

tdepartment of Mathematics, University of Manchester Institute of Science and Technology, Manchester M601QD, UK (na. ornl. gov).

tdepartment of Mathematics, University of Manchester Institute of Science and Technology, Manchester M601QD, UK (na. ornl. gov). SIAM J. ScI. COMPUT. Vol. 17, No. 1, pp. 33-46, January 1996 () 1996 Society for Industrial and Applied Mathematics 3 FAST NONSYMMETRIC ITERATIONS AND PRECONDITIONING FOR NAVIER-STOKES EQUATIONS* HOWARD

More information

Preconditioned inverse iteration and shift-invert Arnoldi method

Preconditioned inverse iteration and shift-invert Arnoldi method Preconditioned inverse iteration and shift-invert Arnoldi method Melina Freitag Department of Mathematical Sciences University of Bath CSC Seminar Max-Planck-Institute for Dynamics of Complex Technical

More information

1. Introduction. Consider the Navier Stokes equations ηu t ν 2 u + (u grad) u + grad p = f div u = 0 (1.1)

1. Introduction. Consider the Navier Stokes equations ηu t ν 2 u + (u grad) u + grad p = f div u = 0 (1.1) University of Maryland Department of Computer Science TR-C4930 University of Maryland Institute for Advanced Computer Studies TR-009-0 BOUNDARY CONDITIONS IN APPROXIMATE COMMUTATOR PRECONDITIONERS FOR

More information

Math 102, Winter Final Exam Review. Chapter 1. Matrices and Gaussian Elimination

Math 102, Winter Final Exam Review. Chapter 1. Matrices and Gaussian Elimination Math 0, Winter 07 Final Exam Review Chapter. Matrices and Gaussian Elimination { x + x =,. Different forms of a system of linear equations. Example: The x + 4x = 4. [ ] [ ] [ ] vector form (or the column

More information

ON A GENERAL CLASS OF PRECONDITIONERS FOR NONSYMMETRIC GENERALIZED SADDLE POINT PROBLEMS

ON A GENERAL CLASS OF PRECONDITIONERS FOR NONSYMMETRIC GENERALIZED SADDLE POINT PROBLEMS U..B. Sci. Bull., Series A, Vol. 78, Iss. 4, 06 ISSN 3-707 ON A GENERAL CLASS OF RECONDIIONERS FOR NONSYMMERIC GENERALIZED SADDLE OIN ROBLE Fatemeh anjeh Ali BEIK his paper deals with applying a class

More information

Chapter 7. Iterative methods for large sparse linear systems. 7.1 Sparse matrix algebra. Large sparse matrices

Chapter 7. Iterative methods for large sparse linear systems. 7.1 Sparse matrix algebra. Large sparse matrices Chapter 7 Iterative methods for large sparse linear systems In this chapter we revisit the problem of solving linear systems of equations, but now in the context of large sparse systems. The price to pay

More information

The amount of work to construct each new guess from the previous one should be a small multiple of the number of nonzeros in A.

The amount of work to construct each new guess from the previous one should be a small multiple of the number of nonzeros in A. AMSC/CMSC 661 Scientific Computing II Spring 2005 Solution of Sparse Linear Systems Part 2: Iterative methods Dianne P. O Leary c 2005 Solving Sparse Linear Systems: Iterative methods The plan: Iterative

More information

University of Maryland Department of Computer Science TR-5009 University of Maryland Institute for Advanced Computer Studies TR April 2012

University of Maryland Department of Computer Science TR-5009 University of Maryland Institute for Advanced Computer Studies TR April 2012 University of Maryland Department of Computer Science TR-5009 University of Maryland Institute for Advanced Computer Studies TR-202-07 April 202 LYAPUNOV INVERSE ITERATION FOR COMPUTING A FEW RIGHTMOST

More information

UNIVERSITY OF MARYLAND A MULTIGRID METHOD ENHANCED BY KRYLOV SUBSPACE ITERATION FOR DISCRETE HELMHOLTZ EQUATIONS CS-TR #4029 / UMIACS TR #99-36

UNIVERSITY OF MARYLAND A MULTIGRID METHOD ENHANCED BY KRYLOV SUBSPACE ITERATION FOR DISCRETE HELMHOLTZ EQUATIONS CS-TR #4029 / UMIACS TR #99-36 UNIVERSITY OF MARYLAND INSTITUTE FOR ADVANCED COMPUTER STUDIES DEPARTMENT OF COMPUTER SCIENCE A MULTIGRID METHOD ENHANCED BY KRYLOV SUBSPACE ITERATION FOR DISCRETE HELMHOLTZ EQUATIONS CS-TR #4029 / UMIACS

More information

Preface to the Second Edition. Preface to the First Edition

Preface to the Second Edition. Preface to the First Edition n page v Preface to the Second Edition Preface to the First Edition xiii xvii 1 Background in Linear Algebra 1 1.1 Matrices................................. 1 1.2 Square Matrices and Eigenvalues....................

More information

A Finite Element Method for an Ill-Posed Problem. Martin-Luther-Universitat, Fachbereich Mathematik/Informatik,Postfach 8, D Halle, Abstract

A Finite Element Method for an Ill-Posed Problem. Martin-Luther-Universitat, Fachbereich Mathematik/Informatik,Postfach 8, D Halle, Abstract A Finite Element Method for an Ill-Posed Problem W. Lucht Martin-Luther-Universitat, Fachbereich Mathematik/Informatik,Postfach 8, D-699 Halle, Germany Abstract For an ill-posed problem which has its origin

More information

The Conjugate Gradient Method

The Conjugate Gradient Method The Conjugate Gradient Method Classical Iterations We have a problem, We assume that the matrix comes from a discretization of a PDE. The best and most popular model problem is, The matrix will be as large

More information

Adaptive preconditioners for nonlinear systems of equations

Adaptive preconditioners for nonlinear systems of equations Adaptive preconditioners for nonlinear systems of equations L. Loghin D. Ruiz A. Touhami CERFACS Technical Report TR/PA/04/42 Also ENSEEIHT-IRIT Technical Report RT/TLSE/04/02 Abstract The use of preconditioned

More information

On the Preconditioning of the Block Tridiagonal Linear System of Equations

On the Preconditioning of the Block Tridiagonal Linear System of Equations On the Preconditioning of the Block Tridiagonal Linear System of Equations Davod Khojasteh Salkuyeh Department of Mathematics, University of Mohaghegh Ardabili, PO Box 179, Ardabil, Iran E-mail: khojaste@umaacir

More information

On the accuracy of saddle point solvers

On the accuracy of saddle point solvers On the accuracy of saddle point solvers Miro Rozložník joint results with Valeria Simoncini and Pavel Jiránek Institute of Computer Science, Czech Academy of Sciences, Prague, Czech Republic Seminar at

More information

Matrix Algorithms. Volume II: Eigensystems. G. W. Stewart H1HJ1L. University of Maryland College Park, Maryland

Matrix Algorithms. Volume II: Eigensystems. G. W. Stewart H1HJ1L. University of Maryland College Park, Maryland Matrix Algorithms Volume II: Eigensystems G. W. Stewart University of Maryland College Park, Maryland H1HJ1L Society for Industrial and Applied Mathematics Philadelphia CONTENTS Algorithms Preface xv xvii

More information

JURJEN DUINTJER TEBBENS

JURJEN DUINTJER TEBBENS Proceedings of ALGORITMY 005 pp. 40 49 GMRES ACCELERATION ANALYSIS FOR A CONVECTION DIFFUSION MODEL PROBLEM JURJEN DUINTJER TEBBENS Abstract. When we apply the GMRES method [4] to linear systems arising

More information

Preconditioners for the incompressible Navier Stokes equations

Preconditioners for the incompressible Navier Stokes equations Preconditioners for the incompressible Navier Stokes equations C. Vuik M. ur Rehman A. Segal Delft Institute of Applied Mathematics, TU Delft, The Netherlands SIAM Conference on Computational Science and

More information

On the influence of eigenvalues on Bi-CG residual norms

On the influence of eigenvalues on Bi-CG residual norms On the influence of eigenvalues on Bi-CG residual norms Jurjen Duintjer Tebbens Institute of Computer Science Academy of Sciences of the Czech Republic duintjertebbens@cs.cas.cz Gérard Meurant 30, rue

More information

Preliminary/Qualifying Exam in Numerical Analysis (Math 502a) Spring 2012

Preliminary/Qualifying Exam in Numerical Analysis (Math 502a) Spring 2012 Instructions Preliminary/Qualifying Exam in Numerical Analysis (Math 502a) Spring 2012 The exam consists of four problems, each having multiple parts. You should attempt to solve all four problems. 1.

More information

1 Number Systems and Errors 1

1 Number Systems and Errors 1 Contents 1 Number Systems and Errors 1 1.1 Introduction................................ 1 1.2 Number Representation and Base of Numbers............. 1 1.2.1 Normalized Floating-point Representation...........

More information

Reduced Synchronization Overhead on. December 3, Abstract. The standard formulation of the conjugate gradient algorithm involves

Reduced Synchronization Overhead on. December 3, Abstract. The standard formulation of the conjugate gradient algorithm involves Lapack Working Note 56 Conjugate Gradient Algorithms with Reduced Synchronization Overhead on Distributed Memory Multiprocessors E. F. D'Azevedo y, V.L. Eijkhout z, C. H. Romine y December 3, 1999 Abstract

More information

Application of turbulence. models to high-lift airfoils. By Georgi Kalitzin

Application of turbulence. models to high-lift airfoils. By Georgi Kalitzin Center for Turbulence Research Annual Research Briefs 997 65 Application of turbulence models to high-lift airfoils By Georgi Kalitzin. Motivation and objectives Accurate prediction of the ow over an airfoil

More information

XIAO-CHUAN CAI AND MAKSYMILIAN DRYJA. strongly elliptic equations discretized by the nite element methods.

XIAO-CHUAN CAI AND MAKSYMILIAN DRYJA. strongly elliptic equations discretized by the nite element methods. Contemporary Mathematics Volume 00, 0000 Domain Decomposition Methods for Monotone Nonlinear Elliptic Problems XIAO-CHUAN CAI AND MAKSYMILIAN DRYJA Abstract. In this paper, we study several overlapping

More information

Key words. conjugate gradients, normwise backward error, incremental norm estimation.

Key words. conjugate gradients, normwise backward error, incremental norm estimation. Proceedings of ALGORITMY 2016 pp. 323 332 ON ERROR ESTIMATION IN THE CONJUGATE GRADIENT METHOD: NORMWISE BACKWARD ERROR PETR TICHÝ Abstract. Using an idea of Duff and Vömel [BIT, 42 (2002), pp. 300 322

More information

Convergence Properties of Preconditioned Hermitian and Skew-Hermitian Splitting Methods for Non-Hermitian Positive Semidefinite Matrices

Convergence Properties of Preconditioned Hermitian and Skew-Hermitian Splitting Methods for Non-Hermitian Positive Semidefinite Matrices Convergence Properties of Preconditioned Hermitian and Skew-Hermitian Splitting Methods for Non-Hermitian Positive Semidefinite Matrices Zhong-Zhi Bai 1 Department of Mathematics, Fudan University Shanghai

More information

AMS526: Numerical Analysis I (Numerical Linear Algebra)

AMS526: Numerical Analysis I (Numerical Linear Algebra) AMS526: Numerical Analysis I (Numerical Linear Algebra) Lecture 1: Course Overview & Matrix-Vector Multiplication Xiangmin Jiao SUNY Stony Brook Xiangmin Jiao Numerical Analysis I 1 / 20 Outline 1 Course

More information

Index. for generalized eigenvalue problem, butterfly form, 211

Index. for generalized eigenvalue problem, butterfly form, 211 Index ad hoc shifts, 165 aggressive early deflation, 205 207 algebraic multiplicity, 35 algebraic Riccati equation, 100 Arnoldi process, 372 block, 418 Hamiltonian skew symmetric, 420 implicitly restarted,

More information

Numerische Mathematik

Numerische Mathematik Numer. Math. (1999) 83: 231 257 Numerische Mathematik c Springer-Verlag 1999 Eigenanalysis of some preconditioned Helmholtz problems Howard C. Elman, Dianne P. O Leary Computer Science Department and Institute

More information

The rate of convergence of the GMRES method

The rate of convergence of the GMRES method The rate of convergence of the GMRES method Report 90-77 C. Vuik Technische Universiteit Delft Delft University of Technology Faculteit der Technische Wiskunde en Informatica Faculty of Technical Mathematics

More information

The antitriangular factorisation of saddle point matrices

The antitriangular factorisation of saddle point matrices The antitriangular factorisation of saddle point matrices J. Pestana and A. J. Wathen August 29, 2013 Abstract Mastronardi and Van Dooren [this journal, 34 (2013) pp. 173 196] recently introduced the block

More information

Unsteady Incompressible Flow Simulation Using Galerkin Finite Elements with Spatial/Temporal Adaptation

Unsteady Incompressible Flow Simulation Using Galerkin Finite Elements with Spatial/Temporal Adaptation Unsteady Incompressible Flow Simulation Using Galerkin Finite Elements with Spatial/Temporal Adaptation Mohamed S. Ebeida Carnegie Mellon University, Pittsburgh, PA 15213 Roger L. Davis and Roland W. Freund

More information

Lecture 3: Inexact inverse iteration with preconditioning

Lecture 3: Inexact inverse iteration with preconditioning Lecture 3: Department of Mathematical Sciences CLAPDE, Durham, July 2008 Joint work with M. Freitag (Bath), and M. Robbé & M. Sadkane (Brest) 1 Introduction 2 Preconditioned GMRES for Inverse Power Method

More information

CONVERGENCE BOUNDS FOR PRECONDITIONED GMRES USING ELEMENT-BY-ELEMENT ESTIMATES OF THE FIELD OF VALUES

CONVERGENCE BOUNDS FOR PRECONDITIONED GMRES USING ELEMENT-BY-ELEMENT ESTIMATES OF THE FIELD OF VALUES European Conference on Computational Fluid Dynamics ECCOMAS CFD 2006 P. Wesseling, E. Oñate and J. Périaux (Eds) c TU Delft, The Netherlands, 2006 CONVERGENCE BOUNDS FOR PRECONDITIONED GMRES USING ELEMENT-BY-ELEMENT

More information

Eigenvalue problems and optimization

Eigenvalue problems and optimization Notes for 2016-04-27 Seeking structure For the past three weeks, we have discussed rather general-purpose optimization methods for nonlinear equation solving and optimization. In practice, of course, we

More information

KEYWORDS. Numerical methods, generalized singular values, products of matrices, quotients of matrices. Introduction The two basic unitary decompositio

KEYWORDS. Numerical methods, generalized singular values, products of matrices, quotients of matrices. Introduction The two basic unitary decompositio COMPUTING THE SVD OF A GENERAL MATRIX PRODUCT/QUOTIENT GENE GOLUB Computer Science Department Stanford University Stanford, CA USA golub@sccm.stanford.edu KNUT SLNA SC-CM Stanford University Stanford,

More information

Synopsis of Numerical Linear Algebra

Synopsis of Numerical Linear Algebra Synopsis of Numerical Linear Algebra Eric de Sturler Department of Mathematics, Virginia Tech sturler@vt.edu http://www.math.vt.edu/people/sturler Iterative Methods for Linear Systems: Basics to Research

More information

The Deflation Accelerated Schwarz Method for CFD

The Deflation Accelerated Schwarz Method for CFD The Deflation Accelerated Schwarz Method for CFD J. Verkaik 1, C. Vuik 2,, B.D. Paarhuis 1, and A. Twerda 1 1 TNO Science and Industry, Stieltjesweg 1, P.O. Box 155, 2600 AD Delft, The Netherlands 2 Delft

More information

1. Introduction Let the least value of an objective function F (x), x2r n, be required, where F (x) can be calculated for any vector of variables x2r

1. Introduction Let the least value of an objective function F (x), x2r n, be required, where F (x) can be calculated for any vector of variables x2r DAMTP 2002/NA08 Least Frobenius norm updating of quadratic models that satisfy interpolation conditions 1 M.J.D. Powell Abstract: Quadratic models of objective functions are highly useful in many optimization

More information

Solution of the Two-Dimensional Steady State Heat Conduction using the Finite Volume Method

Solution of the Two-Dimensional Steady State Heat Conduction using the Finite Volume Method Ninth International Conference on Computational Fluid Dynamics (ICCFD9), Istanbul, Turkey, July 11-15, 2016 ICCFD9-0113 Solution of the Two-Dimensional Steady State Heat Conduction using the Finite Volume

More information

Lecture 11: CMSC 878R/AMSC698R. Iterative Methods An introduction. Outline. Inverse, LU decomposition, Cholesky, SVD, etc.

Lecture 11: CMSC 878R/AMSC698R. Iterative Methods An introduction. Outline. Inverse, LU decomposition, Cholesky, SVD, etc. Lecture 11: CMSC 878R/AMSC698R Iterative Methods An introduction Outline Direct Solution of Linear Systems Inverse, LU decomposition, Cholesky, SVD, etc. Iterative methods for linear systems Why? Matrix

More information

APPLIED NUMERICAL LINEAR ALGEBRA

APPLIED NUMERICAL LINEAR ALGEBRA APPLIED NUMERICAL LINEAR ALGEBRA James W. Demmel University of California Berkeley, California Society for Industrial and Applied Mathematics Philadelphia Contents Preface 1 Introduction 1 1.1 Basic Notation

More information

Numerical Linear Algebra

Numerical Linear Algebra Numerical Linear Algebra The two principal problems in linear algebra are: Linear system Given an n n matrix A and an n-vector b, determine x IR n such that A x = b Eigenvalue problem Given an n n matrix

More information

ANALYSIS OF AUGMENTED LAGRANGIAN-BASED PRECONDITIONERS FOR THE STEADY INCOMPRESSIBLE NAVIER-STOKES EQUATIONS

ANALYSIS OF AUGMENTED LAGRANGIAN-BASED PRECONDITIONERS FOR THE STEADY INCOMPRESSIBLE NAVIER-STOKES EQUATIONS ANALYSIS OF AUGMENTED LAGRANGIAN-BASED PRECONDITIONERS FOR THE STEADY INCOMPRESSIBLE NAVIER-STOKES EQUATIONS MICHELE BENZI AND ZHEN WANG Abstract. We analyze a class of modified augmented Lagrangian-based

More information

for Finite Element Simulation of Incompressible Flow Arnd Meyer Department of Mathematics, Technical University of Chemnitz,

for Finite Element Simulation of Incompressible Flow Arnd Meyer Department of Mathematics, Technical University of Chemnitz, Preconditioning the Pseudo{Laplacian for Finite Element Simulation of Incompressible Flow Arnd Meyer Department of Mathematics, Technical University of Chemnitz, 09107 Chemnitz, Germany Preprint{Reihe

More information

Foundations of Matrix Analysis

Foundations of Matrix Analysis 1 Foundations of Matrix Analysis In this chapter we recall the basic elements of linear algebra which will be employed in the remainder of the text For most of the proofs as well as for the details, the

More information

Math Introduction to Numerical Analysis - Class Notes. Fernando Guevara Vasquez. Version Date: January 17, 2012.

Math Introduction to Numerical Analysis - Class Notes. Fernando Guevara Vasquez. Version Date: January 17, 2012. Math 5620 - Introduction to Numerical Analysis - Class Notes Fernando Guevara Vasquez Version 1990. Date: January 17, 2012. 3 Contents 1. Disclaimer 4 Chapter 1. Iterative methods for solving linear systems

More information

A NOTE ON THE LADYŽENSKAJA-BABUŠKA-BREZZI CONDITION

A NOTE ON THE LADYŽENSKAJA-BABUŠKA-BREZZI CONDITION A NOTE ON THE LADYŽENSKAJA-BABUŠKA-BREZZI CONDITION JOHNNY GUZMÁN, ABNER J. SALGADO, AND FRANCISCO-JAVIER SAYAS Abstract. The analysis of finite-element-like Galerkin discretization techniques for the

More information

On solving linear systems arising from Shishkin mesh discretizations

On solving linear systems arising from Shishkin mesh discretizations On solving linear systems arising from Shishkin mesh discretizations Petr Tichý Faculty of Mathematics and Physics, Charles University joint work with Carlos Echeverría, Jörg Liesen, and Daniel Szyld October

More information

Introduction to Applied Linear Algebra with MATLAB

Introduction to Applied Linear Algebra with MATLAB Sigam Series in Applied Mathematics Volume 7 Rizwan Butt Introduction to Applied Linear Algebra with MATLAB Heldermann Verlag Contents Number Systems and Errors 1 1.1 Introduction 1 1.2 Number Representation

More information

Jos L.M. van Dorsselaer. February Abstract. Continuation methods are a well-known technique for computing several stationary

Jos L.M. van Dorsselaer. February Abstract. Continuation methods are a well-known technique for computing several stationary Computing eigenvalues occurring in continuation methods with the Jacobi-Davidson QZ method Jos L.M. van Dorsselaer February 1997 Abstract. Continuation methods are a well-known technique for computing

More information

Preconditioning Techniques Analysis for CG Method

Preconditioning Techniques Analysis for CG Method Preconditioning Techniques Analysis for CG Method Huaguang Song Department of Computer Science University of California, Davis hso@ucdavis.edu Abstract Matrix computation issue for solve linear system

More information

Contents. Preface... xi. Introduction...

Contents. Preface... xi. Introduction... Contents Preface... xi Introduction... xv Chapter 1. Computer Architectures... 1 1.1. Different types of parallelism... 1 1.1.1. Overlap, concurrency and parallelism... 1 1.1.2. Temporal and spatial parallelism

More information

arxiv: v1 [math.na] 1 Sep 2018

arxiv: v1 [math.na] 1 Sep 2018 On the perturbation of an L -orthogonal projection Xuefeng Xu arxiv:18090000v1 [mathna] 1 Sep 018 September 5 018 Abstract The L -orthogonal projection is an important mathematical tool in scientific computing

More information

Rening the submesh strategy in the two-level nite element method: application to the advection diusion equation

Rening the submesh strategy in the two-level nite element method: application to the advection diusion equation INTERNATIONAL JOURNAL FOR NUMERICAL METHODS IN FLUIDS Int. J. Numer. Meth. Fluids 2002; 39:161 187 (DOI: 10.1002/d.219) Rening the submesh strategy in the two-level nite element method: application to

More information

Properties of Matrices and Operations on Matrices

Properties of Matrices and Operations on Matrices Properties of Matrices and Operations on Matrices A common data structure for statistical analysis is a rectangular array or matris. Rows represent individual observational units, or just observations,

More information

Numerical Methods - Numerical Linear Algebra

Numerical Methods - Numerical Linear Algebra Numerical Methods - Numerical Linear Algebra Y. K. Goh Universiti Tunku Abdul Rahman 2013 Y. K. Goh (UTAR) Numerical Methods - Numerical Linear Algebra I 2013 1 / 62 Outline 1 Motivation 2 Solving Linear

More information