NAVIER-STOKES OPERATORS HOWARD C. ELMAN UMIACS-TR November Revised July 1996
|
|
- Brian Carroll
- 5 years ago
- Views:
Transcription
1 PERTURBATION OF EIGENVALUES OF PRECONDITIONED NAVIER-STOKES OPERATORS HOWARD C. ELMAN Report CS-TR-3559 UMIACS-TR-95- November 995 Revised July 996 Abstract. We study the sensitivity of algebraic eigenvalue problems associated with matrices arising from linearization and discretization of the steady-state Navier-Stokes equations. In particular, for several choices of preconditioners applied to the system of discrete equations, we derive upper bounds on perturbations of eigenvalues as functions of the viscosity and discretization mesh size. The bounds suggest that the sensitivity of the eigenvalues is at worst linear in the inverse of the viscosity and quadratic in the inverse of the mesh size, and that scaling can be used to decrease the sensitivity in some cases. Experimental results supplement these results and conrm the relatively mild dependence on viscosity. They also indicate a dependence on the mesh size of magnitude smaller than the analysis suggests. Department of Computer Science and Institute for Advanced Computer Studies, University of Maryland, College Park, MD 74, elman@cs.umd.edu. This work was supported by the U. S. Army Research Oce under grant DAAL-39-G-6 and the U. S. National Science Foundation under grants ASC and DMS
2 PERTURBATION OF EIGENVALUES OF PRECONDITIONED NAVIER-STOKES OPERATORS HOWARD C. ELMAN Abstract. We study the sensitivity of algebraic eigenvalue problems associated with matrices arising from linearization and discretization of the steady-state Navier-Stokes equations. In particular, for several choices of preconditioners applied to the system of discrete equations, we derive upper bounds on perturbations of eigenvalues as functions of the viscosity and discretization mesh size. The bounds suggest that the sensitivity of the eigenvalues is at worst linear in the inverse of the viscosity and quadratic in the inverse of the mesh size, and that scaling can be used to decrease the sensitivity in some cases. Experimental results supplement these results and conrm the relatively mild dependence on viscosity. They also indicate a dependence on the mesh size of magnitude smaller than the analysis suggests. Key words. eigenvalues, perturbation analysis, Navier-Stokes, preconditioning. AMS(MOS) subject classications. Primary 65F, 65N; secondary 5A6. Introduction. This paper concerns properties of the eigenvalues of matrices arising from the discrete linearized steady-state Navier-Stokes equations. The continuous problem is () u (u grad) u grad p = f in ; together with the incompressibility constraint () div u = in ; subject to suitable boundary conditions where is an open bounded domain in R. These equations constitute a fundamental problem in computational uid dynamics, see e.g., [],[6],[8]. The two-dimensional vector eld u represents the velocity in, p represents pressure, and the scalar is the viscosity, roughly speaking, the ratio of convection to diusion in the system. A methodology for computing the numerical solution is to discretize (){() using nite dierence or nite element methods, and then to solve the resulting nonlinear system by some iterative method. Linearization leads to a set of matrix equations of the form F B T u (3) = f ; B p where u and p now represent discrete versions of velocity and pressure, respectively. We will restrict our attention to the discrete Oseen equations (4) u (w grad) u grad p = f div u = Department of Computer Science and Institute for Advanced Computer Studies, University of Maryland, College Park, MD 74, elman@cs.umd.edu. This work was supported by the U. S. Army Research Oce under grant DAAL-39-G-6 and the U. S. National Science Foundation under grants ASC and DMS
3 where w is given such that div w =. These equations arise from a nonlinear iteration of the form u (m) (u (m ) grad) u (m) grad p (m) = f; div u (m) =, see []. In this case F = A N where A consists of a pair of uncoupled discrete Laplace operators, corresponding to diusion, and N is a skew-symmetric matrix representing convection. We will also assume that the velocity and pressure discretizations are div-stable; see, e.g., [], p. 57, [8], pp., [6]. In matrix notation, this is equivalent to the condition (5) (p; BA B T p) (p; Mp) for all p ; where (; ) denotes the Euclidean inner product, and are constants that are independent of the discretization mesh size h, and for nite elements M is the pressure mass matrix, i.e., the Grammian matrix of basis functions dening the discrete pressure space. For nite dierences on uniform grids, a natural analogue is M = h I. Let L denote the coecient matrix of (3). The following preconditioning matrices were introduced in [3] a block diagonal preconditioner (6) Q D = F M and a block triangular preconditioner (7) Q T = F BT M It was shown in [3] that the eigenvalues of each of the preconditioned matrices A D = LQ D and A T = LQ T are uniformly bounded independent of the mesh size used in the discretization. Numerical experiments also suggested that Krylov subspace iterative methods such as the generalized minimal residual (GMRES) [3] and quasiminimal residual (QMR) methods [5] can be used to solve the preconditioned system with iteration counts independent of the mesh size. We are concerned with the sensitivity of the eigenvalues of the preconditioned Oseen matrix using the two preconditioners (6) and (7). Motivation for studying this lies in the fact that use of either preconditioner in an iteration entails applying the action of the inverse of the matrix of either (6) or (7) to a vector at each step. These in turn require the computation of the action of F, which, if direct methods are used, will dominate the cost. An alternative that was considered in [3] is to approximate the action of F (i.e., compute an approximate solution to systems with coecient matrix F ) using an inner iteration. Unless very stringent stopping criteria are used here, the resulting preconditioned operators can be viewed as perturbations of those of (6) { (7). Thus, we are interested the sensitivity of the eigenvalues to perturbation. If the preconditioned matrix is perturbed by matrix of size, then the perturbations of eigenvalues will depend on and also on parameters associated with the An inequality analogous to (5) also holds, with dierent constants, if M is any matrix spectrally equivalent to the mass matrix; for example, M could be the diagonal matrix consisting of the diagonal of the mass matrix [7]. In the sequel, we will not distinguish among such possibilities for M.
4 underlying problem, specically, the viscosity and mesh size h. In this paper, we examine these dependencies using a combination of analytic and experimental results. The analysis derives from Wilkinson's classical perturbation analysis [8], which shows that if there are no nonlinear elementary divisors, the perturbations are of magnitude O(). The analytic bounds are stated as functions of and h but they also depend on properties of certain matrices associated with the Schur complement BF B T derived from (4). The latter quantities are studied in a series of numerical experiments. The combination of analytic and experimental results indicates that there is an increase in sensitivity to perturbation as the viscosity decreases, with growth roughly linear in =. This eect can be mitigated to some extent by scaling the rst equation of (4) (the momentum equation). The bounds also establish linear dependence on =h with the preconditioner Q T and quadratic dependence with Q D, although the experimental results suggest that perturbations are considerably less sensitive to this parameter. An outline of the paper is as follows. In x, we derive preliminary bounds and relations for several operators associated with the preconditioned matrices. In x3, we derive the analytic perturbation bounds for the block tridiagonal preconditioner, and in x4, we present the analysis for the block diagonal preconditioner. For simplicity, the analysis is done for the case where the coecient matrix of (4) has full rank, although often in practice (and in our experiments) it is rank decient by one because the pressure p is uniquely dened only up to a constant. In x5, we show that the analytic results carry over to this case. In x6, we present the experimental results, and in x7, we show how the analysis applies for the case of inexact computation of the action of F.. Preliminary results. In this section we derive preliminary bounds and relations for several operators associated with the preconditioners (6) { (7). We will assume that the discrete problem (3) arises from a standard nite dierence or loworder nite element scheme on a uniform grid with mesh size h, and that the discrete problem is scaled so that the extreme eigenvalues of the discrete Laplace operators of A are contained in an interval of the form [c h ; c ], where here and below c i denotes a generic constant that is independent of h and. This is a natural scaling for nite elements, and for nite dierences on a uniform grid it corresponds to the ve-point operator with 4 in the diagonal entries and in the o-diagonal entries. With this normalization, BB T =h is also a scaled discrete Laplace operator and its eigenvalues are contained in an interval of the same form. Let the discrete velocity and pressure spaces have dimension n u and n p, respectively. For div-stable discretizations, n u n p, and typically n u is signicantly larger than n p. It will be convenient to use the symbol Q instead of M in the matrices of (6). The preconditioned matrices are then given by (8) A D = F BT B F = Q I G K T for the block diagonal preconditioner, and (9) A T = F BT B F F B T Q = Q I G H For two-dimensional problems, the vector u has two components of grid vectors, and stability considerations often also lead to more grid points for velocity than for pressure [8]. 3
5 for the block triangular preconditioner. The submatrices on the right of (8) { (9) are G = BF ; K T = B T Q ; H = GK T = BF B T Q The identity matrices are of order n u. We recall some results from [3], which give bounds on the eigenvalues of H. Let S = BF B T denote the Schur complement matrix for (3), and let C = B F F T B T ; R = B F F T denote the symmetric and skew-symmetric parts of S, respectively. It is shown in [3] that B T ; () (q; Cq) (q; Qq) ; j(q; Rq)j (q; Qq) ; where and are as in (5) and is the largest eigenvalue of A N, which is also uniformly bounded independent of h [4]. Consequently, Bendixson's theorem [4, p. 48] implies that the eigenvalues of H are contained in the box () " ; # " ; in the complex plane. We rst derive bounds on the singular values of G, which will be used in the perturbation analysisis for the block triangular preconditioner. Theorem.. The largest singular value of G is bounded above by a quantity of magnitude O(=) which is independent of h as h. The smallest singular value of G is bounded below by a quantify of magnitude O(h) which is bounded independent of as. Proof. The singular values of G are the square roots of the eigenvalues of GG T, and the largest and smallest of these eigenvalues are the extrema of (q; GG T q)=(q; q). This Rayleigh quotient can be rewritten as # () (q; GG T q) (q; q) = (F T B T q; F T B T q) (q; q) = (F T B T q; F T B T q) (F T B T q; B T q) (F T B T q; B T q) (q; q) We consider the two terms in the product on the right of () separately. For the rst term, the substitution w = F T B T q gives (F T B T q; F T B T q) (F T B T q; B T q) = (w; w) (w; F T w) = (w; w) (w; F F T w) = (w; w) (w; Aw) Under the assumption on the scaling of the discrete Laplacian operators composing A, it follows that (3) c (F T B T q; F T B T q) c h (F T B T q; B T q) 4
6 The second term of the product in () is (4) (q; BF B T q) (q; q) = (q; Cq) (q; q) (q; Cq) (q; Qq) = (q; Qq) (q; q) It is well known (see [7]) that the pressure mass matrix is spectrally equivalent to h I, so that (5) c h (q; Qq) (q; q) c h Thus, the bounds for the symmetric part in () together with () { (5) imply c h (q; GGT q) c (q; q) The singular values of K will be used to analyze the block diagonal preconditioner. Lemma.. The largest singular value of K is bounded above by a quantity of magnitude O h. The smallest singular value is bounded below by a quantity of magnitude O(). Proof. The largest singular value is kkk = kk T k. Using K T = B T Q and Q = M, we have kk T k kb T k km k But km k = O(h ), and our assumptions on B imply that kb T k = kbb T k = = O(h). The smallest singular value is the inverse of k(kk T ) k =. Then k(kk T ) k kmk k(bb T ) k = c h 4 c h 4 Consider an alternative scaling in problems () and (4) in which the rst equation is multiplied by. This does not change the solutions, but as we will show in xx3{4, it aects the sensitivity of discrete eigenvalues. For (4), scaling gives (6) u (w grad) u grad p = f The new discrete problem is as in (3) except that F, p and f are replaced by ^F = F, p and f, respectively. Let ^Q = Q = M, ^G = B ^F, ^KT = B T ^Q, and ^H = ^G ^KT. The analogues of () and the bounds of Theorem. and Lemma. are given below. The proof follows from the facts that ^G = G, ^K = K and ^H = H. Corollary.3. With the scaling of (6), the eigenvalues of ^H are contained in the box (); the singular values of ^G are bounded above by a quantity of magnitude O() and below by a quantity of magnitude O(h); and the singular values of ^K are bounded above by a quantity of magnitude O and below by a quantity of h magnitude O(). 5
7 In the following, we will not specically identify the matrices associated with this scaling using the \hat"-symbol. Instead, we will use the notation of (8) { (9) to refer generically to both scalings. We will distinguish them as derived from either the \original" formulation (4) or the \scaled" formulation (6) of the Oseen equations. Finally, we will use the notation G;K to denote the secant of the largest principal angle between Range(G T ) and Range(K T ). That is, if Q G and Q K are matrices whose columns represent orthogonal bases of Range(G T ) and Range(K T ), respectively, then (7) G;K = k(q T G Q K) k = min (Q T G Q K) ; where min denotes the smallest singular value (see [7, p. 584]). It is easily shown (e.g., using QR decompositions) that (8) G;K = kk T (GK T ) Gk 3. Analytic bounds for the block triangular preconditioner. It is evident from (9) that the eigenvalues of A T consist of = of multiplicity n u together with the eigenvalues of H. We seek a factorization (9) A T = V T D T V T that provides insight into the sensitivity of these eigenvalues to perturbation. We will look for factors of the form () D T = I n u n p n u n p D nu n p ; V T = V V V V nu n p where the sub-blocks of D T and V T must be determined and the dimensions are as indicated. Let H have Jordan canonical form H = P P. The requirement A T V T = V T D T is satised if () V = I; V = ; G = (I H)V P D ; V = P We distinguish between two cases, (H) and = (H). Suppose rst that H has no eigenvalues equal to. The choice D = in () leads to V = (I H) G. In this case, (9) represents a Jordan form for A T with () V T = I (I H) G P Let A T () = A T E be a perturbation of A T, where E = E E (3) E E ; V T = I P (I H) G P The classical perturbation analysis of Wilkinson based on Gerschgorin theory [8, pp. 7] shows that for every eigenvalue of A T with only linear elementary divisors, perturbations of in (A T ()) are contained in a circle centered at with radius of 6
8 size c where c is independent of. The structure of V T can be used to obtain further insight into the sizes of the perturbations. Let ^E T = V T EV ^E T = ^E (4) ; ^E ^E so that we are concerned with the eigenvalues of (5) V T A T ()V T = D T ^ET = I ^E ^E ^E ^E Here and in the following, the symbol \c" represents a generic constant that is independent of the parameters h, and. Lemma 3.. If 6 (H), then there are n u eigenvalues ^ of A T () (counting multiplicity) satisfying (6) j^ j k ^E k c k ^E k If there are m n p eigenvalues of H with linear elementary divisors, then there are m eigenvalues ^ of D T (), distinct from those of (6), that satisfy (7) j^ j k ^E k c k ^E k Proof. Multiplying the rst block row on the right side of (5) by =, multiplying the rst block column by = and then applying Gerschgorin's theorem leads to the bound ^ ( [ ^E ] ii ) X j6=i [ ^E ] ij = X j [ ^E ] ij ; where is such that the Gerschgorin disks for the rst and second blocks of the scaled matrix are disjoint (see [8, p. 73]). Assertion (6) follows. The argument for (7) is identical, applied to the second block row of (5). Thus, we can restrict our attention to the block diagonal entries of ^E T. Using (){(4) we have (8) ^E = E E (I H) G ; ^E = P (I H) GE P P E P The following result gives bounds on the perturbations of eigenvalues as functions of the viscosity and mesh size h, for a perturbation satisfying kek. 3 Theorem 3.. Assume kek. If 6 (H), then the eigenvalues ^ of A T () that are perturbations of = satisfy (9) j^ j 8 >< > c h O( ) c h O( ) for the original formulation; for the scaled formulation; 3 This assumption is stronger than the inequality je ijj used by Wilkinson; the latter condition follows from our assumption. 7
9 where c = c k(i H) k. For eigenvalues of H with linear elementary divisors, the perturbations ^ (A T ()) satisfy (3) j^ j 8 >< > c h O( ) c h O( ) for the original formulation; for the scaled formulation; where c = c (P ) k(i H) k. Proof. The relations (8) together with standard bounds on matrix l p -norms give k ^E k c h k ^E k c h k(i H) k kgk ; k ^E k c h k ^E k c h (P ) k(i H) k kgk The conclusions follow from the upper bounds on kgk in Theorem.. In the case (H), we use an alternative version of (). By analogy with the analysis above, let V = (I H) y G where (I H) y is the pseudo-inverse. Then V T = I (I H) y G P The similarity (9) then holds for the choice D T = ; V T = I P (I H) y G P I P G where is the orthogonal projection onto the null space of I H T. The perturbation of D T associated with A T () is then V T A T ()V T = D T ^E T = I ^E ^E (3) P G ^E ^E First consider the eigenvalues of H dierent from with linear elementary divisors. Let be a parameter in (; ). Multiplying the second block row of (3) by and the second block column by produces the matrix (3) I ^E ^E P G ^E ^E For any (; ) and small enough, the Gerschgorin disk for a perturbation ^ of (H), 6=, is disjoint from the disks corresponding to perturbations of the eigenvalue. Consequently, if has only linear elementary divisors, then j^ j kp Gk O() For eigenvalues equal to, multiplying the second block row of (3) by = and multiplying the second block column by = produces I ^E = ^E (33) = P G 3= ^E ^E 8
10 Thus, for = with linear or quadratic elementary divisors, the perturbations satisfy j^ j = max k ^E k ; kp Gk O() Bounding the matrix innity norms in (3) and (33) gives the following result. Theorem 3.3. If (H) with linear or quadratic elementary divisors, then the eigenvalues ^ of A T () that are perturbations of = satisfy (34) j^ j = c h max kp k ; kp k kgk O() For (H) dierent from with linear elementary divisors, the perturbations ^ (A T ()) satisfy (35) j^ j c h kp k kgk O() for any (; ). Here kgk = O(=) in the original formulation and kgk = O() for the scaled formulation. Note that for any (A T ) with nonlinear elementary divisors, bounds analogous to (9){(3) and (34){(35) can be obtained in which the dependence on is of the form =r where r is the order of the largest Jordan block for ; see [8, p. 79]. 4. Analytic bounds for the block diagonal preconditioner. For the block diagonal preconditioner, we have results only in the case where the Schur complement matrix H has no nonlinear elementary divisors. As in Theorem 3., the perturbation analysis will be stated in terms of properties of P, a xed matrix of eigenvectors of H. In addition, some of the bounds depend on G;K, the secant of the largest principal angle between Range(G T ) and Range(K T ), as dened in (7). Consider the eigenvalue problem for A D, This leads to the conditions I G K T u p = u p (36) K T p = ( )u; Gu = p on the components of the eigenvectors. One solution corresponds to the eigenvalue = ; the assumption that B and therefore K have full rank then implies that any associated eigenvector satises p =, Gu =. We use two approaches to identify eigenvalues 6= and corresponding eigenvectors.. Substitution of u = K T p into the second equation of (36) gives GK T p = ( ) p That is, any eigenpair (; p) of H = GK T leads to two eigenvalues (37) of A D. The associated eigenvectors are (38) = p 4 ; = p 4 u p ; 9 u p ;
11 where (39) u = K T p; u = K T p. Alternatively, substitution of p = Gu into the rst equation of (36) gives K T Gu = ( ) u K T G has a zero eigenvalue of multiplicity n u n p (the dimension of the null space of G), plus n p nonzero eigenvalues. If (; u) is an eigenpair with 6=, then (37) denes a pair of eigenvalues of A D. In this case, the eigenvectors have the form (4) where (4) p = u ; p u ; p Gu; p = Gu It is straightforward to show that any u or u from (39) is an eigenvector of K T G, and any p or p from (4) is an eigenvector of GK T. Therefore, we will use the symbol \" to refer to the pairs of eigenvalues and eigenvectors of A D, i.e., u will refer to the rst entries of either (38) or (4), p to the second entries, and to the associated eigenvalue. We have established the following result. Theorem 4.. The eigenvectors of A D corresponding to the eigenvalue = have the form u where Gu =. The eigenvectors corresponding to eigenvalues dierent from have the form u, where the associated eigenvalues p satisfy (37), and Hp = p ; K T Gu = u ; p = Gu ; u = K T p Let denote a diagonal matrix with entries f g from (37). If the eigenvalues of H have only linear elementary divisors, then A D V D = V D D D, where B C D D I A n u n p n p ; V D = V U U C A n u ; n p P P n p n u n n p n p n u p n p n p in which the columns of V form an orthogonal basis of the null space of G and (4) P = GU ; U = K T P ( I) We will normalize the matrices U and P as follows n p (43) U = U = U = K T P; P = P ( I); P = P ( I)
12 It is then easily veried that V D = I ( ) ( ) C A P G P P G P C A ; where (44) = V T (I U(GU) G) = V T (I K T (GK T ) G) Now let A D () = A D E be a perturbation of A D where E is as in (3). The eigenvalues of this perturbed matrix are the same as those of (45) V D A D()V D = D D ^ED = I ^E ^E ^E 3 ^E ^E ^E 3 ^E 3 ^E 3 ^E 33 C A ; where ^E D = V D EV D. As in x3, Gerschgorin analysis implies that for small, the eects of perturbation can be bounded using the perturbations on the block diagonal of (45). These are given by (46) ^E = T E V; ^E = ( ) hp G(E U E P ) P (E U E P ) ; ^E 33 = ( ) hp G(E U E P ) P (E U E P ) i i Theorem 4.. If H has no nonlinear elementary divisors, then the eigenvalues ^ of A D () that are perturbations of = satisfy j^ j c h ( G;K) O( ) for both the original and scaled formulations of the discrete Oseen equations. The perturbations of eigenvalues dierent from satisfy (47) j^ j and (48) 8 >< > j^ j c (P ) G;K h c h (P ) G;K 8 >< > c h O( ) O( ) for the original formulation; for the scaled formulation; c (P ) h c O( ) for the original formulation; h (P ) c h O( ) for the scaled formulation Proof. For the perturbations of =, it follows from (8), (44) and (45) that j^ j c k ^E k c h k ^E k c h ( G;K)
13 For the perturbations of eigenvalues dierent from, note that (37) and the bounds () on the eigenvalues of H imply that the norm of each of the diagonal matrices (49) I; I; ( ) ; ; ( I) ; is bounded independent of and h. Gerschgorin analysis gives (5) j^ j c h k ^Ejj k ; where j = corresponds to and j = 3 to. Relations (43) and (46) then imply (5) k ^Ejj k k( ) k kp Gk kp k kk T k k Ik kp k Consider kp Gk. Using the expression for P in (43), we have kp Gk k( I) k kp k kgk ; and combining this with (5) gives k ^E 33 k (P ) k( ) k k( I) k kgk kk T k k Ik Result (48) follows from (5), the boundedness of the matrices of (49), and the bounds on kgk and kk T k in Lemma. and Corollary.3. For kp Gk, rst observe that (4) and (43) imply (P G)(K T P ) = I Let (P G) T = Q R and K T P = Q R be QR factorizations, where R and R are square and nonsingular. Then R T QT Q R = I, and (5) kp Gk = kr k kr k k(q T Q ) k = G;K kr k But kr k is the inverse of the smallest singular value of K T P, so that kr k k P (KK T ) P T k = k k kp k k(kk T ) k = From Lemma., Corollary.3 and the boundedness of k k, the term on the right is bounded by c= kp k for the original formulation and c kp k for the alternative formulation. Result (47) then follows from (5), (5) (with j = ) and (5). Remark 4.. The dierence between (47) and (48) stems from the fact that k( I) k is not independent of, so we cannot bound kp Gk directly. The results of x6 suggest that the perturbations do not behave dierently, but we see no way to avoid including G;K in (47). 5. The rank-decient case. Unless an additional constraint is imposed on the pressure in (){() or (4), the matrix of (3) will be rank decient by one. This is the case for the test problems of x6. We outline here how the analysis above carries over in the rank-decient case.
14 For the block triangular preconditioner, A T V T = V T D T where, for 6 (H), the analogues of the matrices dened in () { () are D T = I I ; V T = ; V y (I H) G P T = I P y (I H) G P y is a square matrix of order n p whose eigenvalues are the nonzero eigenvalues of H, and P spans the associated invariant subspace. In particular, P y replaces P and V y T replaces V T. The analysis of perturbations of the nonzero eigenvalues of A T then carries through verbatim with the inverse of the smallest singular value of P in place of kp k. The case (H) is generalized in a similar manner with (I H) y in place of (I H) For the block diagonal preconditioner, the analogue of the similarity transformation of Theorem 4. is A D V D = V D D D where B D D I n u n p n p n p C A nu np n p ; V D = n p V U U P P n u n p n p Once again, all the analysis of x4 goes through with P referring to the matrix of eigenvectors corresponding to nonzero eigenvalues of H, P y in place of P, and the inverse of the smallest nonzero eigenvalue of KK T in place of k(kk T ) k. 6. Experimental results. We now present the results of numerical experiments that supplement the analysis of xx3{5. Our test problem is a discrete Oseen operator (4) on on = ( ; ) ( ; ), with Dirichlet boundary conditions u = u = on the three xed walls (x =, y =, x = ), and u =, u = on the moving wall (y = ). The coecients of the convection terms describe a circular vortex w = y( x ); w = x( y ) We discretize using bilinear nite elements on a uniform rectangular n n velocity grids of width h = =n, augmented by streamline upwinding [9, p. 85]. To impose div-stability, the pressure discretization uses \macro-elements" of width h; see [8, p. 3]. The hydrostatic pressure is not explicitly specied, so that the matrices (3) are rank-decient by one. Additional details about this problem are given in [3]. All computations were performed in MATLAB on either a Sun SPARC- workstation or a DEC-Alpha 4/75 workstation. We rst show in Figure some sample distributions of eigenvalues of A T and A D, for = = and n = 3. The plot on the left gives an indication of the rectangle enclosing the eigenvalues of H (see ()); the plot on the right represents the result of the mapping 7 p 4 = of (37). Both pictures include the eigenvalue, of multiplicity n u = 78 for A T and n u n p = 89 for A D. We present our results primarily as tabulations of maximum perturbations and other quantities, for various choices of viscosity parameter and grid parameter n. The rows and columns of the tables indicate behavior as either or n gets large (h ). Note that accurate discrete solutions to (4) are obtained only if is not too 3 n p C A n u n p
15 Fig.. Eigenvalues of A T (left) and A D (right), for = = and n = small relative to h. This diculty can be ameliorated to some extent by an appropriate choice of discretization such as the streamline upwinding method used here [],[9, p. 6]. In practical experiments, it is often desired to compute solutions of a xed accuracy for a variety of values of by letting h and simultaneously. In an eort to follow trends in the data, we will consider some combinations of and h that could produce inaccurate solutions. Our main results are in Tables and. Table shows the eects of perturbation of the block tridiagonal preconditioner, for two values of n and various. This data was obtained by computing the eigenvalues of a set of ten perturbed matrices A T () = A T E where = 8 and E is a dense matrix with uniformly distributed random numbers in an interval [ a n ; a n ], where a n = 6=n. (For this choice, kek is approximately constant for all n.) For =, the table presents max j^ j ^ (A T E) where (A T E) is the set of n p eigenvalues that are closest to, and the maximum is over all the perturbations. For 6=, the table shows max j^ j ^ (A T E) where (A T E) denotes the perturbations of (A T ), 6= and 6=. 4 Analogous results for the block diagonal preconditioner are shown in Table. Here we also distinguish between the eigenvalues with positive and negative real parts. Before discussing this data, we present experimental results for three other quantities appearing in the bounds of xx3{4 (P ), k(i H) k, and G;K. These are shown in Tables 3, 4, and 5, respectively. Here (P ) refers to the version used for the 4 These were obtained by sorting (A T ) and (A T E) for each E and comparing the ordered sets. For all tests, similar results were obtained for other values of. 4 /; /;
16 Table Maximum normalized perturbations of eigenvalues of A T. n = 6 = = =3 =5 = Original = Formulation 6= Scaled = Formulation 6= n = 3 Original = Formulation 6= Scaled = Formulation 6= Table Maximum normalized perturbations of eigenvalues of A D. n = 6 = = =3 =5 = Original = Formulation Re() < Re() > Scaled = Formulation Re() < Re() > n = 3 Original = Formulation Re() < Re() > Scaled = Formulation Re() < Re() >
17 Table 3 Condition number of matrix P of eigenvectors of H. = = =3 =5 = n = n = n = e4 Table 4 Norm of (I H). = = =3 =5 = n = n = n = e rank-decient problem as outlined in x5. It was obtained by computing the matrix of eigenvectors P of H, normalizing the columns of P to have unit l -norm, and then computing the ratio of largest to smallest singular values of the submatrix of P corresponding to the nonzero eigenvalues of H. (The normalization ensures that we are within a factor of p n p of the condition number of the optimally scaled version of P [5].) Note that (P ) and G;K exhibit a general tendency to increase as either or n increases, with (P ) being particularly volatile. Now consider the data of Table for the block triangular preconditioner. Several trends are apparent. In the original formulation, the perturbations of both = and 6= are increasing with =. In the scaled formulation, the perturbations of = are insensitive to. This behavior is consistent with the results of Theorem 3... In the scaled formulation, the perturbations of 6= show some growth with = for n = 3, although there is no clear trend. This may derive from growth in the product (P ) k(i H) k. 3. There is little or no increase in perturbation size (and a decrease in some cases) with the change of grid size from n = 6 to n = 3. This contrasts with the bounds from the analysis, which degrade as h. Next, consider Table.. The perturbations of 6= for both the unscaled and scaled problems show some growth with =. The qualitative trends are similar, but the perturbations for the scaled formulations are larger. In contrast, the analysis (Theorem 4.) suggests that the scaled version would be smaller; therefore it appears that the upper bounds of this analysis are not giving a complete indication of dependence on.. The perturbations of = are insensitive to and there is essentially no dierence between the perturbations of the eigenvalues with positive and negative real parts. As noted in Remark 4., we believe the dependence on G;K is an artifact of the proof. 3. The dependence on the mesh size is more pronounced for the triangular preconditioning, although there is no consistent pattern. Thus, the analysis gives upper bounds on perturbation sizes, although it is not possible to completely correlate the analytic bounds and experimental results. This 6
18 Table 5 Secant of largest principal angle between Range(G T ) and Range(K T ) (G;K). = = =3 =5 = n = n = Table 6 Maximum normalized perturbations of eigenvalues of H. = = =3 =5 = n = n = n = is likely due to the lack of a clear pattern in the behavior of (P ) and k(i H) k, together with the use of norm inequalities throughout the analysis that are not necessarily tight. For example, the analysis combines bounds on matrix l -norms derived from the underlying dierential operator with l norms derived from Gerschgorin bounds. We suspect that the factor of =h used to relate these quantities articially inates the dependence of the bounds on the mesh size, although we see no way to avoid introducing this term. In general, both the analysis and experiments indicate a tendency for perturbations of the eigenvalues dierent from to increase with =, but growth appears to be at worst linear and for the block triangular preconditioning it can be reduced by scaling. Finally, Table 6 examines the sensitivity to perturbation of the eigenvalues of the reduced matrix H. The entries are max j^ j ^(HE) The results also indicate that the perturbations increase as the viscosity decreases, and they display some growth as the number of mesh points increase, roughly like that for the block diagonal preconditioner displayed in Table. Note that the dependence on both = and n is much less severe than that of (P ) shown in Table 3. Once again, this stands in contrast with analytic bounds such as those obtainable from the Bauer-Fike theorem [7, p. 34], which suggests perturbations proportional to (P ). 7. Inexact inner iteration. We conclude by examining the eect of inexact computation of the action of F. For brevity, we restrict our attention to the unscaled version of A T ; similar analysis leads to essentially the same conclusions for A D. If the action of F is approximated using an iterative method for each system F w = v, then F can be replaced by F E in (9). The perturbed preconditioned matrix is then A T E where E = F E F EKT BE BEK T =; 7
19 For this analysis it will be useful to consider the complete version of (4), ^E T = F E(I K T (I H) G) P (I (I H) )BE (I K T (I H) G) Suppose the approximate solution ~w satises F EK T P P (I (I H) )BEK T P C A (53) kv F ~wk kvk for some tolerance. Standard inequalities yield the bounds on the relative error kw ~wk kwk kf k kf k ; kw ~wk kwk kf k kek Treating these as approximate equalities gives kek kf k, it follows that k ^E k kf k kf k kk T k k(i H) k kgk ; k ^E k kf k kf k kk T k kp k ; k ^E k kp k ( k(i H) k) kbk kf k k ^E k (P ) ki (I H) k kf k kk T k kk T k k(i H) k kgk ; Using the l -norm, let us consider the dependence of these bounds on the viscosity under the assumption (derived from the experimental results presented above) that the inuence of P and H is not signicant. The bounds on k ^E k, k ^E k, and k ^E k include the product kk T k kgk = O() O(=), which is independent of. Using the coercivity and continuity of the convection-diusion operator, it can be shown that as functions of, kf k = O(); kf k = O(=) (An algebraic proof can be found in [4, Theorem ].) This implies that the bounds on k ^E k and k ^E k grow as = decreases, whereas the bounds on k ^E k and k ^E k are independent of. If is small, then as in x3 we can restrict our attention to the block diagonal entries, which indicates that the perturbations of eigenvalues dierent from in this particular case do not depend on. On the other hand, if is large, then it is not possible to exclude ^E from the Gerschgorin analysis, and the presence of kf k in this bound suggests that perturbations in all eigenvalues grow like =. Consider the implication of these observations on the performance of iterative methods for solving the discrete Oseen equations (3). We demonstrated in [3] that the iteration counts of Krylov subspace methods such as GMRES with the preconditioner Q T (7) are independent of the mesh size, but that there is some deterioration in performance as decreases. Now suppose an inner iteration with stopping criterion (53) is used to approximate the action of F. The analysis given here suggests that if is large, there may be additional degradation of performance for small. In contrast, if is small then only = is sensitive to perturbation, which suggests that (extra) degradation of the inner iteration with decreasing may not be as pronounced. Figure 8
20 Fig.. Iterations of FGMRES with inexact convection-diusion solves. n=6 n= tau= tau=.5 tau=. tau=.d 3 Iterations 8 6 tau=. tau=.d 3 tau= (direct) Iterations 8 6 tau= (direct) /nu /nu shows the results of numerical experiments that corroborate these observations. The gure plots iteration counts of right-preconditioned \exible" GMRES (FGMRES) [] as a function of, for solving the discrete problem (3). The inner iteration for the convection-diusion subproblems F w = v was a line Gauss-Seidel method with stopping criterion (53). 5 The test matrices were those used in x6; the right hand side f consisted of normally distributed random numbers with mean and variance. The outer iteration used a zero initial guess and was stopped when the relative residual in the Euclidean norm was less than or equal to 6. The results with inner iteration are compared with using a direct method for the action of F. They indicate that for the relatively modest tolerance = 3, the inexact inner solves lead to little increase in outer iterations for any. For less stringent, additional outer iterations are required and the number of additional iterations becomes larger as decreases. Remark 7.. It can be shown that for small the perturbations of the eigenvalues of the scaled system behave in the same way as the those for the unscaled system. However, scaling aects the relative weighting given to the two block equations of (3), which in turn may aect the iterative solver. Therefore, we have restricted our attention here to the unscaled system. Acknowledgements. The author acknowledges helpful discussions with Santiago Arteaga, Luca Pavarino and Pete Stewart, and insightful comments from Hans Weinberger. The numerical results with FGMRES used software produced by Tim Kelley []. REFERENCES [] F. Brezzi and M. Fortin. Mixed and Hybrid Finite Element Methods. Springer-Verlag, New York, FGMRES allows the preconditioner to vary during the course of the iteration; it was used here because the number of inner iterations was not generally constant, i.e., the preconditioner is not a xed linear operator. FGMRES is equivalent to GMRES when a direct solve is used for F. 9
21 [] A. Brooks and T. Hughes. Streamline upwind/petrov-galerkin formulations for convection dominated ows with particular emphasis on the incompressible Navier-Stokes equations. Comp. Meth. Appl. Mech. Engng., 399{59, 98. [3] H. Elman and D. Silvester. Fast nonsymmetric iterations and preconditioning for Navier-Stokes equations. SIAM J. Sci. Comput., 733{46, 996. [4] H. C. Elman and M. H. Schultz. Preconditioning by fast direct methods for nonselfadjoint nonseparable elliptic problems. SIAM J. Numer. Anal, 344{57, 986. [5] R. Freund and N. M. Nachtigal. QMR a quasi-minimal residual method for non-hermitian linear systems. Numer. Math., 635{339, 99. [6] R. Glowinski. Numerical Methods for Nonlinear Variational Problems. Springer-Verlag, New York, 984. [7] G. H. Golub and C. F. Van Loan. Matrix Computations. The Johns Hopkins University Press, Baltimore, second edition, 989. [8] M. Gunzburger. Finite Element Methods for Viscous Incompressible Flows. Academic Press, San Diego, 989. [9] C. Johnson. Numerical Solution of Partial Dierential Equations by the Finite Element Method. Cambridge University Press, New York, 987. [] O. A. Karakashian. On a Galerkin-Lagrange multiplier method for the stationary Navier-Stokes equations. SIAM. J. Numer. Anal., 999{93, 98. [] T. Kelley. Iterative Methods for Linear and Nonlinear Equations. SIAM, Philadelphia, 995. [] Y. Saad. A exible inner-outer preconditioned GMRES algorithm. SIAM J. Sci. Comput., 446{469, 993. [3] Y. Saad and M. H. Schultz. GMRES A generalized minimal residual algorithm for solving nonsymmetric linear systems. SIAM J. Sci. Stat. Comput., 7856{869, 986. [4] J. Stoer and R. Bulirsch. Introduction to Numerical Analysis. Springer-Verlag, New York, second edition, 993. [5] A. van der Sluis. Condition numbers and equilibration of matrices. Numer. Math., 44{3, 969. [6] R. Verfurth. A combined conjugate gradient-multigrid algorithm for the numerical solution of the Stokes problem. IMA J. Numer. Anal., 444{455, 984. [7] A. J. Wathen. Realistic eigenvalue bounds for the Galerkin mass matrix. IMA J. Numer. Anal., 7449{457, 987. [8] J. H. Wilkinson. The Algebraic Eigenvalue Problem. Oxford University Press, Oxford, 965.
EQUATIONS WITH LOW VISCOSITY HOWARD C. ELMAN UMIACS-TR November 1996
PRECONDITIONING FOR THE STEADY-STATE NAVIER-STOKES EQUATIONS WITH LOW VISCOSITY HOWARD C. ELMAN Report CS-TR-372 UMIACS-TR-96-82 November 996 Abstract. We introduce a preconditioner for the linearized
More informationUNIVERSITY OF MARYLAND PRECONDITIONERS FOR THE DISCRETE STEADY-STATE NAVIER-STOKES EQUATIONS CS-TR #4164 / UMIACS TR # JULY 2000
UNIVERSITY OF MARYLAND INSTITUTE FOR ADVANCED COMPUTER STUDIES DEPARTMENT OF COMPUTER SCIENCE PERFORMANCE AND ANALYSIS OF SADDLE POINT PRECONDITIONERS FOR THE DISCRETE STEADY-STATE NAVIER-STOKES EQUATIONS
More informationA Review of Preconditioning Techniques for Steady Incompressible Flow
Zeist 2009 p. 1/43 A Review of Preconditioning Techniques for Steady Incompressible Flow David Silvester School of Mathematics University of Manchester Zeist 2009 p. 2/43 PDEs Review : 1984 2005 Update
More informationA MULTIGRID ALGORITHM FOR. Richard E. Ewing and Jian Shen. Institute for Scientic Computation. Texas A&M University. College Station, Texas SUMMARY
A MULTIGRID ALGORITHM FOR THE CELL-CENTERED FINITE DIFFERENCE SCHEME Richard E. Ewing and Jian Shen Institute for Scientic Computation Texas A&M University College Station, Texas SUMMARY In this article,
More informationA Robust Preconditioned Iterative Method for the Navier-Stokes Equations with High Reynolds Numbers
Applied and Computational Mathematics 2017; 6(4): 202-207 http://www.sciencepublishinggroup.com/j/acm doi: 10.11648/j.acm.20170604.18 ISSN: 2328-5605 (Print); ISSN: 2328-5613 (Online) A Robust Preconditioned
More informationFast solvers for steady incompressible flow
ICFD 25 p.1/21 Fast solvers for steady incompressible flow Andy Wathen Oxford University wathen@comlab.ox.ac.uk http://web.comlab.ox.ac.uk/~wathen/ Joint work with: Howard Elman (University of Maryland,
More informationDELFT UNIVERSITY OF TECHNOLOGY
DELFT UNIVERSITY OF TECHNOLOGY REPORT 18-05 Efficient and robust Schur complement approximations in the augmented Lagrangian preconditioner for high Reynolds number laminar flows X. He and C. Vuik ISSN
More informationMathematics and Computer Science
Technical Report TR-2007-002 Block preconditioning for saddle point systems with indefinite (1,1) block by Michele Benzi, Jia Liu Mathematics and Computer Science EMORY UNIVERSITY International Journal
More informationFast Iterative Solution of Saddle Point Problems
Michele Benzi Department of Mathematics and Computer Science Emory University Atlanta, GA Acknowledgments NSF (Computational Mathematics) Maxim Olshanskii (Mech-Math, Moscow State U.) Zhen Wang (PhD student,
More informationNumerische Mathematik
Numer. Math. (2002) 90: 665 688 Digital Object Identifier (DOI) 10.1007/s002110100300 Numerische Mathematik Performance and analysis of saddle point preconditioners for the discrete steady-state Navier-Stokes
More informationEfficient Augmented Lagrangian-type Preconditioning for the Oseen Problem using Grad-Div Stabilization
Efficient Augmented Lagrangian-type Preconditioning for the Oseen Problem using Grad-Div Stabilization Timo Heister, Texas A&M University 2013-02-28 SIAM CSE 2 Setting Stationary, incompressible flow problems
More informationInstitute for Advanced Computer Studies. Department of Computer Science. Iterative methods for solving Ax = b. GMRES/FOM versus QMR/BiCG
University of Maryland Institute for Advanced Computer Studies Department of Computer Science College Park TR{96{2 TR{3587 Iterative methods for solving Ax = b GMRES/FOM versus QMR/BiCG Jane K. Cullum
More information7.4 The Saddle Point Stokes Problem
346 CHAPTER 7. APPLIED FOURIER ANALYSIS 7.4 The Saddle Point Stokes Problem So far the matrix C has been diagonal no trouble to invert. This section jumps to a fluid flow problem that is still linear (simpler
More informationIterative Methods for Problems in Computational Fluid. Howard C. Elman. University of Maryland. David J. Silvester. University of Manchester
Report no. 96/19 Iterative Methods for Problems in Computational Fluid Dynamics Howard C. Elman University of Maryland David J. Silvester University of Manchester Andrew J. Wathen Oxford University We
More informationPRECONDITIONING TECHNIQUES FOR NEWTON S METHOD FOR THE INCOMPRESSIBLE NAVIER STOKES EQUATIONS
BIT Numerical Mathematics 43: 961 974, 2003. 2003 Kluwer Academic Publishers. Printed in the Netherlands. 961 PRECONDITIONING TECHNIQUES FOR NEWTON S METHOD FOR THE INCOMPRESSIBLE NAVIER STOKES EQUATIONS
More informationM.A. Botchev. September 5, 2014
Rome-Moscow school of Matrix Methods and Applied Linear Algebra 2014 A short introduction to Krylov subspaces for linear systems, matrix functions and inexact Newton methods. Plan and exercises. M.A. Botchev
More informationEfficient Solvers for the Navier Stokes Equations in Rotation Form
Efficient Solvers for the Navier Stokes Equations in Rotation Form Computer Research Institute Seminar Purdue University March 4, 2005 Michele Benzi Emory University Atlanta, GA Thanks to: NSF (MPS/Computational
More informationUNIVERSITY OF MARYLAND NAVIER-STOKES EQUATIONS CS-TR #4073 / UMIACS TR #99-66 OCTOBER 1999
UNIVERSITY OF MARYLAND INSTITUTE FOR ADVANCED COMPUTER STUDIES DEPARTMENT OF COMPUTER SCIENCE EFFICIENT PRECONDITIONING OF THE LINEARIZED NAVIER-STOKES EQUATIONS CS-TR #4073 / UMIACS TR #99-66 DAVID SILVESTER
More informationSolving Large Nonlinear Sparse Systems
Solving Large Nonlinear Sparse Systems Fred W. Wubs and Jonas Thies Computational Mechanics & Numerical Mathematics University of Groningen, the Netherlands f.w.wubs@rug.nl Centre for Interdisciplinary
More informationMODIFIED STREAMLINE DIFFUSION SCHEMES FOR CONVECTION-DIFFUSION PROBLEMS. YIN-TZER SHIH AND HOWARD C. ELMAN y
MODIFIED STREAMLINE DIFFUSION SCHEMES FOR CONVECTION-DIFFUSION PROBLEMS YIN-TZER SHIH AND HOWARD C. ELMAN y Abstract. We consider the design of robust and accurate nite element approximation methods for
More informationThe solution of the discretized incompressible Navier-Stokes equations with iterative methods
The solution of the discretized incompressible Navier-Stokes equations with iterative methods Report 93-54 C. Vuik Technische Universiteit Delft Delft University of Technology Faculteit der Technische
More informationPreconditioned GMRES Revisited
Preconditioned GMRES Revisited Roland Herzog Kirk Soodhalter UBC (visiting) RICAM Linz Preconditioning Conference 2017 Vancouver August 01, 2017 Preconditioned GMRES Revisited Vancouver 1 / 32 Table of
More informationUMIACS-TR July CS-TR 2721 Revised March Perturbation Theory for. Rectangular Matrix Pencils. G. W. Stewart.
UMIAS-TR-9-5 July 99 S-TR 272 Revised March 993 Perturbation Theory for Rectangular Matrix Pencils G. W. Stewart abstract The theory of eigenvalues and eigenvectors of rectangular matrix pencils is complicated
More informationNumerical Methods in Matrix Computations
Ake Bjorck Numerical Methods in Matrix Computations Springer Contents 1 Direct Methods for Linear Systems 1 1.1 Elements of Matrix Theory 1 1.1.1 Matrix Algebra 2 1.1.2 Vector Spaces 6 1.1.3 Submatrices
More informationON ORTHOGONAL REDUCTION TO HESSENBERG FORM WITH SMALL BANDWIDTH
ON ORTHOGONAL REDUCTION TO HESSENBERG FORM WITH SMALL BANDWIDTH V. FABER, J. LIESEN, AND P. TICHÝ Abstract. Numerous algorithms in numerical linear algebra are based on the reduction of a given matrix
More informationDomain decomposition on different levels of the Jacobi-Davidson method
hapter 5 Domain decomposition on different levels of the Jacobi-Davidson method Abstract Most computational work of Jacobi-Davidson [46], an iterative method suitable for computing solutions of large dimensional
More informationFinding Rightmost Eigenvalues of Large, Sparse, Nonsymmetric Parameterized Eigenvalue Problems
Finding Rightmost Eigenvalues of Large, Sparse, Nonsymmetric Parameterized Eigenvalue Problems AMSC 663-664 Final Report Minghao Wu AMSC Program mwu@math.umd.edu Dr. Howard Elman Department of Computer
More informationLecture Note III: Least-Squares Method
Lecture Note III: Least-Squares Method Zhiqiang Cai October 4, 004 In this chapter, we shall present least-squares methods for second-order scalar partial differential equations, elastic equations of solids,
More informationON THE GENERALIZED DETERIORATED POSITIVE SEMI-DEFINITE AND SKEW-HERMITIAN SPLITTING PRECONDITIONER *
Journal of Computational Mathematics Vol.xx, No.x, 2x, 6. http://www.global-sci.org/jcm doi:?? ON THE GENERALIZED DETERIORATED POSITIVE SEMI-DEFINITE AND SKEW-HERMITIAN SPLITTING PRECONDITIONER * Davod
More informationOUTLINE 1. Introduction 1.1 Notation 1.2 Special matrices 2. Gaussian Elimination 2.1 Vector and matrix norms 2.2 Finite precision arithmetic 2.3 Fact
Computational Linear Algebra Course: (MATH: 6800, CSCI: 6800) Semester: Fall 1998 Instructors: { Joseph E. Flaherty, aherje@cs.rpi.edu { Franklin T. Luk, luk@cs.rpi.edu { Wesley Turner, turnerw@cs.rpi.edu
More information216 S. Chandrasearan and I.C.F. Isen Our results dier from those of Sun [14] in two asects: we assume that comuted eigenvalues or singular values are
Numer. Math. 68: 215{223 (1994) Numerische Mathemati c Sringer-Verlag 1994 Electronic Edition Bacward errors for eigenvalue and singular value decomositions? S. Chandrasearan??, I.C.F. Isen??? Deartment
More informationA Method for Constructing Diagonally Dominant Preconditioners based on Jacobi Rotations
A Method for Constructing Diagonally Dominant Preconditioners based on Jacobi Rotations Jin Yun Yuan Plamen Y. Yalamov Abstract A method is presented to make a given matrix strictly diagonally dominant
More informationLinear Solvers. Andrew Hazel
Linear Solvers Andrew Hazel Introduction Thus far we have talked about the formulation and discretisation of physical problems...... and stopped when we got to a discrete linear system of equations. Introduction
More informationA Domain Decomposition Based Jacobi-Davidson Algorithm for Quantum Dot Simulation
A Domain Decomposition Based Jacobi-Davidson Algorithm for Quantum Dot Simulation Tao Zhao 1, Feng-Nan Hwang 2 and Xiao-Chuan Cai 3 Abstract In this paper, we develop an overlapping domain decomposition
More informationPerturbation results for nearly uncoupled Markov. chains with applications to iterative methods. Jesse L. Barlow. December 9, 1992.
Perturbation results for nearly uncoupled Markov chains with applications to iterative methods Jesse L. Barlow December 9, 992 Abstract The standard perturbation theory for linear equations states that
More informationETNA Kent State University
Electronic Transactions on Numerical Analysis. Volume 36, pp. 39-53, 009-010. Copyright 009,. ISSN 1068-9613. ETNA P-REGULAR SPLITTING ITERATIVE METHODS FOR NON-HERMITIAN POSITIVE DEFINITE LINEAR SYSTEMS
More informationUMIACS-TR July CS-TR 2494 Revised January An Updating Algorithm for. Subspace Tracking. G. W. Stewart. abstract
UMIACS-TR-9-86 July 199 CS-TR 2494 Revised January 1991 An Updating Algorithm for Subspace Tracking G. W. Stewart abstract In certain signal processing applications it is required to compute the null space
More informationNumerical behavior of inexact linear solvers
Numerical behavior of inexact linear solvers Miro Rozložník joint results with Zhong-zhi Bai and Pavel Jiránek Institute of Computer Science, Czech Academy of Sciences, Prague, Czech Republic The fourth
More informationInstitute for Advanced Computer Studies. Department of Computer Science. On the Perturbation of. LU and Cholesky Factors. G. W.
University of Maryland Institute for Advanced Computer Studies Department of Computer Science College Park TR{95{93 TR{3535 On the Perturbation of LU and Cholesky Factors G. W. Stewart y October, 1995
More informationAMS526: Numerical Analysis I (Numerical Linear Algebra) Lecture 23: GMRES and Other Krylov Subspace Methods; Preconditioning
AMS526: Numerical Analysis I (Numerical Linear Algebra) Lecture 23: GMRES and Other Krylov Subspace Methods; Preconditioning Xiangmin Jiao SUNY Stony Brook Xiangmin Jiao Numerical Analysis I 1 / 18 Outline
More informationThe Eigenvalue Problem: Perturbation Theory
Jim Lambers MAT 610 Summer Session 2009-10 Lecture 13 Notes These notes correspond to Sections 7.2 and 8.1 in the text. The Eigenvalue Problem: Perturbation Theory The Unsymmetric Eigenvalue Problem Just
More informationPROOF OF TWO MATRIX THEOREMS VIA TRIANGULAR FACTORIZATIONS ROY MATHIAS
PROOF OF TWO MATRIX THEOREMS VIA TRIANGULAR FACTORIZATIONS ROY MATHIAS Abstract. We present elementary proofs of the Cauchy-Binet Theorem on determinants and of the fact that the eigenvalues of a matrix
More informationIterative Methods for Solving A x = b
Iterative Methods for Solving A x = b A good (free) online source for iterative methods for solving A x = b is given in the description of a set of iterative solvers called templates found at netlib: http
More informationtdepartment of Mathematics, University of Manchester Institute of Science and Technology, Manchester M601QD, UK (na. ornl. gov).
SIAM J. ScI. COMPUT. Vol. 17, No. 1, pp. 33-46, January 1996 () 1996 Society for Industrial and Applied Mathematics 3 FAST NONSYMMETRIC ITERATIONS AND PRECONDITIONING FOR NAVIER-STOKES EQUATIONS* HOWARD
More informationPreconditioned inverse iteration and shift-invert Arnoldi method
Preconditioned inverse iteration and shift-invert Arnoldi method Melina Freitag Department of Mathematical Sciences University of Bath CSC Seminar Max-Planck-Institute for Dynamics of Complex Technical
More information1. Introduction. Consider the Navier Stokes equations ηu t ν 2 u + (u grad) u + grad p = f div u = 0 (1.1)
University of Maryland Department of Computer Science TR-C4930 University of Maryland Institute for Advanced Computer Studies TR-009-0 BOUNDARY CONDITIONS IN APPROXIMATE COMMUTATOR PRECONDITIONERS FOR
More informationMath 102, Winter Final Exam Review. Chapter 1. Matrices and Gaussian Elimination
Math 0, Winter 07 Final Exam Review Chapter. Matrices and Gaussian Elimination { x + x =,. Different forms of a system of linear equations. Example: The x + 4x = 4. [ ] [ ] [ ] vector form (or the column
More informationON A GENERAL CLASS OF PRECONDITIONERS FOR NONSYMMETRIC GENERALIZED SADDLE POINT PROBLEMS
U..B. Sci. Bull., Series A, Vol. 78, Iss. 4, 06 ISSN 3-707 ON A GENERAL CLASS OF RECONDIIONERS FOR NONSYMMERIC GENERALIZED SADDLE OIN ROBLE Fatemeh anjeh Ali BEIK his paper deals with applying a class
More informationChapter 7. Iterative methods for large sparse linear systems. 7.1 Sparse matrix algebra. Large sparse matrices
Chapter 7 Iterative methods for large sparse linear systems In this chapter we revisit the problem of solving linear systems of equations, but now in the context of large sparse systems. The price to pay
More informationThe amount of work to construct each new guess from the previous one should be a small multiple of the number of nonzeros in A.
AMSC/CMSC 661 Scientific Computing II Spring 2005 Solution of Sparse Linear Systems Part 2: Iterative methods Dianne P. O Leary c 2005 Solving Sparse Linear Systems: Iterative methods The plan: Iterative
More informationUniversity of Maryland Department of Computer Science TR-5009 University of Maryland Institute for Advanced Computer Studies TR April 2012
University of Maryland Department of Computer Science TR-5009 University of Maryland Institute for Advanced Computer Studies TR-202-07 April 202 LYAPUNOV INVERSE ITERATION FOR COMPUTING A FEW RIGHTMOST
More informationUNIVERSITY OF MARYLAND A MULTIGRID METHOD ENHANCED BY KRYLOV SUBSPACE ITERATION FOR DISCRETE HELMHOLTZ EQUATIONS CS-TR #4029 / UMIACS TR #99-36
UNIVERSITY OF MARYLAND INSTITUTE FOR ADVANCED COMPUTER STUDIES DEPARTMENT OF COMPUTER SCIENCE A MULTIGRID METHOD ENHANCED BY KRYLOV SUBSPACE ITERATION FOR DISCRETE HELMHOLTZ EQUATIONS CS-TR #4029 / UMIACS
More informationPreface to the Second Edition. Preface to the First Edition
n page v Preface to the Second Edition Preface to the First Edition xiii xvii 1 Background in Linear Algebra 1 1.1 Matrices................................. 1 1.2 Square Matrices and Eigenvalues....................
More informationA Finite Element Method for an Ill-Posed Problem. Martin-Luther-Universitat, Fachbereich Mathematik/Informatik,Postfach 8, D Halle, Abstract
A Finite Element Method for an Ill-Posed Problem W. Lucht Martin-Luther-Universitat, Fachbereich Mathematik/Informatik,Postfach 8, D-699 Halle, Germany Abstract For an ill-posed problem which has its origin
More informationThe Conjugate Gradient Method
The Conjugate Gradient Method Classical Iterations We have a problem, We assume that the matrix comes from a discretization of a PDE. The best and most popular model problem is, The matrix will be as large
More informationAdaptive preconditioners for nonlinear systems of equations
Adaptive preconditioners for nonlinear systems of equations L. Loghin D. Ruiz A. Touhami CERFACS Technical Report TR/PA/04/42 Also ENSEEIHT-IRIT Technical Report RT/TLSE/04/02 Abstract The use of preconditioned
More informationOn the Preconditioning of the Block Tridiagonal Linear System of Equations
On the Preconditioning of the Block Tridiagonal Linear System of Equations Davod Khojasteh Salkuyeh Department of Mathematics, University of Mohaghegh Ardabili, PO Box 179, Ardabil, Iran E-mail: khojaste@umaacir
More informationOn the accuracy of saddle point solvers
On the accuracy of saddle point solvers Miro Rozložník joint results with Valeria Simoncini and Pavel Jiránek Institute of Computer Science, Czech Academy of Sciences, Prague, Czech Republic Seminar at
More informationMatrix Algorithms. Volume II: Eigensystems. G. W. Stewart H1HJ1L. University of Maryland College Park, Maryland
Matrix Algorithms Volume II: Eigensystems G. W. Stewart University of Maryland College Park, Maryland H1HJ1L Society for Industrial and Applied Mathematics Philadelphia CONTENTS Algorithms Preface xv xvii
More informationJURJEN DUINTJER TEBBENS
Proceedings of ALGORITMY 005 pp. 40 49 GMRES ACCELERATION ANALYSIS FOR A CONVECTION DIFFUSION MODEL PROBLEM JURJEN DUINTJER TEBBENS Abstract. When we apply the GMRES method [4] to linear systems arising
More informationPreconditioners for the incompressible Navier Stokes equations
Preconditioners for the incompressible Navier Stokes equations C. Vuik M. ur Rehman A. Segal Delft Institute of Applied Mathematics, TU Delft, The Netherlands SIAM Conference on Computational Science and
More informationOn the influence of eigenvalues on Bi-CG residual norms
On the influence of eigenvalues on Bi-CG residual norms Jurjen Duintjer Tebbens Institute of Computer Science Academy of Sciences of the Czech Republic duintjertebbens@cs.cas.cz Gérard Meurant 30, rue
More informationPreliminary/Qualifying Exam in Numerical Analysis (Math 502a) Spring 2012
Instructions Preliminary/Qualifying Exam in Numerical Analysis (Math 502a) Spring 2012 The exam consists of four problems, each having multiple parts. You should attempt to solve all four problems. 1.
More information1 Number Systems and Errors 1
Contents 1 Number Systems and Errors 1 1.1 Introduction................................ 1 1.2 Number Representation and Base of Numbers............. 1 1.2.1 Normalized Floating-point Representation...........
More informationReduced Synchronization Overhead on. December 3, Abstract. The standard formulation of the conjugate gradient algorithm involves
Lapack Working Note 56 Conjugate Gradient Algorithms with Reduced Synchronization Overhead on Distributed Memory Multiprocessors E. F. D'Azevedo y, V.L. Eijkhout z, C. H. Romine y December 3, 1999 Abstract
More informationApplication of turbulence. models to high-lift airfoils. By Georgi Kalitzin
Center for Turbulence Research Annual Research Briefs 997 65 Application of turbulence models to high-lift airfoils By Georgi Kalitzin. Motivation and objectives Accurate prediction of the ow over an airfoil
More informationXIAO-CHUAN CAI AND MAKSYMILIAN DRYJA. strongly elliptic equations discretized by the nite element methods.
Contemporary Mathematics Volume 00, 0000 Domain Decomposition Methods for Monotone Nonlinear Elliptic Problems XIAO-CHUAN CAI AND MAKSYMILIAN DRYJA Abstract. In this paper, we study several overlapping
More informationKey words. conjugate gradients, normwise backward error, incremental norm estimation.
Proceedings of ALGORITMY 2016 pp. 323 332 ON ERROR ESTIMATION IN THE CONJUGATE GRADIENT METHOD: NORMWISE BACKWARD ERROR PETR TICHÝ Abstract. Using an idea of Duff and Vömel [BIT, 42 (2002), pp. 300 322
More informationConvergence Properties of Preconditioned Hermitian and Skew-Hermitian Splitting Methods for Non-Hermitian Positive Semidefinite Matrices
Convergence Properties of Preconditioned Hermitian and Skew-Hermitian Splitting Methods for Non-Hermitian Positive Semidefinite Matrices Zhong-Zhi Bai 1 Department of Mathematics, Fudan University Shanghai
More informationAMS526: Numerical Analysis I (Numerical Linear Algebra)
AMS526: Numerical Analysis I (Numerical Linear Algebra) Lecture 1: Course Overview & Matrix-Vector Multiplication Xiangmin Jiao SUNY Stony Brook Xiangmin Jiao Numerical Analysis I 1 / 20 Outline 1 Course
More informationIndex. for generalized eigenvalue problem, butterfly form, 211
Index ad hoc shifts, 165 aggressive early deflation, 205 207 algebraic multiplicity, 35 algebraic Riccati equation, 100 Arnoldi process, 372 block, 418 Hamiltonian skew symmetric, 420 implicitly restarted,
More informationNumerische Mathematik
Numer. Math. (1999) 83: 231 257 Numerische Mathematik c Springer-Verlag 1999 Eigenanalysis of some preconditioned Helmholtz problems Howard C. Elman, Dianne P. O Leary Computer Science Department and Institute
More informationThe rate of convergence of the GMRES method
The rate of convergence of the GMRES method Report 90-77 C. Vuik Technische Universiteit Delft Delft University of Technology Faculteit der Technische Wiskunde en Informatica Faculty of Technical Mathematics
More informationThe antitriangular factorisation of saddle point matrices
The antitriangular factorisation of saddle point matrices J. Pestana and A. J. Wathen August 29, 2013 Abstract Mastronardi and Van Dooren [this journal, 34 (2013) pp. 173 196] recently introduced the block
More informationUnsteady Incompressible Flow Simulation Using Galerkin Finite Elements with Spatial/Temporal Adaptation
Unsteady Incompressible Flow Simulation Using Galerkin Finite Elements with Spatial/Temporal Adaptation Mohamed S. Ebeida Carnegie Mellon University, Pittsburgh, PA 15213 Roger L. Davis and Roland W. Freund
More informationLecture 3: Inexact inverse iteration with preconditioning
Lecture 3: Department of Mathematical Sciences CLAPDE, Durham, July 2008 Joint work with M. Freitag (Bath), and M. Robbé & M. Sadkane (Brest) 1 Introduction 2 Preconditioned GMRES for Inverse Power Method
More informationCONVERGENCE BOUNDS FOR PRECONDITIONED GMRES USING ELEMENT-BY-ELEMENT ESTIMATES OF THE FIELD OF VALUES
European Conference on Computational Fluid Dynamics ECCOMAS CFD 2006 P. Wesseling, E. Oñate and J. Périaux (Eds) c TU Delft, The Netherlands, 2006 CONVERGENCE BOUNDS FOR PRECONDITIONED GMRES USING ELEMENT-BY-ELEMENT
More informationEigenvalue problems and optimization
Notes for 2016-04-27 Seeking structure For the past three weeks, we have discussed rather general-purpose optimization methods for nonlinear equation solving and optimization. In practice, of course, we
More informationKEYWORDS. Numerical methods, generalized singular values, products of matrices, quotients of matrices. Introduction The two basic unitary decompositio
COMPUTING THE SVD OF A GENERAL MATRIX PRODUCT/QUOTIENT GENE GOLUB Computer Science Department Stanford University Stanford, CA USA golub@sccm.stanford.edu KNUT SLNA SC-CM Stanford University Stanford,
More informationSynopsis of Numerical Linear Algebra
Synopsis of Numerical Linear Algebra Eric de Sturler Department of Mathematics, Virginia Tech sturler@vt.edu http://www.math.vt.edu/people/sturler Iterative Methods for Linear Systems: Basics to Research
More informationThe Deflation Accelerated Schwarz Method for CFD
The Deflation Accelerated Schwarz Method for CFD J. Verkaik 1, C. Vuik 2,, B.D. Paarhuis 1, and A. Twerda 1 1 TNO Science and Industry, Stieltjesweg 1, P.O. Box 155, 2600 AD Delft, The Netherlands 2 Delft
More information1. Introduction Let the least value of an objective function F (x), x2r n, be required, where F (x) can be calculated for any vector of variables x2r
DAMTP 2002/NA08 Least Frobenius norm updating of quadratic models that satisfy interpolation conditions 1 M.J.D. Powell Abstract: Quadratic models of objective functions are highly useful in many optimization
More informationSolution of the Two-Dimensional Steady State Heat Conduction using the Finite Volume Method
Ninth International Conference on Computational Fluid Dynamics (ICCFD9), Istanbul, Turkey, July 11-15, 2016 ICCFD9-0113 Solution of the Two-Dimensional Steady State Heat Conduction using the Finite Volume
More informationLecture 11: CMSC 878R/AMSC698R. Iterative Methods An introduction. Outline. Inverse, LU decomposition, Cholesky, SVD, etc.
Lecture 11: CMSC 878R/AMSC698R Iterative Methods An introduction Outline Direct Solution of Linear Systems Inverse, LU decomposition, Cholesky, SVD, etc. Iterative methods for linear systems Why? Matrix
More informationAPPLIED NUMERICAL LINEAR ALGEBRA
APPLIED NUMERICAL LINEAR ALGEBRA James W. Demmel University of California Berkeley, California Society for Industrial and Applied Mathematics Philadelphia Contents Preface 1 Introduction 1 1.1 Basic Notation
More informationNumerical Linear Algebra
Numerical Linear Algebra The two principal problems in linear algebra are: Linear system Given an n n matrix A and an n-vector b, determine x IR n such that A x = b Eigenvalue problem Given an n n matrix
More informationANALYSIS OF AUGMENTED LAGRANGIAN-BASED PRECONDITIONERS FOR THE STEADY INCOMPRESSIBLE NAVIER-STOKES EQUATIONS
ANALYSIS OF AUGMENTED LAGRANGIAN-BASED PRECONDITIONERS FOR THE STEADY INCOMPRESSIBLE NAVIER-STOKES EQUATIONS MICHELE BENZI AND ZHEN WANG Abstract. We analyze a class of modified augmented Lagrangian-based
More informationfor Finite Element Simulation of Incompressible Flow Arnd Meyer Department of Mathematics, Technical University of Chemnitz,
Preconditioning the Pseudo{Laplacian for Finite Element Simulation of Incompressible Flow Arnd Meyer Department of Mathematics, Technical University of Chemnitz, 09107 Chemnitz, Germany Preprint{Reihe
More informationFoundations of Matrix Analysis
1 Foundations of Matrix Analysis In this chapter we recall the basic elements of linear algebra which will be employed in the remainder of the text For most of the proofs as well as for the details, the
More informationMath Introduction to Numerical Analysis - Class Notes. Fernando Guevara Vasquez. Version Date: January 17, 2012.
Math 5620 - Introduction to Numerical Analysis - Class Notes Fernando Guevara Vasquez Version 1990. Date: January 17, 2012. 3 Contents 1. Disclaimer 4 Chapter 1. Iterative methods for solving linear systems
More informationA NOTE ON THE LADYŽENSKAJA-BABUŠKA-BREZZI CONDITION
A NOTE ON THE LADYŽENSKAJA-BABUŠKA-BREZZI CONDITION JOHNNY GUZMÁN, ABNER J. SALGADO, AND FRANCISCO-JAVIER SAYAS Abstract. The analysis of finite-element-like Galerkin discretization techniques for the
More informationOn solving linear systems arising from Shishkin mesh discretizations
On solving linear systems arising from Shishkin mesh discretizations Petr Tichý Faculty of Mathematics and Physics, Charles University joint work with Carlos Echeverría, Jörg Liesen, and Daniel Szyld October
More informationIntroduction to Applied Linear Algebra with MATLAB
Sigam Series in Applied Mathematics Volume 7 Rizwan Butt Introduction to Applied Linear Algebra with MATLAB Heldermann Verlag Contents Number Systems and Errors 1 1.1 Introduction 1 1.2 Number Representation
More informationJos L.M. van Dorsselaer. February Abstract. Continuation methods are a well-known technique for computing several stationary
Computing eigenvalues occurring in continuation methods with the Jacobi-Davidson QZ method Jos L.M. van Dorsselaer February 1997 Abstract. Continuation methods are a well-known technique for computing
More informationPreconditioning Techniques Analysis for CG Method
Preconditioning Techniques Analysis for CG Method Huaguang Song Department of Computer Science University of California, Davis hso@ucdavis.edu Abstract Matrix computation issue for solve linear system
More informationContents. Preface... xi. Introduction...
Contents Preface... xi Introduction... xv Chapter 1. Computer Architectures... 1 1.1. Different types of parallelism... 1 1.1.1. Overlap, concurrency and parallelism... 1 1.1.2. Temporal and spatial parallelism
More informationarxiv: v1 [math.na] 1 Sep 2018
On the perturbation of an L -orthogonal projection Xuefeng Xu arxiv:18090000v1 [mathna] 1 Sep 018 September 5 018 Abstract The L -orthogonal projection is an important mathematical tool in scientific computing
More informationRening the submesh strategy in the two-level nite element method: application to the advection diusion equation
INTERNATIONAL JOURNAL FOR NUMERICAL METHODS IN FLUIDS Int. J. Numer. Meth. Fluids 2002; 39:161 187 (DOI: 10.1002/d.219) Rening the submesh strategy in the two-level nite element method: application to
More informationProperties of Matrices and Operations on Matrices
Properties of Matrices and Operations on Matrices A common data structure for statistical analysis is a rectangular array or matris. Rows represent individual observational units, or just observations,
More informationNumerical Methods - Numerical Linear Algebra
Numerical Methods - Numerical Linear Algebra Y. K. Goh Universiti Tunku Abdul Rahman 2013 Y. K. Goh (UTAR) Numerical Methods - Numerical Linear Algebra I 2013 1 / 62 Outline 1 Motivation 2 Solving Linear
More information