ANALYSIS OF ITERATIVE METHODS FOR SADDLE POINT PROBLEMS: A UNIFIED APPROACH

Size: px
Start display at page:

Download "ANALYSIS OF ITERATIVE METHODS FOR SADDLE POINT PROBLEMS: A UNIFIED APPROACH"

Transcription

1 MATHEMATICS OF COMPUTATION Volume 71, Number 38, Pages S )013- Article electronically published on May 1, 001 ANALYSIS OF ITERATIVE METHODS FOR SADDLE POINT PROBLEMS: A UNIFIED APPROACH WALTER ZULEHNER Abstract. In this paper two classes of iterative methods for saddle point problems are considered: inexact Uzawa algorithms and a class of methods with symmetric preconditioners. In both cases the iteration matrix can be transformed to a symmetric matrix by block diagonal matrices, a simple but essential observation which allows one to estimate the convergence rate of both classes by studying associated eigenvalue problems. The obtained estimates apply for a wider range of situations and are partially sharper than the known estimates in literature. A few numerical tests are given which confirm the sharpness of the estimates. 1. Introduction In this paper we consider systems of linear equations of the form ) A B T u f 1.1) =, B 0 p) g) where A is a symmetric, positive definite n n-matrix, B is a m n-matrix with full rank m n, andb T denotes the transposed matrix of B. Linear systems of the form 1.1) correspond to the Kuhn-Tucker conditions for linearly constrained quadratic programming problems or saddle point problems. Such systems typically result from mixed or hybrid finite element approximations of second-order elliptic problems, elasticity problems or the Stokes equations see e.g. Brezzi, Fortin [5]) and from Lagrange multiplier methods see e.g. Fortin, Glowinski [7]). Under the assumptions mentioned above the coefficient matrix ) A B T K = B 0 is nonsingular and the negative) Schur complement is symmetric and positive definite. C = BA 1 B T Received by the editor March 3, 1998 and, in revised form, February 11, 1999 and May 30, Mathematics Subject Classification. Primary 65N, 65F10. Key words and phrases. Indefinite systems, iterative methods, preconditioners, saddle point problems, Uzawa algorithm. 79 c 001 American Mathematical Society

2 80 WALTER ZULEHNER The system 1.1) can be reformulated in the following way: Au = f B T p, Cp = BA 1 f g. This gives rise to the following theoretical solution approach: Starting with the second equation p can be computed. Then the first equation allows the computation of u. For large scale problems, however, the exact solution of systems of the form Au = b and Cp = c is not possible, in general. At best only approximate solvers for this type of equations are available. The question now is how to construct an efficient method out of the two building blocks of approximate solvers for the systems Au = b and Cp = c. A classical method of this type is the Uzawa algorithm [1], which relies on an exact solver for Au = b and a Jacobi-like iteration for Cp = c. Several modifications have been suggested to avoid the exact solution of Au = b, reaching from Jacobilike iterations Arrow-Hurwics algorithm [1]) to multigrid methods see Verfürth [1]). Methods of this type can be summarized as inexact Uzawa algorithms, whose convergence properties have been investigated, e.g., by Queck [10], Elman and Golub [6] and more recently by Bramble, Pasciak and Vassilev []. A second class of methods has been introduced and analyzed by Bank, Welfert and Yserentant []. There, an additional correction step for u is suggested leading to a symmetric preconditioner of K. The obvious advantage of methods of this type, sometimes called segregated methods, is that they rely only on efficient solvers for separated problems for u and p, which are often quite well understood. They can be implemented very easily as combination of available algorithms. The aim of this paper is to present a convergence analysis for inexact Uzawa algorithms and for the class of methods with symmetric preconditioners on a common theoretical basis. The analysis has been strongly influenced by the theory in []. The main tool is summarized in Lemma 3.1 which gives lower and upper bounds for the eigenvalues of a generalized eigenvalue problem associated with the iteration matrices. The upper bounds yield bounds for the convergence rate of inexact Uzawa algorithms, which generalizes the results in [3], where only preconditioners of A were considered which underestimate A, and in [], where only preconditioners of A were considered which overestimate A. Some of the estimates in these papers are improved. The lower bounds can be used to estimate the convergence of the class of methods with symmetric preconditioners. This approach differs completely from the technique in []. The consequences drawn from this new approach sharpens some of the consequences of the theory presented in [], and special cases are identified which allow the application of an acceleration by a conjugate gradient technique for the total iteration, similar to the results in [3]. The paper is organized as follows. In Section the two classes of iterative methods are shortly reviewed and, for either class, a factorization of the iteration matrix is derived. Section 3 contains the scaling technique and the necessary eigenvalue estimates, which are applied to the two classes of iterative methods in the subsequent Sections and 5, respectively. Each of these two last sections is divided into three parts. The subsections on special cases present the results under assumptions which guarantee a spectrum of real numbers for the iteration matrices, the second part deals with the general case with full freedom with respect to the chosen norm, and, finally, the third part shows the implications if restrictions on the norms are to

3 ITERATIVE METHODS FOR SADDLE POINT PROBLEMS 81 be considered, which are typical in cases of variable preconditioners or acceleration techniques. Finally, Section 6 contains the results of a few numerical tests.. Basic iterative methods Two classes of iterative methods are considered here. For each class a product representation of the iteration matrix is shown..1. Inexact Uzawa algorithms. We start by considering the following class of methods: Âu k1 u k ) = f Au k B T p k, Ĉp k1 p k ) = Bu k1 g, where  and Ĉ are symmetric positive definite matrices. This includes the classical Uzawa algorithm  = A, Ĉ = γi), and the classical Arrow-Hurwicz algorithm  = αi, Ĉ = γi). The method can be seen as a preconditioned Richardson method ) ) uk1 u ˆK k f uk 1 = K p k1 p k g) p k with preconditioner  ) 0 ˆK 1 =. B Ĉ Let u,p ) be the exact solution of 1.1). For the error u k = u k u, p k = p k p we have ) ) uk1 uk = M p 1, k1 p k where M 1 denotes the iteration matrix of the inexact Uzawa algorithm, given by 1 M 1 = I ˆK 1 K =  1  A) )  1 B T Ĉ 1B 1  A) I Ĉ 1B 1 B T. It is easy to see that ) )  M 1 = 1  1 B T Ĉ 1 A  0.1) 1 Ĉ B 1 Ĉ 1B 1 B T Ĉ 1 Ĉ 1 0 Ĉ  1 ) 0 B T ) 1 A = Â) 0 ).) 0 Ĉ 1 B B 1 B T Ĉ 0 I Remark.1. If we reverse the order in dealing with the equations, the following iterative process is obtained: Ĉp k1 p k ) = Bu k g, Âu k1 u k ) = f Au k B T p k1. The method can be seen as a preconditioned Richardson method with preconditioner  ) B ˆK 1 T = T. 0 Ĉ

4 8 WALTER ZULEHNER It is easy to see that this iteration reduces to the previously discussed inexact Uzawa algorithm starting with u 0,p 1 ), where p 1 is given by Ĉp 1 p 0 )=Bu 0 g. So, iteration methods of this type are included in the discussion of inexact Uzawa methods... A class of symmetric preconditioners. The factorization ) ) ) A 0 A 1 0 A B T K = B I 0 C 0 I motivates the use of another preconditioner  ) 1 ) ) 0 0 B T ˆK = B I 0 Ĉ 0 I see []). This is equivalent to the iterative procedure Âû k1 u k ) = f Au k B T p k, Ĉp k1 p k ) = Bû k1 g, Âu k1 û k1 ) = B T p k1 p k ). This method can be viewed as an inexact Uzawa algorithm with an additional correction step for u. It can also be interpreted as a correction method in the following sense: In a first step a preliminary approximation û k1 is determined from Âû k1 u k )=f Au k B T p k in the same way as in the inexact Uzawa algorithm. In order to satisfy the equation Bu = g one uses the ansatz.3) Âu k1 û k1 )= B T p k1 p k ). Ideally speaking, this would lead to the correction equation B 1 B T δp k1 = Bû k1 g. The second step of the iterative method Ĉp k1 p k )=Bû k1 g can now be interpreted as approximating the solution of the correction equation starting from the initial approximation 0 for the correction δp k1.inthissenseĉ approximates the inexact Schur complement B 1 B T. Finally, the correction of u is done according to.3). Simple calculation shows that the iteration matrix is now given by 1 1 M = I ˆK K = ˆK ˆK K)  ) B T 1 ) A  0.) = B B 1 B T Ĉ 0 Ĉ B 1 B T. Summarizing both iteration methods, we have seen that M i = P i N i Q i

5 ITERATIVE METHODS FOR SADDLE POINT PROBLEMS 83 for i =1,, where P i and Q i are block diagonal matrices, and N i is a symmetric matrix. The matrices are given for i = 1 either by.1) or.), for i =by.). In the next section the spectrum of such a matrix is investigated. 3. Scaling and spectral estimates The results presented here apply to both classes of iteration methods. So, for a moment, we drop the subscripts and write M = PNQ with block diagonal matrices P, Q and a symmetric matrix N. Let D and E be two further block matrices, which are also symmetric and positive definite. Then [ M = D 1/ MD 1/ = D 1/ PE 1/][ E 1/ NE 1/][ E 1/ QD 1/] = P N Q. Here and in the sequel we use the following notation: For a real function f : R R and a symmetric matrix M the matrix fm) is defined in the usual way via the spectral decomposition of M. We discuss three different situations: Special cases: If D and E canbechosensuchthat 3.1) P = Q = I, then it follows that M = N, which implies that 1. M is symmetric with respect to the scalar product.,. D,givenby x, y D = Dx, y, where.,. denotes the ordinary Euclidean scalar product;. the spectrum of M and spectrum of N coincide: σm) =σ N ) here σ.) denotes the spectrum); and 3. the convergence rate of the iteration method is given by ρm) = M D = ρ N ). Here, ρ.) denotes the spectral radius and. D the norm associated to the scalar product.,. D. It is easy to see that condition 3.1) can be satisfied if P = I and Q is symmetric and positive definite. Then the transformation matrices are given by D = Q, E = Q 1. The above-mentioned conditions on P and Q correspond to the following special cases for the considered iteration methods: Special case for the inexact Uzawa algorithm, see.1): Â<A. Here and in the sequel, we write M < N, respectively M N, for symmetric matrices M and N if and only if N M is positive definite, respectively positive semidefinite. Special case for the iteration method with symmetric preconditioners: Â<A and Ĉ>BÂ 1 B T.

6 8 WALTER ZULEHNER There is a second special case for this method: If D and E are chosen such that 3.) P = I and Q = I, then M = N with similar implications as before. Condition 3.) can be satisfied if P = I and Q is symmetric and negative definite. Then the transformation matrices are given by This corresponds to the special case D = Q, E = Q 1. Â>A and Ĉ<BÂ 1 B T for the iteration method with symmetric preconditioners. In each of these special cases the convergence analysis reduces to the eigenvalue problem 3.3) N x = ν Ex, respectively to the eigenvalue problem N x = ν Ex. The general case: If one can find transformation matrices D and E with 3.) P 1, Q 1. denotes the spectral norm), then we can still conclude that Condition 3.) leads to So, a necessary condition for E is This is equivalent to 3.5) M D ρ N ). PEP T D 1 and QD 1 Q T E 1. QPEP T Q T E 1. E 1/ QPE 1/ I if QP is symmetric, which is the case for all applications in this paper. Now, if E is chosen such that condition 3.5) is satisfied and if we set D = PEP T ) 1, one easily sees that condition 3.) is satisfied, too. Restrictions on the norm: For some situations, like the analysis of iteration methods with variable preconditioners or nonlinear iterations, it is necessary to partially) prescribe the transformation matrix D. Then it is still possible to satisfy 3.): One easily sees that 3.) follows if E 1 P T DP and E 1 QD 1 Q T. This discussion shows the importance of the generalized eigenvalue problem 3.3) for the analysis of the iteration methods. The next lemma is fundamental for estimating the eigenvalues of 3.3).

7 ITERATIVE METHODS FOR SADDLE POINT PROBLEMS 85 Lemma 3.1. Let  and Q be symmetric, positive definite n n-matrices, B a m n matrix, Ĉ and S symmetric, positive definite m m-matrices. Assume that there are real numbers ρ 1 ρ 0 <ρ 3 ρ <ρ 5 ρ 6 with 3.6) ϕρ 1 ) 0, ϕρ ) 0, ϕρ 5 ) 0, ϕρ 6 ) 0, where ϕµ) =µbµ  ÂQ 1 Â) 1 B T µs Ĉ, and ρ 3 Q  ρ Q. Then we have: If λ is an eigenvalue of the generalized eigenvalue problem  Ĉ) ) ) B T u Qu 3.7) B B 1 B T = λ, p Sp then λ [ρ 1,ρ ] [ρ 3,ρ ] [ρ 5,ρ 6 ]. Proof. Let u, p) R n R m with u, p) 0, 0) and λ R satisfy Âu B T p = λqu, Bu B 1 B T Ĉ)p = λsp. Assume that λ [ρ 3,ρ ]. Then λq  is either positive or negative definite and p 0. In this case we obtain from the first equation u =λq Â) 1 B T p. Then the second equation yields BλQ Â) 1 B T p B 1 B T Ĉ)p = λsp, or, equivalently, λbλ  ÂQ 1 Â) 1 B T p =λs Ĉ)p. Multiplying this equation by p T from the left yields ϕλ)p, p =0 with ϕµ) =µbµ  ÂQ 1 Â) 1 B T µs Ĉ. It is easy to see that ϕµ)p, p is strictly decreasing in µ on each interval outside the set [ρ 3,ρ ]. From the assumptions 3.6) it follows that ϕρ 1 )p, p 0, ϕρ )p, p 0, ϕρ 5 )p, p 0, ϕρ 6 )p, p 0. Then the estimates are direct consequences of the monotony of ϕµ)p, p. Remark 3.. Additionally, one easily verifies that ϕ0)p, p < 0 and ϕ ρ 0 )p, p 0, where ρ 0 is a positive real number satisfying the condition Ĉ ρ 0 S.

8 86 WALTER ZULEHNER This shows that it is always possible to choose ρ, such that ρ < 0, and ρ 1,such that ρ 0 ρ 1. If, additionally, B has full rank m n, then ϕ ρ 0 )p, p > 0, i.e., ρ 1 can even be chosen, such that ρ 0 <ρ 1. Remark 3.3. For the special case  = A, Ĉ = BA 1 B T this lemma immediately yields the spectral estimates by Iliash, Rossi and Toivanen [9], provided spectral inequalities of the form and η 1 Q A η Q θ 1 S BA 1 B T θ S are assumed. Especially for  = Q = A and Ĉ = S = BA 1 B T we obtain the exact bounds ρ 1 = ρ =1 5)/, ρ 3 = ρ =1andρ 5 = ρ 6 =1 5)/, a simple but interesting observation due to Yu. Kuznetsov, Spectral estimates of this kind are used to analyze block diagonal preconditioners see e.g. Silvester and Wathen [11]). So, as a by-product of the convergence theory presented here, the known results on iteration methods with block diagonal preconditioners are reproduced. However, compared to literature, no new information is gained.. Convergence results for inexact Uzawa algorithms The analysis of inexact Uzawa algorithms is based on the setting ) A  0 P = I, Q = 0 Ĉ for the special case see.1)), and on the setting ) P =, Q =  Ĉ 1 for the general case see.)). First, we discuss the  1 A Â) 0 0 I.1. Special case: Â<A. From the discussion in Section 3 we obtain the following scaling matrices in this case: ) ) A  0 A Â) D =, E = Ĉ 0 Ĉ 1, leading to the norm. D, given by the scalar product u v, = A Â)u, v Ĉp,q. p) q).1) D The generalized eigenvalue problem 3.3) leads to a generalized eigenvalue problem of the form 3.7) with ).) λ = ν, Q = ÂA Â) 1 Â, S = Ĉ.

9 ITERATIVE METHODS FOR SADDLE POINT PROBLEMS 87 For the preconditioners  and Ĉ the following spectral inequalities are assumed: There is a constant α R with 1 <α, such that.3) Â<A α Â, and there are constants γ 1,γ R with 0 <γ 1 γ, such that.) γ 1 Ĉ BA 1 B T γ Ĉ. A simple consequence of Lemma 3.1 is Theorem.1. Let A,  be symmetric, positive definite n n-matrices, B a m n matrix, Ĉ a symmetric, positive definite m m matrix, satisfying.3) and.). Then we have: 1. The iteration matrix M 1 of the inexact Uzawa algorithm is symmetric with respect to the scalar product.1).. σm 1 ) [ρ 1,ρ ], 1) with ρ 1 = 1 γ )α [ 1 γ )α ] α 1, ρ = 1 γ 1)α [ 1 γ1 )α ] α If α γ ) <, thenρm 1 )= M 1 D < 1.. ˆK 1 1 K is symmetric and positive definite with respect to the scalar product 1.1) and σ ˆK 1 K) [1 ρ, 1 ρ 1 ] 0, ). Proof. With the notation of Section 3 we have: M1 = N 1,where N 1 is symmetric. This shows the symmetry of M 1 in the corresponding scalar product.1). Anumberν is an eigenvalue of N1 if and only if λ = ν is an eigenvalue of the generalized eigenvalue problem 3.7) with the setting.). The matrix function ϕµ) of Lemma 3.1 is given by ϕµ) =µb[µ 1) A] 1 B T µ 1)Ĉ. Assume that 1 µ 0. Then it follows from.) that ϕµ) 0if µ [µ 1) A] 1 µ 1 A 1, γ 1 or, equivalently, µγ 1 A µ 1)[µ 1) A]. Set Ā =  1/ A 1/. Then this inequality becomes which is satisfied if and only if.5) µγ 1 Ā µ 1)[µ 1)I Ā], µγ 1 ā µ 1)µ 1 ā) for all eigenvalues ā of Ā. From.3) we know that σā) 1,α ]. It is easy to see that.5) is satisfied for all ā 1,α ] if and only if it is satisfied for the extreme value ā = α : µγ 1 α µ 1)µ 1 α ),

10 88 WALTER ZULEHNER which is equivalent to µ µ 1,whereµ 1 is the negative root of the quadratic equation µγ 1 α =µ 1)µ 1 α ). One easily verifies that 1 <µ 1 0. Therefore, we have ϕµ 1 ) 0, which, by Lemma 3.1, implies the bound λ µ 1 for the eigenvalues of 3.7) and the bound ν = λ µ 1 = ρ for the eigenvalues of M 1. Observe that  α 1) Q. Then, for µ>α 1, it follows from.) that ϕµ) 0if µ [µ 1) A] 1 µ 1 A 1, γ or, equivalently, µγ A µ 1)[µ 1) A]. In the same way as before, it can be shown that this inequality is satisfied if µγ α µ 1)µ 1 α ), which is equivalent to µ µ,whereµ is the positive root of the quadratic equation µγ α =µ 1)µ 1 α ). One easily verifies that µ >α 1. Therefore, we have ϕµ ) 0, which, by Lemma 3.1, implies the bound λ µ for the eigenvalues of 3.7) and the bound ν = λ µ = ρ 1 for the eigenvalues of M 1. Statement 3 easily follows from, and is an immediate consequence of 1 and. Remark.. Statement of Theorem.1 justifies the use of a conjugate gradient acceleration for inexact Uzawa algorithms, which was the key observation in [3]. The last part of Theorem.1 provides a conditioning estimate: 1 ρ ) x, x D ˆK 1 1 Kx, x D 1 ρ 1 ) x, x D. Compared to the basic estimate in Theorem 1 in [3], which deals only with the special case Ĉ = BA 1 B T, the upper bounds agree, while the lower bound given here is sharper. Translating our results into the terminology of [3], Theorem 1 in [3] deals with the special case γ 1 = γ = 1 and uses the notations  = A 0, α =1/1 α) withα<1. Then, the lower bound in our theory becomes: 1 ρ = α α 1) α 1= 1 α 1 α. The coinciding upper bound in the notation of [3] reads: 1 ρ 1 = α α 1) α 1= 1 α 1 α. An example in [3] shows the sharpness of these bounds, in particular, the sharpness of the lower bound given here. As far as quantitative conditioning estimates are concerned, Theorem 1 in [3] covers only the case of using the exact Schur complement. In the general case, where the exact Schur complement is replaced by a preconditioner Ĉ one could, of course, derive conditioning estimates by comparing the overall preconditioning

11 ITERATIVE METHODS FOR SADDLE POINT PROBLEMS 89 matrix with the theoretic preconditioner of Theorem 1 in [3] see Remark in [3]). Generally speaking, this approach would lead to less sharp estimates than our approach, which directly provides conditioning estimates without the use of an intermediate theoretic preconditioner... The general case. Now we assume the following spectral inequalities for the preconditioner Â: There are constants α 1,α R with 0 <α 1 1 α, such that.6) α 1  A α Â. Additionally, we assume that at least one of the inequalities α 1 1orα 1  A and at least one of the inequalities 1 α or A α  hold strictly. Furthermore, we introduce q α =max1 α 1,α 1), which is an upper bound of the convergence rate I  1 A  of the Richardson methodforanequationoftheformau = b, preconditioned with Â. As before, for the preconditioner Ĉ the following spectral inequalities are assumed: There are constants γ 1,γ R with 0 <γ 1 γ, such that.7) γ 1 Ĉ BA 1 B T γ Ĉ. According to the discussion in Section 3 we look for a scaling matrix ) Eu 0 E =, 0 E p such that E 1/ QPE 1/ I, or, in details, Eu 1/  1 A Â) 1 Eu 1/ I and Ep 1/ Ĉ 1 Ep 1/ I. Roughly speaking, the larger E is with respect to the ordering of matrices, introduced in Section 3, the smaller are the eigenvalues in 3.3). In this sense E u and E p should be as large as possible. The optimal choice for E p is obviously given by E p = Ĉ. For E u we set.8) E u = Â1/fĀ)Â1/, where Ā =  1/A 1/ and f :0, ) 0, ] is an arbitrary function. Then the condition on E u reads: All eigenvalues λ of the generalized eigenvalue problem  1 A Â) 1 u = λeu 1 u or, equivalently, Ā I)u = λfâ) 1 u have to satisfy the condition: λ 1. From.6) we know: σā) [α 1,α ]. The eigenvalues λ result from the eigenvalues ā of Ā by the formula λ =ā 1)fā).

12 90 WALTER ZULEHNER The condition λ 1 is certainly satisfied if x 1 fx) 1 for all x [α 1,α ]. This condition holds for the function f, givenby fx) 1 =1 α 1 ) α x α 1) x α 1.9) = α 0 x ˆα α α 1 α α 1 with α 0 = α 1 α, ˆα = α 1 α α 1 α. α α 1 α α 1 Observe that x 1 fx) =1 forx = α i, except for the case α i =1,i =1,. So, f is as large as possible at least at the boundary points of the interval which contains the spectrum of Ā. The additional conditions on α 1 and α guarantee that E u is well defined and positive definite. The resulting matrix D is given by ) ˆα Â α D = 0 A 0 0 Ĉ with associated scalar product u v, p) q) Then a simple consequence of Lemma 3.1 is D = ) ˆα Â α 0 A u, v Ĉp,q. Theorem.3. Let A, Â be symmetric, positive definite n n-matrices, B a m n matrix, Ĉ a symmetric, positive definite m m matrix, satisfying.6) and.7). Then we have with the notations of Section and 3: 1. σ N 1 ) [min q α,ρ 1,ρ 3 ), maxρ,ρ )] with q α = max1 α 1,α 1), ρ 1 = 1 γ )α [ 1 γ )α ] α 1, ρ = 1 γ 1)α [ 1 γ1 )α ] α 1, ρ 3 = 1 γ )α 1 1 γ ) α 1 1 α 1, ρ = 1 γ 1)α 1 1 γ1 ) α 1 1 α 1.. M 1 D maxq α, ρ 1, ρ 3,ρ,ρ ). 3. If α γ ) <, then M 1 D < 1. Proof. With the notation of Section 3 we have M 1 D ρ N 1 ) and ν is an eigenvalue of N 1 if and only if λ = ν is an eigenvalue of the generalized eigenvalue problem 3.7) with the setting S = Ĉ, Q = E u see.8),.9)).

13 ITERATIVE METHODS FOR SADDLE POINT PROBLEMS 91 The matrix function ϕµ) of Lemma 3.1 is given by ϕµ) =µbâ 1/ µi fā) 1) 1 Â 1/ B T µ 1)Ĉ. Assume that 1 µ 0. Then it follows from.7) that ϕµ) 0if µ Â 1/ µi fā) 1) 1 Â 1/ µ 1 A 1, γ 1 or, equivalently, if µγ 1 Ā µ 1) µi fā) 1). This means in terms of the eigenvalues ā of Ā:.10) µγ 1 ā µ 1) µ fā) 1). Since σā) [α 1,α ], it suffices to fulfill.10) for all ā [α 1,α ]. It is easy to see that.10) is satisfied for all ā [α 1,α ], if and only if it is satisfied for the extreme value α 1 and α : µγ 1 α 1 µ 1)µ 1α 1 ), and µγ 1 α µ 1)µ α 1), which is equivalent to µ µ 11 and µ µ 1,whereµ 11 is the negative root of the quadratic equation µγ 1 α 1 =µ 1)µ 1α 1 ) and µ 1 is the negative root of the quadratic equation µγ 1 α =µ 1)µ α 1). One easily verifies that 1 <µ 11 0and 1 <µ 1 0. Therefore, we have ϕminµ 11,µ 1 )) 0, which, by Lemma 3.1, implies the bound λ minµ 11,µ 1 ) for the eigenvalues of 3.7) and the bound ν = λ maxρ,ρ ) for the eigenvalues of N1 with ρ = µ 1 and ρ = µ 11. Observe that q α Q Â. Then, for µ>q α, one obtains completely analogously to above that ϕµ) 0ifµ maxµ,µ 1 ), where µ 1 is the positive root of the quadratic equation µγ α 1 =µ 1)µ 1α 1 ) and µ is the positive root of the quadratic equation µγ α =µ 1)µ α 1). Therefore, we have λ maxq α,µ,µ 1 ) for the eigenvalues of 3.7) and ν = λ min q α,ρ 1,ρ 3 ) for the eigenvalues of N1 with ρ 1 = µ and ρ 3 = µ 1. The statements and 3 easily follow from 1. Remark.. This theorem contains the special case α 1 < 1, α =1,γ 1 < 1and γ = 1, discussed in []. For this case we have and ρ 1 =0, ρ =1 γ 1, ρ 3 = 1 α 1 maxq α, ρ 1, ρ 3,ρ,ρ )=ρ,

14 9 WALTER ZULEHNER which gives the estimate M 1 D 1 γ 1)α 1 1 γ1 ) α 1 1 α 1, or in terms of the notations α 1 =1 δ, γ 1 =1 γ used in []: γ1 δ) M 1 D This agrees with the results in []. γ 1 δ) δ..3. The general case with Â-independent norm. The norms introduced so far for analyzing the convergence depend on the preconditioner Â. If the preconditioner is allowed to change during the iteration, these norms change, too. A convergence analysis based on these changing norms is very complicated if not impossible. In this situation it is more advisable to use a norm which is independent of Â. The case of variable preconditioners for A also includes the case of using an inner iteration for solving an equation of the form Au = b, such as a conjugate gradientmethod. InLemma9in[]itisshownhowtorepresentsuchaninner iteration by a variable preconditioner.) Therefore we are now looking for estimates in a norm. D, given by the scalar product u v, = α p) q) 0 Au, v Ĉp,q, D with some scalar factor α 0 > 0. So, in the terminology of Section 3, we prescribe the scaling matrix D by ) α0 A 0 D = 0 Ĉ up to the factor α 0 > 0. Then, according to the discussion in Section 3, E must satisfy the inequalities ) E 1 α0 Â 1AÂ Ĉ 1 and E 1 If we set, as before, ) 1α0 Â 1 A Â)A 1 A Â)Â Ĉ 1 E = ) Eu 0 0 E p with E p = Ĉ and E u = Â1/fĀ)Â1/, these inequalities reduce to α 0 fā) Ā I and 1 fā)ā α Ā 1 I) I. 0 For this it suffices to choose f such that 1 fx) max α 0 x, x 1x )) 1 for all x [α 1,α ]. α 0

15 ITERATIVE METHODS FOR SADDLE POINT PROBLEMS 93 This condition is satisfied for the function f, givenby fx) 1 α x x α 1.11) = q 1 q, α α 1 α α 1 with 1 q 1 = max α 0 α 1, α 1 1 )), α 0 α 1 1 q = max α 0 α, α 1 )). α α 0 Observe that q 1 1 α 1 and q α 1. Then a simple consequence of Lemma 3.1 is Theorem.5. Let A, Â be symmetric, positive definite n n-matrices, B a m n matrix, Ĉ a symmetric, positive definite m m matrix, satisfying.6) and.7). Then we have with the notations of Section and 3: 1. σ N 1 ) [min q A, ρ 1, ρ 3 ), max ρ, ρ )] with q A = maxq 1,q ), ρ 1 = 1 q γ α [1 q γ α ] q, ρ = 1 q γ 1 α [1 q γ 1 α ] q, ρ 3 = 1 q 1 γ α 1 [1 q1 γ α 1 ] q 1, ρ = 1 q 1 γ 1 α 1 [1 q1 γ 1 α 1 ] q 1.. M 1 D max q A, ρ 1, ρ 3, ρ, ρ ). 3. If α γ ) < and 1 α 1 ) α <α 1 α 1 γ ) α γ ),then M 1 D < 1 for values of α 0 with 1 α1 ) max α 1 α 1 γ ), α 1) ) <α 0 < α γ. α α γ ) α Proof. The proof follows along the lines of the proof of the last theorem. The only difference is the definition of the function fx) and the quadratic equations. The upper bound ν max ρ, ρ ) for the eigenvalues of N 1 with ρ = µ 1 and ρ = µ 11 is given by the negative root µ 11 of the quadratic equation µγ 1 α 1 =µ 1)µ q 1 ) and the negative root µ 1 of the quadratic equation µγ 1 α =µ 1)µ q ). The lower bound ν min q A, ρ 1, ρ 3 ) for the eigenvalues of N1 with ρ 1 = µ 1 and ρ 3 = µ 1 is given by the positive root µ 1 of the quadratic equation µγ α 1 =µ 1)µ q 1 ), the positive root µ of the quadratic equation µγ α =µ 1)µ q ),

16 9 WALTER ZULEHNER and the constant q A,whichensures q A Q Â. Remark.6. In [] the special case.1) α 1 =1 δ, α =1δ, γ 1 =1 γ, γ =1 was studied. It could be shown in a very short and elegant way that M 1 D < 1if.13) δ< 1 γ 3 γ for the scaling factor α 0 = δ 1δ. From the theorem given here a convergence rate estimate of the form M 1 D < 1 already follows if.1) δ< 1 3 for scaling factors α 0 with δ 1 δ <α 0 < 1 δ 1 δ). Observe that Condition.1) is weaker than Condition.13). So, in this paper, convergence could be shown in an Â-independent norm under Condition.1), which is weaker than Condition.13) used in []. Furthermore, observe that Condition.1) coincides with the convergence condition 3 in Theorem.1 under the assumptions.1). This is remarkable, because the norm was allowed to depend on  in Theorem.1, while now an Â-independent norm was required, which is more restrictive. 5. Convergence results for symmetric preconditioners The analysis for iteration methods with symmetric preconditioners corresponds to the setting ) A  0 P = I, Q = 0 Ĉ B 1 B T. Again we start by discussing the 5.1. Special case: Â<Aand Ĉ>B 1 B T. In this case we use the scaling ) A  0 A Â) 1 ) 0 D = 0 Ĉ B 1 B T, E = ) 1 0 Ĉ B  1 B T, leading to the norm. D, given by the scalar product u v 5.1), = A p) q) Â)u, v Ĉ B 1 B T )p, q. The generalized eigenvalue problem 3.3) leads to a generalized eigenvalue problem of the form 3.7) with 5.) λ = 1 ν, Q = A Â, S = Ĉ B 1 B T.

17 ITERATIVE METHODS FOR SADDLE POINT PROBLEMS 95 For the preconditioners the following spectral inequalities are assumed. There is a constant α R with α > 1, such that 5.3) Â<A α Â, and there is a constant β 1 R with β 1 < 1, such that 5.) β 1 Ĉ B 1 B T < Ĉ. Then a simple consequence of Lemma 3.1 is Theorem 5.1. Let A,  be symmetric, positive definite n n-matrices, B a m n matrix, Ĉ a symmetric, positive definite m m matrix, satisfying 5.3) and 5.). Then we have: 1. The iteration matrix M of the iteration method with symmetric preconditioner is symmetric with respect to the scalar product 5.1).. σm ) [1 α,ρ ], 1) with ρ = α )1 β 1 ) α ) 1 β 1 ) α 1)1 β 1 ). 3. If α <, thenρm )= M 1 D < 1.. ˆK 1 K is symmetric and positive definite with respect to the scalar product 1 5.1) and σ ˆK K) [1 ρ,α ] 0, ). Proof. With the notation of Section 3 we have: M = N,where N is symmetric. This shows the symmetry of M 1 in the corresponding scalar product.1). We consider the eigenvalue problem 3.7) with the settings 5.). The matrix function ϕµ) of Lemma 3.1 has the form [ 1 ϕµ) =µb µ Â]  ÂA Â) 1 B T µbâ 1 B T µ 1)Ĉ. Assume that µ 1. Then it follows from 5.) that ϕµ) 0if [ ] 1 µ µ  ÂA Â) 1  µ  1 µ 1  1, β 1 or, equivalently, µβ 1  µ 1 µβ 1 )µâ ÂA Â) 1 Â). This means for the transformed matrix Ā =  1/A 1/ : µβ 1 I µ 1 µβ 1 )µi Ā I) 1 ). For µ 1/1 β 1 ), this inequality is satisfied if and only if µβ 1 µ 1 µβ 1 ) µ 1 ). α 1 The negative µ 1 root of the quadratic equation µβ 1 =[1µ 1 β 1 )] µ 1 ) α 1 is given by 1 µ 1 = α )1 β 1 ) α ) 1 β 1 ) α 1)1 β 1 )

18 96 WALTER ZULEHNER and lies in the interval 1/1 β 1 ), 1). Therefore, ϕµ 1 ) 0. Then, by Lemma 3.1, it follows that λ µ 1 < 1 for negative eigenvalues λ and ν = 1/λ ρ with ρ = 1/µ 1 for the eigenvalues ν of N. For estimating positive eigenvalues λ from below by some value ρ we need the condition ρa Â) Â or, equivalently, ρ Ā I) I, which is satisfied if and only if 1 ρ α 1. So, λ 1/α 1) for positive eigenvalues, and, consequently, ν = 1/λ 1 α for the eigenvalues ν of N. Statement 3 easily follows from, and is an immediate consequence of 1 and. The second special case can be analyzed completely analogously: 5.. Special case: Â>Aand Ĉ<BÂ 1 B T. In this case we use the norm. D, given by the scalar product ) ) u v 5.5), = Â p q A)u, v BÂ 1 B T Ĉ)p, q. For the preconditioners the following spectral inequalities are assumed. There is a constant α 1 R with α 1 < 1, such that 5.6) α 1 Â A<Â, and there is a constant β R with β > 1, such that 5.7) Ĉ<BÂ 1 B T β Ĉ. Then a simple consequence of Lemma 3.1 is Theorem 5.. Let A, Â be symmetric, positive definite n n-matrices, B a m n matrix, Ĉ a symmetric, positive definite m m matrix, satisfying 5.6) and 5.7). Then we have: 1. The iteration matrix M of the iteration method with symmetric preconditioner is symmetric with respect to the scalar product 5.5).. σm ) [ ρ 1, 1 α 1 ], 1) with ρ 1 = α 1)β 1) α1 ) β 1) 1 α 1 )β 1). 3. If β < 11/3 α 1 ),thenρm )= M D < 1.. ˆK 1 K is symmetric and positive definite with respect to the scalar product 1 5.5) and σ ˆK K) [α 1, 1ρ 1 ] 0, ). Proof. The proof, which is completely analogous to the proof of the previous theorem, is omitted.

19 ITERATIVE METHODS FOR SADDLE POINT PROBLEMS The general case. For the preconditioner  the following spectral inequalities are assumed: There are constants α 1,α R with 0 <α 1 1 α,such that 5.8) α 1  A α Â. Additionally, it is assumed that α 1 <α. As before, we introduce q α =max1 α 1,α 1). For the preconditioner Ĉ we assume that there are constants β 1,β R with 0 <β 1 1 β, such that 5.9) β 1 Ĉ B 1 B T β Ĉ. Additionally, we assume that at least one of the inequalities β 1 1orβ 1 Ĉ B 1 B T and at least one of the inequalities 1 β or B 1 B T β Ĉ hold strictly. According to the discussion in Section 3 we are looking for a scaling matrix ) Eu 0 E =, 0 E p such that E 1/ QPE 1/ I, or, in details Eu 1/ A Â)E1/ u I and E1/ p Ĉ B 1 B T )Ep 1/ I. Roughly speaking, the larger E is with respect to the ordering of matrices, introduced in Section 3, the smaller are the eigenvalues in 3.3). In this sense E u and E p should be as large as possible. For E u we set 5.10) E u = 1  1. q α It is easy to see that E u satisfies the required estimate. For E p we set 5.11) E p = Ĉ 1/ g C)Ĉ 1/, where C = Ĉ 1/B 1 Ĉ 1/ and g :0, ) 0, ] is an arbitrary function. Then the condition on E p reads: All eigenvalues λ of the generalized eigenvalue problem Ĉ B 1 B T )p = λep 1 p or, equivalently, I C)p 1 = λg C) have to satisfy the condition: λ 1. From 5.9) we know σ C) [β 1,β ]. The eigenvalues λ result from the eigenvalues c of C by the formula λ =1 c)g c). The condition λ 1 is certainly satisfied if x 1 gx) 1 for all x [β 1,β ].

20 98 WALTER ZULEHNER This condition holds for the affine function gx) 1 =1 β 1 ) β x β 1) x β 1 5.1) = β 0 x β β 1 β β ˆβ 1 with β 0 = β 1 β β 1 β β 1 β, ˆβ =. β β 1 β β 1 Observe that x 1 gx) =1 forx = β i, for i =1,, except for the case β i =1. So,g is as large as possible at least at the boundary points of the interval which contains the spectrum of C. The resulting matrix D is given by ) qα Â 0 D = 0 ˆβ Ĉ β 0 BÂ 1 B T with associated scalar product u v, = q p) q) α Âu, ) v ˆβ Ĉ β 0 BÂ 1 B T p, q. D With these settings the generalized eigenvalue problem 3.3) leads to a generalized eigenvalue problem of the form 3.7) with 5.13) λ = 1 ν, Q = q α Â, S = ˆβ Ĉ β 0 BÂ 1 B T. Then a simple consequence of Lemma 3.1 is Theorem 5.3. Let A, Â be symmetric, positive definite n n-matrices, B a m n matrix, Ĉ a symmetric, positive definite m m matrix, satisfying 5.8) and 5.9). Then we have with the notations of Sections and 3: 1. σ N ) [ q α, maxρ 1,ρ )] with q α = max1 α 1,α 1), ρ 1 = 1 q α)1 β 1 ) 1 qα ) 1 β 1 ) q α 1 β 1 ), ρ = 1 q α)β 1) 1 qα ) β 1) q α β 1).. M D maxq α,ρ 1,ρ ). 3. If α < and β < 11/1 q α ),then M D < 1. Proof. Consider the eigenvalue problem 3.7) with the settings 5.13). The matrix function ϕµ) of Lemma 3.1 has the form ) µ ϕµ) = µβ 0 µ 1/q BÂ 1 B T µ ˆβ 1)Ĉ. α It is clear that ϕµ) 0 if and only if ) µ µβ 0 C µ µ 1/q ˆβ 1)I 0. α

21 ITERATIVE METHODS FOR SADDLE POINT PROBLEMS 99 From 5.9) we know σ C) [β 1,β ]. So it suffices to have ) µ µβ 0 β 1 µ µ 1/q ˆβ 1) 0 α and ) µ µβ 0 β µ µ 1/q ˆβ 1) 0. α For µ 0 this leads to µβ 1 [1 µ 1 β 1 )] µ 1 ) and µβ [1 µ β 1)] µ 1 ). q α q α This means µ maxµ 1,µ ), where µ 1 equation given by 1 µ 1 µβ 1 =[1µ 1 β 1 )] µ 1 ), q α = 1 q α)1 β 1 ) is the negative root of the quadratic 1 qα ) 1 β 1 ) q α 1 β 1 ), and µ is the negative root of the quadratic equation µβ =[1µ β 1)] µ 1 ), q α given by 1 µ = 1 q α)β 1) 1 qα ) β 1) q α β 1). 1 So, from Lemma 3.1 we obtain the upper bound λ maxµ 1,µ ) for negative eigenvalues λ. Then, for eigenvalues ν of N, the upper bound ν maxρ 1,ρ )withρ 1 = 1/µ 1 and ρ = 1/µ follows. From Q = q α Â we obtain the lower bound λ 1/q α for positive eigenvalues λ, which gives the lower bound ν q α for the eigenvalues of N. The statements and 3 follow directly from 1. Remark 5.. In [] the special case α 1 =1 α, α =1α, β 1 =1 β, β =1β was studied. In this case Theorem 5.3 gives the bound ) 1 α)β 1 α) β M D max α, αβ, from which convergence follows for α<1andβ<1/1 α). Compare this to the estimate in a different norm) ) β M max α,, 1 β given in [], from which convergence only follows under the more restrictive conditions α<1andβ<1/3.

22 500 WALTER ZULEHNER Our estimate can also be written in the form α if α β M D 1 β, 1 α)β 1 α) β αβ if α< β 1 β. It is easy to see that 1 α)β 1 α) β αβ < β if α< β 1 β 1 β, from which we obtain the weaker estimate ) β M D max α,, 1 β comparable with the estimate in []. A second special case, namely α 1 =1 α, α =1α, β 1 =1 β, β =1, was discussed in []. Here Theorem 5.3 gives the bound ) 1 α)β 1 α) β M D max α, αβ. This estimate can also be written in the form α if α β M D 1β, 1 α)β 1 α) β αβ if α< β 1β. Since 1 α)β 1 α) β αβ < β if α< β 1β 1β, we obtain the weaker estimate ) β M D max α,, 1β comparable with the estimate in []. An example in [] shows the sharpness of the bounds in Theorem The general case with Ĉ-independent norm. The norms introduced so far for analyzing the convergence depend on the preconditioner Ĉ. If the preconditioner for BÂ 1 B T is allowed to change during the iteration, it is advisable to use a norm which is independent of Ĉ. This includes the case of using an inner iteration for solving an equation of the form BÂ 1 B T p = c, such as a conjugate gradient method. We again refer to Lemma 9 in [].) Therefore we are now looking for estimates in a norm. D, given by the scalar product u v, = q p) q) α Âu, v β 0 BÂ1 B T p, q D

23 ITERATIVE METHODS FOR SADDLE POINT PROBLEMS 501 with some scalar factor β 0 > 0. So, in the terminology of Section 3, we prescribe the scaling matrix D by ) qα Â 0 D = 0 β 0 BÂ 1 B T up to the scalar factor β 0. Here we choose ) qα Â 0 E = 0 Ĉ 1/ g C)Ĉ 1/ with 5.1) and gx) 1 = s 1 β x β β 1 s x β 1 β β 1 s 1 = max β 0 β 1, 1β0 β 1 1 )), β 1 s = max β 0 β, 1β0 β 1 )), β and obtain Theorem 5.5. Let A, Â be symmetric, positive definite n n-matrices, B a m n matrix, Ĉ a symmetric, positive definite m m matrix, satisfying 5.8) and 5.9). Then we have with the notations of Sections and 3: 1. σ N ) [ q α, max ρ 1, ρ )] with q α = max1 α 1,α 1), ρ 1 = s 1 q α 1 β 1 ) [s1 q α 1 β 1 )] q α s 1, ρ = s q α β 1) [s q α β 1)] q α s.. M D maxq α, ρ 1, ρ ). 3. If α < and β < 11/1 q α ) and 1 β 1 ) β 1 q α ) <β 1 [1 q α 1 β 1 )][1 q α β 1)], then M D < 1 for values of β 0 with 1 β1 ) 1 q α ) max β 1 [1 q α 1 β 1 )], β 1) ) 1 q α ) β [1 q α β 1)] <β 0 < 1 q αβ 1). β 1 q α ) Proof. The proof follows along the lines of the proof of Theorem.5 and, therefore, is omitted. Remark 5.6. In [] the special case α 1 =1 α, α =1α, β 1 =1 β, β =1β

24 50 WALTER ZULEHNER was studied. In this case Theorem 5.5 gives the bound M D maxq α, ρ 1, ρ ) with q α = α, ρ 1 = s 1 αβ s1 αβ) αs 1, ρ = s αβ s αβ) αs. It turns out that there is a parameter β 0 with maxq α, ρ 1, ρ )=α if and only if α β/1 β). Consider the more interesting case α < β/1 β). For parameters β 0 [β/1 β),β/1 β)] we have: s 1 = 1 β β 0 1 β and s = β 0 1 β). It is easy to see that ρ 1 < ρ for β 0 = β 1β and ρ < ρ 1 < β 1 β for β 0 = β 1 β. From this discussion we can conclude that ) β M D max α,, 1 β if we choose the parameter β 0 =β/1 β). Then convergence follows if α<1and β<1/3. This corresponds up to the norm) to the estimates in []. A sharper estimate is obtained if β 0 [β/1 β),β/1 β)] is chosen such that ρ 1 = ρ. The existence of such a parameter follows from the discussion above the expression ρ ρ 1 changes the sign from β 0 =β/1 β) toβ 0 =β/1 β). Then, according to Theorem 5.5, the convergence conditions read α<1 and β 1 β)1 α) < 1 β)1 α β ). Sufficient for these conditions are α<1andβ<1/3. For small α the condition on β becomes less restrictive. So, e.g., for the limiting case α =0,β must be smaller thanthepositiverootofthecubicequationx 3 x x = 1, or, approximately, β< Numerical experiments We consider the P -P 0 finite element approximation of the Stokes problem u p = f in Ω, u =0 in Ω, u = g on Ω, pdx=0 Ω

25 ITERATIVE METHODS FOR SADDLE POINT PROBLEMS 503 Table 1. Spectral constants level α 0 1 α 0 β1 0 β 0 γ1 0 γ in a bounded domain Ω R. Here, u denotes the velocity and p the pressure of a fluid; g has to satisfy the compatibility condition g ds =0. Ω In particular, we consider the classical driven cavity problem on the unit square, where a unit tangential velocity is prescribed at the top of the square and 0 elsewhere) and f =0. The square is triangulated using a uniform mesh. The level 1 mesh consists of two triangles by connecting the lower left with the upper right corner of the square. The finest mesh considered is the level 8 mesh, obtained by successive uniform refinement, which consists of 3768 triangles. This corresponds to the following numbers of unknowns: n = and m = 3767 in the notation of Section 1. As an a priori preconditioner A 0 for A we chose the classical V-cycle multigrid method with one forward Gauss-Seidel pre-smoothing step and one backward Gauss-Seidel post-smoothing step. As an a priori preconditioner C 0 for the Schur complement BA 1 B T we chose the identity premultiplied by the element area. By the Lanczos method the following spectral inequalities were determined numerically: α 0 1 A 0 A α 0 A 0 and β 0 1 C 0 BA 1 0 BT β 0 C 0 and γ 0 1 C 0 BA 1 B T γ 0 C 0 with α 0 1, α0, β0 1, β0, γ0 1 and γ0 given in Table 1. From these estimates the preconditioner  is defined by This guarantees that  =0.9 α 0 1 A 0. Â<A α  with α = α 0 /0.9 α0 1 ). The cg-accelerated inexact Uzawa algorithm was performed with the preconditioners  and C 0. For the cg-accelerated iteration method with symmetric preconditioner an additional scaling is required for the Schur-complement preconditioner: Ĉ is defined by β 0 Ĉ = α 0 C 0. 1

26 50 WALTER ZULEHNER Table. cg-accelerated iteration methods Unknowns Inexact Uzawa Symm. precond. level n m N ρ ρ th N ρ ρ th This guarantees that β 1 Ĉ B 1 B T < Ĉ with β 1 = β1/1.1 0 β). 0 The cg-method applied to the matrix ˆK 1 K with the scalar product.,. D was used, where K, ˆK and D were chosen according to Subsections.1 and 5.1, with initial guesses u =0andp = 0 and stopping rule e k = ˆK 1 d k D ε ˆK 1 d 0 D = εe 0 with ε =10 8. Here d k denotes the defect of the k-th iterate with respect to original equation 1.1). This corresponds to the classical cg-method applied to the matrix A = D ˆK 1 K with the Euclidean scalar product and the preconditioner  = D. The norm in the stopping rule is the A 1 A-norm. It is easy to see that the evaluation of  and Ĉ is not necessary for performing the cg-method. Only the action of  1 and Ĉ 1 to a known vector is required. Table shows the results of the numerical experiments. For each mesh from level to level 8 it contains the numbers n and m of unknowns, the iteration numbers N and the averaged convergence factor ρ =e k /e 0 ) 1/k in comparison with an upper bound of) the theoretical convergence factor ρ th =[c k /1c k )] 1/k in the A-norm, where c = κ 1)/ κ 1) see, e.g., Hackbusch [8]). Here κ is the upper bound of the condition number of ˆK 1 K in the D-norm, given in Theorems.1 and 5.1, respectively. The numerical tests show a good agreement between the actual and the theoretical convergence rates, a consequence of the sharpness of the estimates. The convergence rates for the iteration method with symmetric preconditioner seem to indicate, as expected, the superiority of the cg-accelerated version given here to the tested method in [], where a cg-acceleration was used only for the inner iteration of the p-equation. The iteration numbers for the inexact Uzawa algorithm and the iteration method with symmetric preconditioner are approximately the same. However, the computational costs for the iteration method with symmetric preconditioner is about twice as high as for the inexact Uzawa method. Of course, all these comparisons must be considered with some caution. We are comparing with respect to different norms. Nevertheless, they seem to confirm the theoretical results. Concluding remarks. It has been shown in this paper that several interesting classes of iteration methods for saddle point problems can be analyzed by the following general strategy: Transform the iteration matrix to a symmetric matrix,

27 ITERATIVE METHODS FOR SADDLE POINT PROBLEMS 505 perform an appropriate scaling and estimate the eigenvalues by Lemma 3.1. There is no problem in using this technique under assumptions different from the assumptions considered here. For example, one could analyze inexact Uzawa algorithms on the basis of spectral inequalities of the form 5.) involving the inexact Schur complement or iteration methods with symmetric preconditioners on the basis of spectral inequalities of the form.) involving the exact Schur complement. It is just a matter of patience and time and it requires only the solution of quadratic equations. It might be also interesting that new cases have been identified, for which it can be shown that the spectrum of the iteration matrix is real. This widens the possibilities of accelerated methods. Acknowledgments. The author is very grateful to Joachim Schöberl, Linz, for providing him with the numerical results. The numerical tests were performed with FEPP, a C-coded finite element package by Joachim Schöberl. References 1. K. Arrow, L. Hurwicz, and H. Uzawa, Studies in nonlinear programming, Stanford University Press, Stanford, CA, MR 1:7115. R.E.Bank,B.D.Welfert,andH.Yserentant,A class of iterative methods for solving saddle point problems, Numer. Math ), MR 91b: J. H. Bramble and J. E. Pasciak, A preconditioning technique for indefinite systems resulting from mixed approximations of elliptic problems, Math. Comp ), MR 89m:65097a. J. H. Bramble, J. E. Pasciak, and A. T. Vassilev, Analysis of the inexact Uzawa algorithm for saddle point problems, SIAM J. Numer. Anal ), MR 98c: F. Brezzi and M. Fortin, Mixed and hybrid finite element methods, Springer-Verlag, MR 9d: H. C. Elman and G. H. Golub, Inexact and preconditioned Uzawa algorithms for saddle point problems, SIAM J. Numer. Anal ), MR 95f: M. Fortin and R. Glowinski, Augmented Lagrangian methods: Applications to the numerical solution of boundary value problems, North Holland, Amsterdam, MR 85a: W. Hackbusch, Iterative solutions of large sparse systems of equations, Springer Verlag, New York, 199. MR 9k: J. Iliash, T. Rossi, and J. Toivanen, Two iterative methods for solving the Stokes problem, Tech.Report,UniversityofJyväskylä, Department of Mathematics, Laboratory of Scientific Computing, W. Queck, The convergence factor of preconditioned algorithms of the Arrow-Hurwicz type, SIAM J. Numer. Anal ), MR 90m: D. Silvester and A. Wathen, Fast iterative solutions of stabilized Stokes systems. Part II: Using general block preconditioners, SIAMJ. Numer. Anal ), MR 95g: R. Verfürth, A combined conjugate gradient multigrid algorithm for the numerical solution fo the Stokes problem, IMA J. Numer. Anal. 198), MR 86f:6500 Institute of Analysis and Computational Mathematics, Johannes Kepler University, A-00 Linz, Austria address: zulehner@numa.uni-linz.ac.at

A Review of Preconditioning Techniques for Steady Incompressible Flow

A Review of Preconditioning Techniques for Steady Incompressible Flow Zeist 2009 p. 1/43 A Review of Preconditioning Techniques for Steady Incompressible Flow David Silvester School of Mathematics University of Manchester Zeist 2009 p. 2/43 PDEs Review : 1984 2005 Update

More information

Schur Complements on Hilbert Spaces and Saddle Point Systems

Schur Complements on Hilbert Spaces and Saddle Point Systems Schur Complements on Hilbert Spaces and Saddle Point Systems Constantin Bacuta Mathematical Sciences, University of Delaware, 5 Ewing Hall 976 Abstract For any continuous bilinear form defined on a pair

More information

SEMI-CONVERGENCE ANALYSIS OF THE INEXACT UZAWA METHOD FOR SINGULAR SADDLE POINT PROBLEMS

SEMI-CONVERGENCE ANALYSIS OF THE INEXACT UZAWA METHOD FOR SINGULAR SADDLE POINT PROBLEMS REVISTA DE LA UNIÓN MATEMÁTICA ARGENTINA Vol. 53, No. 1, 2012, 61 70 SEMI-CONVERGENCE ANALYSIS OF THE INEXACT UZAWA METHOD FOR SINGULAR SADDLE POINT PROBLEMS JIAN-LEI LI AND TING-ZHU HUANG Abstract. Recently,

More information

Numerische Mathematik

Numerische Mathematik Numer. Math. (00) 93: 333 359 Digital Object Identifier (DOI) 10.1007/s00110100386 Numerische Mathematik Two new variants of nonlinear inexact Uzawa algorithms for saddle-point problems Qiya Hu 1,, Jun

More information

Key words. preconditioned conjugate gradient method, saddle point problems, optimal control of PDEs, control and state constraints, multigrid method

Key words. preconditioned conjugate gradient method, saddle point problems, optimal control of PDEs, control and state constraints, multigrid method PRECONDITIONED CONJUGATE GRADIENT METHOD FOR OPTIMAL CONTROL PROBLEMS WITH CONTROL AND STATE CONSTRAINTS ROLAND HERZOG AND EKKEHARD SACHS Abstract. Optimality systems and their linearizations arising in

More information

Fast Iterative Solution of Saddle Point Problems

Fast Iterative Solution of Saddle Point Problems Michele Benzi Department of Mathematics and Computer Science Emory University Atlanta, GA Acknowledgments NSF (Computational Mathematics) Maxim Olshanskii (Mech-Math, Moscow State U.) Zhen Wang (PhD student,

More information

Multilevel Preconditioning of Graph-Laplacians: Polynomial Approximation of the Pivot Blocks Inverses

Multilevel Preconditioning of Graph-Laplacians: Polynomial Approximation of the Pivot Blocks Inverses Multilevel Preconditioning of Graph-Laplacians: Polynomial Approximation of the Pivot Blocks Inverses P. Boyanova 1, I. Georgiev 34, S. Margenov, L. Zikatanov 5 1 Uppsala University, Box 337, 751 05 Uppsala,

More information

On preconditioned Uzawa-type iterations for a saddle point problem with inequality constraints

On preconditioned Uzawa-type iterations for a saddle point problem with inequality constraints On preconditioned Uzawa-type iterations for a saddle point problem with inequality constraints Carsten Gräser and Ralf Kornhuber FU Berlin, FB Mathematik und Informatik (http://www.math.fu-berlin.de/rd/we-02/numerik/)

More information

Journal of Computational and Applied Mathematics. Multigrid method for solving convection-diffusion problems with dominant convection

Journal of Computational and Applied Mathematics. Multigrid method for solving convection-diffusion problems with dominant convection Journal of Computational and Applied Mathematics 226 (2009) 77 83 Contents lists available at ScienceDirect Journal of Computational and Applied Mathematics journal homepage: www.elsevier.com/locate/cam

More information

7.4 The Saddle Point Stokes Problem

7.4 The Saddle Point Stokes Problem 346 CHAPTER 7. APPLIED FOURIER ANALYSIS 7.4 The Saddle Point Stokes Problem So far the matrix C has been diagonal no trouble to invert. This section jumps to a fluid flow problem that is still linear (simpler

More information

ETNA Kent State University

ETNA Kent State University Electronic Transactions on Numerical Analysis. Volume 15, pp. 94-105, 2003. Copyright 2003,. ISSN 1068-9613. ETNA MULTILEVEL PRECONDITIONERS FOR LAGRANGE MULTIPLIERS IN DOMAIN IMBEDDING JANNE MARTIKAINEN,

More information

Chebyshev semi-iteration in Preconditioning

Chebyshev semi-iteration in Preconditioning Report no. 08/14 Chebyshev semi-iteration in Preconditioning Andrew J. Wathen Oxford University Computing Laboratory Tyrone Rees Oxford University Computing Laboratory Dedicated to Victor Pereyra on his

More information

The semi-convergence of GSI method for singular saddle point problems

The semi-convergence of GSI method for singular saddle point problems Bull. Math. Soc. Sci. Math. Roumanie Tome 57(05 No., 04, 93 00 The semi-convergence of GSI method for singular saddle point problems by Shu-Xin Miao Abstract Recently, Miao Wang considered the GSI method

More information

A STOKES INTERFACE PROBLEM: STABILITY, FINITE ELEMENT ANALYSIS AND A ROBUST SOLVER

A STOKES INTERFACE PROBLEM: STABILITY, FINITE ELEMENT ANALYSIS AND A ROBUST SOLVER European Congress on Computational Methods in Applied Sciences and Engineering ECCOMAS 2004 P. Neittaanmäki, T. Rossi, K. Majava, and O. Pironneau (eds.) O. Nevanlinna and R. Rannacher (assoc. eds.) Jyväskylä,

More information

Combination Preconditioning of saddle-point systems for positive definiteness

Combination Preconditioning of saddle-point systems for positive definiteness Combination Preconditioning of saddle-point systems for positive definiteness Andy Wathen Oxford University, UK joint work with Jen Pestana Eindhoven, 2012 p.1/30 compute iterates with residuals Krylov

More information

Multigrid and Iterative Strategies for Optimal Control Problems

Multigrid and Iterative Strategies for Optimal Control Problems Multigrid and Iterative Strategies for Optimal Control Problems John Pearson 1, Stefan Takacs 1 1 Mathematical Institute, 24 29 St. Giles, Oxford, OX1 3LB e-mail: john.pearson@worc.ox.ac.uk, takacs@maths.ox.ac.uk

More information

Structured Preconditioners for Saddle Point Problems

Structured Preconditioners for Saddle Point Problems Structured Preconditioners for Saddle Point Problems V. Simoncini Dipartimento di Matematica Università di Bologna valeria@dm.unibo.it p. 1 Collaborators on this project Mario Arioli, RAL, UK Michele Benzi,

More information

Efficient smoothers for all-at-once multigrid methods for Poisson and Stokes control problems

Efficient smoothers for all-at-once multigrid methods for Poisson and Stokes control problems Efficient smoothers for all-at-once multigrid methods for Poisson and Stoes control problems Stefan Taacs stefan.taacs@numa.uni-linz.ac.at, WWW home page: http://www.numa.uni-linz.ac.at/~stefant/j3362/

More information

Constrained Minimization and Multigrid

Constrained Minimization and Multigrid Constrained Minimization and Multigrid C. Gräser (FU Berlin), R. Kornhuber (FU Berlin), and O. Sander (FU Berlin) Workshop on PDE Constrained Optimization Hamburg, March 27-29, 2008 Matheon Outline Successive

More information

The antitriangular factorisation of saddle point matrices

The antitriangular factorisation of saddle point matrices The antitriangular factorisation of saddle point matrices J. Pestana and A. J. Wathen August 29, 2013 Abstract Mastronardi and Van Dooren [this journal, 34 (2013) pp. 173 196] recently introduced the block

More information

Efficient Augmented Lagrangian-type Preconditioning for the Oseen Problem using Grad-Div Stabilization

Efficient Augmented Lagrangian-type Preconditioning for the Oseen Problem using Grad-Div Stabilization Efficient Augmented Lagrangian-type Preconditioning for the Oseen Problem using Grad-Div Stabilization Timo Heister, Texas A&M University 2013-02-28 SIAM CSE 2 Setting Stationary, incompressible flow problems

More information

Numerical behavior of inexact linear solvers

Numerical behavior of inexact linear solvers Numerical behavior of inexact linear solvers Miro Rozložník joint results with Zhong-zhi Bai and Pavel Jiránek Institute of Computer Science, Czech Academy of Sciences, Prague, Czech Republic The fourth

More information

Convergence Properties of Preconditioned Hermitian and Skew-Hermitian Splitting Methods for Non-Hermitian Positive Semidefinite Matrices

Convergence Properties of Preconditioned Hermitian and Skew-Hermitian Splitting Methods for Non-Hermitian Positive Semidefinite Matrices Convergence Properties of Preconditioned Hermitian and Skew-Hermitian Splitting Methods for Non-Hermitian Positive Semidefinite Matrices Zhong-Zhi Bai 1 Department of Mathematics, Fudan University Shanghai

More information

Preconditioners for reduced saddle point systems arising in elliptic PDE-constrained optimization problems

Preconditioners for reduced saddle point systems arising in elliptic PDE-constrained optimization problems Zeng et al. Journal of Inequalities and Applications 205 205:355 DOI 0.86/s3660-05-0879-x RESEARCH Open Access Preconditioners for reduced saddle point systems arising in elliptic PDE-constrained optimization

More information

ON A GENERAL CLASS OF PRECONDITIONERS FOR NONSYMMETRIC GENERALIZED SADDLE POINT PROBLEMS

ON A GENERAL CLASS OF PRECONDITIONERS FOR NONSYMMETRIC GENERALIZED SADDLE POINT PROBLEMS U..B. Sci. Bull., Series A, Vol. 78, Iss. 4, 06 ISSN 3-707 ON A GENERAL CLASS OF RECONDIIONERS FOR NONSYMMERIC GENERALIZED SADDLE OIN ROBLE Fatemeh anjeh Ali BEIK his paper deals with applying a class

More information

Multigrid Based Optimal Shape and Topology Design in Magnetostatics

Multigrid Based Optimal Shape and Topology Design in Magnetostatics Multigrid Based Optimal Shape and Topology Design in Magnetostatics Dalibor Lukáš 1 Department of Applied Mathematics, VŠB Technical University of Ostrava, 17. listopadu 15, 708 33 Ostrava Poruba, Czech

More information

Indefinite Preconditioners for PDE-constrained optimization problems. V. Simoncini

Indefinite Preconditioners for PDE-constrained optimization problems. V. Simoncini Indefinite Preconditioners for PDE-constrained optimization problems V. Simoncini Dipartimento di Matematica, Università di Bologna, Italy valeria.simoncini@unibo.it Partly joint work with Debora Sesana,

More information

Adaptive methods for control problems with finite-dimensional control space

Adaptive methods for control problems with finite-dimensional control space Adaptive methods for control problems with finite-dimensional control space Saheed Akindeinde and Daniel Wachsmuth Johann Radon Institute for Computational and Applied Mathematics (RICAM) Austrian Academy

More information

Mathematics and Computer Science

Mathematics and Computer Science Technical Report TR-2007-002 Block preconditioning for saddle point systems with indefinite (1,1) block by Michele Benzi, Jia Liu Mathematics and Computer Science EMORY UNIVERSITY International Journal

More information

On convergence rate of the augmented Lagrangian algorithm for nonsymmetric saddle point problems

On convergence rate of the augmented Lagrangian algorithm for nonsymmetric saddle point problems Applied Numerical athematics 54 2005) 122 134 wwwelseviercom/locate/apnum On convergence rate of the augmented Lagrangian algorithm for nonsymmetric saddle point problems G Awanou a,,jlai b a Institute

More information

MULTIGRID PRECONDITIONING IN H(div) ON NON-CONVEX POLYGONS* Dedicated to Professor Jim Douglas, Jr. on the occasion of his seventieth birthday.

MULTIGRID PRECONDITIONING IN H(div) ON NON-CONVEX POLYGONS* Dedicated to Professor Jim Douglas, Jr. on the occasion of his seventieth birthday. MULTIGRID PRECONDITIONING IN H(div) ON NON-CONVEX POLYGONS* DOUGLAS N ARNOLD, RICHARD S FALK, and RAGNAR WINTHER Dedicated to Professor Jim Douglas, Jr on the occasion of his seventieth birthday Abstract

More information

Preliminary/Qualifying Exam in Numerical Analysis (Math 502a) Spring 2012

Preliminary/Qualifying Exam in Numerical Analysis (Math 502a) Spring 2012 Instructions Preliminary/Qualifying Exam in Numerical Analysis (Math 502a) Spring 2012 The exam consists of four problems, each having multiple parts. You should attempt to solve all four problems. 1.

More information

Fast solvers for steady incompressible flow

Fast solvers for steady incompressible flow ICFD 25 p.1/21 Fast solvers for steady incompressible flow Andy Wathen Oxford University wathen@comlab.ox.ac.uk http://web.comlab.ox.ac.uk/~wathen/ Joint work with: Howard Elman (University of Maryland,

More information

Math Introduction to Numerical Analysis - Class Notes. Fernando Guevara Vasquez. Version Date: January 17, 2012.

Math Introduction to Numerical Analysis - Class Notes. Fernando Guevara Vasquez. Version Date: January 17, 2012. Math 5620 - Introduction to Numerical Analysis - Class Notes Fernando Guevara Vasquez Version 1990. Date: January 17, 2012. 3 Contents 1. Disclaimer 4 Chapter 1. Iterative methods for solving linear systems

More information

9.1 Preconditioned Krylov Subspace Methods

9.1 Preconditioned Krylov Subspace Methods Chapter 9 PRECONDITIONING 9.1 Preconditioned Krylov Subspace Methods 9.2 Preconditioned Conjugate Gradient 9.3 Preconditioned Generalized Minimal Residual 9.4 Relaxation Method Preconditioners 9.5 Incomplete

More information

A Robust Preconditioned Iterative Method for the Navier-Stokes Equations with High Reynolds Numbers

A Robust Preconditioned Iterative Method for the Navier-Stokes Equations with High Reynolds Numbers Applied and Computational Mathematics 2017; 6(4): 202-207 http://www.sciencepublishinggroup.com/j/acm doi: 10.11648/j.acm.20170604.18 ISSN: 2328-5605 (Print); ISSN: 2328-5613 (Online) A Robust Preconditioned

More information

20. A Dual-Primal FETI Method for solving Stokes/Navier-Stokes Equations

20. A Dual-Primal FETI Method for solving Stokes/Navier-Stokes Equations Fourteenth International Conference on Domain Decomposition Methods Editors: Ismael Herrera, David E. Keyes, Olof B. Widlund, Robert Yates c 23 DDM.org 2. A Dual-Primal FEI Method for solving Stokes/Navier-Stokes

More information

Lecture Note III: Least-Squares Method

Lecture Note III: Least-Squares Method Lecture Note III: Least-Squares Method Zhiqiang Cai October 4, 004 In this chapter, we shall present least-squares methods for second-order scalar partial differential equations, elastic equations of solids,

More information

Linear algebra issues in Interior Point methods for bound-constrained least-squares problems

Linear algebra issues in Interior Point methods for bound-constrained least-squares problems Linear algebra issues in Interior Point methods for bound-constrained least-squares problems Stefania Bellavia Dipartimento di Energetica S. Stecco Università degli Studi di Firenze Joint work with Jacek

More information

The Conjugate Gradient Method

The Conjugate Gradient Method The Conjugate Gradient Method Classical Iterations We have a problem, We assume that the matrix comes from a discretization of a PDE. The best and most popular model problem is, The matrix will be as large

More information

ON THE ROLE OF COMMUTATOR ARGUMENTS IN THE DEVELOPMENT OF PARAMETER-ROBUST PRECONDITIONERS FOR STOKES CONTROL PROBLEMS

ON THE ROLE OF COMMUTATOR ARGUMENTS IN THE DEVELOPMENT OF PARAMETER-ROBUST PRECONDITIONERS FOR STOKES CONTROL PROBLEMS ON THE ROLE OF COUTATOR ARGUENTS IN THE DEVELOPENT OF PARAETER-ROBUST PRECONDITIONERS FOR STOKES CONTROL PROBLES JOHN W. PEARSON Abstract. The development of preconditioners for PDE-constrained optimization

More information

Convergence of a Class of Stationary Iterative Methods for Saddle Point Problems

Convergence of a Class of Stationary Iterative Methods for Saddle Point Problems Convergence of a Class of Stationary Iterative Methods for Saddle Point Problems Yin Zhang 张寅 August, 2010 Abstract A unified convergence result is derived for an entire class of stationary iterative methods

More information

Journal of Computational and Applied Mathematics. Optimization of the parameterized Uzawa preconditioners for saddle point matrices

Journal of Computational and Applied Mathematics. Optimization of the parameterized Uzawa preconditioners for saddle point matrices Journal of Computational Applied Mathematics 6 (009) 136 154 Contents lists available at ScienceDirect Journal of Computational Applied Mathematics journal homepage: wwwelseviercom/locate/cam Optimization

More information

UZAWA TYPE ALGORITHMS FOR NONSYMMETRIC SADDLE POINT PROBLEMS

UZAWA TYPE ALGORITHMS FOR NONSYMMETRIC SADDLE POINT PROBLEMS MATHEMATICS OF COMPUTATION Volume 69, Number 230, Pages 667 689 S 0025-5718(99)01152-7 Article electronically published on May 17, 1999 UZAWA TYPE ALGORITHMS FOR NONSYMMETRIC SADDLE POINT PROBLEMS JAMES

More information

On matrix equations X ± A X 2 A = I

On matrix equations X ± A X 2 A = I Linear Algebra and its Applications 326 21 27 44 www.elsevier.com/locate/laa On matrix equations X ± A X 2 A = I I.G. Ivanov,V.I.Hasanov,B.V.Minchev Faculty of Mathematics and Informatics, Shoumen University,

More information

JOHANNES KEPLER UNIVERSITY LINZ. Institute of Computational Mathematics

JOHANNES KEPLER UNIVERSITY LINZ. Institute of Computational Mathematics JOHANNES KEPLER UNIVERSITY LINZ Institute of Computational Mathematics Schur Complement Preconditioners for Multiple Saddle Point Problems of Block Tridiagonal Form with Application to Optimization Problems

More information

Algebraic Multigrid Methods for the Oseen Problem

Algebraic Multigrid Methods for the Oseen Problem Algebraic Multigrid Methods for the Oseen Problem Markus Wabro Joint work with: Walter Zulehner, Linz www.numa.uni-linz.ac.at This work has been supported by the Austrian Science Foundation Fonds zur Förderung

More information

A MULTIGRID ALGORITHM FOR. Richard E. Ewing and Jian Shen. Institute for Scientic Computation. Texas A&M University. College Station, Texas SUMMARY

A MULTIGRID ALGORITHM FOR. Richard E. Ewing and Jian Shen. Institute for Scientic Computation. Texas A&M University. College Station, Texas SUMMARY A MULTIGRID ALGORITHM FOR THE CELL-CENTERED FINITE DIFFERENCE SCHEME Richard E. Ewing and Jian Shen Institute for Scientic Computation Texas A&M University College Station, Texas SUMMARY In this article,

More information

Local Inexact Newton Multilevel FEM for Nonlinear Elliptic Problems

Local Inexact Newton Multilevel FEM for Nonlinear Elliptic Problems Konrad-Zuse-Zentrum für Informationstechnik Berlin Heilbronner Str. 10, D-10711 Berlin-Wilmersdorf Peter Deuflhard Martin Weiser Local Inexact Newton Multilevel FEM for Nonlinear Elliptic Problems Preprint

More information

ON THE GENERALIZED DETERIORATED POSITIVE SEMI-DEFINITE AND SKEW-HERMITIAN SPLITTING PRECONDITIONER *

ON THE GENERALIZED DETERIORATED POSITIVE SEMI-DEFINITE AND SKEW-HERMITIAN SPLITTING PRECONDITIONER * Journal of Computational Mathematics Vol.xx, No.x, 2x, 6. http://www.global-sci.org/jcm doi:?? ON THE GENERALIZED DETERIORATED POSITIVE SEMI-DEFINITE AND SKEW-HERMITIAN SPLITTING PRECONDITIONER * Davod

More information

Key words. inf-sup constant, iterative solvers, preconditioning, saddle point problems

Key words. inf-sup constant, iterative solvers, preconditioning, saddle point problems NATURAL PRECONDITIONING AND ITERATIVE METHODS FOR SADDLE POINT SYSTEMS JENNIFER PESTANA AND ANDREW J. WATHEN Abstract. The solution of quadratic or locally quadratic extremum problems subject to linear(ized)

More information

Introduction to Iterative Solvers of Linear Systems

Introduction to Iterative Solvers of Linear Systems Introduction to Iterative Solvers of Linear Systems SFB Training Event January 2012 Prof. Dr. Andreas Frommer Typeset by Lukas Krämer, Simon-Wolfgang Mages and Rudolf Rödl 1 Classes of Matrices and their

More information

Geometric Multigrid Methods

Geometric Multigrid Methods Geometric Multigrid Methods Susanne C. Brenner Department of Mathematics and Center for Computation & Technology Louisiana State University IMA Tutorial: Fast Solution Techniques November 28, 2010 Ideas

More information

Iterative solvers for saddle point algebraic linear systems: tools of the trade. V. Simoncini

Iterative solvers for saddle point algebraic linear systems: tools of the trade. V. Simoncini Iterative solvers for saddle point algebraic linear systems: tools of the trade V. Simoncini Dipartimento di Matematica, Università di Bologna and CIRSA, Ravenna, Italy valeria@dm.unibo.it 1 The problem

More information

SUPERCONVERGENCE PROPERTIES FOR OPTIMAL CONTROL PROBLEMS DISCRETIZED BY PIECEWISE LINEAR AND DISCONTINUOUS FUNCTIONS

SUPERCONVERGENCE PROPERTIES FOR OPTIMAL CONTROL PROBLEMS DISCRETIZED BY PIECEWISE LINEAR AND DISCONTINUOUS FUNCTIONS SUPERCONVERGENCE PROPERTIES FOR OPTIMAL CONTROL PROBLEMS DISCRETIZED BY PIECEWISE LINEAR AND DISCONTINUOUS FUNCTIONS A. RÖSCH AND R. SIMON Abstract. An optimal control problem for an elliptic equation

More information

Kasetsart University Workshop. Multigrid methods: An introduction

Kasetsart University Workshop. Multigrid methods: An introduction Kasetsart University Workshop Multigrid methods: An introduction Dr. Anand Pardhanani Mathematics Department Earlham College Richmond, Indiana USA pardhan@earlham.edu A copy of these slides is available

More information

OUTLINE ffl CFD: elliptic pde's! Ax = b ffl Basic iterative methods ffl Krylov subspace methods ffl Preconditioning techniques: Iterative methods ILU

OUTLINE ffl CFD: elliptic pde's! Ax = b ffl Basic iterative methods ffl Krylov subspace methods ffl Preconditioning techniques: Iterative methods ILU Preconditioning Techniques for Solving Large Sparse Linear Systems Arnold Reusken Institut für Geometrie und Praktische Mathematik RWTH-Aachen OUTLINE ffl CFD: elliptic pde's! Ax = b ffl Basic iterative

More information

Chapter 7 Iterative Techniques in Matrix Algebra

Chapter 7 Iterative Techniques in Matrix Algebra Chapter 7 Iterative Techniques in Matrix Algebra Per-Olof Persson persson@berkeley.edu Department of Mathematics University of California, Berkeley Math 128B Numerical Analysis Vector Norms Definition

More information

A Robust Preconditioner for the Hessian System in Elliptic Optimal Control Problems

A Robust Preconditioner for the Hessian System in Elliptic Optimal Control Problems A Robust Preconditioner for the Hessian System in Elliptic Optimal Control Problems Etereldes Gonçalves 1, Tarek P. Mathew 1, Markus Sarkis 1,2, and Christian E. Schaerer 1 1 Instituto de Matemática Pura

More information

The All-floating BETI Method: Numerical Results

The All-floating BETI Method: Numerical Results The All-floating BETI Method: Numerical Results Günther Of Institute of Computational Mathematics, Graz University of Technology, Steyrergasse 30, A-8010 Graz, Austria, of@tugraz.at Summary. The all-floating

More information

Linear Algebra Massoud Malek

Linear Algebra Massoud Malek CSUEB Linear Algebra Massoud Malek Inner Product and Normed Space In all that follows, the n n identity matrix is denoted by I n, the n n zero matrix by Z n, and the zero vector by θ n An inner product

More information

Structured Preconditioners for Saddle Point Problems

Structured Preconditioners for Saddle Point Problems Structured Preconditioners for Saddle Point Problems V. Simoncini Dipartimento di Matematica Università di Bologna valeria@dm.unibo.it. p.1/16 Collaborators on this project Mario Arioli, RAL, UK Michele

More information

In order to solve the linear system KL M N when K is nonsymmetric, we can solve the equivalent system

In order to solve the linear system KL M N when K is nonsymmetric, we can solve the equivalent system !"#$% "&!#' (%)!#" *# %)%(! #! %)!#" +, %"!"#$ %*&%! $#&*! *# %)%! -. -/ 0 -. 12 "**3! * $!#%+,!2!#% 44" #% &#33 # 4"!#" "%! "5"#!!#6 -. - #% " 7% "3#!#3! - + 87&2! * $!#% 44" ) 3( $! # % %#!!#%+ 9332!

More information

Optimal multilevel preconditioning of strongly anisotropic problems.part II: non-conforming FEM. p. 1/36

Optimal multilevel preconditioning of strongly anisotropic problems.part II: non-conforming FEM. p. 1/36 Optimal multilevel preconditioning of strongly anisotropic problems. Part II: non-conforming FEM. Svetozar Margenov margenov@parallel.bas.bg Institute for Parallel Processing, Bulgarian Academy of Sciences,

More information

Termination criteria for inexact fixed point methods

Termination criteria for inexact fixed point methods Termination criteria for inexact fixed point methods Philipp Birken 1 October 1, 2013 1 Institute of Mathematics, University of Kassel, Heinrich-Plett-Str. 40, D-34132 Kassel, Germany Department of Mathematics/Computer

More information

Spectral Properties of Saddle Point Linear Systems and Relations to Iterative Solvers Part I: Spectral Properties. V. Simoncini

Spectral Properties of Saddle Point Linear Systems and Relations to Iterative Solvers Part I: Spectral Properties. V. Simoncini Spectral Properties of Saddle Point Linear Systems and Relations to Iterative Solvers Part I: Spectral Properties V. Simoncini Dipartimento di Matematica, Università di ologna valeria@dm.unibo.it 1 Outline

More information

PARTITION OF UNITY FOR THE STOKES PROBLEM ON NONMATCHING GRIDS

PARTITION OF UNITY FOR THE STOKES PROBLEM ON NONMATCHING GRIDS PARTITION OF UNITY FOR THE STOES PROBLEM ON NONMATCHING GRIDS CONSTANTIN BACUTA AND JINCHAO XU Abstract. We consider the Stokes Problem on a plane polygonal domain Ω R 2. We propose a finite element method

More information

Numerische Mathematik

Numerische Mathematik Numer. Math. 000) 86: 443 470 Digital Object Identifier DOI) 0.007/s00000060 Numerische Mathematik Effective preconditioning of Uzawa type schemes for a generalized Stokes problem Georgij M. Kobelkov,

More information

Jae Heon Yun and Yu Du Han

Jae Heon Yun and Yu Du Han Bull. Korean Math. Soc. 39 (2002), No. 3, pp. 495 509 MODIFIED INCOMPLETE CHOLESKY FACTORIZATION PRECONDITIONERS FOR A SYMMETRIC POSITIVE DEFINITE MATRIX Jae Heon Yun and Yu Du Han Abstract. We propose

More information

Contraction Methods for Convex Optimization and monotone variational inequalities No.12

Contraction Methods for Convex Optimization and monotone variational inequalities No.12 XII - 1 Contraction Methods for Convex Optimization and monotone variational inequalities No.12 Linearized alternating direction methods of multipliers for separable convex programming Bingsheng He Department

More information

INTERGRID OPERATORS FOR THE CELL CENTERED FINITE DIFFERENCE MULTIGRID ALGORITHM ON RECTANGULAR GRIDS. 1. Introduction

INTERGRID OPERATORS FOR THE CELL CENTERED FINITE DIFFERENCE MULTIGRID ALGORITHM ON RECTANGULAR GRIDS. 1. Introduction Trends in Mathematics Information Center for Mathematical Sciences Volume 9 Number 2 December 2006 Pages 0 INTERGRID OPERATORS FOR THE CELL CENTERED FINITE DIFFERENCE MULTIGRID ALGORITHM ON RECTANGULAR

More information

Iterative Methods for Solving A x = b

Iterative Methods for Solving A x = b Iterative Methods for Solving A x = b A good (free) online source for iterative methods for solving A x = b is given in the description of a set of iterative solvers called templates found at netlib: http

More information

Efficient Solvers for the Navier Stokes Equations in Rotation Form

Efficient Solvers for the Navier Stokes Equations in Rotation Form Efficient Solvers for the Navier Stokes Equations in Rotation Form Computer Research Institute Seminar Purdue University March 4, 2005 Michele Benzi Emory University Atlanta, GA Thanks to: NSF (MPS/Computational

More information

ITERATIVE METHODS FOR DOUBLE SADDLE POINT SYSTEMS

ITERATIVE METHODS FOR DOUBLE SADDLE POINT SYSTEMS IERAIVE MEHODS FOR DOULE SADDLE POIN SYSEMS FAEMEH PANJEH ALI EIK 1 AND MIHELE ENZI Abstract. We consider several iterative methods for solving a class of linear systems with double saddle point structure.

More information

STOPPING CRITERIA FOR MIXED FINITE ELEMENT PROBLEMS

STOPPING CRITERIA FOR MIXED FINITE ELEMENT PROBLEMS STOPPING CRITERIA FOR MIXED FINITE ELEMENT PROBLEMS M. ARIOLI 1,3 AND D. LOGHIN 2 Abstract. We study stopping criteria that are suitable in the solution by Krylov space based methods of linear and non

More information

From the Boundary Element DDM to local Trefftz Finite Element Methods on Polyhedral Meshes

From the Boundary Element DDM to local Trefftz Finite Element Methods on Polyhedral Meshes www.oeaw.ac.at From the Boundary Element DDM to local Trefftz Finite Element Methods on Polyhedral Meshes D. Copeland, U. Langer, D. Pusch RICAM-Report 2008-10 www.ricam.oeaw.ac.at From the Boundary Element

More information

AMS subject classifications. Primary, 65N15, 65N30, 76D07; Secondary, 35B45, 35J50

AMS subject classifications. Primary, 65N15, 65N30, 76D07; Secondary, 35B45, 35J50 A SIMPLE FINITE ELEMENT METHOD FOR THE STOKES EQUATIONS LIN MU AND XIU YE Abstract. The goal of this paper is to introduce a simple finite element method to solve the Stokes equations. This method is in

More information

PRECONDITIONING TECHNIQUES FOR NEWTON S METHOD FOR THE INCOMPRESSIBLE NAVIER STOKES EQUATIONS

PRECONDITIONING TECHNIQUES FOR NEWTON S METHOD FOR THE INCOMPRESSIBLE NAVIER STOKES EQUATIONS BIT Numerical Mathematics 43: 961 974, 2003. 2003 Kluwer Academic Publishers. Printed in the Netherlands. 961 PRECONDITIONING TECHNIQUES FOR NEWTON S METHOD FOR THE INCOMPRESSIBLE NAVIER STOKES EQUATIONS

More information

Short title: Total FETI. Corresponding author: Zdenek Dostal, VŠB-Technical University of Ostrava, 17 listopadu 15, CZ Ostrava, Czech Republic

Short title: Total FETI. Corresponding author: Zdenek Dostal, VŠB-Technical University of Ostrava, 17 listopadu 15, CZ Ostrava, Czech Republic Short title: Total FETI Corresponding author: Zdenek Dostal, VŠB-Technical University of Ostrava, 17 listopadu 15, CZ-70833 Ostrava, Czech Republic mail: zdenek.dostal@vsb.cz fax +420 596 919 597 phone

More information

Introduction. Chapter One

Introduction. Chapter One Chapter One Introduction The aim of this book is to describe and explain the beautiful mathematical relationships between matrices, moments, orthogonal polynomials, quadrature rules and the Lanczos and

More information

A NOTE ON THE LADYŽENSKAJA-BABUŠKA-BREZZI CONDITION

A NOTE ON THE LADYŽENSKAJA-BABUŠKA-BREZZI CONDITION A NOTE ON THE LADYŽENSKAJA-BABUŠKA-BREZZI CONDITION JOHNNY GUZMÁN, ABNER J. SALGADO, AND FRANCISCO-JAVIER SAYAS Abstract. The analysis of finite-element-like Galerkin discretization techniques for the

More information

Preface to the Second Edition. Preface to the First Edition

Preface to the Second Edition. Preface to the First Edition n page v Preface to the Second Edition Preface to the First Edition xiii xvii 1 Background in Linear Algebra 1 1.1 Matrices................................. 1 1.2 Square Matrices and Eigenvalues....................

More information

A MULTILEVEL SUCCESSIVE ITERATION METHOD FOR NONLINEAR ELLIPTIC PROBLEMS

A MULTILEVEL SUCCESSIVE ITERATION METHOD FOR NONLINEAR ELLIPTIC PROBLEMS MATHEMATICS OF COMPUTATION Volume 73, Number 246, Pages 525 539 S 0025-5718(03)01566-7 Article electronically published on July 14, 2003 A MULTILEVEL SUCCESSIVE ITERATION METHOD FOR NONLINEAR ELLIPTIC

More information

Multigrid Methods for Saddle Point Problems

Multigrid Methods for Saddle Point Problems Multigrid Methods for Saddle Point Problems Susanne C. Brenner Department of Mathematics and Center for Computation & Technology Louisiana State University Advances in Mathematics of Finite Elements (In

More information

Primal-dual relationship between Levenberg-Marquardt and central trajectories for linearly constrained convex optimization

Primal-dual relationship between Levenberg-Marquardt and central trajectories for linearly constrained convex optimization Primal-dual relationship between Levenberg-Marquardt and central trajectories for linearly constrained convex optimization Roger Behling a, Clovis Gonzaga b and Gabriel Haeser c March 21, 2013 a Department

More information

Efficient Solvers for Stochastic Finite Element Saddle Point Problems

Efficient Solvers for Stochastic Finite Element Saddle Point Problems Efficient Solvers for Stochastic Finite Element Saddle Point Problems Catherine E. Powell c.powell@manchester.ac.uk School of Mathematics University of Manchester, UK Efficient Solvers for Stochastic Finite

More information

XIAO-CHUAN CAI AND MAKSYMILIAN DRYJA. strongly elliptic equations discretized by the nite element methods.

XIAO-CHUAN CAI AND MAKSYMILIAN DRYJA. strongly elliptic equations discretized by the nite element methods. Contemporary Mathematics Volume 00, 0000 Domain Decomposition Methods for Monotone Nonlinear Elliptic Problems XIAO-CHUAN CAI AND MAKSYMILIAN DRYJA Abstract. In this paper, we study several overlapping

More information

SOLVING MESH EIGENPROBLEMS WITH MULTIGRID EFFICIENCY

SOLVING MESH EIGENPROBLEMS WITH MULTIGRID EFFICIENCY SOLVING MESH EIGENPROBLEMS WITH MULTIGRID EFFICIENCY KLAUS NEYMEYR ABSTRACT. Multigrid techniques can successfully be applied to mesh eigenvalue problems for elliptic differential operators. They allow

More information

Exam in TMA4215 December 7th 2012

Exam in TMA4215 December 7th 2012 Norwegian University of Science and Technology Department of Mathematical Sciences Page of 9 Contact during the exam: Elena Celledoni, tlf. 7359354, cell phone 48238584 Exam in TMA425 December 7th 22 Allowed

More information

AN AUGMENTED LAGRANGIAN APPROACH TO LINEARIZED PROBLEMS IN HYDRODYNAMIC STABILITY

AN AUGMENTED LAGRANGIAN APPROACH TO LINEARIZED PROBLEMS IN HYDRODYNAMIC STABILITY AN AUGMENTED LAGRANGIAN APPROACH TO LINEARIZED PROBLEMS IN HYDRODYNAMIC STABILITY MAXIM A. OLSHANSKII AND MICHELE BENZI Abstract. The solution of linear systems arising from the linear stability analysis

More information

Numerical Methods in Matrix Computations

Numerical Methods in Matrix Computations Ake Bjorck Numerical Methods in Matrix Computations Springer Contents 1 Direct Methods for Linear Systems 1 1.1 Elements of Matrix Theory 1 1.1.1 Matrix Algebra 2 1.1.2 Vector Spaces 6 1.1.3 Submatrices

More information

AMS526: Numerical Analysis I (Numerical Linear Algebra) Lecture 23: GMRES and Other Krylov Subspace Methods; Preconditioning

AMS526: Numerical Analysis I (Numerical Linear Algebra) Lecture 23: GMRES and Other Krylov Subspace Methods; Preconditioning AMS526: Numerical Analysis I (Numerical Linear Algebra) Lecture 23: GMRES and Other Krylov Subspace Methods; Preconditioning Xiangmin Jiao SUNY Stony Brook Xiangmin Jiao Numerical Analysis I 1 / 18 Outline

More information

4.6 Iterative Solvers for Linear Systems

4.6 Iterative Solvers for Linear Systems 4.6 Iterative Solvers for Linear Systems Why use iterative methods? Virtually all direct methods for solving Ax = b require O(n 3 ) floating point operations. In practical applications the matrix A often

More information

ON A SPLITTING PRECONDITIONER FOR SADDLE POINT PROBLEMS

ON A SPLITTING PRECONDITIONER FOR SADDLE POINT PROBLEMS J. Appl. Math. & Informatics Vol. 36(208, No. 5-6, pp. 459-474 https://doi.org/0.437/jami.208.459 ON A SPLITTING PRECONDITIONER FOR SADDLE POINT PROBLEMS DAVOD KHOJASTEH SALKUYEH, MARYAM ABDOLMALEKI, SAEED

More information

Domain Decomposition Algorithms for an Indefinite Hypersingular Integral Equation in Three Dimensions

Domain Decomposition Algorithms for an Indefinite Hypersingular Integral Equation in Three Dimensions Domain Decomposition Algorithms for an Indefinite Hypersingular Integral Equation in Three Dimensions Ernst P. Stephan 1, Matthias Maischak 2, and Thanh Tran 3 1 Institut für Angewandte Mathematik, Leibniz

More information

A Nonoverlapping Subdomain Algorithm with Lagrange Multipliers and its Object Oriented Implementation for Interface Problems

A Nonoverlapping Subdomain Algorithm with Lagrange Multipliers and its Object Oriented Implementation for Interface Problems Contemporary Mathematics Volume 8, 998 B 0-88-0988--03030- A Nonoverlapping Subdomain Algorithm with Lagrange Multipliers and its Object Oriented Implementation for Interface Problems Daoqi Yang. Introduction

More information

A Finite Element Method for an Ill-Posed Problem. Martin-Luther-Universitat, Fachbereich Mathematik/Informatik,Postfach 8, D Halle, Abstract

A Finite Element Method for an Ill-Posed Problem. Martin-Luther-Universitat, Fachbereich Mathematik/Informatik,Postfach 8, D Halle, Abstract A Finite Element Method for an Ill-Posed Problem W. Lucht Martin-Luther-Universitat, Fachbereich Mathematik/Informatik,Postfach 8, D-699 Halle, Germany Abstract For an ill-posed problem which has its origin

More information

FIRST-ORDER SYSTEM LEAST SQUARES (FOSLS) FOR PLANAR LINEAR ELASTICITY: PURE TRACTION PROBLEM

FIRST-ORDER SYSTEM LEAST SQUARES (FOSLS) FOR PLANAR LINEAR ELASTICITY: PURE TRACTION PROBLEM SIAM J. NUMER. ANAL. c 1998 Society for Industrial Applied Mathematics Vol. 35, No. 1, pp. 320 335, February 1998 016 FIRST-ORDER SYSTEM LEAST SQUARES (FOSLS) FOR PLANAR LINEAR ELASTICITY: PURE TRACTION

More information

ETNA Kent State University

ETNA Kent State University Electronic Transactions on Numerical Analysis. Volume 36, pp. 39-53, 009-010. Copyright 009,. ISSN 1068-9613. ETNA P-REGULAR SPLITTING ITERATIVE METHODS FOR NON-HERMITIAN POSITIVE DEFINITE LINEAR SYSTEMS

More information

The Mixed Finite Element Multigrid Preconditioned Minimum Residual Method for Stokes Equations

The Mixed Finite Element Multigrid Preconditioned Minimum Residual Method for Stokes Equations The Mixed Finite Element Multigrid Preconditioned Minimum Residual Method for Stokes Equations K. Muzhinji, S. Shateyi, and S, S. Motsa 2 University of Venda, Department of Mathematics, P Bag X5050, Thohoyandou

More information