PRECONDITIONING ITERATIVE METHODS FOR THE OPTIMAL CONTROL OF THE STOKES EQUATIONS

Similar documents
Chebyshev semi-iteration in Preconditioning

ON THE ROLE OF COMMUTATOR ARGUMENTS IN THE DEVELOPMENT OF PARAMETER-ROBUST PRECONDITIONERS FOR STOKES CONTROL PROBLEMS

A Review of Preconditioning Techniques for Steady Incompressible Flow

Optimal solvers for PDE-Constrained Optimization

Combination Preconditioning of saddle-point systems for positive definiteness

OPTIMAL SOLVERS FOR PDE-CONSTRAINED OPTIMIZATION

Fast solvers for steady incompressible flow

Block-triangular preconditioners for PDE-constrained optimization

Preconditioners for reduced saddle point systems arising in elliptic PDE-constrained optimization problems

A Robust Preconditioned Iterative Method for the Navier-Stokes Equations with High Reynolds Numbers

Fast Iterative Solution of Saddle Point Problems

Efficient Solvers for the Navier Stokes Equations in Rotation Form

On the accuracy of saddle point solvers

A MULTIGRID ALGORITHM FOR. Richard E. Ewing and Jian Shen. Institute for Scientic Computation. Texas A&M University. College Station, Texas SUMMARY

PDE-constrained Optimization and Beyond (PDE-constrained Optimal Control)

Numerical behavior of inexact linear solvers

Key words. preconditioned conjugate gradient method, saddle point problems, optimal control of PDEs, control and state constraints, multigrid method

Department of Computer Science, University of Illinois at Urbana-Champaign

Multigrid and Iterative Strategies for Optimal Control Problems

Efficient Augmented Lagrangian-type Preconditioning for the Oseen Problem using Grad-Div Stabilization

STOPPING CRITERIA FOR MIXED FINITE ELEMENT PROBLEMS

The Mixed Finite Element Multigrid Preconditioned Minimum Residual Method for Stokes Equations

Journal of Computational and Applied Mathematics. Optimization of the parameterized Uzawa preconditioners for saddle point matrices

Parallel Algorithms for Four-Dimensional Variational Data Assimilation

Preconditioning for Nonsymmetry and Time-dependence

On the Superlinear Convergence of MINRES. Valeria Simoncini and Daniel B. Szyld. Report January 2012

Multigrid absolute value preconditioning

AMS526: Numerical Analysis I (Numerical Linear Algebra) Lecture 23: GMRES and Other Krylov Subspace Methods; Preconditioning

A Tuned Preconditioner for Inexact Inverse Iteration for Generalised Eigenvalue Problems

Notes for CS542G (Iterative Solvers for Linear Systems)

The Conjugate Gradient Method

The Mixed Finite Element Multigrid Preconditioned MINRES Method for Stokes Equations

The antitriangular factorisation of saddle point matrices

7.4 The Saddle Point Stokes Problem

Fast Solvers for Unsteady Incompressible Flow

A Robust Preconditioner for the Hessian System in Elliptic Optimal Control Problems

Journal of Computational and Applied Mathematics. Multigrid method for solving convection-diffusion problems with dominant convection

Mathematics and Computer Science

PDE Solvers for Fluid Flow

On the interplay between discretization and preconditioning of Krylov subspace methods

Parallelizing large scale time domain electromagnetic inverse problem

Solving Symmetric Indefinite Systems with Symmetric Positive Definite Preconditioners

Key words. inf-sup constant, iterative solvers, preconditioning, saddle point problems

IN this paper, we investigate spectral properties of block

Chapter 7 Iterative Techniques in Matrix Algebra

DELFT UNIVERSITY OF TECHNOLOGY

Algebraic Multigrid as Solvers and as Preconditioner

Preconditioners for the incompressible Navier Stokes equations

Structured Preconditioners for Saddle Point Problems

Stabilization and Acceleration of Algebraic Multigrid Method

Adaptive algebraic multigrid methods in lattice computations

Solving PDEs with Multigrid Methods p.1

Discrete Projection Methods for Incompressible Fluid Flow Problems and Application to a Fluid-Structure Interaction

Preconditioning Techniques Analysis for CG Method

Linear and Non-Linear Preconditioning

Multigrid Based Optimal Shape and Topology Design in Magnetostatics

Key words. conjugate gradients, normwise backward error, incremental norm estimation.

Termination criteria for inexact fixed point methods

Convergence of a Class of Stationary Iterative Methods for Saddle Point Problems

Performance Comparison of Relaxation Methods with Singular and Nonsingular Preconditioners for Singular Saddle Point Problems

Key words. PDE-constrained optimization, space-time methods, preconditioning, Schur complement, domain decomposition, parallel computing.

Iterative Methods for Sparse Linear Systems

Iterative Methods for Solving A x = b

Indefinite Preconditioners for PDE-constrained optimization problems. V. Simoncini

Constrained Minimization and Multigrid

Preconditioned inverse iteration and shift-invert Arnoldi method

Efficient smoothers for all-at-once multigrid methods for Poisson and Stokes control problems

Preconditioners for state constrained optimal control problems with Moreau-Yosida penalty function

Preconditioning Iterative Methods for PDE Constrained Optimization

ON A GENERAL CLASS OF PRECONDITIONERS FOR NONSYMMETRIC GENERALIZED SADDLE POINT PROBLEMS

Contribution of Wo¹niakowski, Strako²,... The conjugate gradient method in nite precision computa

A Recursive Trust-Region Method for Non-Convex Constrained Minimization

ANALYSIS OF THE MINIMAL RESIDUAL METHOD APPLIED TO ILL-POSED OPTIMALITY SYSTEMS

Domain Decomposition Preconditioners for Spectral Nédélec Elements in Two and Three Dimensions

The Lanczos and conjugate gradient algorithms

CONVERGENCE BOUNDS FOR PRECONDITIONED GMRES USING ELEMENT-BY-ELEMENT ESTIMATES OF THE FIELD OF VALUES

Preface to the Second Edition. Preface to the First Edition

ETNA Kent State University

Indefinite and physics-based preconditioning

The Nullspace free eigenvalue problem and the inexact Shift and invert Lanczos method. V. Simoncini. Dipartimento di Matematica, Università di Bologna

Iterative methods for Linear System of Equations. Joint Advanced Student School (JASS-2009)

Adaptive methods for control problems with finite-dimensional control space

Efficient Solvers for Stochastic Finite Element Saddle Point Problems

Preconditioned GMRES Revisited

SOLVING MESH EIGENPROBLEMS WITH MULTIGRID EFFICIENCY

Preconditioning of Saddle Point Systems by Substructuring and a Penalty Approach

ON THE GENERALIZED DETERIORATED POSITIVE SEMI-DEFINITE AND SKEW-HERMITIAN SPLITTING PRECONDITIONER *

Jae Heon Yun and Yu Du Han

Finding Rightmost Eigenvalues of Large, Sparse, Nonsymmetric Parameterized Eigenvalue Problems

Spectral element agglomerate AMGe

Inexact Inverse Iteration for Symmetric Matrices

The amount of work to construct each new guess from the previous one should be a small multiple of the number of nonzeros in A.

Trust-Region SQP Methods with Inexact Linear System Solves for Large-Scale Optimization

Applied Mathematics 205. Unit V: Eigenvalue Problems. Lecturer: Dr. David Knezevic

Introduction to Scientific Computing

An advanced ILU preconditioner for the incompressible Navier-Stokes equations

Schur Complement Matrix And Its (Elementwise) Approximation: A Spectral Analysis Based On GLT Sequences

A STOKES INTERFACE PROBLEM: STABILITY, FINITE ELEMENT ANALYSIS AND A ROBUST SOLVER

Numerische Mathematik

arxiv: v1 [math.na] 18 Dec 2017

Transcription:

PRECONDITIONING ITERATIVE METHODS FOR THE OPTIMAL CONTROL OF THE STOKES EQUATIONS TYRONE REES, ANDREW J. WATHEN Abstract. Solving problems regarding the optimal control of partial differential equations (PDEs) also known as PDE-constrained optimization is a frontier area of numerical analysis. Of particular interest is the problem of flow control, where one would like to effect some desired flow by exerting, for example, an external force. The bottleneck in many current algorithms is the solution of the optimality system a system of equations in saddle point form that is usually very large and ill-conditioned. In this paper we describe two preconditioners a block-diagonal preconditioner for the minimal residual method and a block-lower triangular preconditioner for a non-standard conjugate gradient method which can be effective when applied to such problems where the PDEs are the Stokes equations. We consider only distributed control here, although other problems for example boundary control could be treated in the same way. We give numerical results, and compare these with those obtained by solving the equivalent forward problem using similar techniques.. Introduction. Suppose that we have a flow that satisfies the Stokes equations in some domain Ω with some given boundary condition, and that we have some mechanism for example, the application of a magnetic field to change the forcing term on the right hand side of the PDE. Let v and ˆp be given functions which are called the desired states. Then the question is how do we choose the forcing term such that the velocity v and pressure p are as close as possible to v and ˆp, in some sense, while still satisfying the Stokes equations. One way of formulating this problem is by minimizing a cost functional of trackingtype with the Stokes equations as a constraint, as follows: min v,p,u v v L (Ω) + δ p ˆp L (Ω) + β u L (Ω) (.) s.t. v + p = u in Ω v = in Ω, v = w on Ω. Here u denotes the forcing term on the right hand side, which is known as the control. In order for the problem to be well-posed we also include the control in the cost functional, together with a Tikhonov regularization parameter β, which is usually chosen a priori. A constant δ is added in front of the desired pressure to enable us to penalize the pressure. We would normally take ˆp =. We specify a Dirichlet boundary condition with v taking some value w which may or may not be taken from the desired state on the boundary. There are two methods with which one can discretize this problem we can either discretize the equations first and then optimize that system, or alternatively carry out the optimization first and then discretize the resulting optimality system. Since the Stokes equations are self-adjoint we will get the same discrete optimality system either way, provided the discretization methods are consistent between equations in Department of Computer Science, University of British Columbia, Vancouver, BritishColumbia, V6T Z4, Canada (tyronere@cs.ubc.ca) Mathematical Institute, University of Oxford, 4-9 St Giles, Oxford, OX 3LB, United Kingdom (wathen@maths.ox.ac.uk)

T. Rees, A.J. Wathen the optimize-then-discretize technique. We will therefore only consider the discretizethen-optimize approach here. Let{ φ j }, j =,...,n v +n and {ψ k }, k =,...,n p besetsoffiniteelement basis functions that form a stable mixed finite element discretization for the Stokes equations see, for example, [, Chapter 5] for further details and let v h = n v+n i= V iφi and p h = n p i= P iψ i be finite-dimensional approximations to v and p. Furthermore, let us also approximate the control from the velocity space, so u h = n v i= U iφ i. The discrete Stokes equation is of the form [ ] K B T B ][ v p = [ Q v ] [ f u+ g where v, p and u are the coefficient vectors in the expansions of v h, p h and u h respectively, K = [ Ω φ i : φ j ], B = [ Ω ψ k φ j ], Q v = [ Ωφ i φ j ], f = [ n u+n j=n V u+ j Ω φ i : φ j ] and g = [ n u+n j=n V u+ j Ω ψ i φ j ]. Note that the coefficents V j, j = n u+,...,n u +n are fixed so that v h interpolates the boundary data w. On discretizing, the cost functional becomes ], min vt Q v v v T b+ δ pt Q p p δp T d+ β ut Q v u where Q p = [ Ω ψ iψ j ], b = [ Ω v φ i ] and d = [ Ω pψ i]. Let us introduce two vectors of Lagrange multipliers, λ and µ. Then finding a critical point of the Lagrangian function gives the discrete optimality system of the form Q v K B T δq p B βq v Q T v K B T Q v B v p u λ µ = b δd f g. (.) It will be useful to relabel this system so that it becomes Q K βq v Q T y u = c, (.3) K Q ξ h [ ] K B T where Q = blkdiag(q v,δq p ), K =, Q = [Q v ] B T and the vectors y, ξ, c and h take their obvious definitions. For more detail on the practicalities of discretizing control problems of this type see, for example, Rees, Stoll and Wathen [3]. Findinganefficientmethod tosolvethissystemwillbethetopicoftheremainder of the paper. In Section we introduce two preconditioners that can be applied to this problem; one block diagonal, which we apply using the minimal residual method (MINRES) of Paige and Saunders [], and one block lower triangular, which we use with the Conjugate Gradient method (CG) of Hestenes and Steifel [5] applied with a nonstandard inner product. Both of these methods rely on good approximations to the (, )-block and the Schur complement, and we discuss suitable choices in Sections.3 and.4 respectively. Finally, in Section 3 we give numerical results.

Preconditioners for Stokes control 3. Solution methods. The matrix in (.3) is of saddle point form, that is [ A C T A = C ], (.) wherea = blkdiag(q,βq v ) andc = [K Q]. The matrixais, ingeneral, verylarge the discrete Stokes equations are just one of it s components yet is sparse. A good choice for solving such systems are iterative methods in particular Krylov subspace methods. We will consider two such methods here: MINRES and Conjugate Gradients in a non-standardinner product, and extend the workofrees, Dollarand Wathen [] and Rees and Stoll [] respectively to the case where the PDE is Stokes equations; significant complications arise here which are not present for simpler problems. We comment that there are a large number of papers in the literature which deal with solving problems for the optimal control of PDEs. Below we comment on a few of these which share the philosophy of this paper. Most of these consider the model problem of the optimal control of Poisson s equation; it is not clear how easily they would be applied to the control of the Stokes equations and the additional difficulty this poses. Schöberl and Zulehner [4] developed a preconditioner which is both optimal with respect to the problem size and with respect to the choice of regularization parameter, β. This method was recently generalized slightly by Herzog and Sachs [4]. A multigrid-based preconditioner has also been developed by Biros and Dogan [3] which has both h and β independent convergence properties, but it is not clear how their method would generalize to Stokes control. We note that the approximate reduced Hessian approximation used by Haber and Asher [3] and Biros and Ghattas [4] also leads to a preconditioner with h independence. Other solution methods employing multigrid for this and similar classes of problems were described by Borzi [5], Asher and Haber [] and Engel and Griebel []... Block diagonal preconditioners. It is well known that matrices of the form A are indefinite, and one choice of solution method for such systems is MINRES. For MINRES to be efficient for such a matrix we need to combine the method with a good preconditioner i.e. a matrix P which is cheap to invert and which clusters the eigenvalues of P A. One method that is often used see [, Section..] and the references therein is to look for a block diagonal preconditioner of the form P = [ ] A. S Preconditioners of this form for the optimal control of Poisson s equation were discussed by Rees, Dollar and Wathen []. It is well known (see, for example, [, Theorem 6.6]) that if A, A, CA C T and S are positive definite matrices such that there exist constants δ,, φ and Φ such that the generalized Rayleigh quotients satisfy δ xt Ax x T A x, φ yt CA C T y y T Φ S y for all vectors x R nv+np and y R nv+np, x, y, then the eigenvalues λ of

4 T. Rees, A.J. Wathen P A are real, and satisfy or δ δ +4 Φ λ +4φδ, δ λ, δ + δ +4δφ λ + +4Φ. Therefore, if we can find matrices A and S that are cheap to invert and are good approximations to A and the Schur complement CA C T in the sense defined above, then we will have a good preconditioner, since the eigenvalues of P A will be in three distinct clusters bounded away from. In the ideal case where A = A and S = CA C T we have δ = = φ = Φ =. Then the preconditioned system will have precisely three eigenvalues,, + 5 and 5, so MINRES would converge in three iterations [9]... Block lower-triangular preconditioners. Instead of MINRES we may want to use a conjugate gradient method to solve a saddle point problem of the form (.). Since (.) is not positive definite, the standard conjugate gradient algorithm cannot be used. However, the matrix [ ] [ A A C T C S C is self-adjoint with respect to the inner product defined by u,v H := u T Hv, where H = [ ] A A, S provided that this defines an inner product i.e. when A A and S are positive definite. Therefore can we apply the conjugate gradient algorithm with this inner product, along with preconditioner [ ] A P =. C S This method was first described by Bramble and Pasciak in[8], and has since generated a lot of interest see, for example, [, 6, 8, 4, 7, 7, 9]. This method was used in a control context by Rees and Stoll []. Convergence of this method again depends on the eigenvalue distribution of the preconditioned system the clustering of the eigenvalues is given by, e.g., Rees and Stoll [, Theorem 3.], and the relevant result is stated below in Section.4. Note that in order to apply this preconditioner only solves with A and S are needed, hence an implicit approximation, for example multigrid, can be used; for more detail see e.g. Stoll [6]. One drawback of this method is that you need A A to be positive definite; this means that not just any approximation to A will do. This requirement usually results in having to find the eigenvalues of A A for a candidate A, and then adding an appropriate scaling γ so that A > γa we will discuss this point further once we ve described possible approximations A in the following section. ]

Preconditioners for Stokes control 5.3. Approximation of the (,) block. Suppose, for simplicity, that our domain Ω R. If, as is usual, we use the same element space for all components in the velocity vector, and this has basis {φ i }. Then Q v = blkdiag(q v,q v ), where Q v = [ Ω φ iφ j ]. Then the matrix A is just a block diagonal matrix composed of the mass matrices in the bases {φ i } or {ψ i }. Wathen [3] showed that for a general mass matrix, Q, if D := diag(q), then it is possible to calculate constants ξ and Ξ such that ξ λ(d Q) Ξ. The constants depend on the elements used for example, for Q elements ξ = /4, Ξ = 9/4 and for Q elements ξ =/4, Ξ = 5/6. The diagonal itself would therefore be a reasonable approximation to A. However, as A is in a sense easy to invert, it would help to have as good an approximation here as we can. Using the bounds described above we have all the information we need to use the relaxed Jacobi method accelerated by the Chebyshevsemi iteration. This is a method that is very cheap to use and, as demonstrated by Wathen and Rees in [3], is particularly effective in this case. In particular, since the eigenvalues of D Q are evenly distributed, there is very little difference between the convergence of this method and the (non-linear) conjugate gradient method preconditioned with D. Note that since the conjugate gradient algorithm is non-linear, we cannot use it as a preconditioner for a stationary Krylov subspace method such as MINRES, unless run to convergence. The Chebyshev semi-iteration, on the other hand, is a linear method. Suppose we use it to solve Qx = b for some right hand side b. Then we can write every iteration as x (m) = Tm b, for some matrix T m implicitly defined by the method which is independent of b. Increasing m makes T m a better approximation to Q in the sense defined above. The upper and lower eigenvalue bounds can be obtained analytically for example, Table I in Rees and Stoll [] gives the upper and lower bounds for each m from ) to for a Q discretization. Therefore, for the problem (.), if δm v λ ((Tm) v Q v ( ) v m and δp m λ (Tm p) Q p p m, then δ m xt Ax x T A x m, (.) where A = blkdiag(tm v,tv m,tp m,βtv m,βtv m ) and δ m = min(δm v,δp m ) and m = max( v m, p m), both independent of the mesh size, h. We therefore have an inexpensive way to make the bounds on λ(a A) as close to unity as required. Note that, since we can work out these bounds accurately and inexpensively, the scaling parameter which needs to be calculated to ensure that A A is positive definite which is a requirement for CG in a non-standard inner product. can be easily chosen; see Rees and Stoll [] for more details..4. Approximation of the Schur complement. Now consider the Schur complement, β QQ v Q T +KQ K =: S. The dominant term in this sum, for all but the smallest values of β, is KQ K the term that contains the PDE. Figure. shows the eigenvalue distribution for this approximation of S for a relatively coarse Q Q discretization with β =.. As we can see from the figure, the eigenvalues of (KQ K) S are nicely clustered, and so we could expect good convergence of MINRES if we took S as KQ K. The effect of varying β is described in, e.g., [9].

6 T. Rees, A.J. Wathen 7 6 5 4 3 4 6 8 4 Fig... Eigenvalues of (KQ K) S However, a preconditioner must be easy to invert, and solving a system with KQ K requires two solves with the discrete Stokes matrix, which is not cheap. We therefore would like some matrix, K, such that KQ K approximates KQ K. Note that the mass matrices are not really significant in this context, and it is sufficient that K K T approximates K. In order to achieve such an approximation, Braess and Peisker [7] show that it is not sufficient that K approximates K. Indeed, for the Stokes equations, Silvester and Wathen [5] showed that an ideal preconditioner is K = blkdiag(k,m p ),wherek isamultigridcycle,buttheeigenvaluesof( K K T ) K arenotatallclustered,andtheapproximationofk isapooroneinthiscase. Suppose we wish to solve the equation Kx = b, for some right hand side vector b. Braess and Peiskerhowevergoontoshowthat ifwetakeanapproximationk m whichisimplicitly defined by an iteration x (m) = Km b, say, which converges to the solution x in the sense that x (m) x η m x, then η m = K m K I, and one can show [7, Section 4] ( η) xt K x x T K T mk m x (+η). (.3) Hence, approximation of K by K T mk m would be suitable in this case. Note that MINRES cannot be used to approximate K, unless run until convergence, since like CG MINRES is a Krylov subspace method, and hence nonlinear. We would therefore have to use a flexible outer method if we were to make use of an inner Krylov process as an approximation for the Stokes operator. As before, consider a simple iteration of the form x (m+) = x (m) +M Kr (m), (.4) where r (m) is the residual at the m th step, and with a block lower-triangular splitting matrix [ ] K M :=, (.5) B Q where K approximates K and Q approximates Q p, which is itself spectrally equivalent to the Schur complement for the Stokes problem [, Section 6.]. By the result

Preconditioners for Stokes control 7 of Braess and Peisker, we just need to show that this iteration converges i.e. that ρ(i M K) <, where ρ denotes the spectral radius in order that this defines a good approximation to the square. We ignore the one zero eigenvalue of K which is due to the hydrostatic pressure here, and in what follows, since if we start an iteration orthogonal to this kernel, we will remain orthogonal to the kernel [, Section.3]. Consider two cases K K positive definite, and K K indefinite. In the first case, it can be shown [, Theorem 3.], [3, Theorem 4.] that if K and Q are positive definite matrices such that υ xt Kx x T K x Υ, ψ yt BK B T y y T Ψ, (.6) Q y then λ is real and positive, and moreover satisfies (+ψ)υ (+ψ) Υ 4ψΥ λ (+Ψ)υ (+Ψ) υ 4Ψυ υ λ Υ or (+ψ)υ + (+ψ) υ 4ψυ λ (+Ψ)Υ+ (+Ψ) Υ 4ΨΥ. We would like to put some numbers to these bounds in order to see what this means for a simple iteration based on a splitting with the block lower triangular matrix (.5). It is well known that a multigrid iteration is a good approximation to the action of the inverse of K, and we can scale such an iteration so that xt Kx xk x +ρm ρ m, where m is the number of V-cycles. A realistic value for ρ is.5 (see [, pp. 94-95], for example), and experimentation shows m = gives reasonable performance. Using Q p for the Schur complement approximation we have γ xt BK B T x x T Q p x Γ, x, where for D Q elements, γ =., Γ =. Approximating this by steps of the Chebyshev semi-iteration will weaken these bounds by a factor of.96 in the lower bound and Θ =.4 in the upper. With these numbers, we have that λ(m A) [.9,.9], and hence ρ(i M A) =.8 <. Therefore the simple iteration (.4) with the splitting (.5) will converge. Although we ve assumed K K in the analysis above, experiments show we still have good convergence properties even if this isn t true. In the case where K K is indefinite the situation is more complicated, as now the eigenvalues will in general be complex. Consider the generalized eigenvalue problem: [ ] K B T B ][ x y = λ [ K B Q ][ x y ]. (.7) We still assume that K, K and Q are positive definite. If Bx =, it is clear that λ = xt Kx x T K x, and hence these eigenvalues must be real with υ λ Υ.

8 T. Rees, A.J. Wathen Suppose now that Bx. We can rearrange the second equation to give y = λ λ Q Bx, and substituting this into the first equation and rearranging gives If we define then we can write this as or, alternatively, λ = λ xt K x B T Q x T Kx +( λ)xt Bx x T. Kx κ := κ(x) = xt Kx x T K x, Therefore the eigenvalues satisfy σ := σ(x) = xt B T Q Bx x T, Kx λ /κ+( λ)σ λ =, λ (σ +)κλ+σκ =. λ = (σ +)κ± (σ +) κ 4σκ. We know from above that if κ = xt Kx x T K x Note that (σ +) κ 4σκ = κ = or κ =, then all the eigenvalues are real. 4σ (+σ), 4σ the last inequality being since (+σ) has a maximum value of which occurs when 4σ σ =. This tells us that for κ [, (+σ) ], λ C. In this case, λ = (σ +)κ±i 4σκ (σ +) κ λ = (σ +) κ +4σκ (σ +) κ 4 = σκ. Therefore the complex eigenvalues satisfy υψ λ Ψ. (.8) Moreover, Re(λ) = (σ+)κ >, so all the complex eigenvalues live in the right-hand plane. Also, Im(λ) 4σκ (σ Re(λ) = +) κ 4σκ (σ +) κ = (σ +)κ. (σ +)κ

Preconditioners for Stokes control 9 If we define F(σ,κ) := 4σκ (σ +) κ, (σ +)κ then so F σ = (σ ) (σ +) (4 κ)κσ κ (σ +), F σ = σ =. This critical point is clearly a maximum. This means that, for any fixed κ, F(σ,κ) has it s maximum at σ =. Therefore Im(λ) Re(λ) = F(σ,κ) κ κ κ = κ υ. Therefore, putting this together with (.8) above, the complex eigenvalues satisfy { λ z = re iθ C : υψ r Ψ, tan ( υ ) θ tan ( υ )}. (.9) For κ >, the result given above for K K positive definite still hold, and we have λ R which satisfy (ψ +)Υ (ψ +) Υ 4ψΥ λ (Ψ+)Υ+ (Ψ+) Υ 4ΨΥ. [ ] 4σ What about κ (+σ),? In this case, too, the bounds above hold, since in the derivation of these bounds we required no information about δ, all that is assumed is that λ R see [, Theorem 3.]. Verifying the inner bounds required that δ >, so these do not carry over, but there is no such problem with the outer bounds. We have proved the following theorem: Theorem.. Let λ be an eigenvalue associated with the generalized eigenvalue problem [ ][ ] [ ][ ] K B T x K x = λ, B y B Q y where K, K and Q are positive definite and satisfy (.6). If λ R, then it satisfies (+ψ)υ (+ψ) Υ 4ψΥ λ (+Ψ)Υ+ (+Ψ) Υ 4ΨΥ or υ λ Υ, and if λ C, then λ = re iθ, where r and θ satisfy υψ r Ψ, tan ( υ ) θ tan ( υ ).

T. Rees, A.J. Wathen Togetboundsforρ(I M A)wehavetobemorecarefulbecauseofthepresence of the complex eigenvalues. Figure. is a relevant diagram for the situation here. All the complex eigenvalues will be contained in the unit circle if the line d labelled on the diagram is less than unity. By the cosine rule: d = +Ψ Ψcosθ, where tanθ = υ. Therefore all the complex eigenvalues are in the unit circle if Ψ < cosθ. Note that, using the same argument, the distance from the origin to the point where the circle of radius ψυ and centre - touches the ray that makes an angle θ with the x-axis is +ψυ ψυcosθ. Φ d θ Fig... Diagram of the geometry containing the complex eigenvalues. θ = υ and d is the unknown length.

Preconditioners for Stokes control There follows: Corollary.. Suppose that the eigenvalues of the generalized eigenvalue problem (.7) are as described in Theorem.. Define { ξ := max υ,υ, (+ψ)υ (+ψ) Υ 4ψΥ, (+Ψ)Υ+ (+Ψ) Υ 4ΨΥ, +Ψ Ψcosθ, +ψυ } ψυcosθ. Then a simple iteration with splitting matrix [ ] K M = B Q will converge if ξ <, with the asymptotic convergence rate being ξ. Zulehner also derived an approximation to the convergence factor [3, Theorem 4.3]. Note that Corollary. differs slightly from the result in Zulehner this is because neither the result given here nor in [3] are sharp with regards to the complex eigenvalues. The two results are obtained in very different ways, and neither can be said to be a better approximation than the other one. Figure.3 shows the bounds predicted above and the actual eigenvalues for a number of approximations to the matrix K. This shows that we will get asymptotic convergence, but in practice we see good results from the first iteration. Also, the theory above is equally valid for the block upper-triangular approximation to the discrete Stokes matrix, whereas in practice we observe that it takes far more iterations with this upper-triangular splitting to converge. Let us again return to the case where K K is positive definite. Then we know from Section. that [ ] [ ] M K K B T K = B Q B is self adjoint in the inner product defined by [ ] K K H =. Q If we define K := M K, then we have that K is H normal, i.e. K K = K K, where K = H KT H. The iteration matrix I M K is therefore H normal, and so which tells us that I M K H = ρ(i M K), x k x H ρ k x H,

T. Rees, A.J. Wathen.8.6.4...4.6.8.8.6.4...4.6.8.5.5.5.5 (a) h =.5, K given by AMG V-cycle with pre- and post-smoothing step (b) h =.5, K given by AMG V-cycle with pre- and post-smoothing steps.8.6.4...4.6.8.8.6.4...4.6.8.5.5.5.5 (c) h =.5, K given by AMG V-cycles with pre- and post-smoothing steps (d) h =.5, K given by AMG V-cycle with pre- and post-smoothing step Fig..3. * s denote computed eigenvalues. Lines, from left to right, are at, (ψ+)υ (ψ+) Υ 4ψΥ, υ, Υ and (Ψ+)Υ+ (Ψ+) Υ 4ΨΥ, (the last two virtually coincide here). Dashed region is the bounds of Theorem. for the complex eigenvalues. Also shown is the unit circle centred at z =.. where ρ = ρ(i M K), the spectral radius of the iteration matrix. To apply the result of Braess and Peisker (.3) we need a constant η k such that the error converges the norm, i.e. x k x η k x. We know that over a finite dimensional vector space all norms are equivalent, though the equivalence constants may be h dependent for a discretized PDE problem. Thus there exist positive constants γ and Γ such that γ x x H Γ x,

Preconditioners for Stokes control 3 and hence x k x x k x H / γ ρm γ x H Γρ m γ x. (.) We now need to know the values of the constants γ and Γ. Recalling standard bounds for two dimensional finite element matrices see e.g. Theorems.3 and.9 in [] we have that, under mild assumptions, there exist positive constants d, D, c and C such that: Then x T Hx Γx T x would mean that dh x T x x T Kx Dx T x ch x T x x T Q p x Ch x T x. y T (K K )y+z T Q z Γ(y T y+z T z). Therefore if we have constants Γ and Γ such that y T (K K )y Γ y T y and z T Q z Γ z T z then we could take Γ = max(γ,γ ). First, note that from (.6) Therefore x T Kx Υx T K x Υx T Kx (Υ )x T Kx Υx T K x Υ(x T Kx x T K x) (Υ )x T Kx x T (K K )x D(Υ ) x T x Υ xt (K K )x x T x Γ = (Υ )D. Υ D(Υ ). Υ Let Q = Tm p represent m steps of the Chebyshev semi-iteration, as defined in Section.3, where Then δ p m xt Q p x x T T p mx p m. z T Q z z T z = zt Tm pz z T z = zt Tmz p z T Q p z zt Q p z z T z C ph δm p.

4 T. Rees, A.J. Wathen Therefore we can take Γ = C p h, and hence ( (Υ )D Γ = max, C ph ) Υ δm p satisfies x T Hx Γx T x. Now we turn our attention to a lower bound. Similarly to above, we take γ = min(γ,γ ), where Again, we have from (.6): γ y T y y T (K K )y and γ z T z z T Q z. υy T K y y T Ky = υy T Ky+( υ)y T Ky (υ )y T Ky υy T (K K )y (υ )dh υ yt (K K )y y T. y Again arguing as above, z T Q z z T z = zt Tmz p z T Q p z xt Q p x x T x c ph p. m Therefore we can take ( (υ )c u dh γ = min, c ph ) υ p, m which satisfies γx T x x T Hx. By equation(.) the contraction constant for convergence in the -norm is given by ρ m Γ/ γ. It is clear that γ = νh, where ν is a constant. For the numerator, in general, we will have Γ = (Υ )D Υ, as h is small. This would mean that Γ = O ( h ), γ i.e. the contraction constant would be dependent upon h. However, we have control over the value of Υ, as this measures the accuracy of the approximation to K. Recall that K is a good approximation to K if Υ is close to unity. As K is a multigrid process we can make this parameter as close to as required by simply taking more V-cycles, better smoothing, etc. If this approximation is good enough, and (Υ )D Υ is smaller than Cph, we will get a constant number of δm p iterations, at least up to some value of h. Note that we have knowledge of all the parameters involved, so given a smallest required value of h which one will know a priori one can pick an approximation K which gives a reasonable method. The

Preconditioners for Stokes control 5 quantity ρ m also appears in the numerator, so convergencecan be improved by taking more inexact Uzawa iterations. Even though the above argument only holds when K K is positive definite, we see the same behaviour in practice for the general case. Since solving the approximation to K is particularly expensive here it is worth getting the approximation to the mass matrix, Q, as close to Q as possible. Therefore, in the results that follow we take Q to be defined implicity by steps of the Chebyshev semi-iteration applied to the appropriate mass matrix. The inexact Uzawa method can be improved with the introduction of a parameter τ in front of the approximation to the Schur complement []. In the inexact case the optimal parameter is hard to obtain, but a good approximation is (φ+φ)/, where λ(s S) [φ,φ]. For Q elements and a Dirichlet problem, λ(q p S) [.,] [, p. 7], so we take our scaling parameter as τ = 3/5. We therefore advocate a practical splitting matrix for inexact Uzawa of [ ] K M =. B τq A matrix of the form [ A P := K m Q Km T ], where A is composed of Chebyshev approximations and K m is a simple iteration based on the splitting matrix M, should therefore be an effective preconditioner for the matrix A. 3. Numerical Results. First, consider the following forward problem, which sets the boundary conditions that we will use for the control problem. This is a classic test problem in fluid dynamics called leaky cavity flow, and a discussion is given by Elman, Silvester and Wathen [, Example 5..3]. Example 3.. Let Ω = [,], and let i and j denote unit vectors in the direction of the x and y axis respectively. Let v and p satisfy the Stokes equations v + p = in Ω v = in Ω, and let v = on the boundary except for on x =, y, where v = j. We discretize the Stokes problem using Q Q elements and solve the resulting linear system using MINRES []. As a preconditioner we use the block diagonal matrix blkdiag( K,T ), following Silvester and Wathen [5], where K denotes one AMG V-cycle(using the HSL MI AMG routine[6] applied via a MATLAB interface) and T is twenty steps of the Chebyshev semi-iteration applied with the pressure mass matrix. The problem was solved using MATLAB R9b, and the number of iterations and the time taken for different mesh sizes is given in Table 3.. The constant number of iterations independent of h and linear growth in CPU time (i.e. linear complexity of the solver) are well-understood for this problem see[, Chapter 6]. Figure 3. shows the streamlines and the pressure of the solution obtained. Note the small recirculations present in the lower corners these are Moffatt eddies. Adding a forcing term that reduces these eddies will be the object of our control problem, Example 3..

6 T. Rees, A.J. Wathen h size CPU time (s) Iterations 87.5 5 3 659.9 7 4,467.76 8 5 9,539.349 3 6 37,57.54 3 7 48,739 6.66 3 Table 3. Number of MINRES iterations and time taken to solve the forward problem in Example 3. Pressure Streamlines 5.8.6.4 5.5.5..5 Fig. 3.. Solution of Example 3. Example 3.. Let Ω = [,], and consider an optimal control problem of the form (.), with Dirichlet boundary conditions as given in Example 3. (leaky cavity flow). Take the desired pressure as p = and let v = y i x j. The exponentially distributed streamlines of the desired velocity are shown in Figure 3.. We discretize(.) usingq Q elements, alsousingq elements forthe control. Table 3. shows the results for solving the problem using MINRES, with right hand side as in Example 3. and with β = and δ =. As a preconditioner we use the block diagonal preconditioner, with K approximated by m steps of the simple iteration with splitting matrix M = [ K B S where S = BK B T is the exact Schur complement of the Stokes equation. K is given by k HSL MI AMG V-cycles. This is not a practical preconditioner, since it includes the exact Schur complement of the Stokes matrix, but we can see clearly that if the approximation K is not good enough we do not even in this idealized case get an optimal preconditioner. This phenomenon is explained by the theory in Section.4. It is therefore vital that the approximation K is close enough to ],

Preconditioners for Stokes control 7 Streamlines.9.8.7.6.5.4.3....4.6.8 Fig. 3.. K, in the sense defined in the previous section, in order to get an effective practical preconditioner. Table 3. Comparison of solution methods for solving Example 3. using MINRES preconditioned with the block diagonal preconditioner with m steps of inexact Uzawa approximating K and k AMG V-cycles approximating K. h size Exact, m= m=, k= m=, k= m=, k=3 m=, k=4 time its time its time its time its time its 344.89 5.9 9.79 7.8 7.85 7 3 5.38 7.43 35.35 7.365 7.38 7 4 6344 3.9 5 7.359 65 3.79 7 3.35 7 3.96 7 5 599 6.63 5 43.933 79 7.858 3 64.8 7 64.55 7 h size Exact, m= m=, k= m=, k= m=, k=3 m=, k=4 time its time its time its time its time its 344.73. 7.99 5.96 3. 3 3 5.48 3.49 9.4 5.43 5.45 5 4 6344 3.466 3 3.954 3 3.347 5 3.93 3 3.39 3 5 599 57.84 98.885 39 65.489 5 6.5 3 6.398 3 As we saw in Section.4, a practical preconditioner can be obtained by replacing the exact Stokes Schur complement by the pressure mass matrix or more generally, by something that approximates the pressure mass matrix. We take this to be steps of the Chebyshev semi-iteration applied to the relevant matrix, as described in Section.3. Experimentation suggests that taking two steps of the inexact Uzawa method, in which K is given by three HSL MI AMG V-cycles, will give a good preconditioner. In the results that follow we take β =, δ = and solve to a tolerance of 6 in the appropriate norm. As we see from Table 3.3, the overall technique which we have described seems to be a good method for solving the Stokes control problem. Comparing the results here with those to solve the forward problem in Table 3. the iteration numbers aren t that much more, and they do not increase significantly with the mesh size; the solution times also scale roughly linearly. Solving the control problem using the blocktriangular preconditioner is just over a factor of ten more expensive that solving a

8 T. Rees, A.J. Wathen Table 3.3 Comparison of solution methods for solving Example 3. using MINRES and BPCG preconditioned with the block diagonal and block lower triangular preconditioners respectively with steps of inexact Uzawa approximating K and 3 AMG V-cycles approximating K. h size MINRES BPCG backslash time its time its time 344.89 5.83 4.6 3 5.358 3.94 7.59 4 6344.76 33.679 8.6 5 599 4.965 33 3.33 7.3 6 54.74 35 4.584 single forward problem for every grid size an overhead that seems reasonable, given the increased complexity of the control problem in comparison to the forward problem. Figures 3.3 and 3.4 show the number of iterations taken to solve this problem for different values of β and δ in (.) respectively. These show that as we might expect from the theory decreasing β and increasing δ increases the number of iterations required to solve the system using our methods. From the plots in Figures 3.5 and 3.6 it seems that the value δ = gives a pressure of the same order as the uncontrolled problem, the solution of which is shown in Figure 3.. However, one can conceive of situations where we require a tighter bound on the pressure, and hence a higher value of δ. 4 β = β = β = 3 β = 4 β = 5 β = 6 9 8 7 6 β = β = β = 3 β = 4 β = 5 β = 6 No. of iter. 8 6 No. of iter. 5 4 4 3 3 4 5 6 7 log (N) 3 4 5 6 7 log (N) (a) MINRES (b) BPCG Fig. 3.3. Plot of problem size vs iterations needed for different β, where δ =. We have only presented a simple distributed control problem here. It is possible to solve other types of control problem using the same method see [] for a discussion in the simpler case of Poisson control. It is also possible to use this method together with bound constraints on the control Stoll and Wathen [8] discuss this approach in consideration of the Poisson control problem. 4. Conclusions. In this paper we have presented two preconditioners one for MINRES, and one for CG in a non-standard inner product that can be used to solve problems in Stokes control. These both rely on effective approximations to the (,) block, which is composed of mass matrices, and to the Schur complement. We advocate using the Chebyshev semi-iteration used to accelerate a relaxed Jacobi iteration as an approximation to the (,) block, and an inexact Uzawa based approx-

Preconditioners for Stokes control 9 35 3 5 δ = δ = δ = δ = δ = δ = 3 8 6 4 δ = δ = δ = δ = δ = δ = 3 No. of iter. 5 No. of iter. 8 6 5 4 3 4 5 6 7 log (N) 3 4 5 6 7 log (N) (a) MINRES (b) BPCG Fig. 3.4. Plot of problem size vs iterations needed for different δ, where β =. Pressure Streamlines.8.6.4.5.5. (a) Computed state for β =, δ =.5 Pressure Streamlines.8.6.4.5.5. (b) Computed state for β =, δ =.5 Fig. 3.5. Computed states for Example 3. in two dimensions, β =. imation for the Schur complement. We have given some theoretical justification for the effectiveness of such preconditioners and have given some numerical results.

T. Rees, A.J. Wathen Pressure Streamlines 4.8.6.4 4.5.5..5 (a) Computed state for δ = Pressure Streamlines 4.8.6.4 4.5.5..5 (b) Computed state for δ = Fig. 3.6. Computed states for Example 3. in two dimensions, β = 5. We compared these results with those for solving the equivalent forward problem, and the iteration count is only marginally higher in the control case, and behaves in broadly the same way as the iterations taken the solve the forward problem as the mesh size decreases. These approximations therefore seem reasonable for problems of this type. Furthermore, the ideas presented here have the potential to be extended to develop preconditioners for a variety of problems, with the additional constraints and features that real-world applications require. REFERENCES [] Uri M. Asher and Eldad Haber, A multigrid method for distributed parameter estimation problems, Electron. Trans. Numer. Anal., 5 (3), pp. 7. [] Michele Benzi, Gene H. Golub, and Jörg Liesen, Numerical solution of saddle point problems, Acta Numer., 4 (5), pp. 37. [3] George Biros and Günay Dogan, A multilevel algorithm for inverse problems with elliptic PDE constraints, Inverse Problems, 4 (8), p. 34 (8pp). [4] G. Biros and O. Ghattas, Parallel Lagrange-Newton-Krylov-Schur methods for PDEconstrained optimization. Part I: The Krylov-Schur solver, SIAM J. Sci. Comput., 7 ().

Preconditioners for Stokes control [5] A. Borzì and V. Schulz, Multigrid methods for PDE optimization, SIAM Rev., 5 (9), pp. 36 395. [6] J. Boyle, M. D. Mihajlovic, and J. A. Scott, HSL MI: an efficient amg preconditioner, Tech. Report RAL-TR-7-, Department of Computational and Applied Mathematics, Rutherford Appleton Laboratory, 7. [7] D. Braess and D. Peisker, On the numerical solution of the biharmonic equation and the role of squaring matrices for preconditioning, IMA J Numer. Anal., 6 (986), pp. 393 44. [8] J.H. Bramble and J. E. Pasciak, A preconditioning technique for indefinite systems resulting from mixed approximations of elliptic problems, Math. Comp., 5 (988), pp. 7. [9] H. S. Dollar, Nicholas I. M. Gould, Martin Stoll, and Andrew J. Wathen, Preconditioning saddle-point systems with applications in optimization, SIAM J. Sci. Comput., 3 (), pp. 49 7. [] H.C. Elman, Multigrid and Krylov subspace methods for the discrete Stokes equations, Internat. J. Numer. Methods Fluids, (995), pp. 755 77. [] Howard Elman, David Silvester, and Andy Wathen, Finite elements and fast iterative solvers: with applications in incompressible fluid dynamics, Numerical Mathematics and Scientific Computation, Oxford University Press, Oxford, 5. [] M. Engel and M. Griebel, A multigrid method for constrained optimal control problems, tech. report, SFB-Preprint 46, Sonderforschungsbereich 6, Rheinische Friedrich-Wilhelms- Universität Bonn, 8. [3] E. Haber and U. Ascher, Preconditioned all at once methods for large sparse parameter estimation problems, Inverse Problems, 7 (), pp. 847 864. [4] R. Herzog and E. Sachs, Preconditioned conjugate gradient method for optimal control problems with control and state constraints, Tech. Report Preprint-Number SPP53-88, Deutsche Forschungsgemeinschaft, Priority Program 53, 9. [5] M. R. Hestenes and E. Stiefel, Methods of conjugate gradients for solving linear systems, J. Res. Nat. Bur. Stand., 49 (95), pp. 49 436. [6] Axel Klawonn, Block-triangular preconditioners for saddle point problems with a penalty term, SIAM J. Sci. Comput., 9 (998), pp. 7 84. [7] J. Liesen and B. N. Parlett, On nonsymmetric saddle point matrices that allow conjugate gradient iterations, Numer. Math., 8 (8), pp. 65 64. [8] A. Meyer and T. Steidten, Improvement and experiments on the Bramble-Pasciak type CG for mixed problems in elasticity, tech. report, TU Chemnitz, Germany,. [9] Malcolm F. Murphy, Gene H. Golub, and Andrew J. Wathen, A note on preconditioning for indefinite linear systems, SIAM J. Sci. Comput., (), pp. 969 97. [] C. C. Paige and M. A. Saunders, Solution of sparse indefinite systems of linear equations, SIAM J. Numer. Anal., (975), pp. 67 69. [] Tyrone Rees, H. Sue Dollar, and Andrew J. Wathen, Optimal solvers for PDEconstrained optimization, SIAM J. Sci. Comput., 3 (), pp. 7 98. [] Tyrone Rees and Martin Stoll, Block triangular preconditioners for PDE-constrained optimization, Tech. Report 5/9, OCCAM, Oxford University Mathematical Institute, March 9. (to appear in NLAA). [3] Tyrone Rees, Martin Stoll, and Andrew J. Wathen, All-at-once preconditioning in PDEconstrained optimization, August 9. (to appear in Kybernetika for special issue on Algorithmy meeting, Podbansk, Slovakia). [4] Joachim Schöberl and Walter Zulehner, Symmetric indefinite preconditioners for saddle point problems with applications to PDE-constrained optimization problems, SIAM J. Matrix Anal. Appl., 9 (7), pp. 75 773. [5] D.J. Silvester and A.J. Wathen, Fast iterative solution of stabilised Stokes systems Part II: Using general block preconditioners, SIAM J. Numer. Anal., 3 (994), pp. 35 367. [6] Martin Stoll, Solving Linear Systems using the Adjoint, PhD thesis, University of Oxford, 9. [7] Martin Stoll and Andy Wathen, Combination preconditioning and the Bramble Pasciak + preconditioner, SIAM J. Matrix Anal. Appl., 3 (8), pp. 58 68. [8], Preconditioning for active set and projected gradient methods as semi-smooth Newton methods for PDE-constrained optimization with control constraints, Tech. Report 9/5, Oxford Centre for Collaborative Applied Mathematics, 9. [9] H.S. Thorne, Properties of linear systems in PDE-constrained optimization. Part i: Distributed control, Technical Report RAL-TR-9-7, Rutherford Appleton Laboratory, 9. [3] A. J. Wathen, Realistic eigenvalue bounds for the Galerkin mass matrix, IMA J Numer Anal, 7 (987), pp. 449 457.

T. Rees, A.J. Wathen [3] A. J. Wathen and T. Rees, Chebyshev semi-iteration in preconditioning for problems including the mass matrix, Electronic Transactions on Numerical Analysis, 34 (9), pp. 5 35. [3] Walter Zulehner, Analysis of iterative methods for saddle point problems: a unified approach, Math. Comput., 7 (), pp. 479 55.