Fast Iterative Solution of Saddle Point Problems

Similar documents
Efficient Solvers for the Navier Stokes Equations in Rotation Form

Efficient Augmented Lagrangian-type Preconditioning for the Oseen Problem using Grad-Div Stabilization

A Review of Preconditioning Techniques for Steady Incompressible Flow

ANALYSIS OF AUGMENTED LAGRANGIAN-BASED PRECONDITIONERS FOR THE STEADY INCOMPRESSIBLE NAVIER-STOKES EQUATIONS

Mathematics and Computer Science

Preconditioners for the incompressible Navier Stokes equations

A Robust Preconditioned Iterative Method for the Navier-Stokes Equations with High Reynolds Numbers

Fast solvers for steady incompressible flow

c 2011 Society for Industrial and Applied Mathematics

7.4 The Saddle Point Stokes Problem

AN AUGMENTED LAGRANGIAN APPROACH TO LINEARIZED PROBLEMS IN HYDRODYNAMIC STABILITY

c 2006 Society for Industrial and Applied Mathematics

Spectral Properties of Saddle Point Linear Systems and Relations to Iterative Solvers Part I: Spectral Properties. V. Simoncini

An advanced ILU preconditioner for the incompressible Navier-Stokes equations

A comparison of preconditioners for incompressible Navier Stokes solvers

ON THE ROLE OF COMMUTATOR ARGUMENTS IN THE DEVELOPMENT OF PARAMETER-ROBUST PRECONDITIONERS FOR STOKES CONTROL PROBLEMS

Department of Computer Science, University of Illinois at Urbana-Champaign

DELFT UNIVERSITY OF TECHNOLOGY

Iterative solvers for saddle point algebraic linear systems: tools of the trade. V. Simoncini

OUTLINE ffl CFD: elliptic pde's! Ax = b ffl Basic iterative methods ffl Krylov subspace methods ffl Preconditioning techniques: Iterative methods ILU

ON AUGMENTED LAGRANGIAN METHODS FOR SADDLE-POINT LINEAR SYSTEMS WITH SINGULAR OR SEMIDEFINITE (1,1) BLOCKS * 1. Introduction

Key words. inf-sup constant, iterative solvers, preconditioning, saddle point problems

Indefinite and physics-based preconditioning

AMS526: Numerical Analysis I (Numerical Linear Algebra) Lecture 23: GMRES and Other Krylov Subspace Methods; Preconditioning

Finding Rightmost Eigenvalues of Large, Sparse, Nonsymmetric Parameterized Eigenvalue Problems

DELFT UNIVERSITY OF TECHNOLOGY

New multigrid smoothers for the Oseen problem

Some Preconditioning Techniques for Saddle Point Problems

Numerical behavior of inexact linear solvers

ON A GENERAL CLASS OF PRECONDITIONERS FOR NONSYMMETRIC GENERALIZED SADDLE POINT PROBLEMS

Structured Preconditioners for Saddle Point Problems

Preconditioning for Nonsymmetry and Time-dependence

Algebraic Multigrid Methods for the Oseen Problem

X. He and M. Neytcheva. Preconditioning the incompressible Navier-Stokes equations with variable viscosity. J. Comput. Math., in press, 2012.

Indefinite Preconditioners for PDE-constrained optimization problems. V. Simoncini

DELFT UNIVERSITY OF TECHNOLOGY

Preconditioning Techniques for Large Linear Systems Part III: General-Purpose Algebraic Preconditioners

The antitriangular factorisation of saddle point matrices

c 2011 Society for Industrial and Applied Mathematics

Block preconditioners for saddle point systems arising from liquid crystal directors modeling

PRECONDITIONING TECHNIQUES FOR NEWTON S METHOD FOR THE INCOMPRESSIBLE NAVIER STOKES EQUATIONS

ON THE GENERALIZED DETERIORATED POSITIVE SEMI-DEFINITE AND SKEW-HERMITIAN SPLITTING PRECONDITIONER *

Solving Large Nonlinear Sparse Systems

Efficient Solvers for Stochastic Finite Element Saddle Point Problems

Structured Preconditioners for Saddle Point Problems

20. A Dual-Primal FETI Method for solving Stokes/Navier-Stokes Equations

1. Introduction. We consider the solution of systems of linear equations with the following block 2 2 structure:

Multigrid and Iterative Strategies for Optimal Control Problems

Numerische Mathematik

Journal of Computational and Applied Mathematics. Optimization of the parameterized Uzawa preconditioners for saddle point matrices

Block triangular preconditioner for static Maxwell equations*

Vector and scalar penalty-projection methods

On the accuracy of saddle point solvers

1. Introduction. Consider the Navier Stokes equations ηu t ν 2 u + (u grad) u + grad p = f div u = 0 (1.1)

Regularized HSS iteration methods for saddle-point linear systems

PDE-constrained Optimization and Beyond (PDE-constrained Optimal Control)

SEMI-CONVERGENCE ANALYSIS OF THE INEXACT UZAWA METHOD FOR SINGULAR SADDLE POINT PROBLEMS

Adaptive preconditioners for nonlinear systems of equations

PALADINS: Scalable Time-Adaptive Algebraic Splitting and Preconditioners for the Navier-Stokes Equations

Chebyshev semi-iteration in Preconditioning

The semi-convergence of GSI method for singular saddle point problems

Assessment of a vorticity based solver for the Navier-Stokes equations

Preconditioning the incompressible Navier-Stokes equations with variable viscosity

Splitting Iteration Methods for Positive Definite Linear Systems

The Conjugate Gradient Method

STOPPING CRITERIA FOR MIXED FINITE ELEMENT PROBLEMS

The Nullspace free eigenvalue problem and the inexact Shift and invert Lanczos method. V. Simoncini. Dipartimento di Matematica, Università di Bologna

Preface to the Second Edition. Preface to the First Edition

Preconditioned GMRES Revisited

ETNA Kent State University

On solving linear systems arising from Shishkin mesh discretizations

Trust-Region SQP Methods with Inexact Linear System Solves for Large-Scale Optimization

Convergence Properties of Preconditioned Hermitian and Skew-Hermitian Splitting Methods for Non-Hermitian Positive Semidefinite Matrices

Multigrid absolute value preconditioning

CONVERGENCE BOUNDS FOR PRECONDITIONED GMRES USING ELEMENT-BY-ELEMENT ESTIMATES OF THE FIELD OF VALUES

A STOKES INTERFACE PROBLEM: STABILITY, FINITE ELEMENT ANALYSIS AND A ROBUST SOLVER

IN this paper, we investigate spectral properties of block

Preconditioners for reduced saddle point systems arising in elliptic PDE-constrained optimization problems

Linear Solvers. Andrew Hazel

On the Superlinear Convergence of MINRES. Valeria Simoncini and Daniel B. Szyld. Report January 2012

Theoretical advances. To illustrate our approach, consider the scalar ODE model,

FEniCS Course. Lecture 0: Introduction to FEM. Contributors Anders Logg, Kent-Andre Mardal

UNIVERSITY OF MARYLAND PRECONDITIONERS FOR THE DISCRETE STEADY-STATE NAVIER-STOKES EQUATIONS CS-TR #4164 / UMIACS TR # JULY 2000

Fast Solvers for Unsteady Incompressible Flow

2.29 Numerical Fluid Mechanics Spring 2015 Lecture 9

AMS526: Numerical Analysis I (Numerical Linear Algebra)

Schur Complements on Hilbert Spaces and Saddle Point Systems

arxiv: v1 [math.na] 26 Dec 2013

Multilevel low-rank approximation preconditioners Yousef Saad Department of Computer Science and Engineering University of Minnesota

M.A. Botchev. September 5, 2014

On the interplay between discretization and preconditioning of Krylov subspace methods

Efficient iterative algorithms for linear stability analysis of incompressible flows

A Review of Solution Techniques for Unsteady Incompressible Flow

ETNA Kent State University

Linear algebra issues in Interior Point methods for bound-constrained least-squares problems

Solving Symmetric Indefinite Systems with Symmetric Positive Definite Preconditioners

A note on accurate and efficient higher order Galerkin time stepping schemes for the nonstationary Stokes equations

Preconditioning of Saddle Point Systems by Substructuring and a Penalty Approach

1 Computing with constraints

Part 4: Active-set methods for linearly constrained optimization. Nick Gould (RAL)

Transcription:

Michele Benzi Department of Mathematics and Computer Science Emory University Atlanta, GA

Acknowledgments NSF (Computational Mathematics) Maxim Olshanskii (Mech-Math, Moscow State U.) Zhen Wang (PhD student, Emory) Thanks also to Valeria Simoncini (U. of Bologna, Italy)

Motivation and goals Let A be a real, symmetric, n n matrix and let f R n be given. Let, denote the standard inner product in R n. Consider the following two problems: 1 Solve Au = f 2 Minimize the function J(u) = 1 Au, u f, u 2

Motivation and goals Let A be a real, symmetric, n n matrix and let f R n be given. Let, denote the standard inner product in R n. Consider the following two problems: 1 Solve Au = f 2 Minimize the function J(u) = 1 Au, u f, u 2 Note that J(u) = Au f. Hence, if A is positive definite (SPD), the two problems are equivalent, and there exists a unique solution u = A 1 f.

Motivation and goals Let A be a real, symmetric, n n matrix and let f R n be given. Let, denote the standard inner product in R n. Consider the following two problems: 1 Solve Au = f 2 Minimize the function J(u) = 1 Au, u f, u 2 Note that J(u) = Au f. Hence, if A is positive definite (SPD), the two problems are equivalent, and there exists a unique solution u = A 1 f. Many algorithms exist for solving SPD linear systems: Cholesky, Preconditioned Conjugate Gradients, AMG, etc.

Motivation and goals (cont.) Now we add a set of linear constraints:

Motivation and goals (cont.) Now we add a set of linear constraints: Minimize J(u) = 1 Au, u f, u 2 subject to Bu = g where A is n n, symmetric B is m n, with m < n f R n, g R m are given (either f or g could be 0, but not both)

Motivation and goals (cont.) Now we add a set of linear constraints: Minimize J(u) = 1 Au, u f, u 2 subject to Bu = g where A is n n, symmetric B is m n, with m < n f R n, g R m are given (either f or g could be 0, but not both) Standard approach: Introduce Lagrange multipliers, p R m Lagrangian L(u, p) = 1 Au, u f, u + p, Bu g 2 First-order optimality conditions: u L = 0, p L = 0

Motivation and goals (cont.) Optimality conditions: u L = Au + B T p f = 0, p L = Bu g = 0

Motivation and goals (cont.) Optimality conditions: u L = Au + B T p f = 0, p L = Bu g = 0 or, A B T B O «u p «= f g «(1)

Motivation and goals (cont.) Optimality conditions: u L = Au + B T p f = 0, p L = Bu g = 0 or, A B T B O «u p «= f g «(1) System (1) is a saddle point problem. Its solutions (u, p ) are saddle points for the Lagrangian L(u, p): min u max p L(u, p) = L(u, p ) = max min L(u, p) u p

Motivation and goals (cont.) Optimality conditions: u L = Au + B T p f = 0, p L = Bu g = 0 or, A B T B O «u p «= f g «(1) System (1) is a saddle point problem. Its solutions (u, p ) are saddle points for the Lagrangian L(u, p): min u max p L(u, p) = L(u, p ) = max min L(u, p) u p Also called a KKT system (Karush Kuhn Tucker), or equilibrium equations.

Motivation and goals (cont.) Optimality conditions: u L = Au + B T p f = 0, p L = Bu g = 0 or, A B T B O «u p «= f g «(1) System (1) is a saddle point problem. Its solutions (u, p ) are saddle points for the Lagrangian L(u, p): min u max p L(u, p) = L(u, p ) = max min L(u, p) u p Also called a KKT system (Karush Kuhn Tucker), or equilibrium equations. Gil Strang calls (1) the fundamental problem of scientific computing.

Motivation and goals (cont.) Saddle point problems do occur frequently, e.g.:

Motivation and goals (cont.) Saddle point problems do occur frequently, e.g.: Incompressible flow problems (Stokes, linearized Navier Stokes) Mixed FEM formulations of 2nd- and 4th-order elliptic PDEs Time-harmonic Maxwell equations PDE-constrained optimization (e.g., variational data assimilation) SQP and IP methods for nonlinear constrained optimization Structural analysis Resistive networks, power network analysis Certain economic models

Motivation and goals (cont.) Saddle point problems do occur frequently, e.g.: Incompressible flow problems (Stokes, linearized Navier Stokes) Mixed FEM formulations of 2nd- and 4th-order elliptic PDEs Time-harmonic Maxwell equations PDE-constrained optimization (e.g., variational data assimilation) SQP and IP methods for nonlinear constrained optimization Structural analysis Resistive networks, power network analysis Certain economic models Comprehensive survey: M. Benzi, G. Golub and J. Liesen, Numerical solution of saddle point problems, Acta Numerica 14 (2005), pp. 1 137.

Motivation and goals (cont.) Saddle point problems do occur frequently, e.g.: Incompressible flow problems (Stokes, linearized Navier Stokes) Mixed FEM formulations of 2nd- and 4th-order elliptic PDEs Time-harmonic Maxwell equations PDE-constrained optimization (e.g., variational data assimilation) SQP and IP methods for nonlinear constrained optimization Structural analysis Resistive networks, power network analysis Certain economic models Comprehensive survey: M. Benzi, G. Golub and J. Liesen, Numerical solution of saddle point problems, Acta Numerica 14 (2005), pp. 1 137. The bibliography in this paper contains 535 items.

Motivation and goals (cont.) Saddle point problems do occur frequently, e.g.: Incompressible flow problems (Stokes, linearized Navier Stokes) Mixed FEM formulations of 2nd- and 4th-order elliptic PDEs Time-harmonic Maxwell equations PDE-constrained optimization (e.g., variational data assimilation) SQP and IP methods for nonlinear constrained optimization Structural analysis Resistive networks, power network analysis Certain economic models Comprehensive survey: M. Benzi, G. Golub and J. Liesen, Numerical solution of saddle point problems, Acta Numerica 14 (2005), pp. 1 137. The bibliography in this paper contains 535 items. Google Scholar reports 480 citations to date.

Motivation and goals (cont.) The aim of this talk:

Motivation and goals (cont.) The aim of this talk: To briefly review the basic properties of saddle point systems

Motivation and goals (cont.) The aim of this talk: To briefly review the basic properties of saddle point systems To give an overview of solution algorithms for specific problems

Motivation and goals (cont.) The aim of this talk: To briefly review the basic properties of saddle point systems To give an overview of solution algorithms for specific problems To point out some current challenges and recent developments

Motivation and goals (cont.) The aim of this talk: To briefly review the basic properties of saddle point systems To give an overview of solution algorithms for specific problems To point out some current challenges and recent developments The emphasis of the talk will be on preconditioned iterative solvers for large, sparse saddle point problems, with a focus on our own recent work on preconditioners for incompressible flow problems.

Motivation and goals (cont.) The aim of this talk: To briefly review the basic properties of saddle point systems To give an overview of solution algorithms for specific problems To point out some current challenges and recent developments The emphasis of the talk will be on preconditioned iterative solvers for large, sparse saddle point problems, with a focus on our own recent work on preconditioners for incompressible flow problems. The ultimate goal: to develop robust preconditioners that perform uniformly well independently of discretization details and problem parameters.

Motivation and goals (cont.) The aim of this talk: To briefly review the basic properties of saddle point systems To give an overview of solution algorithms for specific problems To point out some current challenges and recent developments The emphasis of the talk will be on preconditioned iterative solvers for large, sparse saddle point problems, with a focus on our own recent work on preconditioners for incompressible flow problems. The ultimate goal: to develop robust preconditioners that perform uniformly well independently of discretization details and problem parameters. For flow problems, we would like to have solvers that converge fast regardless of mesh size, viscosity, etc. Moreover, the cost per iteration should be linear in the number of unknowns.

Outline 1 Properties of saddle point matrices

Outline 1 Properties of saddle point matrices 2 Examples: Incompressible flow problems

Outline 1 Properties of saddle point matrices 2 Examples: Incompressible flow problems 3 Some solution algorithms

Outline 1 Properties of saddle point matrices 2 Examples: Incompressible flow problems 3 Some solution algorithms 4 The Augmented Lagrangian (AL) approach

Outline 1 Properties of saddle point matrices 2 Examples: Incompressible flow problems 3 Some solution algorithms 4 The Augmented Lagrangian (AL) approach 5 The modified Augmented Lagrangian-based preconditioner

Outline 1 Properties of saddle point matrices 2 Examples: Incompressible flow problems 3 Some solution algorithms 4 The Augmented Lagrangian (AL) approach 5 The modified Augmented Lagrangian-based preconditioner 6 Conclusions

Properties of saddle point matrices Outline 1 Properties of saddle point matrices 2 Examples: Incompressible flow problems 3 Some solution algorithms 4 The Augmented Lagrangian (AL) approach 5 The modified Augmented Lagrangian-based preconditioner 6 Conclusions

Properties of saddle point matrices Solvability of saddle point problems The following result establishes necessary and sufficient conditions for the unique solvability of the saddle point problem (1).

Properties of saddle point matrices Solvability of saddle point problems The following result establishes necessary and sufficient conditions for the unique solvability of the saddle point problem (1). Theorem. Assume that A is symmetric positive semidefinite n n B has full rank: rank (B) = m

Properties of saddle point matrices Solvability of saddle point problems The following result establishes necessary and sufficient conditions for the unique solvability of the saddle point problem (1). Theorem. Assume that A is symmetric positive semidefinite n n B has full rank: rank (B) = m Then the coefficient matrix «A B T A = B O is nonsingular Null (A) Null (B) = {0}.

Properties of saddle point matrices Solvability of saddle point problems The following result establishes necessary and sufficient conditions for the unique solvability of the saddle point problem (1). Theorem. Assume that A is symmetric positive semidefinite n n B has full rank: rank (B) = m Then the coefficient matrix «A B T A = B O is nonsingular Null (A) Null (B) = {0}. Furthermore, A is indefinite, with n positive and m negative eigenvalues.

Properties of saddle point matrices Solvability of saddle point problems The following result establishes necessary and sufficient conditions for the unique solvability of the saddle point problem (1). Theorem. Assume that A is symmetric positive semidefinite n n B has full rank: rank (B) = m Then the coefficient matrix «A B T A = B O is nonsingular Null (A) Null (B) = {0}. Furthermore, A is indefinite, with n positive and m negative eigenvalues. In particular, A is invertible if A is SPD and B has full rank ( standard case ).

Properties of saddle point matrices Generalizations, I In some cases, a stabilization (or regularization) term needs to be added in the (2,2) position, leading to linear systems of the form

Properties of saddle point matrices Generalizations, I In some cases, a stabilization (or regularization) term needs to be added in the (2,2) position, leading to linear systems of the form A B T B βc «u p «= f g «(2) where β > 0 is a small parameter and the m m matrix C is symmetric positive semidefinite, and often singular, with C 2 = 1.

Properties of saddle point matrices Generalizations, I In some cases, a stabilization (or regularization) term needs to be added in the (2,2) position, leading to linear systems of the form A B T B βc «u p «= f g «(2) where β > 0 is a small parameter and the m m matrix C is symmetric positive semidefinite, and often singular, with C 2 = 1. This type of system arises, for example, from the stabilization of FEM pairs that do not satisfy the LBB ( inf-sup ) condition.

Properties of saddle point matrices Generalizations, I In some cases, a stabilization (or regularization) term needs to be added in the (2,2) position, leading to linear systems of the form A B T B βc «u p «= f g «(2) where β > 0 is a small parameter and the m m matrix C is symmetric positive semidefinite, and often singular, with C 2 = 1. This type of system arises, for example, from the stabilization of FEM pairs that do not satisfy the LBB ( inf-sup ) condition. Another important example is the discretization of the Reissner Mindlin plate model in linear elasticity. In this case β is related to the thickness of the plate; the limit case β = 0 can be seen as a reformulation of the biharmonic problem.

Properties of saddle point matrices Generalizations, II In other cases, the matrix A is not symmetric: A A T. In this case, the saddle point system does not arise from a constrained minimization problem.

Properties of saddle point matrices Generalizations, II In other cases, the matrix A is not symmetric: A A T. In this case, the saddle point system does not arise from a constrained minimization problem. The most important examples of this case are linear systems arising from the Picard and Newton linearizations of the steady incompressible Navier Stokes equations. The following result is applicable to the Picard linearization (Oseen problem):

Properties of saddle point matrices Generalizations, II In other cases, the matrix A is not symmetric: A A T. In this case, the saddle point system does not arise from a constrained minimization problem. The most important examples of this case are linear systems arising from the Picard and Newton linearizations of the steady incompressible Navier Stokes equations. The following result is applicable to the Picard linearization (Oseen problem): Theorem. Assume that H = 1 2 (A + AT ) is symmetric positive semidefinite n n B has full rank: rank (B) = m

Properties of saddle point matrices Generalizations, II In other cases, the matrix A is not symmetric: A A T. In this case, the saddle point system does not arise from a constrained minimization problem. The most important examples of this case are linear systems arising from the Picard and Newton linearizations of the steady incompressible Navier Stokes equations. The following result is applicable to the Picard linearization (Oseen problem): Theorem. Assume that Then H = 1 2 (A + AT ) is symmetric positive semidefinite n n B has full rank: rank (B) = m Null(H) Null(B) = {0} A invertible A invertible Null(A) Null(B) = {0}.

Properties of saddle point matrices Nonsymmetric, positive definite form Consider the following equivalent formulation:

Properties of saddle point matrices Nonsymmetric, positive definite form Consider the following equivalent formulation: A B T B O «u p «= f g «

Properties of saddle point matrices Nonsymmetric, positive definite form Consider the following equivalent formulation: A B T B O «u p «= f g «Theorem. Assume B has full rank. If H = 1 (A + 2 AT ) is positive definite, then the spectrum of «A B T A := B O lies entirely in the open right-half plane Re(z) > 0. Moreover, if A is SPD and the following condition holds: λ min (A) > 4 λ max(s) where S = BA 1 B T ( Schur complement ), then A is diagonalizable with real positive eigenvalues.

Properties of saddle point matrices Nonsymmetric, positive definite form Consider the following equivalent formulation: A B T B O «u p «= f g «Theorem. Assume B has full rank. If H = 1 (A + 2 AT ) is positive definite, then the spectrum of «A B T A := B O lies entirely in the open right-half plane Re(z) > 0. Moreover, if A is SPD and the following condition holds: λ min (A) > 4 λ max(s) where S = BA 1 B T ( Schur complement ), then A is diagonalizable with real positive eigenvalues. In this case, there exists a non-standard inner product on R n+m in which A is self-adjoint and positive definite, and a corresponding conjugate gradient method (B./Simoncini, NM 2006).

Examples: Incompressible flow problems Outline 1 Properties of saddle point matrices 2 Examples: Incompressible flow problems 3 Some solution algorithms 4 The Augmented Lagrangian (AL) approach 5 The modified Augmented Lagrangian-based preconditioner 6 Conclusions

Examples: Incompressible flow problems Example 1: the generalized Stokes problem Let Ω be a domain in R d and let α 0, ν > 0. Consider the system

Examples: Incompressible flow problems Example 1: the generalized Stokes problem Let Ω be a domain in R d and let α 0, ν > 0. Consider the system αu ν u + p = f in Ω,

Examples: Incompressible flow problems Example 1: the generalized Stokes problem Let Ω be a domain in R d and let α 0, ν > 0. Consider the system αu ν u + p = f in Ω, div u = 0 in Ω,

Examples: Incompressible flow problems Example 1: the generalized Stokes problem Let Ω be a domain in R d and let α 0, ν > 0. Consider the system αu ν u + p = f in Ω, div u = 0 in Ω, u = 0 on Ω.

Examples: Incompressible flow problems Example 1: the generalized Stokes problem Let Ω be a domain in R d and let α 0, ν > 0. Consider the system αu ν u + p = f in Ω, div u = 0 in Ω, u = 0 on Ω. Weak formulation: Find (u, p) (H 1 0 (Ω)) d L 2 0(Ω) such that

Examples: Incompressible flow problems Example 1: the generalized Stokes problem Let Ω be a domain in R d and let α 0, ν > 0. Consider the system αu ν u + p = f in Ω, div u = 0 in Ω, u = 0 on Ω. Weak formulation: Find (u, p) (H 1 0 (Ω)) d L 2 0(Ω) such that α u, v + u, v p, div v = f, v, v (H 1 0 (Ω)) d,

Examples: Incompressible flow problems Example 1: the generalized Stokes problem Let Ω be a domain in R d and let α 0, ν > 0. Consider the system αu ν u + p = f in Ω, div u = 0 in Ω, u = 0 on Ω. Weak formulation: Find (u, p) (H 1 0 (Ω)) d L 2 0(Ω) such that α u, v + u, v p, div v = f, v, v (H 1 0 (Ω)) d, q, div u = 0, q L 2 0(Ω), where, denotes the L 2 inner product.

Examples: Incompressible flow problems Example 1: the generalized Stokes problem Let Ω be a domain in R d and let α 0, ν > 0. Consider the system αu ν u + p = f in Ω, div u = 0 in Ω, u = 0 on Ω. Weak formulation: Find (u, p) (H 1 0 (Ω)) d L 2 0(Ω) such that α u, v + u, v p, div v = f, v, v (H 1 0 (Ω)) d, q, div u = 0, q L 2 0(Ω), where, denotes the L 2 inner product. The standard Stokes problem is obtained for α = 0 (steady case). In this case we can assume ν = 1.

Examples: Incompressible flow problems Example 1: the generalized Stokes problem (cont.) Discretization using LBB-stable finite element pairs or other div-stable scheme leads to an algebraic saddle point problem:

Examples: Incompressible flow problems Example 1: the generalized Stokes problem (cont.) Discretization using LBB-stable finite element pairs or other div-stable scheme leads to an algebraic saddle point problem: A B T B O «u p «= f 0 «

Examples: Incompressible flow problems Example 1: the generalized Stokes problem (cont.) Discretization using LBB-stable finite element pairs or other div-stable scheme leads to an algebraic saddle point problem: A B T B O «u p «= f 0 «Here A is a discrete reaction-diffusion operator, B T the discrete gradient, and B the discrete (negative) divergence. For α = 0, A is just the discrete vector Laplacian.

Examples: Incompressible flow problems Example 1: the generalized Stokes problem (cont.) Discretization using LBB-stable finite element pairs or other div-stable scheme leads to an algebraic saddle point problem: A B T B O «u p «= f 0 «Here A is a discrete reaction-diffusion operator, B T the discrete gradient, and B the discrete (negative) divergence. For α = 0, A is just the discrete vector Laplacian. If an unstable FEM pair is used, then a regularization term βc is added in the (2, 2) block of A. The specific choice of β and C depends on the particular discretization used.

Examples: Incompressible flow problems Example 1: the generalized Stokes problem (cont.) Discretization using LBB-stable finite element pairs or other div-stable scheme leads to an algebraic saddle point problem: A B T B O «u p «= f 0 «Here A is a discrete reaction-diffusion operator, B T the discrete gradient, and B the discrete (negative) divergence. For α = 0, A is just the discrete vector Laplacian. If an unstable FEM pair is used, then a regularization term βc is added in the (2, 2) block of A. The specific choice of β and C depends on the particular discretization used. Robust, optimal solvers have been developed for this problem: Cahouet Chabard for α > 0; Silvester Wathen for α = 0.

Examples: Incompressible flow problems Sparsity pattern: 2D stokes (Q1-P0) Without stabilization (C = O)

Examples: Incompressible flow problems Sparsity pattern: 2D stokes (Q1-P0) With stabilization (C O)

Examples: Incompressible flow problems Example 2: the generalized Oseen problem Let Ω be a domain in R d and let α 0 and ν > 0. Also, let w be a divergence-free vector field on Ω. Consider the system

Examples: Incompressible flow problems Example 2: the generalized Oseen problem Let Ω be a domain in R d and let α 0 and ν > 0. Also, let w be a divergence-free vector field on Ω. Consider the system αu ν u + (w ) u + p = f in Ω,

Examples: Incompressible flow problems Example 2: the generalized Oseen problem Let Ω be a domain in R d and let α 0 and ν > 0. Also, let w be a divergence-free vector field on Ω. Consider the system αu ν u + (w ) u + p = f in Ω, div u = 0 in Ω,

Examples: Incompressible flow problems Example 2: the generalized Oseen problem Let Ω be a domain in R d and let α 0 and ν > 0. Also, let w be a divergence-free vector field on Ω. Consider the system αu ν u + (w ) u + p = f in Ω, div u = 0 in Ω, u = 0 on Ω. Note that for w = 0 we recover the generalized Stokes problem.

Examples: Incompressible flow problems Example 2: the generalized Oseen problem Let Ω be a domain in R d and let α 0 and ν > 0. Also, let w be a divergence-free vector field on Ω. Consider the system αu ν u + (w ) u + p = f in Ω, div u = 0 in Ω, u = 0 on Ω. Note that for w = 0 we recover the generalized Stokes problem. Weak formulation: Find (u, p) (H 1 0 (Ω)) d L 2 0(Ω) such that

Examples: Incompressible flow problems Example 2: the generalized Oseen problem Let Ω be a domain in R d and let α 0 and ν > 0. Also, let w be a divergence-free vector field on Ω. Consider the system αu ν u + (w ) u + p = f in Ω, div u = 0 in Ω, u = 0 on Ω. Note that for w = 0 we recover the generalized Stokes problem. Weak formulation: Find (u, p) (H 1 0 (Ω)) d L 2 0(Ω) such that α u, v + ν u, v + (w ) u, v p, div v = f, v, v (H 1 0 (Ω)) d,

Examples: Incompressible flow problems Example 2: the generalized Oseen problem Let Ω be a domain in R d and let α 0 and ν > 0. Also, let w be a divergence-free vector field on Ω. Consider the system αu ν u + (w ) u + p = f in Ω, div u = 0 in Ω, u = 0 on Ω. Note that for w = 0 we recover the generalized Stokes problem. Weak formulation: Find (u, p) (H 1 0 (Ω)) d L 2 0(Ω) such that α u, v + ν u, v + (w ) u, v p, div v = f, v, v (H 1 0 (Ω)) d, q, div u = 0, q L 2 0(Ω).

Examples: Incompressible flow problems Example 2: the generalized Oseen problem Let Ω be a domain in R d and let α 0 and ν > 0. Also, let w be a divergence-free vector field on Ω. Consider the system αu ν u + (w ) u + p = f in Ω, div u = 0 in Ω, u = 0 on Ω. Note that for w = 0 we recover the generalized Stokes problem. Weak formulation: Find (u, p) (H 1 0 (Ω)) d L 2 0(Ω) such that α u, v + ν u, v + (w ) u, v p, div v = f, v, v (H 1 0 (Ω)) d, q, div u = 0, q L 2 0(Ω). The standard Oseen problem is obtained for α = 0 (steady case).

Examples: Incompressible flow problems Example 2: the generalized Oseen problem (cont.) Discretization using LBB-stable finite element pairs or other div-stable scheme leads to an algebraic saddle point problem:

Examples: Incompressible flow problems Example 2: the generalized Oseen problem (cont.) Discretization using LBB-stable finite element pairs or other div-stable scheme leads to an algebraic saddle point problem: A B T B O «u p «= f 0 «

Examples: Incompressible flow problems Example 2: the generalized Oseen problem (cont.) Discretization using LBB-stable finite element pairs or other div-stable scheme leads to an algebraic saddle point problem: A B T B O «u p «= f 0 «Now A is a discrete reaction-convection-diffusion operator. For α = 0, A is just a discrete vector convection-diffusion operator. Note that now A A T.

Examples: Incompressible flow problems Example 2: the generalized Oseen problem (cont.) Discretization using LBB-stable finite element pairs or other div-stable scheme leads to an algebraic saddle point problem: A B T B O «u p «= f 0 «Now A is a discrete reaction-convection-diffusion operator. For α = 0, A is just a discrete vector convection-diffusion operator. Note that now A A T. The Oseen problem arises from Picard iteration applied to the steady incompressible Navier Stokes equations, and from fully implicit schemes applied to the unsteady NSE. The wind w represents an approximation of the solution u obtained from the previous Picard step, or from time-lagging.

Examples: Incompressible flow problems Example 2: the generalized Oseen problem (cont.) Discretization using LBB-stable finite element pairs or other div-stable scheme leads to an algebraic saddle point problem: A B T B O «u p «= f 0 «Now A is a discrete reaction-convection-diffusion operator. For α = 0, A is just a discrete vector convection-diffusion operator. Note that now A A T. The Oseen problem arises from Picard iteration applied to the steady incompressible Navier Stokes equations, and from fully implicit schemes applied to the unsteady NSE. The wind w represents an approximation of the solution u obtained from the previous Picard step, or from time-lagging. As we will see, this problem can be very challenging to solve, especially for small values of the viscosity ν and on stretched meshes.

Examples: Incompressible flow problems Eigenvalues of discrete Oseen problem (ν = 0.01), indefinite form A B T Eigenvalues of Oseen matrix A = B O «, MAC discretization. 4 3 2 1 0 1 2 3 4 50 0 50 100 150 200 250 Note the different scales in the x and y axes.

Examples: Incompressible flow problems Eigenvalues of discrete Oseen problem (ν = 0.01), positive definite form A B T Eigenvalues of Oseen matrix A = B O «, MAC discretization. 6 4 2 0 2 4 6 0 20 40 60 80 100 120 140 160 180 200 Note the different scales in the x and y axes.

Some solution algorithms Outline 1 Properties of saddle point matrices 2 Examples: Incompressible flow problems 3 Some solution algorithms 4 The Augmented Lagrangian (AL) approach 5 The modified Augmented Lagrangian-based preconditioner 6 Conclusions

Some solution algorithms Overview of available solvers Two main classes of solvers exist: 1 Direct methods: based on factorization of A

Some solution algorithms Overview of available solvers Two main classes of solvers exist: 1 Direct methods: based on factorization of A High-quality software exists (Duff et al.; Demmel et al.)

Some solution algorithms Overview of available solvers Two main classes of solvers exist: 1 Direct methods: based on factorization of A High-quality software exists (Duff et al.; Demmel et al.) Quite popular in some areas

Some solution algorithms Overview of available solvers Two main classes of solvers exist: 1 Direct methods: based on factorization of A High-quality software exists (Duff et al.; Demmel et al.) Quite popular in some areas Stability issues (indefiniteness)

Some solution algorithms Overview of available solvers Two main classes of solvers exist: 1 Direct methods: based on factorization of A High-quality software exists (Duff et al.; Demmel et al.) Quite popular in some areas Stability issues (indefiniteness) Large amounts of fill-in

Some solution algorithms Overview of available solvers Two main classes of solvers exist: 1 Direct methods: based on factorization of A High-quality software exists (Duff et al.; Demmel et al.) Quite popular in some areas Stability issues (indefiniteness) Large amounts of fill-in Not feasible for 3D problems

Some solution algorithms Overview of available solvers Two main classes of solvers exist: 1 Direct methods: based on factorization of A High-quality software exists (Duff et al.; Demmel et al.) Quite popular in some areas Stability issues (indefiniteness) Large amounts of fill-in Not feasible for 3D problems Difficult to parallelize

Some solution algorithms Overview of available solvers Two main classes of solvers exist: 1 Direct methods: based on factorization of A High-quality software exists (Duff et al.; Demmel et al.) Quite popular in some areas Stability issues (indefiniteness) Large amounts of fill-in Not feasible for 3D problems Difficult to parallelize 2 Krylov subspace methods (MINRES, GMRES, Bi-CGSTAB,...)

Some solution algorithms Overview of available solvers Two main classes of solvers exist: 1 Direct methods: based on factorization of A High-quality software exists (Duff et al.; Demmel et al.) Quite popular in some areas Stability issues (indefiniteness) Large amounts of fill-in Not feasible for 3D problems Difficult to parallelize 2 Krylov subspace methods (MINRES, GMRES, Bi-CGSTAB,...) Appropriate for large, sparse problems

Some solution algorithms Overview of available solvers Two main classes of solvers exist: 1 Direct methods: based on factorization of A High-quality software exists (Duff et al.; Demmel et al.) Quite popular in some areas Stability issues (indefiniteness) Large amounts of fill-in Not feasible for 3D problems Difficult to parallelize 2 Krylov subspace methods (MINRES, GMRES, Bi-CGSTAB,...) Appropriate for large, sparse problems Tend to converge slowly

Some solution algorithms Overview of available solvers Two main classes of solvers exist: 1 Direct methods: based on factorization of A High-quality software exists (Duff et al.; Demmel et al.) Quite popular in some areas Stability issues (indefiniteness) Large amounts of fill-in Not feasible for 3D problems Difficult to parallelize 2 Krylov subspace methods (MINRES, GMRES, Bi-CGSTAB,...) Appropriate for large, sparse problems Tend to converge slowly Number of iterations increases as problem size grows

Some solution algorithms Overview of available solvers Two main classes of solvers exist: 1 Direct methods: based on factorization of A High-quality software exists (Duff et al.; Demmel et al.) Quite popular in some areas Stability issues (indefiniteness) Large amounts of fill-in Not feasible for 3D problems Difficult to parallelize 2 Krylov subspace methods (MINRES, GMRES, Bi-CGSTAB,...) Appropriate for large, sparse problems Tend to converge slowly Number of iterations increases as problem size grows Effective preconditioners a must

Some solution algorithms Overview of available solvers Two main classes of solvers exist: 1 Direct methods: based on factorization of A High-quality software exists (Duff et al.; Demmel et al.) Quite popular in some areas Stability issues (indefiniteness) Large amounts of fill-in Not feasible for 3D problems Difficult to parallelize 2 Krylov subspace methods (MINRES, GMRES, Bi-CGSTAB,...) Appropriate for large, sparse problems Tend to converge slowly Number of iterations increases as problem size grows Effective preconditioners a must Much effort has been put into developing preconditioners, with optimality and robustness w.r.t. parameters as the ultimate goals. Parallelizability also needs to be taken into account.

Some solution algorithms Preconditioners Preconditioning: Find an invertible matrix P such that Krylov methods applied to the preconditioned system P 1 A x = P 1 b will converge rapidly (possibly, independently of the discretization parameter h).

Some solution algorithms Preconditioners Preconditioning: Find an invertible matrix P such that Krylov methods applied to the preconditioned system P 1 A x = P 1 b will converge rapidly (possibly, independently of the discretization parameter h). In practice, fast convergence is typically observed when the eigenvalues of the preconditioned matrix P 1 A are clustered away from zero. However, it is not an easy matter to characterize the rate of convergence, in general.

Some solution algorithms Preconditioners Preconditioning: Find an invertible matrix P such that Krylov methods applied to the preconditioned system P 1 A x = P 1 b will converge rapidly (possibly, independently of the discretization parameter h). In practice, fast convergence is typically observed when the eigenvalues of the preconditioned matrix P 1 A are clustered away from zero. However, it is not an easy matter to characterize the rate of convergence, in general. To be effective, a preconditioner must significantly reduce the total amount of work:

Some solution algorithms Preconditioners Preconditioning: Find an invertible matrix P such that Krylov methods applied to the preconditioned system P 1 A x = P 1 b will converge rapidly (possibly, independently of the discretization parameter h). In practice, fast convergence is typically observed when the eigenvalues of the preconditioned matrix P 1 A are clustered away from zero. However, it is not an easy matter to characterize the rate of convergence, in general. To be effective, a preconditioner must significantly reduce the total amount of work: Setting up P must be inexpensive

Some solution algorithms Preconditioners Preconditioning: Find an invertible matrix P such that Krylov methods applied to the preconditioned system P 1 A x = P 1 b will converge rapidly (possibly, independently of the discretization parameter h). In practice, fast convergence is typically observed when the eigenvalues of the preconditioned matrix P 1 A are clustered away from zero. However, it is not an easy matter to characterize the rate of convergence, in general. To be effective, a preconditioner must significantly reduce the total amount of work: Setting up P must be inexpensive Evaluating z = P 1 r must be inexpensive

Some solution algorithms Preconditioners Preconditioning: Find an invertible matrix P such that Krylov methods applied to the preconditioned system P 1 A x = P 1 b will converge rapidly (possibly, independently of the discretization parameter h). In practice, fast convergence is typically observed when the eigenvalues of the preconditioned matrix P 1 A are clustered away from zero. However, it is not an easy matter to characterize the rate of convergence, in general. To be effective, a preconditioner must significantly reduce the total amount of work: Setting up P must be inexpensive Evaluating z = P 1 r must be inexpensive Convergence must be rapid

Some solution algorithms Preconditioners Options include:

Some solution algorithms Preconditioners Options include: 1 ILU preconditioners

Some solution algorithms Preconditioners Options include: 1 ILU preconditioners 2 Coupled multigrid methods (geometric and algebraic; Vanka-type)

Some solution algorithms Preconditioners Options include: 1 ILU preconditioners 2 Coupled multigrid methods (geometric and algebraic; Vanka-type) 3 Schur complement-based methods ( segregated approach)

Some solution algorithms Preconditioners Options include: 1 ILU preconditioners 2 Coupled multigrid methods (geometric and algebraic; Vanka-type) 3 Schur complement-based methods ( segregated approach) Block diagonal preconditioning

Some solution algorithms Preconditioners Options include: 1 ILU preconditioners 2 Coupled multigrid methods (geometric and algebraic; Vanka-type) 3 Schur complement-based methods ( segregated approach) Block diagonal preconditioning Block triangular preconditioning (Elman et al.)

Some solution algorithms Preconditioners Options include: 1 ILU preconditioners 2 Coupled multigrid methods (geometric and algebraic; Vanka-type) 3 Schur complement-based methods ( segregated approach) Block diagonal preconditioning Block triangular preconditioning (Elman et al.) Uzawa, SIMPLE,...

Some solution algorithms Preconditioners Options include: 1 ILU preconditioners 2 Coupled multigrid methods (geometric and algebraic; Vanka-type) 3 Schur complement-based methods ( segregated approach) Block diagonal preconditioning Block triangular preconditioning (Elman et al.) Uzawa, SIMPLE,... 4 Constraint preconditioning ( null space methods )

Some solution algorithms Preconditioners Options include: 1 ILU preconditioners 2 Coupled multigrid methods (geometric and algebraic; Vanka-type) 3 Schur complement-based methods ( segregated approach) Block diagonal preconditioning Block triangular preconditioning (Elman et al.) Uzawa, SIMPLE,... 4 Constraint preconditioning ( null space methods ) 5 Augmented Lagrangian-based techniques (AL)

Some solution algorithms Preconditioners Options include: 1 ILU preconditioners 2 Coupled multigrid methods (geometric and algebraic; Vanka-type) 3 Schur complement-based methods ( segregated approach) Block diagonal preconditioning Block triangular preconditioning (Elman et al.) Uzawa, SIMPLE,... 4 Constraint preconditioning ( null space methods ) 5 Augmented Lagrangian-based techniques (AL) The choice of an appropriate preconditioner is highly problem-dependent.

Some solution algorithms Preconditioners (cont.) Example: The Silvester Wathen preconditioner for the Stokes problem is

Some solution algorithms Preconditioners (cont.) Example: The Silvester Wathen preconditioner for the Stokes problem is P = ba O O M b p! where b A 1 is given by a multigrid V-cycle applied to linear systems with coefficient matrix A and b M p is the diagonal of the pressure mass matrix. This preconditioner is provably optimal:

Some solution algorithms Preconditioners (cont.) Example: The Silvester Wathen preconditioner for the Stokes problem is P = ba O O M b p! where b A 1 is given by a multigrid V-cycle applied to linear systems with coefficient matrix A and b M p is the diagonal of the pressure mass matrix. This preconditioner is provably optimal: MINRES preconditioned with P converges at a rate independent of the mesh size h

Some solution algorithms Preconditioners (cont.) Example: The Silvester Wathen preconditioner for the Stokes problem is P = ba O O M b p! where b A 1 is given by a multigrid V-cycle applied to linear systems with coefficient matrix A and b M p is the diagonal of the pressure mass matrix. This preconditioner is provably optimal: MINRES preconditioned with P converges at a rate independent of the mesh size h Each preconditioned MINRES iteration costs O(n + m) flops

Some solution algorithms Preconditioners (cont.) Example: The Silvester Wathen preconditioner for the Stokes problem is P = ba O O M b p! where b A 1 is given by a multigrid V-cycle applied to linear systems with coefficient matrix A and b M p is the diagonal of the pressure mass matrix. This preconditioner is provably optimal: MINRES preconditioned with P converges at a rate independent of the mesh size h Each preconditioned MINRES iteration costs O(n + m) flops Efficient parallelization is possible

Some solution algorithms Preconditioners (cont.) Example: The Silvester Wathen preconditioner for the Stokes problem is P = ba O O M b p! where b A 1 is given by a multigrid V-cycle applied to linear systems with coefficient matrix A and b M p is the diagonal of the pressure mass matrix. This preconditioner is provably optimal: MINRES preconditioned with P converges at a rate independent of the mesh size h Each preconditioned MINRES iteration costs O(n + m) flops Efficient parallelization is possible But what about more difficult problems?

Some solution algorithms Block preconditioners If A is invertible, A has the block LU factorization

Some solution algorithms Block preconditioners If A is invertible, A has the block LU factorization ««A B T I O A B T A = = B O BA 1 I O S where S = BA 1 B T (Schur complement). «,

Some solution algorithms Block preconditioners If A is invertible, A has the block LU factorization ««A B T I O A B T A = = B O BA 1 I O S where S = BA 1 B T (Schur complement). Let «A O A B T P D =, P O S T = O S then j The spectrum of P 1 A is σ(p 1 A) = D D 1, 1± 5 2 ff «, «,

Some solution algorithms Block preconditioners If A is invertible, A has the block LU factorization ««A B T I O A B T A = = B O BA 1 I O S where S = BA 1 B T (Schur complement). Let «A O A B T P D =, P O S T = O S then j The spectrum of P 1 A is σ(p 1 A) = D D 1, 1± 5 2 ff «, «, The spectrum of P 1 A is σ(p 1A) = {1} T T

Some solution algorithms Block preconditioners If A is invertible, A has the block LU factorization ««A B T I O A B T A = = B O BA 1 I O S where S = BA 1 B T (Schur complement). Let «A O A B T P D =, P O S T = O S then j The spectrum of P 1 A is σ(p 1 A) = D D 1, 1± 5 2 ff «, «, The spectrum of P 1 A is σ(p 1A) = {1} T T GMRES converges in three iterations with P D, and in two iterations with P T.

Some solution algorithms Block preconditioners (cont.) In practice, it is necessary to replace A and S with easily invertible approximations:

Some solution algorithms Block preconditioners (cont.) In practice, it is necessary to replace A and S with easily invertible approximations:!! ba O ba B T P D = O S b, P T = O S b

Some solution algorithms Block preconditioners (cont.) In practice, it is necessary to replace A and S with easily invertible approximations:!! ba O ba B T P D = O S b, P T = O S b ba should be spectrally equivalent to A: that is, we want cond( b A 1 A) c for some constant c independent of h

Some solution algorithms Block preconditioners (cont.) In practice, it is necessary to replace A and S with easily invertible approximations:!! ba O ba B T P D = O S b, P T = O S b ba should be spectrally equivalent to A: that is, we want cond( b A 1 A) c for some constant c independent of h Often a small, fixed number of multigrid V-cycles will do

Some solution algorithms Block preconditioners (cont.) In practice, it is necessary to replace A and S with easily invertible approximations:!! ba O ba B T P D = O S b, P T = O S b ba should be spectrally equivalent to A: that is, we want cond( b A 1 A) c for some constant c independent of h Often a small, fixed number of multigrid V-cycles will do Approximating S is more involved, except in special situations; for example, in the case of Stokes we can use the pressure mass matrix ( b S = M p) or its diagonal, assuming the LBB condition holds. This is the Silvester Wathen preconditioner.

Some solution algorithms Block preconditioners (cont.) In practice, it is necessary to replace A and S with easily invertible approximations:!! ba O ba B T P D = O S b, P T = O S b ba should be spectrally equivalent to A: that is, we want cond( b A 1 A) c for some constant c independent of h Often a small, fixed number of multigrid V-cycles will do Approximating S is more involved, except in special situations; for example, in the case of Stokes we can use the pressure mass matrix ( b S = M p) or its diagonal, assuming the LBB condition holds. This is the Silvester Wathen preconditioner. For the Oseen problem this does not work, except for very small Reynolds.

Some solution algorithms Block preconditioners (cont.) Recall that S = BA 1 B T is a discretization of the operator S = div( ν + w ) 1

Some solution algorithms Block preconditioners (cont.) Recall that S = BA 1 B T is a discretization of the operator S = div( ν + w ) 1 A plausible (if non-rigorous) approximation of the inverse of this operator is bs 1 := 1 ( ν + w ) p where the subscript p indicated that the convection-diffusion operator acts on the pressure space. Hence, the action of S 1 can be approximated by a matrix-vector multiply with a discrete pressure convection-diffusion operator, followed by a Poisson solve.

Some solution algorithms Block preconditioners (cont.) Recall that S = BA 1 B T is a discretization of the operator S = div( ν + w ) 1 A plausible (if non-rigorous) approximation of the inverse of this operator is bs 1 := 1 ( ν + w ) p where the subscript p indicated that the convection-diffusion operator acts on the pressure space. Hence, the action of S 1 can be approximated by a matrix-vector multiply with a discrete pressure convection-diffusion operator, followed by a Poisson solve. This is known as the pressure convection-diffusion preconditioner (PCD), introduced and analyzed by Kay, Loghin, and Wathen (SISC, 2001).

Some solution algorithms Block preconditioners (cont.) Recall that S = BA 1 B T is a discretization of the operator S = div( ν + w ) 1 A plausible (if non-rigorous) approximation of the inverse of this operator is bs 1 := 1 ( ν + w ) p where the subscript p indicated that the convection-diffusion operator acts on the pressure space. Hence, the action of S 1 can be approximated by a matrix-vector multiply with a discrete pressure convection-diffusion operator, followed by a Poisson solve. This is known as the pressure convection-diffusion preconditioner (PCD), introduced and analyzed by Kay, Loghin, and Wathen (SISC, 2001). This preconditioner performs well for small or moderate Reynolds numbers.

Some solution algorithms Results for Kay, Loghin and Wathen preconditioner Test problems: steady Oseen, homogeneous Dirichlet BCs, two choices of the wind function.

Some solution algorithms Results for Kay, Loghin and Wathen preconditioner Test problems: steady Oseen, homogeneous Dirichlet BCs, two choices of the wind function. A constant wind problem: w = 1 0 «

Some solution algorithms Results for Kay, Loghin and Wathen preconditioner Test problems: steady Oseen, homogeneous Dirichlet BCs, two choices of the wind function. A constant wind problem: w = 1 0 A recirculating flow (vortex) problem: w = «4(2y 1)(1 x)x 4(2x 1)(1 y)y «

Some solution algorithms Results for Kay, Loghin and Wathen preconditioner Test problems: steady Oseen, homogeneous Dirichlet BCs, two choices of the wind function. A constant wind problem: w = 1 0 A recirculating flow (vortex) problem: w = «4(2y 1)(1 x)x 4(2x 1)(1 y)y Uniform FEM discretizations: isop2-p0 and isop2-p1. These discretizations satisfy the inf-sup condition: no pressure stabilization is needed. SUPG stabilization is used for the velocities. «

Some solution algorithms Results for Kay, Loghin and Wathen preconditioner Test problems: steady Oseen, homogeneous Dirichlet BCs, two choices of the wind function. A constant wind problem: w = 1 0 A recirculating flow (vortex) problem: w = «4(2y 1)(1 x)x 4(2x 1)(1 y)y Uniform FEM discretizations: isop2-p0 and isop2-p1. These discretizations satisfy the inf-sup condition: no pressure stabilization is needed. SUPG stabilization is used for the velocities. The Krylov subspace method used is Bi-CGSTAB. This method requires two matrix-vector multiplies with A and two applications of the preconditioner at each iteration. «

Some solution algorithms Results for Kay, Loghin and Wathen preconditioner Test problems: steady Oseen, homogeneous Dirichlet BCs, two choices of the wind function. A constant wind problem: w = 1 0 A recirculating flow (vortex) problem: w = «4(2y 1)(1 x)x 4(2x 1)(1 y)y Uniform FEM discretizations: isop2-p0 and isop2-p1. These discretizations satisfy the inf-sup condition: no pressure stabilization is needed. SUPG stabilization is used for the velocities. The Krylov subspace method used is Bi-CGSTAB. This method requires two matrix-vector multiplies with A and two applications of the preconditioner at each iteration. A preconditioning step requires two convection-diffusion solves (three in 3D) and one Poisson solve at each iteration, plus some mat-vecs. «

Some solution algorithms Results for Kay, Loghin, and Wathen preconditioner (cont.) mesh size h constant wind viscosity ν 1 0.1 0.01 0.001 0.0001 1/16 6 / 12 8 / 16 12 / 24 30 / 34 100 / 80 1/32 6 / 10 10 / 16 14 / 24 24 / 28 86 / 92 1/64 6 / 10 8 / 14 16 / 24 22 / 32 64 / 66 1/128 6 / 10 8 / 12 16 / 26 24 / 36 64 / 58 rotating vortex 1/16 6 / 8 10 / 12 30 / 40 > 400 / 188 1/32 6 / 8 10 / 12 30 / 40 > 400 / 378 1/64 4 / 6 8 / 12 26 / 40 > 400 / > 400 1/128 4 / 6 8 / 10 22 / 44 228 / > 400 Number of Bi-CGSTAB iterations (Note: exact solves used throughout. Stopping criterion: b Ax k 2 < 10 6 b 2 ).

The Augmented Lagrangian (AL) approach Outline 1 Properties of saddle point matrices 2 Examples: Incompressible flow problems 3 Some solution algorithms 4 The Augmented Lagrangian (AL) approach 5 The modified Augmented Lagrangian-based preconditioner 6 Conclusions

The Augmented Lagrangian (AL) approach Augmented Lagrangian formulation of saddle point problems Consider the equivalent augmented Lagrangian formulation (Fortin, Glowinski, 1982) given by «««A + γb T W 1 B B T u f + γb T W 1 g =, (3) B O p g where γ > 0 and W is symmetric positive definite.

The Augmented Lagrangian (AL) approach Augmented Lagrangian formulation of saddle point problems Consider the equivalent augmented Lagrangian formulation (Fortin, Glowinski, 1982) given by «««A + γb T W 1 B B T u f + γb T W 1 g =, (3) B O p g where γ > 0 and W is symmetric positive definite. Letting A γ := A + γb T W 1 B and f γ := f + γb T W 1 g, «««Aγ B T u fγ =, or A b x = ˆb. (4) B O p g

The Augmented Lagrangian (AL) approach Augmented Lagrangian formulation of saddle point problems Consider the equivalent augmented Lagrangian formulation (Fortin, Glowinski, 1982) given by «««A + γb T W 1 B B T u f + γb T W 1 g =, (3) B O p g where γ > 0 and W is symmetric positive definite. Letting A γ := A + γb T W 1 B and f γ := f + γb T W 1 g, «««Aγ B T u fγ =, or A b x = ˆb. (4) B O p g B. and Olshanskii introduced the following block preconditioner for (4) Aγ B T «P = O S b, S b 1 = νm b p 1 γw 1. (5)

The Augmented Lagrangian (AL) approach Analysis for the Oseen problem Theorem (B./Olshanskii, SISC 2006) Setting W = M p, the preconditioned matrix P 1 b A has the eigenvalue 1 of multiplicity n; the remaining m eigenvalues are contained in a rectangle in the right half plane with sides independent of the mesh size h, and bounded away from 0. Moreover, for γ = O(ν 1 ) the rectangle does not depend on ν. When γ, all the eigenvalues tend to 1.