Outline. 1 Partial Differential Equations. 2 Numerical Methods for PDEs. 3 Sparse Linear Systems

Similar documents
Scientific Computing: An Introductory Survey

Outline. 1 Boundary Value Problems. 2 Numerical Methods for BVPs. Boundary Value Problems Numerical Methods for BVPs

Scientific Computing: An Introductory Survey

Kasetsart University Workshop. Multigrid methods: An introduction

Index. higher order methods, 52 nonlinear, 36 with variable coefficients, 34 Burgers equation, 234 BVP, see boundary value problems

AMS526: Numerical Analysis I (Numerical Linear Algebra for Computational and Data Sciences)

Finite difference methods for the diffusion equation

Partial Differential Equations

7 Hyperbolic Differential Equations

PDE Solvers for Fluid Flow

Finite Difference Methods for Boundary Value Problems

Multigrid Methods and their application in CFD

Additive Manufacturing Module 8

Iterative Methods and Multigrid

AMS526: Numerical Analysis I (Numerical Linear Algebra)

PDEs, part 1: Introduction and elliptic PDEs

Applied Mathematics 205. Unit III: Numerical Calculus. Lecturer: Dr. David Knezevic

Elliptic Problems / Multigrid. PHY 604: Computational Methods for Physics and Astrophysics II

Tutorial 2. Introduction to numerical schemes

2.29 Numerical Fluid Mechanics Spring 2015 Lecture 9

1. Fast Iterative Solvers of SLE

Computation Fluid Dynamics

AM 205: lecture 14. Last time: Boundary value problems Today: Numerical solution of PDEs

An Efficient Low Memory Implicit DG Algorithm for Time Dependent Problems

Scientific Computing with Case Studies SIAM Press, Lecture Notes for Unit VII Sparse Matrix

Stabilization and Acceleration of Algebraic Multigrid Method

ECE539 - Advanced Theory of Semiconductors and Semiconductor Devices. Numerical Methods and Simulation / Umberto Ravaioli

Multi-Factor Finite Differences

Computational Linear Algebra

Solving Symmetric Indefinite Systems with Symmetric Positive Definite Preconditioners

Finite Differences for Differential Equations 28 PART II. Finite Difference Methods for Differential Equations

Lecture 9 Approximations of Laplace s Equation, Finite Element Method. Mathématiques appliquées (MATH0504-1) B. Dewals, C.

Chapter 5. Methods for Solving Elliptic Equations

Aspects of Multigrid

Preface to the Second Edition. Preface to the First Edition

Multigrid absolute value preconditioning

Notes for CS542G (Iterative Solvers for Linear Systems)

High Performance Nonlinear Solvers

A STUDY OF MULTIGRID SMOOTHERS USED IN COMPRESSIBLE CFD BASED ON THE CONVECTION DIFFUSION EQUATION

Solving PDEs with Multigrid Methods p.1

Chapter 7 Iterative Techniques in Matrix Algebra

Diffusion / Parabolic Equations. PHY 688: Numerical Methods for (Astro)Physics

Solving linear systems (6 lectures)

Numerical Methods for PDEs

Chapter Two: Numerical Methods for Elliptic PDEs. 1 Finite Difference Methods for Elliptic PDEs

AIMS Exercise Set # 1

Boundary Value Problems - Solving 3-D Finite-Difference problems Jacob White

The Conjugate Gradient Method

Time-dependent variational forms

Preliminary Examination, Numerical Analysis, August 2016

Numerical Solutions to Partial Differential Equations

Adaptive algebraic multigrid methods in lattice computations

ME Computational Fluid Mechanics Lecture 5

A Linear Multigrid Preconditioner for the solution of the Navier-Stokes Equations using a Discontinuous Galerkin Discretization. Laslo Tibor Diosady

Implicit Solution of Viscous Aerodynamic Flows using the Discontinuous Galerkin Method

Introduction to Partial Differential Equations

Lecture 18 Classical Iterative Methods

An Introduction to the Discontinuous Galerkin Method

Qualifying Examination

Constrained Minimization and Multigrid

Part 1. The diffusion equation

Domain decomposition on different levels of the Jacobi-Davidson method

First, Second, and Third Order Finite-Volume Schemes for Diffusion

Iterative Methods for Linear Systems

The method of lines (MOL) for the diffusion equation

Chapter 3 Second Order Linear Equations

Notes on PCG for Sparse Linear Systems

Numerical Methods for Engineers and Scientists

Numerical Methods for PDEs

Bindel, Fall 2016 Matrix Computations (CS 6210) Notes for

2.2. Methods for Obtaining FD Expressions. There are several methods, and we will look at a few:

Iterative Methods and Multigrid

Numerical Analysis of Differential Equations Numerical Solution of Parabolic Equations

x n+1 = x n f(x n) f (x n ), n 0.

Time stepping methods

Introduction to PDEs and Numerical Methods: Exam 1

Understand the existence and uniqueness theorems and what they tell you about solutions to initial value problems.

2.29 Numerical Fluid Mechanics Spring 2015 Lecture 13

Applied Numerical Analysis

INTRODUCTION TO PDEs

Finite Difference and Finite Element Methods

AMS 147 Computational Methods and Applications Lecture 17 Copyright by Hongyun Wang, UCSC

Algebraic Multigrid as Solvers and as Preconditioner

Iterative Solvers. Lab 6. Iterative Methods

Spectral Processing. Misha Kazhdan

AMS526: Numerical Analysis I (Numerical Linear Algebra) Lecture 23: GMRES and Other Krylov Subspace Methods; Preconditioning

MULTIGRID METHODS FOR NONLINEAR PROBLEMS: AN OVERVIEW

256 Summary. D n f(x j ) = f j+n f j n 2n x. j n=1. α m n = 2( 1) n (m!) 2 (m n)!(m + n)!. PPW = 2π k x 2 N + 1. i=0?d i,j. N/2} N + 1-dim.

The amount of work to construct each new guess from the previous one should be a small multiple of the number of nonzeros in A.

An Introduction to Algebraic Multigrid (AMG) Algorithms Derrick Cerwinsky and Craig C. Douglas 1/84

Some notes about PDEs. -Bill Green Nov. 2015

SOLVING SPARSE LINEAR SYSTEMS OF EQUATIONS. Chao Yang Computational Research Division Lawrence Berkeley National Laboratory Berkeley, CA, USA

Linear Systems of Equations. ChEn 2450

1 Extrapolation: A Hint of Things to Come

Solving Sparse Linear Systems: Iterative methods

Solving Sparse Linear Systems: Iterative methods

Numerical Analysis of Differential Equations Numerical Solution of Elliptic Boundary Value

An Efficient Algorithm Based on Quadratic Spline Collocation and Finite Difference Methods for Parabolic Partial Differential Equations.

Linear Solvers. Andrew Hazel

Finite Differences: Consistency, Stability and Convergence

Transcription:

Outline Partial Differential Equations Numerical Methods for PDEs Sparse Linear Systems 1 Partial Differential Equations 2 Numerical Methods for PDEs 3 Sparse Linear Systems Michael T Heath Scientific omputing 2 / 105

Partial Differential Equations Numerical Methods for PDEs Sparse Linear Systems Partial Differential Equations Partial Differential Equations haracteristics lassification Partial differential equations (PDEs) involve partial derivatives with respect to more than one independent variable Independent variables typically include one or more space dimensions and possibly time dimension as well More dimensions complicate problem formulation: we can have pure initial value problem, pure boundary value problem, or mixture of both Equation and boundary data may be defined over irregular domain Michael T Heath Scientific omputing 3 / 105

Partial Differential Equations Numerical Methods for PDEs Sparse Linear Systems Partial Differential Equations haracteristics lassification Partial Differential Equations, continued For simplicity, we will deal only with single PDEs (as opposed to systems of several PDEs) with only two independent variables, either two space variables, denoted by x and y, or one space variable denoted by x and one time variable denoted by t Partial derivatives with respect to independent variables are denoted by subscripts, for example u t = @u/@t u xy = @ 2 u/@x@y Michael T Heath Scientific omputing 4 / 105

Partial Differential Equations Numerical Methods for PDEs Sparse Linear Systems lassification of PDEs Partial Differential Equations haracteristics lassification Order of PDE is order of highest-order partial derivative appearing in equation For example, advection equation is first order Important second-order PDEs include Heat equation : u t = u xx Wave equation : u tt = u xx Laplace equation : u xx + u yy =0 Michael T Heath Scientific omputing 8 / 105

Partial Differential Equations Numerical Methods for PDEs Sparse Linear Systems lassification of PDEs, continued Partial Differential Equations haracteristics lassification Second-order linear PDEs of general form au xx + bu xy + cu yy + du x + eu y + fu+ g =0 are classified by value of discriminant b 2 4ac b 2 b 2 b 2 4ac > 0: hyperbolic (eg, wave equation) 4ac =0: parabolic (eg, heat equation) 4ac < 0: elliptic (eg, Laplace equation) Michael T Heath Scientific omputing 9 / 105

Partial Differential Equations Numerical Methods for PDEs Sparse Linear Systems lassification of PDEs, continued Partial Differential Equations haracteristics lassification lassification of more general PDEs is not so clean and simple, but roughly speaking Hyperbolic PDEs describe time-dependent, conservative physical processes, such as convection, that are not evolving toward steady state Parabolic PDEs describe time-dependent, dissipative physical processes, such as diffusion, that are evolving toward steady state Elliptic PDEs describe processes that have already reached steady state, and hence are time-independent Michael T Heath Scientific omputing 10 / 105

Partial Differential Equations Numerical Methods for PDEs Sparse Linear Systems Time-Dependent Problems Time-Dependent Problems Time-Independent Problems Time-dependent PDEs usually involve both initial values and boundary values Michael T Heath Scientific omputing 11 / 105

Partial Differential Equations Numerical Methods for PDEs Sparse Linear Systems Time-Dependent Problems Time-Dependent Problems Time-Independent Problems Time-dependent PDEs usually involve both initial values and boundary values Time Space Michael T Heath Scientific omputing 11 / 105

Example: Poisson Equation in 2D @ 2 u @x + @2 u 2 @y 2 = f(x, y)in @ u = 0on@ u =0 Ex 1: If f(x, y) =sin x sin y, Ex 2: If f(x, y) =1, u(x, y) = 1 sin x sin y 2 2 u(x, y) = 1,1 X k,l odd 16 sin k x sin l y 2 kl(k 2 + l 2 ) Q: How large must k and l be for exact solution to be correct to M? Spectral collocation would yield u = u exact ± M by N 15

Numerical Solution: Finite Di erences n y +1 4 3 2 u i 1,j u i,j+1 u ij u i,j 1 u i+1,j @ 2 u @x 2 + @2 u @y 2 - ui+1,j 2u ij u i 1,j x 2 + u i,j+1 2u ij u i,j 1 y 2 = f ij 1 i =1 n x j =0 i =0 1 2 3 4 n x +1 5-point finite-di erence stencil j =1 n y Here, the unknowns are u =[u 11,u 21,,u nx,n y ] T This particular (so-called natural or lexicographical) ordering gives rise to abandedsystemmatrixforu As in the 1D case, the error is O( x 2 )+O( y 2 )=O(h 2 )ifwetake x = y =: h Assuming for simplicity that N = n x = n y,wehaven = N 2 unknowns

For i, j 2 [1,,N] 2, the governing finite di erence equations are ui+1,j 2u i,j + u i 1,j x 2 + u i,j+1 2u i,j + u i,j 1 y 2 = f ij Assuming a lexicographical ordering in which the i- (x-) index advances fastest, the system matrix has the form 1 h 2 0 @ 4 1 1 1 4 1 1 1 1 1 4 1 1 4 1 1 1 4 1 1 1 1 1 4 1 1 1 1 4 1 1 1 4 1 1 1 4 1 A {z } A 0 @ u 11 u 21 u N1 u 12 u 22 u N2 u 1N u 2N u NN 1 A {z } u = 0 @ f 11 f 21 f N1 f 12 f 22 f N2 f 1N f 2N f NN 1 A {z } f

The system matrix A is sparse, with 5 nonzeros per row (good) and has a bandwith N (bad) The di culty is that solving Au = f using Gaussian elimination results in signifcant fill eachofthefactorsl and U have N 3 = n 3/2 nonzeros Worse, for 3D problems with N 3 unknowns, u =[u 111,u 211,,u nx,n y,n z ] T, A is sparse, with 7 nonzeros per row (good) and has a bandwith N 2 (awful) In 3D, LU decomposition yields N 5 =n 5/3 nonzeros in L and U The situation can be rescued in 2D with a reordering of the unknowns (eg, via nesteddissection) to yield O(n log n) nonzerosinl and U In 3D, nested-dissection yields O(n 3/2 ) nonzeros in the factors Direct solution is not scalable for more than two space dimensions The following Matlab examples illustrate the issue of fill: fd poisson 2dm fd poisson 3dm

Matrix-Fill for 2D and 3D Poisson, Lexicographical Ordering nnz 3D As expected, the error scales like h 2 1/N 2 in both 2D and 3D The resepctive storage costs (and work per rhs) are N 3 and N 5 nnz 2D Alternative orderings are asymptotically better, but the constants tend to be large error 2D,3D

Matrix-Fill for 2D Poisson, symamd Ordering N 3 nnz 2D error 2D We see for N =80(n =6400)a5 reduction in number of nonzeros by reording with matlab s symamd function The requirements for indirect addressing to access elements of the complacty-stored matrix further adds to overhead Gains tend to be realized only for very large N and are even less beneficial in 3D Despite this, it s still a reasonable idea to reorder in matlab because it s available and easy to use

Iterative Solvers The curse of dimensionality for d>2 resulted in a move towards iterative (rather than direct-, LU-based) linear solvers once computers became fast enough to tackle 3D problems in the mid-80s With iterative solvers, factorization is replaced by, say, Au = f =) u = A 1 f = U 1 L 1 f u k+1 = u k + M 1 (f Au k ), which only requires matrix-vector products With e k := u u k, we have e k+1 = I M 1 A e k, (as we ve seen before) This is known as Richardson iteration For the particular case M = D =diag(a), it is Jacobi iteration We can derive Jacobi iteration (and multigrid by looking at a parabolic PDE, known as the (unsteady) heat equation (The Poisson equation is sometimes referred to as the steady-state heat equation)

The intrinsic advantage of iterative solvers is that there is no fill associated with matrix factorization Often one does not even construct the matrix Rather, we simply evaluate the residual r k := f Au k and set u k+1 = u k + M 1 r k For a sparse matrix A, the operation count is O(n) periteration Assuming the preconditioner cost is also sparse, the overall cost is O(nk max ), where k max is the number of iterations required to reach a desired tolerance The choice of iteration (Richardson, conjugate gradient, GMRES) can greatly influence k max Even more significant is the choice of M Usually, one seeks an M such that the cost of solving Mz = r is O(n) andthat k max = O(1) That is, the iteration count is bounded, independent of n The overall algorithm is therefore O(n), which is optimal

Iterative Solvers - Linear Elliptic Problems PDEs give rise to large sparse linear systems of the form Au = f Here, we ll take A to be the (SPD) matrix arising from finite di erences applied to the Poisson equation @ 2 u @x + @2 u = f(x, y) x, y 2 [0, 1] 2, u =0on@ 2 @y 2 2 u x 2 + 2 u y 2 ij f ij, Assuming uniform spacing in x and y we have 2 u x := u i+1,j 2u ij + u i 1,j 2 h 2 and 2 u y := u i,j+1 2u ij + u i,j 1 2 h 2

Our finite di erence formula is thus, 1 h 2 (u i+1,j + u i 1,j 4u ij + u i,j+1 + u i,j 1 )=f ij Rearranging, we can solve for u ij : 4 h 2u ij = f ij + 1 h 2 (u i+1,j + u i 1,j + u i,j+1 + u i,j 1 ) u ij = h2 4 f ij + + 1 4 (u i+1,j + u i 1,j + u i,j+1 + u i,j 1 )

Jacobi iteration uses the preceding expression as a fixed-point iteration: u k+1 ij = h2 4 f ij + 1 4 uk i+1,j + u k i 1,j + u k i,j+1 + u k i,j 1 Note that this is analogous to u k+1 ij = u k ij + h2 4 = h2 4 f ij + average of current neighbor values apple f ij + 1 h 2 u k i+1,j + u k i 1,j 4u k ij + u k i,j+1 + u k i,j 1 u k+1 = u k + t (f Au k ), t := h2 4, which is Euler forward applied to du dt = Au + f We note that we have stability if t < 2

Recall that the eigenvalues for the 1D di usion operator are j = 2 h 2 (1 cos j x) < 4 h 2 In 2D, we pick up contributions from both 2 u x 2 and 2 u y 2, so and we have stability since max < 8 h 2 max t < 8 h 2 h 2 4 =2 So, Jacobi iteration is equivalent to solving Au = f by time marching du dt = Au + f using EF with maximum allowable timestep size, t = h2 4

Jacobi Iteration in Matrix Form Our unsteady heat equation has the matrix form u k+1 = u k + t (f Au k ) For variable diagonal entries, Richardson iteration is u k+1 = u k + M 1 (f Au k ) If =1andM = D 1 =diag(a) [d ii =1/a ii, d ij =0,i6= j], we have standard Jacobi iteration If < 1 we have damped Jacobi M is generally known as a smoother or a preconditioner, depending on context

Rate of onvergence for Jacobi Iteration Let e k := u u k Since Au = f, wehave u k+1 = u k + t (Au Au k ) u = u e k+1 = e k tae k e k+1 = (I ta) e k e k = (I ta) k e 0 = (I ta) k u if u 0 =0 If < 1, then the high wavenumber error components will decay because t will be well within the stability region for EF

The low-wavenumber components of the solution (and error) evolve like e tk, because these will be well-resolved in time by Euler forward Thus, we can anticipate If with min 2 2 (for 2D) 1, we have e k u e min tk e k u e 2 2 (h 2 /4)k apple tol Example, find the number of iterations when tol=10 12 e ( 2 h 2 /4)k 10 12 ( 2 h 2 /4)k ln 10 12 24 (276) k 28 2 2 h 2 6N 2 Here, N=number of points in each direction

Recap Low-wavenumber components decay at a fixed rate: e min tk Stability mandates t<h 2 /4=1/4(N +1) 2 Number of steps scales like N 2 Note, if = 1, thenhighest and lowest wavenumber components decay at same rate If 1 2 < < 1, high wavenumber components of error decay very fast We say that damped Jacobi iteration is a smoother

Example: 1D Jacobi Iteration Solution after 1 iteration Error after 1 iteration Solution after 5 itierations Error after 5 itierations

Observations: Error, e k is smooth after just a few iterations: Error components are û j e j2 kh 2 2 /4 sin k x j, and components for j>1rapidlygotozero Exact solution is u = u k + e k (e k unknown, but smooth) Error satisfies, and can be computed from, Ae k = r k (:= f Au k = Au Au k = Ae k ) These observations suggest that the error can be well approximated on a coarser grid and added back to u k to improve the current guess The two steps, smooth and coarse-grid correction are at the heart of one of the fastest iteration strategies, known as multigrid

Multigrid: Solve Ae k = r k approximately on a coarse grid and set ũ k = u k + ẽ k Approximation strategy is similar to least squares Let ẽ k = V e c, and AV e c r, where V is an n n c matrix with n c n/2 Typically, columns of V interpolate coarse point values to their midpoints Most common approach (for A SPD) is to require e c to solve V T [AV e c r] = 0 =) ẽ k = V V T AV 1 V T r = V V T AV 1 V T A e k For A SPD, ẽ k is the A-orthogonal projection of e k onto R(V )

An example of V for n =5andn c =2 is V = 2 6 6 6 6 4 1 2 1 1 2 1 2 1 1 2 3 7 7 7 7 5 + + + + + i =0 1 2 3 4 5 n x +1 oarse-to-fine interpolation poisson_mgm demo

Example: Damped Jacobi (Richardson) Iteration Solution after 1 iteration Error after 1 iteration Solution after 5 itierations Error after 5 itierations

Multigrid Summary Main Ideas Solution after 5 iterations Error after 5 iterations Take a few damped-jacobi steps (smoothing the error), to get u k Approximate this smooth error, e k := u u k, on a coarser grid Exact error satisfies Ae k = Au Au k = f Au =: r k Let e f := V e c be the interpolant of e c, the coarse-grid approximation to e k e f is closest element in R(V )toe k (in the A-norm), given by the projection: e f = V V T AV 1 V T Ae k = V A c ) 1 V T r k Update u k with the coarse-grid correction: u k u k + e f Smooth again and repeat

Example: Two-Level Multigrid Solution after 1 iteration Error after 1 iteration Solution after 5 itierations Error after 5 itierations

Example: Two-Level Multigrid Solution after 1 iteration Error after 1 iteration Iteration History Error after 5 itierations

Multigrid omments Smoothing can be improved using under-relaxation ( ¾ = 2/3 is optimal for 1D case) asically want more of the high-end error spectrum to be damped System in A c is less expensive to solve, but is typically best solved by repeating the smooth/ coarse-grid correct pair on yet another level down an recur until n c ~ 1, at which point system is easy to solve Typical MG complexity is O(n) or O(n log n), with very good constants in higher space dimensions (N c = N/2 à n c = n/8 in 3D) For high aspect-ratio cells, variable coefficients, etc, smoothing and coarsening strategies require more care, so this continues to be an active research area

Stability Region for Euler s Method Unstable Stable -2-1 Region where 1 + h < 1

Growth Factors for Real e t G G t t t Each growth factor approximates e t for t à 0 For EF, G is not bounded by 1 For Trapezoidal Rule, local (small t) approximation is O( t 2 ), but G à -1 as t à -1 [ Trapezoid method is not L-stable ] DF2 will give 2 nd -order accuracy, stability, and G à0 as t à -1

Time Dependent Problems We ll consider two examples: di usion (heat equation) and advection heat equation: advection: @u @t @u @t = @2 u @x 2 = c @u @x + s and I + s and I

Heat Equation: @u @t = @2 u @x, > 0 2 For the heat equation, the solution evolves in the direction of local curvature If the the solution is locally concave down, u decreases there If the the solution is concave up, u increases @ 2 u @x 2 @u < 0 =) @t < 0 @ 2 u @x 2 > 0 +

Example Solutions (eigenfunctions): u t = u xx,u(0) = u(1) = 0 u(x, t) = û(t)sin x @u @t = dû dt dû dt sin x = 2 = 2 û û = e 2t û(0) û sin x u(x, t) = û(t)sin10 x @u @t = dû dt dû dt sin x = 100 2 = 100 2 û û = e 100 2t û(0)! Very rapid decay û sin x (1) + 05(2)

Solution of Partial Di erential Equations Unsteady Heat Equation: u t = u xx + q(x, t), u(x =0,t)=u(x = L, t) =0, u(x, t =0)=u 0 (x) Discretize in space: Finite di erence Weighted residual technique (FEM, Galerkin + high-order polynomials, etc)

Finite Di erence ase: u j 1 u j u j+1 0=:x 0 x 1 x 2 x j 1 x j x j+1 x n+1 := 1 In ODE form: du i dt du dt = (Au) i + q i, i =1,,n = Au + q, u(t =0)=u 0 0 1

Here, Eigenvalues: x =1/(n +1)andA is the SPD tridiagonal matrix 0 1 2 1 A = 1 1 2 1 x 2 1 @ 1 A 1 2 (A) = 2 x (1 cos(k x)) 2 2 2 (1 + O( x 2 )), 4(n +1) 2 2 2 (1 + O( x 2 4 )), x 2

an view this semi-discrete form as a system of ODEs: du dt = f(u) := Au + q(x,t) Jacobian df i du j = a ij J = A Stability is determined by the eigenvalues of J and by the choice of timestepper Some possible explicit timesteppers EF: u k+1 = u k + t f k 23 A3: u k+1 = u k + t 12 f k 16 12 f k 1 + 5 12 f k 2 Stable, as long as (J) t in the stability region

Stability: (J) = (A) = 2 (1 cos k x) x2 Worst case is (J) 4 x 2 For Euler forward (EF), require Stability Regions, EF, A2, A3 t (J) < 2 or t < 2 x2 4 = x2 2, which is a very severe timestep restriction Question: What is the maximum allowable timestep size for A3 in this case?

Question: What is the maximum allowable timestep size for A3 in this case? Stability Regions, EF, A2, A3

Severity of explicit timestep restriction: Suppose =1andyouwanterror 10 6! x 10 3! t 10 6, just for stability This is an example of a sti system High wavenumbers ( (A)) are uninteresting but restrict the timestep size For this reason, the heat equation is most often treated implicitly

Possible Implicit Approaches: 8 < Examples: du dt = f(u) : E Trapezoid (aka rank-nicolson) DF2 or DF3 E: u k+1 = u k + t Au k+1 + q(x,t k+1 ) N: u k+1 t u k = + 1 2 Au k+1 + q k+1 Au k + q k DF2: 3u k+1 4u k + u k 1 2 t = Au k+1 + q(x,t k+1 )

E Example: u k+1 + tau k+1 = u k + tq k+1 [I + ta] u k+1 = u k + tq k+1 Hu k+1 = u k + tq k+1 Here, H := [I + ta] is SPD, tridiagonal, and strongly diagonally dominant (In all number of space dimensions) Hu = f is easier to solve than Au = f Jacobi- (diagonal-) preconditioned conjugate gradient iteration is often the best choice of solver, particularly in higher space dimensions Note that all the implicit solvers end up with the form Hu = f and generally have the same costs for the linear heat equation considered here Note that N (aka trapezoid method) is not L-stable and will have potential di culties noted in our discussion of IVPs

Discretization ased on Weighted Residual Technique in Space oming back to the heat equation (with s/is), u t = u xx + q(x, t), WRT - residual orthogonal to test functions Z v( u xx + q(x, t) u t ) dx = 0 8 vx N 0 If u = LHS: nx j=1 Z u j (t) j (x) and v = v @u @t dx = nx j=1 i i (x), then j dx! u j (t) = du dt, with the mass matrix having entries Z ij := i(x) j (x) dx

On the right, we have Z RHS = = Z v @2 u @x dx + 2 Z @v @u @x @x dx + vq dx Z vq dx Setting v = i and u = P j ju j (t), RHS = nx j=1 Z d i dx d i dx dx u j (t) + Z iqdx = Au + b, 8 >< >: Z d i d i a ij := Z dx dx dx b i := iqdx

In summary, the WRT formulation is, Find u(x, t) 2 X0 N such that, Z v @u Z Z @v @t dx = @u @x @x dx + vq dx 8 v 2 X0 N, which leads to the ODE du dt In standard form, = Au + b, plus initial condition u(t =0)=u 0 du dt = 1 Au + 1 b, Stability is thus governed by (J) = ( 1 A), not just (A) Presence of in front of du dt must not be ignored hoice of timestepper motivated by same concerns as for finite-di erences: (J) O( x 2 ) Implicit timestepping generally preferred SPD systems Jacobi (diagonal) preconditioned conjugate gradient iteration is generally the solver of choice

Time Stepping for Di usion Equation: Recall, with boundary conditions u(0) = u(1) = 0, the finite di erence operator Au = h 2 [u j+1 u j u j 1 ] with h := 1/(n + 1) has eigenvalues in the interval [0,M] with M = max k k = max k 2 h 2 [1 cos k h] 4 h 2 Our ODE is u t = Au, so we are concerned with k With Euler Forward, we require t < 2forstability,! t< h2 2 no matter how smooth the initial condition This intrinsic sti ness motivates the use of implicit methods for the heat equation (DF2 is a good one) heat1d_efm matlab example: and heat1d_ebm heat1dm

Partial Differential Equations Numerical Methods for PDEs Sparse Linear Systems Example: Advection Equation Partial Differential Equations haracteristics lassification Advection equation u t = cu x where c is nonzero constant Unique solution is determined by initial condition u(0,x)=u 0 (x), 1 <x<1 where u 0 is given function defined on R We seek solution u(t, x) for t 0 and all x 2 R From chain rule, solution is given by u(t, x) =u 0 (x ct) Solution is initial function u 0 shifted by ctto right if c>0, or to left if c<0 Michael T Heath Scientific omputing 5 / 105

Partial Differential Equations Numerical Methods for PDEs Sparse Linear Systems Example, continued Partial Differential Equations haracteristics lassification Typical solution of advection equation, with initial function advected (shifted) over time < interactive example > Michael T Heath Scientific omputing 6 / 105

haracteristics Partial Differential Equations Numerical Methods for PDEs Sparse Linear Systems Partial Differential Equations haracteristics lassification haracteristics for PDE are level curves of solution For advection equation u t = straight lines of slope c cu x, characteristics are haracteristics determine where boundary conditions can or must be imposed for problem to be well-posed Michael T Heath Scientific omputing 7 / 105

Matlab Demo: onvection

Matlab Demo: onvection

Time Stepping for Advection Equation: @u @t = c @u @x Unlike the di usion equation, which smears out the initial condition (with high wavenumber components decaying particularly fast), the advection equation simply moves things around, with no decay This property is evidenced by the spatial operator having purely (or close to purely) imaginary eigenvalues Preserving high-wavenumber content (in space) for all time makes this problem particularly challenging There is always some spatial discretization error Its e ects accumulate over time (with no decay of the error) For su ciently large final time T any fixed grid (ie, fixed n) simulation for general problems will eventually have too much error Long time-integrations, therefore, typically require relatively fine meshes and/or high-order spatial discretizations

FL, Eigenvalues, and Stability: Fourier Analysis onsider: u t = cu x, u(0) = u(1) (periodic s) entered di erence formula in space: du j dt = c 2 x (u j+1 u j 1 ) = u j = 1 2 x 2 6 4 3 0 1 1 1 0 1 1 1 7 5 1 1 0 Periodic Matrix

Periodic Domain Periodicity: u 0 u n u j 1 u j u j+1 x 0 x 1 x 2 x j 1 x j x j+1 x n Allows us to run for long times without having to have averylongdomain Allows us to analyze the properties of our spatial discretization

FL, Eigenvalues, and Stability: Fourier Analysis onsider: u t = cu x, u(0) = u(1) (periodic s) entered di erence formula in space: du j dt = c 2 x (u j+1 u j 1 ) = u j Eigenvector: u j = e i2 kx j Eigenvalue: u j = c 2 x ei2 k x e i2 k x e i2 kx j = 2ic 2 x e i2 k x e i2 k x 2i u j = k u j k = ic x sin(2 k x)

FL, Eigenvalues, and Stability: Fourier Analysis Eigenvalue: u j = = c 2 x ei2 k x e i2 k x e i2 kx j 2ic 2 x = k u j e i2 k x e i2 k x 2i u j k = ic x sin(2 k x) Eigenvalues are purely imaginary, max modulus is For constant c and equation as max k k = c x x, wedefinetheflfortheadvection FL = t c x ourant Number

ourant Number, Eigenvalues, and Stability: Fourier Analysis For constant c and x, wedefinetheflfortheadvection equation as FL = t c x FL=1 would correspond to a timestep size where a particle moving at speed c would move one grid spacing in a single timestep For centered finite di erences in space, FL=1 also corresponds t =1 From our IVP stability analysis, we know that we need t <7236 for A3 and < 2828 for RK4 This would correspond to FL <7236 and 2828, respectively

FL, Eigenvalues, and Stability: Fourier Analysis MATLA EXAMPLE: conv_ab3m

Advection For advection, no decay in physical solution Solution is persistent Numerical method is either dispersive, dissipative, or both If = T, discrete operator is skew-symmetric (imaginary eigenvalues) and numerical method has no decay (due to spatial error, at least) ut it will be dispersive We come back to dissipative shortly

Long time-integration! accumulation of error Second-order, O( x 2 ), accuracy is not su cient Modulo boundary conditions (or with periodicity), we can easily extend our 2nd-order centered-di erence formula to O( x 4 )throughrichardson extrapolation Let and h u j := c 2 x [u j+1 u j 1 ] 2h u j := c 4 x [u j+2 u j 2 ] for j =1,,n (with wrap for periodic ends)

Instead of now use du dt = du dt = h u apple 4 3 hu 1 3 2hu For A3, say, 23 u k+1 = u k + t 12 f k 16 12 f k 1 + 5 12 f k 2 apple 4 f k = 3 hu k 1 3 2hu k Don t re-evaluate f k 1 or f k 2 Just re-use the previously computed values

Numerical Dissipation Numerical Dissipation So far, we ve consider only central di erence formulas Upwind discretizations o er more stability, through the introduction of numerical dissipation You must be very careful about the wind direction!