TMA4212 Numerical solution of partial differential equations with finite difference methods

Size: px
Start display at page:

Download "TMA4212 Numerical solution of partial differential equations with finite difference methods"

Transcription

1 TMA4 Numerical solution of partial differential equations with finite difference methods Brynjulf Owren January 3, 07 Translated and amended by E. Celledoni

2 Preface The preparation of this note started in the winter of 004. The note is a teaching aid for the first half of the course TMA40 "Numerical solution of partial differential equations with difference methods". In the winter of 006 the note was updated and several new sections were added to adapt it to the course TMA4. I want to thank all the students who followed the courses during these semesters. They have been a source of inspiration for the writing and they helped me in the correction of many typos and mistakes in the earlier versions of this note. i

3 ii

4 Contents Introduction Background material 3. Background on matrix theory Jordan form Symmetric matrices Positive definite matrices Gershgorin s theorem Vector and matrix norms Consistent and subordinate matrix norms Matrix norms and spectral radius Difference formulae Taylor expansion Big O-notation Difference approximations to the derivatives Difference operators and other operators Differential operator Boundary value problems 5 3. A simple case example norm stability for the case example Neumann boundary conditions Linear boundary value problems A self-adjoint case A nonlinear example Discretization of the heat equation 7 4. On the derivation of the heat equation Numerical solution of the initial/boundary value problem Numerical approximation on a grid Euler, Backward Euler and Crank Nicolson Solution of the linear systems in Backward Euler s method and Crank Nicolson Solution of linear systems in Matlab The θ-method Semi-discretization Semi-discretization of the heat equation Semidiscretization principle in general General approach iii

5 iv CONTENTS u t = Lu with different choices of L Boundary conditions involving derivatives Different types of boundary conditions Discretization of the boundary conditions Nonlinear parabolic differential equations Stability, consistency and convergence Properties of the continuous problem Convergence of a numerical method Domain of dependence of a numerical method Proof of convergence for the Euler s method on the (I/BVP) with r Stability on unbounded time interval (F -stability) Stability on [0, T ] when h 0, k Stability and roundoff error Consistency and Lax equivalence theorem von Neumann s stability criterion Elliptic differential equations Elliptic equation on the plane Difference methods derived using Taylor series Discretization of a self-adjoint equation Boundary conditions of Neumann and Robin type Grid-like net and variable step-size General rectangular net Discretization using Taylor expansion on a completely general net Difference formulae derived by integration Net based on triangles Difference equations Convergence of the methods for elliptic equations Convergence for the 5-point formula on a Dirichlet problem Some general comments on convergence A discussion on the solution of large linear systems Hyperbolic equations 8 7. Examples Characteristics Explicit difference formulae for u t + au x = Stability Implicit methods for u t + au x = Hyperbolic systems of first order equations Dissipation and dispersion

6 Chapter Introduction The numerical approximation of partial differential equations is an important component in the simulation of natural processes. Examples where simulation techniques are useful are chemical processes, fluid mechanics, structural dynamics, quantum physical processes, electromagnetism, finance, etc. When we talk about partial differential equations (PDEs) we mean equations where the solution is a function (or a vector of functions) of at least two variables, which are called independent variables. The equation describes a relation involving the solution and its partial derivatives. But the specification of a mathematical model in applications involves much more than just this relation. First of all the model is associated to a geometry. This means that we specify a domain in the space of the independent variables where the differential equation should be satisfied. This domain can be finite or infinite. In two dimensions such domain can be a subset of the plane, but also the surface of a cylinder or a sphere. Typically a PDE in itself has infinitely many solutions. The specification of a PDE on a domain must be supplemented with the specification of boundary conditions. There are many different types of boundary conditions, but in general they specify the solution or its derivatives on the boundary of the domain. If one of the independent variables is physical time, then the boundary conditions at the initial time are called initial conditions, and this is just a particular type of boundary conditions. Most PDEs are such that it is not possible to write their solution in a closed form. To understand the behavior of the solutions, it is therefore necessary to use approximation methods and computer simulations. There are several numerical methods which can be used for this purpose, among them we will focus mostly on difference methods. Other techniques are spectral methods and finite element methods. These can be studied in other courses (TMA40). Partial differential equations can be linear or nonlinear. Since this is a introductory course we will mostly consider the linear case. Linear PDEs are divided in three subclasses, parabolic, elliptic and hyperbolic differential equations. The PhD course MA803 considers nonlinear PDEs with particular focus on hyperbolic PDEs. A typical prototype of parabolic partial differential equation is the heat equation, this is the main subject of chapter 3. In this chapter we work most of the time with independent variables, time and one space dimension. For this reason the discussion about the geometry of the problems is somewhat limited in this chapter. In chapter 4 we consider general topics regarding difference approximation of PDEs which are of interest for all the three types of PDEs. This chapter is about convergence of the numerical approximations to the exact solution of the PDEs. The concepts of stability

7 CHAPTER. INTRODUCTION and consistency will be also introduced as important tools for showing convergence of a numerical scheme. In chapter 5 we discuss elliptic equations and typical examples in this case are the Laplace s equation and the Poisson s equation. We work with -dimensional domains, where both the independent variables are space variables. In this case it is interesting to look at more complex geometries compared to the parabolic case, and the domain is not necessarily rectangular like in chapter 3. As a consequence most of the discussion is about how to approximate different boundary conditions for domains which are not rectangular. The first chapter of this note reviews some background material on matrix theory and Taylor expansion, in this chapter we will also set the notation on differential operators.

8 Chapter Background material. Background on matrix theory Let A be a n n-matrix with real (or complex) entries, we write A R n n (or A C n n ). We say that A is diagonalizable if it exist a matrix X C n n such that λ Λ = X λ AX =... λn. Every λ i C is called egenvalue of A. The matrix X has n columns denoted by X = [x,..., x n ], and every x i C n is called egenvector (associated to the egenvalue λ i ). We write a diagonal matrix as Λ above, in the following form Λ = diag(λ,..., λ n )... Jordan form For any A R n n (or A C n n ) it exists a matrix M C n n such that J M J AM = J =..., (block-diagonal). (.) Jk Here J i is a m i m i -matrix, and k i= m i = n. The Jordan-blocks J i have the form λ i. J i = λ.. i..., if m i λi and J i = [λ i ] if m i =. If all m i =, then k = n and the matrix is diagonalizable. If A has n distinct egenvalues, it is always diagonalizable. The converse is not true, that is a matrix can be diagonalizable even if it has multiple eigenvalues. 3

9 4 CHAPTER. BACKGROUND MATERIAL.. Symmetric matrices When we talk about symmetric matrices, we mean normally real symmetric matrices. The transpose A T of a m n-matrix A, is a n m-matrix with a ji as the (ij)-element (a matrix whose columns are the rows of A). A n n matrix is symmetric if A T = A. A symmetric n n matrix has real eigenvalues λ,..., λ n and a set of real orthonormal eigenvectors x,..., x n. Let, denote the standard inner-product on C n, then x i, x j = δ ij (Kronecker-delta). A consequence of this is that the matrix of eigenvectors X = [x,..., x n ] is real and orthogonal and its inverse is therefore the transpose The diagonalization of A is given by X = X T. Λ = diag(λ,..., λ n ), X = [x,..., x n ], X T X = I, X T AX = Λ A = XΛX T..3 Positive definite matrices If A is symmetric and x, Ax = x T Ax > 0 for all 0 x R n A is called positive definite. A (symmetric) is positive semi-definite if x, Ax 0 for all x R n and x, Ax = 0 for at least a x 0. A positive definite A has only positive eigenvalues. A positive semi-definite A has only non-negative eigenvalues, and at least a 0-eigenvalue...4 Gershgorin s theorem Gershgorin s theorem. Is given A = (a ik ) C n n. Define n disks S j in the complex plane by S j = z C : z a jj a jk. k j The union S = n S j contains all the eigenvalues of A. For every eigenvalue λ of A there j= is a j such that λ S j. Example. A = + i

10 .. BACKGROUND ON MATRIX THEORY 5 3 Im z Re z, Proof of Gershgorin s theorem: Let λ be a eigenvalue with associate eigenvector x = [ξ,..., ξ n ] T 0. Choose l among the indexes,..., n such that ξ l ξ k, k =,..., n, and so ξ l > 0. The equation Ax = λx has component l: n k= a lk ξ k = λ ξ l (λ a ll )ξ l = k l a lk ξ k Divide by ξ l on each side and take the absolute value λ a ll = ξ k a lk ξ l a lk ξ k ξ l Then we get λ S l. k l k l k l a lk Example. Diagonally dominant matrices with positive diagonal elements are positive definite. Why?..5 Vector and matrix norms Consider a vector space X (real or complex). A norm : X R satisfies the following axioms. x 0 for all x, x = 0 x = 0.. αx = α x (α R (C)) 3. x + y x + y Examples. x = (ξ k ), X = R n. x = ( n n ) / ξ k, x = ξ k, x = max ξ k. k n k= k=

11 6 CHAPTER. BACKGROUND MATERIAL The matrix spaces R n n and C n n are also vector spaces over R (C). We say that is a matrix norm if for all A, B R n n (C n n ). A > 0 for all A, A = 0 A = 0,. αa = α A, (α R (C)), 3. A + B A + B, 4. A B A B. Remark. The last point requires that a matrix-matrix product is defined (this operation is not defined in a general vector space). In abstract terms the axioms 4 give an example of Banach-algebra. Example. The Frobenius-norm of a matrix is defined as n A F = j= k= n a jk..6 Consistent and subordinate matrix norms A given matrix norm is consistent with a given vector norm on R n if / Ax A x for all A R n n, x R n. A given matrix norm is subordinate to a given vector norm on R n if A = max x 0 Ax x. Examples. We give here as examples some of the most common subordinate matrix norms. We look for matrix norms subordinate to the three vector norms, and.. Let be the matrix norm subordinate to the vector norm. One can show that A R n n (C n n ) is n A = max a ik. k n In other words we can say that A is the maximal column-sum in A.. To find the matrix norm subordinate to the vector norm we must define the spectral radius of a matrix M R n n (C n n ). If λ,..., λ n are the eigenvalues of M, we denote the spectral radius of M by ρ(m), and it is defined as i=. ρ(m) = max k n λ k. (.) If we plot the eigenvalues of M in the complex plane, the spectral radius is the minimal radius of a circle centered in the origin and containing all eigenvalues of M. We define now the -norm of a matrix A as A = ρ(a T A).

12 .. BACKGROUND ON MATRIX THEORY 7 Note that A T A is positive (semi)definite, so all the eigenvalues are real and positive. Taking the square root of the biggest eigenvalue, we obtain A. Note also that the spectral radius of A can be very different from (the square root of ) the spectral radius of A T A. On the other hand if A is symmetric then A = ρ(a). 3. Let be the matrix norm subordinate to the vector norm. We have A = max i n k= n a ik. That is A is the maximal row-sum in A. Observe also that A = A T...7 Matrix norms and spectral radius. For any matrix norm it is true that A ρ(a). (.3) Proof: Let x be a eigenvector of A associated to an eigenvalue λ such that Ax = λx. Let y C n be arbitrary. Then we have A(xy T ) = (Ax)y T = λ(xy T ), such that As a consequence A(xy T ) A xy T. λ xy T = λ(xy T ) = A(xy T ) A xy T. Therefore λ A, and since this is true for every eigenvalue of A, it must be ρ(a) A. Question to the reader: What is wrong with the following line of reasoning λ x = λ x = A x A x etc. Convergent matrix. A matrix A is said to be convergent (to zero) if A k 0 when k. A sufficient criterion. If A < for a particular matrix norm, A is convergent. Proof. A k = A A k A A k A k 0 if A < Necessary and sufficient criterion. A is convergent if and only if the spectral radius ρ(a), defined by (.), satisfies ρ(a) <.

13 8 CHAPTER. BACKGROUND MATERIAL Proof: We use Jordan form, and let A = MJM where M C n n and J is like in (.). Then we have A = MJM MJM = MJ M, and by induction we get A k = MJ k M. Now A k 0 if and only if J k 0. And J k 0 if and only if every Jordan block Ji k 0. Assume such a Jordan block has diagonal element λ and the m i m i -matrix F has its (j, j +) elements, for j =,..., m i, equal to, and the other elements equal to zero. Then J i = λi + F where I is the identity matrix. The matrix F is nihilpotent, i.e. F m = 0, m n. We assume that k n and compute J k i = (λi + F ) k = k m=0 ( ) n k λ k m F m = m m=0 ( ) n k λ k m F m = m m=0 ϕ (m) k (λ)f m where ϕ k (λ) = λ k /k!. When k then ϕ (m) k (λ) 0 for 0 m n if and only if λ <. This must be true for all Jordan blocks (i.e. eigenvalues of A) and this concludes the proof.. Difference formulae.. Taylor expansion free variable. Let u C n+ (I) where I R is a interval of the real line. This means that the n + -th derivative of u exists and is continuous on the interval I. Then the following formula is valid. Taylor s formula with reminder. With x I, x + h I is u(x + h) = n m=0 h m m! u(m) (x) + r n where r n = hn+ (n + )! u(n+) (x + θh), 0 < θ <. free variables. Assume now that u C n+ (Ω) where Ω R. It is convenient to use a operatornotation for the partial derivatives. We write h = [h, k], and let h := h x + k y i.e. h u = h u x + k u y The operator produces the derivative of a function in the direction h = [h, k], and we find the directional derivative. We can also define powers of the operator by for example ( (h ) = h x + k ) = h y x + hk x y + k y The extension to the m-th power is obvious. then we can write

14 .. DIFFERENCE FORMULAE 9 Taylor s formula with reminder for functions of two variables. u(x + h, y + k) = n m=0 ( h m! x + k ) m u(x, y) + r n (.4) y where r n = ( h (n + )! x + k ) n+ u(x + θh, y + θk), 0 < θ <. y We have here assumed that the line segment between (x, y) and (x + h, y + k) is included in Ω. Derivation of the previous formula. We look at a function of one variable µ(t) = u(x + th, y + tk) for fixed x, y, h, k. Using Taylor s expansion with reminder for the case of one variable for µ(t) around t = 0, we obtain the two variables formula by setting t =... Big O-notation Let φ be a function of h and p a positive integer. Then we have φ(h) = O(h p ) when h 0 if there exist two constants C, H > 0 such that φ(h) C h p when 0 < h < H. If this holds, we say that φ(h) is of order p in the variable h. The typical use of the big O-notation is in connection with the local truncation error in numerical methods. For example in the Taylor expansion in one variable r n = h n+ (n + )! u(n+) (x + θh) M (n + )! h n+, M = max y I u(n+) (y) where we know that the maximum exists if I is a closed, and bounded interval. So in this case we have r n = O(h n+ ). Note that with the definition above for positive integers p we have φ(h) = O(h p+ ) φ(h) = O(h p ). Therefore sometimes when we write φ(h) = O(h p ) we mean that p is the biggest possible integer such that this is true. Often it is convenient to write O(h p ) in formulae with sums, like for example in the Taylor expansion of u above we replace r n with O(h n+ ) such that u(x + h) = n k=0 h k k! u(m) (x) + O(h n+ ). In general we have that φ(h) = O(h p φ) and ψ(h) = O(h p ψ), then ψ(h) + φ(h) = O(h q ), where q = min(p φ, p ψ ). But sometimes one can get higher powers. An obvious example is when φ(h) = h, ψ(h) = h 3 h, each of them is O(h ) but their sum is O(h 3 ). If you multiply a function φ(h) by a constant ( 0), the order does not change.

15 0 CHAPTER. BACKGROUND MATERIAL..3 Difference approximations to the derivatives We introduce a grid on R i.e. a monotone sequence of real numbers {x n } where x n R. x n x n x n+ Assume that u(x) is a given function, u C q (I), for a q which we will specify later. Let u n := u(x n ), u (m) n := u (m) (x n ). Assume the grid points x n are equidistant, i.e. x n+ = x n + h for all n, where h R is called step-size. We want to approximate u (m) n with expressions of the type q a l u n+l l=p p q are integers, and typically p 0 and q 0. Truncation error. We define τ n (h) = q l=p a l u n+l u (m) n. The strategy is to choose p and q, and then compute the q p + parameters a p,..., a q such that τ n is small. By Taylor expansion we obtain u n+l = u(x n + lh) = ν (lh) k k=0 k! u (k) n + r ν, where r ν = O(h ν+ ) and ν m, such that q ν τ n = a l k! (lh)k u (k) n u (m) n + O(h ν+ ), l=p k=0 which can be rearranged in the form ν h k q τ n = a l l k u (k) n k! k=0 l=p u (m) n + O(h ν+ ). We want that τ n = O(h r ) with r as big as possible. To approximate u (m) n we need to impose conditions on p and q. Set j := q p. In order to get consistent approximation formulae (i.e. such that τ n (h) 0 when h 0), we must require j m. We choose then a p,..., a q such that h k k! q 0 0 k m, l k a l = k = m, l=p 0 m + k j. (.5) Note that we have q p+ = j+ free parameters a p,..., a q we can use, and the conditions in (.5) must be satisfied for 0 k j, this means a total of j + conditions. The system of equations has a unique solution for h 0. If we choose a l from (.5), and assume ν j, we obtain the following truncation error τ n = ν k=j+ h k k! u(k) n q a l l k + O(h ν+ ). This methods is called the method of undetermined coefficients. l=p

16 .. DIFFERENCE FORMULAE Example. m = (u n). Choose p =, q =, j =. We want to find a, a 0, a. We write j + = 3 equations i.e. k = 0,, i (.5). k = 0 a + a 0 + a = 0 a = h k = h a + 0 a 0 + h a = a 0 = 0 k = h a + 0 a 0 + h a = 0 a = h Looking at the first terms in the local truncation error we obtain τ n = h k ( k! u(k) n h ( )k + ) h k = u (s+) n (s + )! hs. k=3 s= In the last equality, we have used the fact that the terms with even k disappear, such that we can put k = s + and let s =,,.... We have omitted the upper limit value for the index s on purpose because the number of terms we include in the remainder depend on the circumstances. Since the first term in the expression for τ n is of type h, we say that the formula is of order. Some more formulae. Other popular difference approximations are m = m = u n+ u n h u n u n h = u n +! h u n + = u n! h u n + m = Which order have these formulae? Exercises u n+ u n + u n h = u n + h u (4) n +. Assume m =, p =, q = 0 use the method of undetermined coefficients to obtain an approximation of u () n of the second order in h.. Assume m =, p =, q = use the method of undetermined coefficients to obtain an approximation of u () n of the third order in h. 3. Consider m =, p = 0, q =, so that j = q p = < m. Try to construct an approximation formula for u () n using the method of undetermined coefficients. What happens?..4 Difference operators and other operators Forward difference: Backward difference: u(x) = u(x + h) u(x). u(x) = u(x) u(x h). Central difference: δu(x) = u(x + h ) u(x h ). Mean value: µu(x) = ( u(x + h ) + u(x h )). Shift: Unity operator Eu(x) = u(x + h). u(x) = u(x).

17 CHAPTER. BACKGROUND MATERIAL Linearity. All the operators,, δ, µ, E,, are linear. This means that for α R, and with functions u(x) and v(x) we have F (αu(x) + v(x)) = αf u(x) + F v(x), where F can be any of the operators above. Let us verify this for F =. (αu(x) + v(x)) = (αu(x + h) + v(x + h)) (αu(x) + v(x)) Powers of the operators. powers of F as follows Example. = α(u(x + h) u(x)) + (v(x + h) v(x)) = α u(x) + v(x). Let F be one of the above defined operators. We can define F 0 =, δu(x) = u(x + h ) u(x h ), F k u(x) = F (F k u(x)). δ u(x) = δ(δu(x)) = δu(x + h ) δu(x h ) = u(x + h) u(x) (u(x) u(x h)), = u(x + h) u(x) + u(x h). Another interesting example is the shift operator. We observe that E k u(x) = u(x+kh). In this case it is easy to extend the definition to include all possible real powers, simply by defining E s u(x) = u(x + sh) for all s R. For example we have E u(x) = u(x h) and this is the inverse of E since Eu(x h) = E u(x + h) = u(x). Relations between the difference operators. u(x) = u(x + h) u(x) = Eu(x) u(x) = (E ) u(x), u(x) = u(x) u(x h) = u(x) E u(x) = ( E ) u(x), δu(x) = u(x + h ) u(x h ) = (E/ E / ) u(x), µu(x) = (u(x + h ) + u(x h ) ) = (E/ + E / ) u(x). In a more compact notation we have And now we have for example such that k u(x) = k l=0 = (E ), = ( E ), δ = (E / E / ), µ = (E/ + E / ). k = (E ) k = k l=0 ( ) k ( ) k l E l u(x) = l ( ) k ( ) k l E l, l k l=0 ( ) k ( ) k l u(x + lh). l

18 .. DIFFERENCE FORMULAE 3..5 Differential operator. Define Let m=0 D = d dx so that Du(x) = u (x). D m u(x) = u (m) (x). If u(x) is analytic in an interval containing x, x + h we have ( h m ) u(x + h) = m! Dm u(x) = m! (hd)m u(x) = e hd u(x). We think of this only as a notation. We have and then E = e hd. m=0 Eu(x) = e hd u(x), Relation between D and the other operators. = E = e hd = = hd +! (hd) + m= m! (hd)m, We will see that under the extra assumption that u is analytic we can make manipulations with analytic functions in the way we are used to. The meaning is always that the final result is expanded with a Taylor expansion and is interpreted as a sum of powers of operators which are applied to a smooth function. The analyticity requirement can always be relaxed by considering a Taylor expansion with reminder, and requiring the function to be differentiable only a finite number of times. We consider powers of, and we obtain or ( ) k k (hd) m = = h k D k + k m!! hk+ D k+ + m= k u(x) = h k D k u(x) + k! hk+ D k+ u(x) + showing that k /h k is a first order approximation (truncation error O(h)) of the operator D k. Note that for s R we have E s (sh) k u(x) = u(x + sh) = D k u(x) = e shd u(x) k! k=0 which reflects known computational rules. For central differences we can therefore write δ = E / E / = e hd e hd = sinh hd. By analytic function on an interval we simply mean that its Taylor expansion converges in a neighborhood of any point of the interval.

19 4 CHAPTER. BACKGROUND MATERIAL We can also compute that is δ k = ( sinh hd ) k = ( hd + 3! ( ) ) hd 3 k + = (hd) k + k 4 (hd)k+ + δ k u(x) = h k D k u(x) + k 4 hk+ D k+ u(x) + this shows that δ k /h k is a second order approximation of D k. In particular we find as before that δ u(x) = u(x + h) u(x) + u(x h) = h u (x) + h4 u (4) (x) + (.6) It is tempting to manipulate further with analytic functions. We have seen that We write therefore formally δ = sinh hd. D = h sinh δ. It is possible to expand sinh z in a Taylor expansion sinh z = z 6 z z5 5 z7 + so z = δ/ and by multiplying by /h we obtain D = ( δ h 4 δ δ5 5 ) 768 δ7 +. Since we know that δ k = O(h k ) we see that we can find approximations to the differential operator D of arbitrary high order by including enough terms in the expansion. The manipulation we have carried out is not rigorously justified here, but it turns out to be correct. For a detailed discussion on algebraic manipulations with differential operators, see the textbook by Arieh Iserles, A first course in the numerical analysis of differential equations, published by Cambridge University Press.

20 Chapter 3 Boundary value problems 3. A simple case example We consider the boundary value problem u xx = f(x), 0 < x <, u(0) = α, u() = β, (3.) the exact solution can be obtained by integrating twice on both sides between 0 and, and then imposing the boundary conditions. We want to use this simple test problem to illustrate some of the basic features of finite difference discretization methods. To obtain a finite difference discretization for this problem we consider the grid x m = m h, m = 0,..., M +, h = M +, and the notation u m := u(x m ) such that u 0 = α and u M+ = β. We denote with capital letters U m u m the numerical approximation to u(x) at the grid point x = x m. By replacing derivatives with central difference approximations to the left hand side of (3.) we obtain the so called discrete problem whose solution is the numerical approximation that we are seeking, this is This is a linear system of equations h (U m U m + U m+ ) = f m, m =,..., M. (3.) A h U = F, (3.3) where A h is a M M matrix U R M and F R M and A h := h , U := U. U M, F := f α h f. f M f M β h. We know that u (x m ) = h δ u(x m ) + O ( h ) and we want to deduce similar information about the error e m := U m u m, e 0 = 0, e M+ = 0. 5

21 6 CHAPTER 3. BOUNDARY VALUE PROBLEMS Let the error vector e h be e h := U u, u := u. u M. We will in the sequel associate such vector to a piecewise constant function e h (x) defined on the interval [0, ] as follows e h (x) = e m, x [x m, x m+ ), m =,..., M. Because we are approximating a function, u(x) solution of (3.), it is appropriate to think of the numerical solution as a piecewise constant function approximating u(x) and similarly for the error. We are therefore interested in measuring the norm of this piecewise constant error function rather than the norm of the corresponding error vector, however the two are closely related. In fact we can see that the following relationships hold: e h = max 0 x e h (x) = max m M e m = e h ; e h = 0 e h(x) dx = M xm+ m= x m e h (x) dx = h M m= e m = h e h ; e h = ( ) 0 e h(x) dx = ( h ) M m= e m = h e h ; and we see that in these three popular cases the vector norm is related to the function norm of the corresponding piecewise constant function (with respect to the assumed grid) simply by a scaling factor. A similar result is true for the case of q. Truncation error. The truncation error is the vector that by definition has components τ m := h (u m u m + u m+ ) f m, m =,..., M, τ h := By using that u (x m ) = δ u(x h m ) h u (4) m + O ( h 4) and u m = f m, we obtain τ m = u m + h u (4) m + O ( h 4) f m = h u (4) m + O ( h 4). Equation for the error. The relationship between the error e h and the truncation error is easily obtained: recall that by definition τ h = A h u F, rearranging and subtracting this from A h U = F we obtain the important relation which can be also written componentwise as τ. τ M. Ae h = τ h (3.4) h (e m e m + e m+ ) = τ m, m =,..., M.

22 3.. A SIMPLE CASE EXAMPLE 7 Definition. A method for the boundary value problem (3.) is said to be consistent with the boundary value problem with respect to the norm if and only if τ h 0, when h 0. Consistence in the vector norm implies that the corresponding piecewise constant function tends to zero as h tends to zero in the corresponding function norm, this is because of the relationship between vector and function norms as we have seen for, and. Definition. A method for the boundary value problem (3.) is said to be convergent in the (function) norm if and only if e h 0, when h 0. Definition. Stability. Assume a difference method for the boundary value problem (3.) is given by the discrete equation A h U = F, where h is the step-size of discretization. The method is stable in the norm if there exist constants C > 0 and H > 0 such that. A h exists for all h < H;. A h C for all h < H. The matrix norm in which we should prove stability is the one subordinate to the chosen function/vector norm in which we want to prove convergence. Proposition. For the boundary value problem (3.), stability and consistence with respect to the norm imply convergence in the same norm. Proof. We use (3.4) to obtain a bound for the norm of the error. Since we have stability A h is invertible for all h < H and therfore Then e h = A h τ h. e h q h q A h q τ h q = A h q τ h q C τ h q, and we conclude that e h q 0 as h 0 because so does τ h q. For the case the vector norm of v R M and the function norm of the corresponding piecewise constant function defined on the grid coincide, so to include this case we can conveniently adopt the notation h := lim q h q =.

23 8 CHAPTER 3. BOUNDARY VALUE PROBLEMS 3.. -norm stability for the case example Observe that the eigenvalues of the matrix (3.3) are λ m = (cos(m π h) ), m =,..., M, h and the corresponding eigenvectors v m have components v m j = sin(m π j h), j =,..., M. Since by definition A h = ρ(a T h A h) then because A h is symmetric A h = ρ(a h ). Denote with σ(b) the spectrum of the M M matrix B (the collection of all the eigenvalues of B), then A h = max λ σ(ah ) λ. Analogously A h = ρ(a h ) = max λ σ(a h ) λ = Using the series expansion cos(x) = x for the m-th eigenvalue we get + x4 min λ σ(ah ) λ. 4! + O(x6 ), with x = m π h in the expression λ m = ( ) (m π h) (m π h)4 h + + O(h 6 ) = m π + O(h ), 4! and such that min λ = λ = π + O(h ), λ σ(a h ) A h = λ, when h 0, π and so there exist C > 0 and H > 0 such that A h = λ < C for all h < H, which proves stability of the proposed difference scheme in the -norm. Exercise. Using the estimates for A h and τ h obtained in this section, prove that assume f is twice differentiable. e h π h.5 f + O(h 3.5 ), 3.. Neumann boundary conditions We consider now the boundary value problem u xx = f(x), 0 < x <, u (0) = σ, u() = β, (3.5) and we propose three different discretizations of the left boundary condition which combined with the earlier consider discretization of the second derivative will lead to three different linear systems. In this case the matrices we obtain are no longer symmetric.

24 3.. A SIMPLE CASE EXAMPLE 9 Case We approximate u (0) = σ by U U 0 h and combine this discrete equation with (3.), this leads to the (M + ) (M + ) linear system = σ, A h U = F, where A h := h h h , U := U 0. U M, F := σ f. f M f M β h. This leads to an overall method of order of consistency. Case To obtain order two we introduce a fictitious node external to the domain to the left of x 0 which we call x. We approximate u (0) = σ by U U h = σ. We use this equation together with (3.) for m = 0, i.e. and eliminate U. We are left with the equation h (U U 0 + U ) = f 0, U U 0 h = f 0h + σ which we now combine with (3.), this leads to a (M + ) (M + ) linear system A h U = F, where A h is as in the previous case and F := f 0 h + σ f. f M f M β h.

25 0 CHAPTER 3. BOUNDARY VALUE PROBLEMS Case 3 In this discretization we want to obtain order without the use of fictitious nodes. We then consider the approximation of u (0) = σ by 3 U 0 U + U h where the left hand side gives a discretization of order two of the first derivative of u at zero. Proceeding as in case, we obtain the (M + ) (M + ) linear system A h U = F, = σ, where A h := h 3 h h h , U := U 0 U. U M, F := σ f. f M f M β h. Exercise. Prove with the method of undetermined coefficients that 3 u(0) u(h) + u(h) = u (0) + O(h ). h Neumann boundary conditions on both sides We consider now the boundary value problem u xx = f(x), 0 < x <, u (0) = σ 0, u () = σ, (3.6) and we propose a second order discretization of the boundary conditions using two fictitious nodes external to [0, ], x and x M+. The approximated boundary conditions become U U h = σ 0, U M+ U M h = σ, which we combine respectively with (3.) for m = 0 and m = M + and obtain the linear system of equations (M + ) (M + ) linear system A h U = F, where A h := h h h , U := h h U 0. U M+, F := σ 0 + h f 0 f. f M f M+ h σ. But now the matrix A h is singular, in fact it is easy to verify that v := [,..., ] T R M+ is an eigenvector of A h with eigenvalue 0. We want to determine the condition of existence of solutions for this linear system.

26 3.. A SIMPLE CASE EXAMPLE Before we do that we consider the solution of (3.6). Integrating on both sides between 0 and we obtain that is u () u (0) = σ σ 0 = 0 0 f(x) dx, f(x) dx, (3.7) this is a necessary condition for the existence of solutions of the boundary value problem (3.6). We now turn to the analysis of the solutions of the discrete boundary value problem A h U = F. Recall that Range(A h ) := span{a h e,... A h e M+ }, where we denote with e j j =,..., M+ the canonical basis of R M+ and then A h e,... A h e M+ are the columns of A h. Recall that Ker(A T h ) = {z RM+ A T h z = 0}, and Ker(A T h ) is a vector space. Proposition. A h U = F admits a solution if and only if F Range(A h ) Proposition. F Range(A h ) if and only if F Ker(A T h ). Proof The proof of these two propositions is by exercise. Since F Ker(A T h ) if and only if z R M+, s.t. A T h z = 0, FT z = 0, we need to characterize z = [z 0,..., z M+ ] T generic element of Ker(A T h ), this amounts at finding the solutions of the homogeneous linear system A T h z = 0, h h h h h z 0... z M+ 0. =.,. 0 and we find that the solution is z = [z 0, hz 0,..., hz 0, z 0 ] T R M+ for any value of z 0 R. Then F T h M z = 0 z 0 (f 0 + σ h 0) + hz 0 f m + z 0 (f M+ + σ ) = 0, m=

27 CHAPTER 3. BOUNDARY VALUE PROBLEMS for all z 0 and for z 0 0 we can divide out by z 0 to obtain the condition σ σ 0 = f 0 h + h M m= f m + f M+ h. We easily recognize that this condition is the discrete analogous of (3.7) and in fact the right hand side is the composite trapezoidal rule applied to the integral of f(x) on the discretization grid. 3. Linear boundary value problems Consider the general linear boundary value problem a(x)u (x) + b(x)u (x) + c(x)u(x) = f(x), a x b, u(a) = α, u(b) = β, a discretization of this problem on the grid x m = a + m h, h = b a M+ is a m U m U m + U m+ h + b m U m+ U m h + c m U m = f m, m =,..., M and a m := a(x m ), b m := b(x m ), c m := c(x m ) and U m u m := u(x m ). The corresponding linear system of equations has the form h h c a a + h b a h b h c a a + h b a m h b M a M + h b M h c M a M The linear system we obtain will in general be not symmetric. 3.. A self-adjoint case Suppose now we want to discretize (κ(x)u (x)) = f(x), a x b, u(a) = α, u(b) = β, U. U M = we could differentiate out the left hand side to obtain a problem of the class described above, i.e κ(x)u (x) + κ (x)u (x) = f(x) this is however not the best strategy because, as we have seen, this leads in general to non symmetric linear systems. Consider now the differential operator L defined as follows Lu(x) := (κ(x)u (x)) this operator is self-adjoint with respect to the L ([a, b]) inner product, f, g := b a f(x)g(x) dx, f, g L ([a, b]). This motivates the search for a symmetric discretization A h of L. f α( a h f. b h ) f M f M β( a M h + b M h ).

28 3.3. A NONLINEAR EXAMPLE 3 To achieve this we consider the mid-points of the grid x m x m + h, we then consider the following approximations = x m h and x m+ = and We use these two to produce κ m u m κ m+ u m+ (κ u ) xm = [κ h m u m (κ m u m u m = κ m + O(h ) h u m+ u m = κ m+ + O(h ). h + κ m+ ] )u m + κ m+ u m+ + O(h ). The methods is of second order and the linear system we have to solve is h (κ + κ 3 ) κ 3 κ 3 (κ 3 + κ 5 ) κ κm κ M (κ M + κ M+ ) U.. U M = f α κ h f. f M f M β κ M+ h. 3.3 A nonlinear example We consider the pendulum equations θ (t) + sin(θ(t)) = 0, 0 < t < T, θ(0) = α, θ(t ) = β, (3.8) with numerical discretization on the grid t m = m h, m = 0,..., M + and h = leading to the discretized equations h (Θ m Θ m + Θ m+ ) + sin(θ m ) = 0, m =,,..., M. The system of equations is now nonlinear and we denote it in a vector form as G h (Θ) = 0, Θ = Θ. Θ M, T M+ where G h (Θ) = A h Θ + sin(θ), sin(θ) := [sin(θ ),..., sin(θ M )] T and A h is the same as in (3.3). The numerical solution of this system must be addressed iteratively, for example with a Newton method: Θ [k+] = Θ [k] J h (Θ [k] ) G h (Θ [k] ), where J h (Θ) is the Jacobian of G h (Θ), and it is easily seen that J h (Θ) is a tridiagonal symmetric matrix such that for j = i or j = i +, h (J h ) i,j (Θ) = + cos(θ h i ) for j = i, 0 otherwise,

29 4 CHAPTER 3. BOUNDARY VALUE PROBLEMS so J h (Θ) = A h + C(Θ), C(Θ) := diag(cos(θ ),..., cos(θ M )). The truncation error is τ h := G h ( θ ), θ := θ. θ M, θ j := θ(t j ). It is easily shown by Taylor expansion that the components of τ h satisfy τ j = h θ (4) (t j ) + O(h 4 ), j =,..., M, this ensures that the method is consistent of order. As usual we want to use the connection between error E h := Θ θ and truncation error τ h in order to prove convergence. In general this connection is a bit less manageable in the case of nonlinear problems, but in this particular example it is not too complicated. We combine the discrete equation G h (Θ) = 0 and the equation for the truncation error G h ( θ ) = τ h to obtain G h (Θ) G h ( θ ) = τ h. (3.9) From (3.9), using G h (Θ) = A h Θ + sin(θ) we get and by Taylor expansion A h E h + sin(θ) sin( θ ) = τ h, sin(θ) = sin( θ ) + C(ˆθ)E h, where C(ˆθ) is the diagonal matrix earlier defined, and the components ˆθ i of the vector ˆθ belong to the open intervals (Θ i, θ(t i )). Due to the stability already proven for the linear case, A h is invertible for all h < H and so E h + A h C(ˆθ)E h = A h τ h. We here use the same notation for the vectors E h and τ h and the corresponding piecewise constant functions defined on the discretization grid, the norms we consider in the sequel are function norms. Assuming we operate in the -norm, we get ] E h A h [ C(ˆθ)E h + τ h A h [ max m M cos(ˆθ m ) E h + τ h so ( A h ) E h A h τ h. We know from earlier analysis that A h = λ where λ is the eigenvalue of A h with minimum absolute value. We also obtained the estimate λ = π + O(h ) and so we can also deduce that A h = + O(h ) and for h small enough A π h < and we get E h A h A h τ h. Using again the estimate for A h we can obtain A h A h = π + O(h ). ]

30 3.3. A NONLINEAR EXAMPLE 5 E h π τ h + O(h ) and recalling that τ h = h θ (4) + O(h 4 ) we get finally E h ( ) π h θ (4) + O(h 4 ) which guarantees convergence when h goes to zero, as it is easy to see that θ (4) is bounded by differentiating the equation twice.

31 6 CHAPTER 3. BOUNDARY VALUE PROBLEMS

32 Chapter 4 Discretization of the heat equation 4. On the derivation of the heat equation We will use a standard example to describe the numerical schemes for parabolic differential equations throughout the course. We are talking about the linear heat equation in one space dimension, which for example can be used to model the flow of heat on a straight homogeneous rod over time. The rod is insulated everywhere except at the two ends The rod in the picture has length L =, and let the coordinate x describes a point along the rod. At time t 0 the rod has temperature u(x, t) at the point x. We can derive the differential equation by using Fourier s law. The flux of heat φ through a cross section of the rod at x is proportional to the temperature gradient, such that φ = λu x, λ > 0, and the following conservation law holds true ρcu t + φ x = 0, ρ is the rod s density. These two equations imply together that u t = a u xx, a = λ ρc. By introducing scales for time, space and temperature w = u u 0, y = x L, τ = at L, where w, y and τ are dimensionless variables, u 0 is the characteristic temperature,and L is the rod s length, we get w τ = w yy, 0 < y <. We have seen that after scaling it is always possible to assume that the space interval is [0, ] and the coefficient a can be set equal to. From now on we will usually look at the problem u t = u xx, 0 < x <, t > 0. 7

33 8 CHAPTER 4. DISCRETIZATION OF THE HEAT EQUATION Together with the differential equation we need to provide appropriate boundary conditions and initial conditions. The kind of boundary and initial conditions necessary and sufficient to get a unique solution vary from differential equation to differential equation. We will consider few options for the heat equations. Pure initial value problem. In this case we assume the rod is infinitely long. u t = u xx, x R, t > 0, u(x, 0) = f(x), x R. Initial/Boundary value problem (I/BVP). This case includes the situation of heat transport in a homogeneous rod of length. We must consider an initial function and boundary conditions at the two ends of the rod. t u t = u xx, 0 < x <, t > 0, u(x, 0) = f(x), 0 x, u(0, t) = g 0 (t), t > 0, u = g 0 u = g u(, t) = g (t), t > 0. (4.) u = f x 4. Numerical solution of the initial/boundary value problem 4.. Numerical approximation on a grid. We adopt a step-size h in the x-direction, and one in the t-direction which we denote k. We assume at first that h = /(M + ) for a given integer M. We then define gridpoints or nodes (x m, t n ) by x m = mh, 0 m M +, t n = nk, n = 0,,,.... Observe that this means that x 0 = 0 and x M+ = are the boundary points. The solution in the point (x m, t n ) is denoted u n m := u(x m, t n ). And from now on we denote with U n m the approximation to the solution in (x m, t n ) produced by the numerical method.

34 4.. NUMERICAL SOLUTION OF THE INITIAL/BOUNDARY VALUE PROBLEM9 t : : u known u unknown 4.. Euler, Backward Euler and Crank Nicolson We present three different difference schemes for the heat equation. x The Euler s method. We adopt a simpler notation for the derivatives and set x u = u x = u x, k xu = k u x k, tu = u t = u t. We expand u n+ m = u(x m, t n + k) for constant x = x m, around t = t n, and get u n+ m = u n m + k t u n m + ϕ n m, ϕ n m = k t u n m +. But we can now use the heat equation ensuring t u n m = x u n m, we then approximate this second derivative with central differences as in (.6) u n+ m = u n m + k h δ xu n m ψ n m + ϕ n m where the index on δ means that we apply this operator in the x-direction i.e. From the expression in (.6) we find that Summarizing we have δ x u n m = u n m+ u n m + u n m. ψ n m = k h 4 x u n m +. where u n+ m = u n m + k h δ x u n m + kτ n m = u n m + k h (un m+ u n m + u n m ) + kτ n m kτ n m = ϕ n m ψ n m = ( k t ) k h x 4 u n m + the Euler s formula is obtained by replacing all the exact u-values (small letters) with approximate values U (capital letters) in the above formula and discard the term kτ n m.

35 30 CHAPTER 4. DISCRETIZATION OF THE HEAT EQUATION Euler s method m, n + (4.) U n+ m = U n m + r δ x U n m r = k h m, n m, n m +, n The picture above to the right is called computational molecule, it is a sort of local chart over the grid, indicating which grid-points are used in the formula. The idea is now to start at n = 0, corresponding to t 0 = 0 where u(x, t 0 ) = u(x, 0) = f(x) which is known. It is then possible to order the values U 0 m = f(x m ), m = 0,..., M +. Then we set n = and use first the boundary values to get U 0 = g 0(k) and U M+ = g (k). For the remaining values we use the formula (4.) above. It is possible to see that the grid-point at level t n+ = t in the computational molecule can be computed using known values. Algorithm (Euler s method for the heat equation) Um 0 := f(x m ), m = 0,..., M + for n = 0,,,... U0 n+ := g 0 (t n+ ) U n+ M+ := g (t n+ ) Um n+ := Um n + r δx Um, n m =,..., M end Example. u t = u xx 0 < x <, t > 0, { x, 0 x f(x) =, ( x), < x, g 0 (t) = g (t) = 0, t > 0. u x In the picture above we display the initial function. By running a simulation in Matlab based on this example, where we let h = 0. (M = 9), and k = The reason why we take k so small compared to h will be explained later on. Figure 4. shows the numerical solution in the grid-points as small rings, at time t = 0, 5, 0 and 0.

36 4.. NUMERICAL SOLUTION OF THE INITIAL/BOUNDARY VALUE PROBLEM3 n=0 n= n=0 n= Figure 4.: Matlab simulation of the heat equation using the Euler method Backward Euler. We expand now instead u n m around x = x m, t = t n+, and obtain where u n m = u(x m, t n+ k) = u n+ m k t u n+ m + k t u n+ m + = u n+ m k x u n+ m + k t u n+ m + = u n+ m k ( h δ x u n+ m h 4 x u n+ m + ) + k t u n+ m + = u n+ m rδ x u n+ m + kτ n m, kτ n m = ( k h x 4 + ) k t u n+ m +. By replacing the u s med U s and discard the term kτ n m we obtain Backward Euler s method U n+ m r δ x U n+ m = U n m, r = k h. (4.3) m, n + m, n + m +, n + m, n The Backward Euler method is an implicit method. This means that at each time step we must solve a system of linear equations to compute Um n+, m =,..., M. We

37 3 CHAPTER 4. DISCRETIZATION OF THE HEAT EQUATION will discuss later on the solution of linear systems. We now will present another implicit method. Crank Nicolsons method. This method is based on the trapezoidal rule, consider the following expansion of the error for the trapezoidal rule for quadrature k f(t) dt = k (f(0) + f(k)) ( ) k k3 f +. 0 To derive the method we use the obvious formula u(x m, t n+ ) u(x m, t n ) = tn+ t n u t (x m, t) dt, and approximate the integral with the trapezoidal rule where we use the notation u n+/ m = u(x m, t n + k). u n+ m = u n m + k ( tu n m + t u n+ m ) k3 3 t u n+/ m + = u n m + k ( xu n m + xu n+ m ) k3 t 3 u n+/ m + = u n m + ( k h δ xu n m + ) h δ xu n+ m ( k h xu 4 n m + ) h xu 4 n+ m + k3 3 t u n+/ m +. We simplify and summarize u n+ m = u n m + r (δ xu n m + δ xu n+ m ) + kτ n m, where we have used that kτ n m = k3 3 t u n+/ m ( 4 xu n m + xu 4 n+ m k h 4 xu n+/ m +, ) = 4 xu n+/ m + O(k ). Crank Nicolsons method m, n + m, n + m +, n + ( r δ x) U n+ m = ( + r δ x) U n m, r = k h. (4.4) m, n m, n m +, n We summarize by writing all the formulae in a compact form (E) U n+ m = ( + r δ x) U n m or k tu n m = h δ xu n m, (BE) ( r δ x) U n+ m = U n m or k tu n+ m = h δ xu n+ m, (CN) ( r δ x) U n+ m = ( + r δ x) U n m or k δ tu n+/ m = h δ xµ t U n+/ m. Note that (E) is explicit while both (BE) and (CN) are implicit.

38 4.. NUMERICAL SOLUTION OF THE INITIAL/BOUNDARY VALUE PROBLEM33 The local the truncation error. In the methods presented we have used the symbol τm, n this is the local truncation error in the point (x m, t n ). Given a finite difference formula, to find the corresponding local truncation error one inserts the exact solution evaluated in the grid points in the finite difference formula. Since the exact solution does not satisfy the finite difference formula, then an error term appears, this is the local truncation error. So for example for the forward Euler method applied to the heat equation we have the finite difference formula Um n+ Um n = U m n U m n + Um+ n k h, replacing Um n with the exact solution in (x m, t n ), u n m we get so u n+ m u n m k = un m un m + u n m+ h + τ n m, τm n := un+ m u n m un m un m + u n m k h. The local truncation error can expressed in powers of the step-sizes h and k and the derivatives of the exact solution by using the Taylor expansion Solution of the linear systems in Backward Euler s method and Crank Nicolson Our starting point is that we know Um, n 0 m M +, together with U0 n+ and U n+ M+ given by the boundary conditions. We need to compute Um n+, m M. Let us consider Crank Nicolson first. The right hand side (r.h.s.) of the equations is known, we set ( d n+ m = + r ) δ x Um n = r U m n + ( r) Um n + r U m+, n m M. For the left hand side (l.h.s.) we get component-wise ( r ) δ x Um n+ = r U m n+ n+ + ( + r) Um r U m+ n+, m M, where we substitute U0 n+ = g0 n+ = g 0 (t n+ ) and U n+ M+ = gn+ = g (t n+ ). We can now express the equation in matrix-vector form + r r U n+ d n+ + r gn+ 0 r + r r U n+ d n+ =.. r + r r U n+ M d n+ M + r r U n+ M d n+ M + r gn+ In a similar way we get the following linear system for Bakward-Euler + r r r + r r r + r r r + r U n+ U n+. U n+ M U n+ M = U n + r gn+ 0 U n. U n M U n M + r gn+.

Numerical Solutions to Partial Differential Equations

Numerical Solutions to Partial Differential Equations Numerical Solutions to Partial Differential Equations Zhiping Li LMAM and School of Mathematical Sciences Peking University Numerical Methods for Partial Differential Equations Finite Difference Methods

More information

Numerical Solutions to Partial Differential Equations

Numerical Solutions to Partial Differential Equations Numerical Solutions to Partial Differential Equations Zhiping Li LMAM and School of Mathematical Sciences Peking University The Implicit Schemes for the Model Problem The Crank-Nicolson scheme and θ-scheme

More information

Finite difference method for elliptic problems: I

Finite difference method for elliptic problems: I Finite difference method for elliptic problems: I Praveen. C praveen@math.tifrbng.res.in Tata Institute of Fundamental Research Center for Applicable Mathematics Bangalore 560065 http://math.tifrbng.res.in/~praveen

More information

Foundations of Matrix Analysis

Foundations of Matrix Analysis 1 Foundations of Matrix Analysis In this chapter we recall the basic elements of linear algebra which will be employed in the remainder of the text For most of the proofs as well as for the details, the

More information

The following definition is fundamental.

The following definition is fundamental. 1. Some Basics from Linear Algebra With these notes, I will try and clarify certain topics that I only quickly mention in class. First and foremost, I will assume that you are familiar with many basic

More information

CLASSIFICATION AND PRINCIPLE OF SUPERPOSITION FOR SECOND ORDER LINEAR PDE

CLASSIFICATION AND PRINCIPLE OF SUPERPOSITION FOR SECOND ORDER LINEAR PDE CLASSIFICATION AND PRINCIPLE OF SUPERPOSITION FOR SECOND ORDER LINEAR PDE 1. Linear Partial Differential Equations A partial differential equation (PDE) is an equation, for an unknown function u, that

More information

[2] (a) Develop and describe the piecewise linear Galerkin finite element approximation of,

[2] (a) Develop and describe the piecewise linear Galerkin finite element approximation of, 269 C, Vese Practice problems [1] Write the differential equation u + u = f(x, y), (x, y) Ω u = 1 (x, y) Ω 1 n + u = x (x, y) Ω 2, Ω = {(x, y) x 2 + y 2 < 1}, Ω 1 = {(x, y) x 2 + y 2 = 1, x 0}, Ω 2 = {(x,

More information

UNIVERSITY OF MANITOBA

UNIVERSITY OF MANITOBA Question Points Score INSTRUCTIONS TO STUDENTS: This is a 6 hour examination. No extra time will be given. No texts, notes, or other aids are permitted. There are no calculators, cellphones or electronic

More information

The method of lines (MOL) for the diffusion equation

The method of lines (MOL) for the diffusion equation Chapter 1 The method of lines (MOL) for the diffusion equation The method of lines refers to an approximation of one or more partial differential equations with ordinary differential equations in just

More information

Implicit Scheme for the Heat Equation

Implicit Scheme for the Heat Equation Implicit Scheme for the Heat Equation Implicit scheme for the one-dimensional heat equation Once again we consider the one-dimensional heat equation where we seek a u(x, t) satisfying u t = νu xx + f(x,

More information

Finite difference method for heat equation

Finite difference method for heat equation Finite difference method for heat equation Praveen. C praveen@math.tifrbng.res.in Tata Institute of Fundamental Research Center for Applicable Mathematics Bangalore 560065 http://math.tifrbng.res.in/~praveen

More information

In this chapter we study elliptical PDEs. That is, PDEs of the form. 2 u = lots,

In this chapter we study elliptical PDEs. That is, PDEs of the form. 2 u = lots, Chapter 8 Elliptic PDEs In this chapter we study elliptical PDEs. That is, PDEs of the form 2 u = lots, where lots means lower-order terms (u x, u y,..., u, f). Here are some ways to think about the physical

More information

G1110 & 852G1 Numerical Linear Algebra

G1110 & 852G1 Numerical Linear Algebra The University of Sussex Department of Mathematics G & 85G Numerical Linear Algebra Lecture Notes Autumn Term Kerstin Hesse (w aw S w a w w (w aw H(wa = (w aw + w Figure : Geometric explanation of the

More information

Tutorial 2. Introduction to numerical schemes

Tutorial 2. Introduction to numerical schemes 236861 Numerical Geometry of Images Tutorial 2 Introduction to numerical schemes c 2012 Classifying PDEs Looking at the PDE Au xx + 2Bu xy + Cu yy + Du x + Eu y + Fu +.. = 0, and its discriminant, B 2

More information

Finite Differences for Differential Equations 28 PART II. Finite Difference Methods for Differential Equations

Finite Differences for Differential Equations 28 PART II. Finite Difference Methods for Differential Equations Finite Differences for Differential Equations 28 PART II Finite Difference Methods for Differential Equations Finite Differences for Differential Equations 29 BOUNDARY VALUE PROBLEMS (I) Solving a TWO

More information

Lecture Introduction

Lecture Introduction Lecture 1 1.1 Introduction The theory of Partial Differential Equations (PDEs) is central to mathematics, both pure and applied. The main difference between the theory of PDEs and the theory of Ordinary

More information

Finite Difference Methods for Boundary Value Problems

Finite Difference Methods for Boundary Value Problems Finite Difference Methods for Boundary Value Problems October 2, 2013 () Finite Differences October 2, 2013 1 / 52 Goals Learn steps to approximate BVPs using the Finite Difference Method Start with two-point

More information

13 PDEs on spatially bounded domains: initial boundary value problems (IBVPs)

13 PDEs on spatially bounded domains: initial boundary value problems (IBVPs) 13 PDEs on spatially bounded domains: initial boundary value problems (IBVPs) A prototypical problem we will discuss in detail is the 1D diffusion equation u t = Du xx < x < l, t > finite-length rod u(x,

More information

Introduction to numerical schemes

Introduction to numerical schemes 236861 Numerical Geometry of Images Tutorial 2 Introduction to numerical schemes Heat equation The simple parabolic PDE with the initial values u t = K 2 u 2 x u(0, x) = u 0 (x) and some boundary conditions

More information

Numerical Analysis of Differential Equations Numerical Solution of Parabolic Equations

Numerical Analysis of Differential Equations Numerical Solution of Parabolic Equations Numerical Analysis of Differential Equations 215 6 Numerical Solution of Parabolic Equations 6 Numerical Solution of Parabolic Equations TU Bergakademie Freiberg, SS 2012 Numerical Analysis of Differential

More information

Elementary Linear Algebra

Elementary Linear Algebra Matrices J MUSCAT Elementary Linear Algebra Matrices Definition Dr J Muscat 2002 A matrix is a rectangular array of numbers, arranged in rows and columns a a 2 a 3 a n a 2 a 22 a 23 a 2n A = a m a mn We

More information

Finite difference methods for the diffusion equation

Finite difference methods for the diffusion equation Finite difference methods for the diffusion equation D150, Tillämpade numeriska metoder II Olof Runborg May 0, 003 These notes summarize a part of the material in Chapter 13 of Iserles. They are based

More information

Chapter Two: Numerical Methods for Elliptic PDEs. 1 Finite Difference Methods for Elliptic PDEs

Chapter Two: Numerical Methods for Elliptic PDEs. 1 Finite Difference Methods for Elliptic PDEs Chapter Two: Numerical Methods for Elliptic PDEs Finite Difference Methods for Elliptic PDEs.. Finite difference scheme. We consider a simple example u := subject to Dirichlet boundary conditions ( ) u

More information

Separation of Variables in Linear PDE: One-Dimensional Problems

Separation of Variables in Linear PDE: One-Dimensional Problems Separation of Variables in Linear PDE: One-Dimensional Problems Now we apply the theory of Hilbert spaces to linear differential equations with partial derivatives (PDE). We start with a particular example,

More information

DS-GA 1002 Lecture notes 0 Fall Linear Algebra. These notes provide a review of basic concepts in linear algebra.

DS-GA 1002 Lecture notes 0 Fall Linear Algebra. These notes provide a review of basic concepts in linear algebra. DS-GA 1002 Lecture notes 0 Fall 2016 Linear Algebra These notes provide a review of basic concepts in linear algebra. 1 Vector spaces You are no doubt familiar with vectors in R 2 or R 3, i.e. [ ] 1.1

More information

Numerical Solutions to Partial Differential Equations

Numerical Solutions to Partial Differential Equations Numerical Solutions to Partial Differential Equations Zhiping Li LMAM and School of Mathematical Sciences Peking University A Model Problem in a 2D Box Region Let us consider a model problem of parabolic

More information

INTRODUCTION TO PDEs

INTRODUCTION TO PDEs INTRODUCTION TO PDEs In this course we are interested in the numerical approximation of PDEs using finite difference methods (FDM). We will use some simple prototype boundary value problems (BVP) and initial

More information

n 1 f n 1 c 1 n+1 = c 1 n $ c 1 n 1. After taking logs, this becomes

n 1 f n 1 c 1 n+1 = c 1 n $ c 1 n 1. After taking logs, this becomes Root finding: 1 a The points {x n+1, }, {x n, f n }, {x n 1, f n 1 } should be co-linear Say they lie on the line x + y = This gives the relations x n+1 + = x n +f n = x n 1 +f n 1 = Eliminating α and

More information

Numerical Solutions to Partial Differential Equations

Numerical Solutions to Partial Differential Equations Numerical Solutions to Partial Differential Equations Zhiping Li LMAM and School of Mathematical Sciences Peking University A Model Problem and Its Difference Approximations 1-D Initial Boundary Value

More information

Lecture 4: Numerical solution of ordinary differential equations

Lecture 4: Numerical solution of ordinary differential equations Lecture 4: Numerical solution of ordinary differential equations Department of Mathematics, ETH Zürich General explicit one-step method: Consistency; Stability; Convergence. High-order methods: Taylor

More information

FDM for parabolic equations

FDM for parabolic equations FDM for parabolic equations Consider the heat equation where Well-posed problem Existence & Uniqueness Mass & Energy decreasing FDM for parabolic equations CNFD Crank-Nicolson + 2 nd order finite difference

More information

Topic 1: Matrix diagonalization

Topic 1: Matrix diagonalization Topic : Matrix diagonalization Review of Matrices and Determinants Definition A matrix is a rectangular array of real numbers a a a m a A = a a m a n a n a nm The matrix is said to be of order n m if it

More information

Solutions Preliminary Examination in Numerical Analysis January, 2017

Solutions Preliminary Examination in Numerical Analysis January, 2017 Solutions Preliminary Examination in Numerical Analysis January, 07 Root Finding The roots are -,0, a) First consider x 0 > Let x n+ = + ε and x n = + δ with δ > 0 The iteration gives 0 < ε δ < 3, which

More information

Simple Examples on Rectangular Domains

Simple Examples on Rectangular Domains 84 Chapter 5 Simple Examples on Rectangular Domains In this chapter we consider simple elliptic boundary value problems in rectangular domains in R 2 or R 3 ; our prototype example is the Poisson equation

More information

Second Order Elliptic PDE

Second Order Elliptic PDE Second Order Elliptic PDE T. Muthukumar tmk@iitk.ac.in December 16, 2014 Contents 1 A Quick Introduction to PDE 1 2 Classification of Second Order PDE 3 3 Linear Second Order Elliptic Operators 4 4 Periodic

More information

Time-dependent variational forms

Time-dependent variational forms Time-dependent variational forms Hans Petter Langtangen 1,2 1 Center for Biomedical Computing, Simula Research Laboratory 2 Department of Informatics, University of Oslo Oct 30, 2015 PRELIMINARY VERSION

More information

Applied Mathematics 205. Unit III: Numerical Calculus. Lecturer: Dr. David Knezevic

Applied Mathematics 205. Unit III: Numerical Calculus. Lecturer: Dr. David Knezevic Applied Mathematics 205 Unit III: Numerical Calculus Lecturer: Dr. David Knezevic Unit III: Numerical Calculus Chapter III.3: Boundary Value Problems and PDEs 2 / 96 ODE Boundary Value Problems 3 / 96

More information

Computation Fluid Dynamics

Computation Fluid Dynamics Computation Fluid Dynamics CFD I Jitesh Gajjar Maths Dept Manchester University Computation Fluid Dynamics p.1/189 Garbage In, Garbage Out We will begin with a discussion of errors. Useful to understand

More information

3 2 6 Solve the initial value problem u ( t) 3. a- If A has eigenvalues λ =, λ = 1 and corresponding eigenvectors 1

3 2 6 Solve the initial value problem u ( t) 3. a- If A has eigenvalues λ =, λ = 1 and corresponding eigenvectors 1 Math Problem a- If A has eigenvalues λ =, λ = 1 and corresponding eigenvectors 1 3 6 Solve the initial value problem u ( t) = Au( t) with u (0) =. 3 1 u 1 =, u 1 3 = b- True or false and why 1. if A is

More information

FIXED POINT ITERATIONS

FIXED POINT ITERATIONS FIXED POINT ITERATIONS MARKUS GRASMAIR 1. Fixed Point Iteration for Non-linear Equations Our goal is the solution of an equation (1) F (x) = 0, where F : R n R n is a continuous vector valued mapping in

More information

Introduction - Motivation. Many phenomena (physical, chemical, biological, etc.) are model by differential equations. f f(x + h) f(x) (x) = lim

Introduction - Motivation. Many phenomena (physical, chemical, biological, etc.) are model by differential equations. f f(x + h) f(x) (x) = lim Introduction - Motivation Many phenomena (physical, chemical, biological, etc.) are model by differential equations. Recall the definition of the derivative of f(x) f f(x + h) f(x) (x) = lim. h 0 h Its

More information

Index. higher order methods, 52 nonlinear, 36 with variable coefficients, 34 Burgers equation, 234 BVP, see boundary value problems

Index. higher order methods, 52 nonlinear, 36 with variable coefficients, 34 Burgers equation, 234 BVP, see boundary value problems Index A-conjugate directions, 83 A-stability, 171 A( )-stability, 171 absolute error, 243 absolute stability, 149 for systems of equations, 154 absorbing boundary conditions, 228 Adams Bashforth methods,

More information

Numerical Solutions to Partial Differential Equations

Numerical Solutions to Partial Differential Equations Numerical Solutions to Partial Differential Equations Zhiping Li LMAM and School of Mathematical Sciences Peking University Discretization of Boundary Conditions Discretization of Boundary Conditions On

More information

Numerical Analysis of Differential Equations Numerical Solution of Elliptic Boundary Value

Numerical Analysis of Differential Equations Numerical Solution of Elliptic Boundary Value Numerical Analysis of Differential Equations 188 5 Numerical Solution of Elliptic Boundary Value Problems 5 Numerical Solution of Elliptic Boundary Value Problems TU Bergakademie Freiberg, SS 2012 Numerical

More information

Lecture 2: Linear Algebra Review

Lecture 2: Linear Algebra Review EE 227A: Convex Optimization and Applications January 19 Lecture 2: Linear Algebra Review Lecturer: Mert Pilanci Reading assignment: Appendix C of BV. Sections 2-6 of the web textbook 1 2.1 Vectors 2.1.1

More information

NATIONAL BOARD FOR HIGHER MATHEMATICS. Research Scholarships Screening Test. Saturday, January 20, Time Allowed: 150 Minutes Maximum Marks: 40

NATIONAL BOARD FOR HIGHER MATHEMATICS. Research Scholarships Screening Test. Saturday, January 20, Time Allowed: 150 Minutes Maximum Marks: 40 NATIONAL BOARD FOR HIGHER MATHEMATICS Research Scholarships Screening Test Saturday, January 2, 218 Time Allowed: 15 Minutes Maximum Marks: 4 Please read, carefully, the instructions that follow. INSTRUCTIONS

More information

u xx + u yy = 0. (5.1)

u xx + u yy = 0. (5.1) Chapter 5 Laplace Equation The following equation is called Laplace equation in two independent variables x, y: The non-homogeneous problem u xx + u yy =. (5.1) u xx + u yy = F, (5.) where F is a function

More information

Numerical Methods for PDEs

Numerical Methods for PDEs Numerical Methods for PDEs Problems 1. Numerical Differentiation. Find the best approximation to the second drivative d 2 f(x)/dx 2 at x = x you can of a function f(x) using (a) the Taylor series approach

More information

Finite Difference Methods (FDMs) 1

Finite Difference Methods (FDMs) 1 Finite Difference Methods (FDMs) 1 1 st - order Approxima9on Recall Taylor series expansion: Forward difference: Backward difference: Central difference: 2 nd - order Approxima9on Forward difference: Backward

More information

x n+1 = x n f(x n) f (x n ), n 0.

x n+1 = x n f(x n) f (x n ), n 0. 1. Nonlinear Equations Given scalar equation, f(x) = 0, (a) Describe I) Newtons Method, II) Secant Method for approximating the solution. (b) State sufficient conditions for Newton and Secant to converge.

More information

Linear Algebra: Matrix Eigenvalue Problems

Linear Algebra: Matrix Eigenvalue Problems CHAPTER8 Linear Algebra: Matrix Eigenvalue Problems Chapter 8 p1 A matrix eigenvalue problem considers the vector equation (1) Ax = λx. 8.0 Linear Algebra: Matrix Eigenvalue Problems Here A is a given

More information

Linear Algebra March 16, 2019

Linear Algebra March 16, 2019 Linear Algebra March 16, 2019 2 Contents 0.1 Notation................................ 4 1 Systems of linear equations, and matrices 5 1.1 Systems of linear equations..................... 5 1.2 Augmented

More information

Numerical Methods for Partial Differential Equations

Numerical Methods for Partial Differential Equations Numerical Methods for Partial Differential Equations Steffen Börm Compiled July 12, 2018, 12:01 All rights reserved. Contents 1. Introduction 5 2. Finite difference methods 7 2.1. Potential equation.............................

More information

Problem set 3: Solutions Math 207B, Winter Suppose that u(x) is a non-zero solution of the eigenvalue problem. (u ) 2 dx, u 2 dx.

Problem set 3: Solutions Math 207B, Winter Suppose that u(x) is a non-zero solution of the eigenvalue problem. (u ) 2 dx, u 2 dx. Problem set 3: Solutions Math 27B, Winter 216 1. Suppose that u(x) is a non-zero solution of the eigenvalue problem u = λu < x < 1, u() =, u(1) =. Show that λ = (u ) 2 dx u2 dx. Deduce that every eigenvalue

More information

COMP 558 lecture 18 Nov. 15, 2010

COMP 558 lecture 18 Nov. 15, 2010 Least squares We have seen several least squares problems thus far, and we will see more in the upcoming lectures. For this reason it is good to have a more general picture of these problems and how to

More information

Finite Difference and Finite Element Methods

Finite Difference and Finite Element Methods Finite Difference and Finite Element Methods Georgy Gimel farb COMPSCI 369 Computational Science 1 / 39 1 Finite Differences Difference Equations 3 Finite Difference Methods: Euler FDMs 4 Finite Element

More information

MIT (Spring 2014)

MIT (Spring 2014) 18.311 MIT (Spring 014) Rodolfo R. Rosales May 6, 014. Problem Set # 08. Due: Last day of lectures. IMPORTANT: Turn in the regular and the special problems stapled in two SEPARATE packages. Print your

More information

Existence Theory: Green s Functions

Existence Theory: Green s Functions Chapter 5 Existence Theory: Green s Functions In this chapter we describe a method for constructing a Green s Function The method outlined is formal (not rigorous) When we find a solution to a PDE by constructing

More information

PDEs, part 1: Introduction and elliptic PDEs

PDEs, part 1: Introduction and elliptic PDEs PDEs, part 1: Introduction and elliptic PDEs Anna-Karin Tornberg Mathematical Models, Analysis and Simulation Fall semester, 2013 Partial di erential equations The solution depends on several variables,

More information

AM 205: lecture 14. Last time: Boundary value problems Today: Numerical solution of PDEs

AM 205: lecture 14. Last time: Boundary value problems Today: Numerical solution of PDEs AM 205: lecture 14 Last time: Boundary value problems Today: Numerical solution of PDEs ODE BVPs A more general approach is to formulate a coupled system of equations for the BVP based on a finite difference

More information

u n 2 4 u n 36 u n 1, n 1.

u n 2 4 u n 36 u n 1, n 1. Exercise 1 Let (u n ) be the sequence defined by Set v n = u n 1 x+ u n and f (x) = 4 x. 1. Solve the equations f (x) = 1 and f (x) =. u 0 = 0, n Z +, u n+1 = u n + 4 u n.. Prove that if u n < 1, then

More information

Linear Algebra and Matrix Inversion

Linear Algebra and Matrix Inversion Jim Lambers MAT 46/56 Spring Semester 29- Lecture 2 Notes These notes correspond to Section 63 in the text Linear Algebra and Matrix Inversion Vector Spaces and Linear Transformations Matrices are much

More information

Introduction to Quantitative Techniques for MSc Programmes SCHOOL OF ECONOMICS, MATHEMATICS AND STATISTICS MALET STREET LONDON WC1E 7HX

Introduction to Quantitative Techniques for MSc Programmes SCHOOL OF ECONOMICS, MATHEMATICS AND STATISTICS MALET STREET LONDON WC1E 7HX Introduction to Quantitative Techniques for MSc Programmes SCHOOL OF ECONOMICS, MATHEMATICS AND STATISTICS MALET STREET LONDON WC1E 7HX September 2007 MSc Sep Intro QT 1 Who are these course for? The September

More information

Last Update: April 7, 201 0

Last Update: April 7, 201 0 M ath E S W inter Last Update: April 7, Introduction to Partial Differential Equations Disclaimer: his lecture note tries to provide an alternative approach to the material in Sections.. 5 in the textbook.

More information

LECTURE NOTES ELEMENTARY NUMERICAL METHODS. Eusebius Doedel

LECTURE NOTES ELEMENTARY NUMERICAL METHODS. Eusebius Doedel LECTURE NOTES on ELEMENTARY NUMERICAL METHODS Eusebius Doedel TABLE OF CONTENTS Vector and Matrix Norms 1 Banach Lemma 20 The Numerical Solution of Linear Systems 25 Gauss Elimination 25 Operation Count

More information

Linear Operators, Eigenvalues, and Green s Operator

Linear Operators, Eigenvalues, and Green s Operator Chapter 10 Linear Operators, Eigenvalues, and Green s Operator We begin with a reminder of facts which should be known from previous courses. 10.1 Inner Product Space A vector space is a collection of

More information

Math 7824 Spring 2010 Numerical solution of partial differential equations Classroom notes and homework

Math 7824 Spring 2010 Numerical solution of partial differential equations Classroom notes and homework Math 7824 Spring 2010 Numerical solution of partial differential equations Classroom notes and homework Jan Mandel University of Colorado Denver May 12, 2010 1/20/09: Sec. 1.1, 1.2. Hw 1 due 1/27: problems

More information

AIMS Exercise Set # 1

AIMS Exercise Set # 1 AIMS Exercise Set #. Determine the form of the single precision floating point arithmetic used in the computers at AIMS. What is the largest number that can be accurately represented? What is the smallest

More information

Knowledge Discovery and Data Mining 1 (VO) ( )

Knowledge Discovery and Data Mining 1 (VO) ( ) Knowledge Discovery and Data Mining 1 (VO) (707.003) Review of Linear Algebra Denis Helic KTI, TU Graz Oct 9, 2014 Denis Helic (KTI, TU Graz) KDDM1 Oct 9, 2014 1 / 74 Big picture: KDDM Probability Theory

More information

November 18, 2013 ANALYTIC FUNCTIONAL CALCULUS

November 18, 2013 ANALYTIC FUNCTIONAL CALCULUS November 8, 203 ANALYTIC FUNCTIONAL CALCULUS RODICA D. COSTIN Contents. The spectral projection theorem. Functional calculus 2.. The spectral projection theorem for self-adjoint matrices 2.2. The spectral

More information

Analysis-3 lecture schemes

Analysis-3 lecture schemes Analysis-3 lecture schemes (with Homeworks) 1 Csörgő István November, 2015 1 A jegyzet az ELTE Informatikai Kar 2015. évi Jegyzetpályázatának támogatásával készült Contents 1. Lesson 1 4 1.1. The Space

More information

MATH 583A REVIEW SESSION #1

MATH 583A REVIEW SESSION #1 MATH 583A REVIEW SESSION #1 BOJAN DURICKOVIC 1. Vector Spaces Very quick review of the basic linear algebra concepts (see any linear algebra textbook): (finite dimensional) vector space (or linear space),

More information

Review of Basic Concepts in Linear Algebra

Review of Basic Concepts in Linear Algebra Review of Basic Concepts in Linear Algebra Grady B Wright Department of Mathematics Boise State University September 7, 2017 Math 565 Linear Algebra Review September 7, 2017 1 / 40 Numerical Linear Algebra

More information

Definition 5.1. A vector field v on a manifold M is map M T M such that for all x M, v(x) T x M.

Definition 5.1. A vector field v on a manifold M is map M T M such that for all x M, v(x) T x M. 5 Vector fields Last updated: March 12, 2012. 5.1 Definition and general properties We first need to define what a vector field is. Definition 5.1. A vector field v on a manifold M is map M T M such that

More information

Chem 3502/4502 Physical Chemistry II (Quantum Mechanics) 3 Credits Fall Semester 2006 Christopher J. Cramer. Lecture 5, January 27, 2006

Chem 3502/4502 Physical Chemistry II (Quantum Mechanics) 3 Credits Fall Semester 2006 Christopher J. Cramer. Lecture 5, January 27, 2006 Chem 3502/4502 Physical Chemistry II (Quantum Mechanics) 3 Credits Fall Semester 2006 Christopher J. Cramer Lecture 5, January 27, 2006 Solved Homework (Homework for grading is also due today) We are told

More information

Partial Differential Equations

Partial Differential Equations Partial Differential Equations Introduction Deng Li Discretization Methods Chunfang Chen, Danny Thorne, Adam Zornes CS521 Feb.,7, 2006 What do You Stand For? A PDE is a Partial Differential Equation This

More information

Notes for Expansions/Series and Differential Equations

Notes for Expansions/Series and Differential Equations Notes for Expansions/Series and Differential Equations In the last discussion, we considered perturbation methods for constructing solutions/roots of algebraic equations. Three types of problems were illustrated

More information

Linear Algebra Review (Course Notes for Math 308H - Spring 2016)

Linear Algebra Review (Course Notes for Math 308H - Spring 2016) Linear Algebra Review (Course Notes for Math 308H - Spring 2016) Dr. Michael S. Pilant February 12, 2016 1 Background: We begin with one of the most fundamental notions in R 2, distance. Letting (x 1,

More information

Introduction to Partial Differential Equations

Introduction to Partial Differential Equations Introduction to Partial Differential Equations Philippe B. Laval KSU Current Semester Philippe B. Laval (KSU) Key Concepts Current Semester 1 / 25 Introduction The purpose of this section is to define

More information

Introduction to the Numerical Solution of IVP for ODE

Introduction to the Numerical Solution of IVP for ODE Introduction to the Numerical Solution of IVP for ODE 45 Introduction to the Numerical Solution of IVP for ODE Consider the IVP: DE x = f(t, x), IC x(a) = x a. For simplicity, we will assume here that

More information

1 Directional Derivatives and Differentiability

1 Directional Derivatives and Differentiability Wednesday, January 18, 2012 1 Directional Derivatives and Differentiability Let E R N, let f : E R and let x 0 E. Given a direction v R N, let L be the line through x 0 in the direction v, that is, L :=

More information

j=1 x j p, if 1 p <, x i ξ : x i < ξ} 0 as p.

j=1 x j p, if 1 p <, x i ξ : x i < ξ} 0 as p. LINEAR ALGEBRA Fall 203 The final exam Almost all of the problems solved Exercise Let (V, ) be a normed vector space. Prove x y x y for all x, y V. Everybody knows how to do this! Exercise 2 If V is a

More information

SINC PACK, and Separation of Variables

SINC PACK, and Separation of Variables SINC PACK, and Separation of Variables Frank Stenger Abstract This talk consists of a proof of part of Stenger s SINC-PACK computer package (an approx. 400-page tutorial + about 250 Matlab programs) that

More information

Elementary linear algebra

Elementary linear algebra Chapter 1 Elementary linear algebra 1.1 Vector spaces Vector spaces owe their importance to the fact that so many models arising in the solutions of specific problems turn out to be vector spaces. The

More information

Scientific Computing WS 2018/2019. Lecture 9. Jürgen Fuhrmann Lecture 9 Slide 1

Scientific Computing WS 2018/2019. Lecture 9. Jürgen Fuhrmann Lecture 9 Slide 1 Scientific Computing WS 2018/2019 Lecture 9 Jürgen Fuhrmann juergen.fuhrmann@wias-berlin.de Lecture 9 Slide 1 Lecture 9 Slide 2 Simple iteration with preconditioning Idea: Aû = b iterative scheme û = û

More information

Fundamentals of Linear Algebra. Marcel B. Finan Arkansas Tech University c All Rights Reserved

Fundamentals of Linear Algebra. Marcel B. Finan Arkansas Tech University c All Rights Reserved Fundamentals of Linear Algebra Marcel B. Finan Arkansas Tech University c All Rights Reserved 2 PREFACE Linear algebra has evolved as a branch of mathematics with wide range of applications to the natural

More information

1.2 Derivation. d p f = d p f(x(p)) = x fd p x (= f x x p ). (1) Second, g x x p + g p = 0. d p f = f x g 1. The expression f x gx

1.2 Derivation. d p f = d p f(x(p)) = x fd p x (= f x x p ). (1) Second, g x x p + g p = 0. d p f = f x g 1. The expression f x gx PDE-constrained optimization and the adjoint method Andrew M. Bradley November 16, 21 PDE-constrained optimization and the adjoint method for solving these and related problems appear in a wide range of

More information

2tdt 1 y = t2 + C y = which implies C = 1 and the solution is y = 1

2tdt 1 y = t2 + C y = which implies C = 1 and the solution is y = 1 Lectures - Week 11 General First Order ODEs & Numerical Methods for IVPs In general, nonlinear problems are much more difficult to solve than linear ones. Unfortunately many phenomena exhibit nonlinear

More information

Numerical Methods for Inverse Kinematics

Numerical Methods for Inverse Kinematics Numerical Methods for Inverse Kinematics Niels Joubert, UC Berkeley, CS184 2008-11-25 Inverse Kinematics is used to pose models by specifying endpoints of segments rather than individual joint angles.

More information

forms Christopher Engström November 14, 2014 MAA704: Matrix factorization and canonical forms Matrix properties Matrix factorization Canonical forms

forms Christopher Engström November 14, 2014 MAA704: Matrix factorization and canonical forms Matrix properties Matrix factorization Canonical forms Christopher Engström November 14, 2014 Hermitian LU QR echelon Contents of todays lecture Some interesting / useful / important of matrices Hermitian LU QR echelon Rewriting a as a product of several matrices.

More information

AN ELEMENTARY PROOF OF THE SPECTRAL RADIUS FORMULA FOR MATRICES

AN ELEMENTARY PROOF OF THE SPECTRAL RADIUS FORMULA FOR MATRICES AN ELEMENTARY PROOF OF THE SPECTRAL RADIUS FORMULA FOR MATRICES JOEL A. TROPP Abstract. We present an elementary proof that the spectral radius of a matrix A may be obtained using the formula ρ(a) lim

More information

Fundamentals of Engineering Analysis (650163)

Fundamentals of Engineering Analysis (650163) Philadelphia University Faculty of Engineering Communications and Electronics Engineering Fundamentals of Engineering Analysis (6563) Part Dr. Omar R Daoud Matrices: Introduction DEFINITION A matrix is

More information

Linear Algebra Primer

Linear Algebra Primer Introduction Linear Algebra Primer Daniel S. Stutts, Ph.D. Original Edition: 2/99 Current Edition: 4//4 This primer was written to provide a brief overview of the main concepts and methods in elementary

More information

A Brief Outline of Math 355

A Brief Outline of Math 355 A Brief Outline of Math 355 Lecture 1 The geometry of linear equations; elimination with matrices A system of m linear equations with n unknowns can be thought of geometrically as m hyperplanes intersecting

More information

Lecture 2: Review of Prerequisites. Table of contents

Lecture 2: Review of Prerequisites. Table of contents Math 348 Fall 217 Lecture 2: Review of Prerequisites Disclaimer. As we have a textbook, this lecture note is for guidance and supplement only. It should not be relied on when preparing for exams. In this

More information

PDEs, part 3: Hyperbolic PDEs

PDEs, part 3: Hyperbolic PDEs PDEs, part 3: Hyperbolic PDEs Anna-Karin Tornberg Mathematical Models, Analysis and Simulation Fall semester, 2011 Hyperbolic equations (Sections 6.4 and 6.5 of Strang). Consider the model problem (the

More information

ELEMENTARY LINEAR ALGEBRA

ELEMENTARY LINEAR ALGEBRA ELEMENTARY LINEAR ALGEBRA K R MATTHEWS DEPARTMENT OF MATHEMATICS UNIVERSITY OF QUEENSLAND First Printing, 99 Chapter LINEAR EQUATIONS Introduction to linear equations A linear equation in n unknowns x,

More information

Partial differential equation for temperature u(x, t) in a heat conducting insulated rod along the x-axis is given by the Heat equation:

Partial differential equation for temperature u(x, t) in a heat conducting insulated rod along the x-axis is given by the Heat equation: Chapter 7 Heat Equation Partial differential equation for temperature u(x, t) in a heat conducting insulated rod along the x-axis is given by the Heat equation: u t = ku x x, x, t > (7.1) Here k is a constant

More information

Scientific Computing I

Scientific Computing I Scientific Computing I Module 8: An Introduction to Finite Element Methods Tobias Neckel Winter 2013/2014 Module 8: An Introduction to Finite Element Methods, Winter 2013/2014 1 Part I: Introduction to

More information

Numerical Analysis Preliminary Exam 10 am to 1 pm, August 20, 2018

Numerical Analysis Preliminary Exam 10 am to 1 pm, August 20, 2018 Numerical Analysis Preliminary Exam 1 am to 1 pm, August 2, 218 Instructions. You have three hours to complete this exam. Submit solutions to four (and no more) of the following six problems. Please start

More information