MA2501 Numerical Methods Spring 2015

Similar documents
MA2501 Numerical Methods Spring 2015

you expect to encounter difficulties when trying to solve A x = b? 4. A composite quadrature rule has error associated with it in the following form

(0, 0), (1, ), (2, ), (3, ), (4, ), (5, ), (6, ).

Math 578: Assignment 2

Numerical Methods. King Saud University

INTERPOLATION. and y i = cos x i, i = 0, 1, 2 This gives us the three points. Now find a quadratic polynomial. p(x) = a 0 + a 1 x + a 2 x 2.

NUMERICAL METHODS. x n+1 = 2x n x 2 n. In particular: which of them gives faster convergence, and why? [Work to four decimal places.

We consider the problem of finding a polynomial that interpolates a given set of values:

1 Solutions to selected problems

Homework and Computer Problems for Math*2130 (W17).

A first order divided difference

1 Lecture 8: Interpolating polynomials.

8.7 Taylor s Inequality Math 2300 Section 005 Calculus II. f(x) = ln(1 + x) f(0) = 0

Polynomial Interpolation Part II

Numerical Analysis: Interpolation Part 1

ECONOMICS 207 SPRING 2006 LABORATORY EXERCISE 5 KEY. 8 = 10(5x 2) = 9(3x + 8), x 50x 20 = 27x x = 92 x = 4. 8x 2 22x + 15 = 0 (2x 3)(4x 5) = 0

Integration, differentiation, and root finding. Phys 420/580 Lecture 7

Romberg Rule of Integration

COURSE Numerical integration of functions (continuation) 3.3. The Romberg s iterative generation method

Worksheet 9. Topics: Taylor series; using Taylor polynomials for approximate computations. Polar coordinates.

MA2501 Numerical Methods Spring 2015

Taylor and Maclaurin Series. Approximating functions using Polynomials.

Scientific Computing

1 Solutions to selected problems

Exam in Numerical Methods (MA2501)

Taylor series. Chapter Introduction From geometric series to Taylor polynomials

Chapter 3: Root Finding. September 26, 2005

Midterm for Introduction to Numerical Analysis I, AMSC/CMSC 466, on 10/29/2015

Jim Lambers MAT 460/560 Fall Semester Practice Final Exam

TAYLOR POLYNOMIALS DARYL DEFORD

EXAMPLE OF ONE-STEP METHOD

CS 450 Numerical Analysis. Chapter 8: Numerical Integration and Differentiation

Applied Numerical Analysis (AE2220-I) R. Klees and R.P. Dwight

Notation Nodes are data points at which functional values are available or at which you wish to compute functional values At the nodes fx i

Numerical interpolation, extrapolation and fi tting of data

Lectures 9-10: Polynomial and piecewise polynomial interpolation

Floating Point Number Systems. Simon Fraser University Surrey Campus MACM 316 Spring 2005 Instructor: Ha Le

Lecture Note 3: Interpolation and Polynomial Approximation. Xiaoqun Zhang Shanghai Jiao Tong University

Math Numerical Analysis

Numerical Analysis Exam with Solutions

8.8 Applications of Taylor Polynomials

1. Method 1: bisection. The bisection methods starts from two points a 0 and b 0 such that

Romberg Integration. MATH 375 Numerical Analysis. J. Robert Buchanan. Spring Department of Mathematics

Math Review for Exam Answer each of the following questions as either True or False. Circle the correct answer.

x x2 2 + x3 3 x4 3. Use the divided-difference method to find a polynomial of least degree that fits the values shown: (b)

FIXED POINT ITERATION

Power series and Taylor series

Numerical Integration (Quadrature) Another application for our interpolation tools!

CHALLENGE! (0) = 5. Construct a polynomial with the following behavior at x = 0:

Math 113 (Calculus 2) Exam 4

Do not turn over until you are told to do so by the Invigilator.

Review I: Interpolation

Taylor Series and Numerical Approximations

Chapter 4: Interpolation and Approximation. October 28, 2005

Example 1 Which of these functions are polynomials in x? In the case(s) where f is a polynomial,

Announcements. Topics: Homework: - sections , 6.1 (extreme values) * Read these sections and study solved examples in your textbook!

Taylor and Maclaurin Series. Approximating functions using Polynomials.

AIMS Exercise Set # 1

Solutions 2. January 23, x x

Properties of a Taylor Polynomial

CHAPTER 4. Interpolation

Lecture Note 3: Polynomial Interpolation. Xiaoqun Zhang Shanghai Jiao Tong University

Announcements. Related Rates (last week), Linear approximations (today) l Hôpital s Rule (today) Newton s Method Curve sketching Optimization problems

Extrapolation in Numerical Integration. Romberg Integration

7.0: Minimax approximations

Numerical Methods of Approximation

Interpolation. P. Sam Johnson. January 30, P. Sam Johnson (NITK) Interpolation January 30, / 75

JUST THE MATHS UNIT NUMBER ORDINARY DIFFERENTIAL EQUATIONS 1 (First order equations (A)) A.J.Hobson

Chapter 8: Taylor s theorem and L Hospital s rule

THE SECANT METHOD. q(x) = a 0 + a 1 x. with

Zeros of Functions. Chapter 10

CHAPTER 10 Zeros of Functions

Additional exercises with Numerieke Analyse

15 Nonlinear Equations and Zero-Finders

Power series solutions for 2nd order linear ODE s (not necessarily with constant coefficients) a n z n. n=0

Errata List Numerical Mathematics and Computing, 7th Edition Ward Cheney & David Kincaid Cengage Learning (c) March 2013

X. Numerical Methods

Analysis of Algorithms I: Asymptotic Notation, Induction, and MergeSort

Week 1: need to know. November 14, / 20

1 Backward and Forward Error

Examination paper for TMA4215 Numerical Mathematics

MATH 1242 FINAL EXAM Spring,

Section 4.2: The Mean Value Theorem

JUST THE MATHS UNIT NUMBER DIFFERENTIATION APPLICATIONS 5 (Maclaurin s and Taylor s series) A.J.Hobson

p. 10, Def. 1.6B: read: Any such C...

Final Examination. CS 205A: Mathematical Methods for Robotics, Vision, and Graphics (Spring 2016), Stanford University

Examples of solutions to the examination in Computational Physics (FYTN03), October 24, 2016

Mathematics for Engineers. Numerical mathematics

Regression and Nonlinear Axes

Outline. 1 Interpolation. 2 Polynomial Interpolation. 3 Piecewise Polynomial Interpolation

Logarithmic Functions

Limit. Chapter Introduction

1.2 Finite Precision Arithmetic

CS 323: Numerical Analysis and Computing

17.2 Nonhomogeneous Linear Equations. 27 September 2007

14 Lecture 14 Local Extrema of Function

Ma 530 Power Series II

TAYLOR AND MACLAURIN SERIES

November 20, Interpolation, Extrapolation & Polynomial Approximation

This practice exam is intended to help you prepare for the final exam for MTH 142 Calculus II.

Transcription:

Norwegian University of Science and Technology Department of Mathematics MA2501 Numerical Methods Spring 2015 Solutions to exercise set 7 1 Cf. Cheney and Kincaid, Exercise 4.1.9 Consider the data points x i 0 1 2 4 6 f(x i ) 1 9 23 93 259 a) Find the interpolation polynomial through these data points using Newton interpolation, and compute an approximation of f at x = 3. b) Do the same using only the first four interpolation points. a) We obtain the following table for the divided differences: 0 1 8 1 9 3 14 1 2 23 7 0 35 1 4 93 12 83 6 259 The interpolation polynomial (in nested Newton form) is therefore p(x) = 1 + x(8 + (x 1)(3 + (x 2) + 0 (x 4))). (Obviously, one can omit the term 0 (x 4) in that expression). Evaluating this polynomial, we see that p(3) = 49. b) For the interpolation polynomial of degree three through the first four interpolation points, we can simply reuse the previously computed table of divided differences and read off the result from there using only the first four columns. In this particular case, we obtain the same interpolation polynomial p as before (therefore, obviously, also its value at the point x = 3 is the same as above). March 8, 2015 Page 1 of 10

2 Suppose we have the nodes x 0 = 2, x 1 = 1, x 2 = 1 and x 3 = 4, and know the divided differences f[x 3 ] = 11, f[x 2, x 3 ] = 5, f[x 2, x 0, x 1 ] = 2, and f[x 0, x 2, x 1, x 3 ] = 0.6. What is f(-1)? To start we exploit the fact that divided differences are invariant under permutations of the nodes. Thus, sorting the nodes in increasing order, we have: f[x 3 ] = 11, f[x 2, x 3 ] = 5, f[x 0, x 1, x 2 ] = 2, and f[x 0, x 1, x 2, x 3 ] = 0.6. We seek f( 1) = f(x 1 ) = f[x 1 ]. From the recursive definition of divided differences: f[x 0, x 1, x 2, x 3 ] = f[x 1, x 2, x 3 ] f[x 0, x 1, x 2 ] x 3 x 0 which rearranged to isolate the unknown f[x 1, x 2, x 3 ] becomes f[x 1, x 2, x 3 ] = f[x 0, x 1, x 2 ] + (x 3 x 0 )f[x 0, x 1, x 2, x 3 ] = 2 + (4 ( 2))0.6 = 1.6. Similarly we find f[x 1, x 2 ] as f[x 2 ] as f[x 1, x 2 ] = f[x 2, x 3 ] (x 3 x 1 )f[x 1, x 2, x 3 ] = 5 (4 ( 1))1.6 = 3, and finally f[x 1 ] = f( 1) as f[x 2 ] = f[x 3 ] (x 3 x 2 )f[x 2, x 3 ] = 11 (4 1)5 = 4, f[x 1 ] = f[x 2 ] (x 2 x 1 )f[x 1, x 2 ] = 4 (1 ( 1))( 3) = 2. March 8, 2015 Page 2 of 10

3 Use the interpolation error formula to find a bound for the error, and compare the bound to the actual error for the case n = 2 for Exercise 6 in Exercise set 6. a) For f(x) = x ln x we first calculate f (x) = ln x+1, f (x) = 1/x and f (3) (x) = 1/x 2. Since f (3) (x) = 1/x 2 is a strictly decreasing function of x for positive x we find a bound M on the interval (8.3, 8.7) for n = 2 by M = 1/8.3 2. Employing the First Interpolation Error Theorem in the textbook with n = 2 and x = 8.4 on this interval gives the error bound: f(8.4) p 2 (8.4) = 1 1 (8.4 8.3) (8.4 8.6) (8.4 8.7) (2 + 1)! 8.32 0.1 0.2 0.3 6 8.3 2 1.46 10 5. The true error is 1.38 10 5 correct to the digits given. As expected this is lower than the error bound. b) For f(x) = x 4 x 3 + x 2 x + 1 we again calculate f (x) = 4x 3 3x 2 + 2x 1, f (x) = 12x 2 6x + 2 and f (3) (x) = 24x 6. Since f (3) (x) = 24x 6 is easily observed to be a strictly decreasing function of x for negative x, we find a bound M on the interval ( 0.5, 0) for n = 2 by M = f (3) ( 0.5) = 18. Employing the First Interpolation Error Theorem in the textbook with n = 2 and x = 1/3 on this interval gives the error bound: 1 f( 1/3) p 2 ( 1/3) 18 ( 1/3 + 1/2) ( 1/3 + 1/4) ( 1/3 0) (2 + 1)! 18 = 6 6 12 3 = 1 72 1.4 10 2. The true error is 9.6 10 3 correct to the digits given. As expected this is lower than the error bound. 4 Given the function f(x) = e x sin x on the interval [ 4, 2]. a) Show by induction f (m) (x) = dm dx m f(x) = 2m/2 e x sin(x + mπ/4). b) Let p n (x) be the polynomial that interpolates f(x) in n+1 uniformly distributed nodes (including the endpoints). Find an upper bound for the interpolation error on this interval, expressed using n. What must n be to guarantee an error less than 10 5? Use trial and error or calculate it using Matlab. c) Use Matlab to verify that the value of n found in b), indeed results in an error less than 10 5. March 8, 2015 Page 3 of 10

a) It is trivial to see that the formula is valid for m = 0. Assume now it is correct for some m 0. Then f (m+1) (x) = d dx f (m) (x) = 2 m/2 [e x sin (x + mπ/4) + e x cos (x + mπ/4)] = 2 m/2 e x [sin (x + mπ/4) + cos (x + mπ/4)] = 2 m/2 e x [ 1 2 2 sin (x + mπ/4) + 1 ] cos (x + mπ/4) 2 = 2 m/2 e x 2 1/2 [cos (π/4) sin (x + mπ/4) + sin (π/4) cos (x + mπ/4)] = 2 (m+1)/2 e x sin ((x + mπ/4) + π/4) = 2 (m+1)/2 e x sin (x + (m + 1)π/4). This shows that the formula holds for m + 1 as well, assuming it holds for m. The proof by induction is then complete, and the stated formula is seen to hold for all m 0 b) Our function f has continous derivatives of any order on the entire real line and on the interval [ 4, 2] it satisfies f (n+1) (x) = 2 (n+1)/2 e x sin(x + (n + 1)π/4) 2 (n+1)/2 e 2 = M, where we have used that e x is a strictly increasing positiv function and that sin(x) 1 for all x. The Second Interpolation Error Theorem is applicable and we have for x [ 4, 2] 1 f(x) p(x) 4(n + 1) Mhn+1 = ( e 2 6 ) n+1 2 =. 4(n + 1) n 1 4(n + 1) 2(n+1)/2 e 2 ( ) 6 n+1 n By differentiating, this expression is readliy seen to decrease with increasing positive n. Using trial and error we find that the expression has the value 1.27 10 5 for n = 15 and 2.26 10 6 for n = 16. Thus n = 16 is required to guarantee an accuracy of 10 5. c) Using Matlab we have plotted below f(x) p(x) over the interval, where p(x) is the interpolating polynomial from b) with n = 16. We observe that the error stays several orders of magnitude lower than our requirement of 10 5. This is to be expected. Recall that we were in fact guaranteed an error less than 2.26 10 6, and the theorem only gives fairly loose upper bounds. Also observe that the errors near the endpoints dominate the errors towards the middle of the interval. This is typical, and illustrates why Chebyshev nodes, which become more clustered as we approach the endpoints, usually gives siginificantly smaller errors for the same number of nodes. March 8, 2015 Page 4 of 10

f(x) p(x) 8 x 10 9 7 6 5 4 3 2 1 0 4 3 2 1 0 1 2 x 5 Consider the function f(x) = (x 2 + 1)e x. a) Use central differences, i.e., the formula f (x) f(x + h) f(x h), 2h with step sizes h = 1, h = 1/2, h = 1/4, h = 1/8, in order to approximate f (0). b) Use Richardson extrapolation for obtaining a better approximation of f (0) from the values you have already computed. a) The approximations of f (0) for the different step sizes are: h = 1 : 2.350402387287603, h = 1 2 : 1.302738263734369, h = 1 4 : 1.073602346434715, h = 1 8 : 1.018271923834063. b) The first step in Richardson extrapolation uses the approximations of f (0) calculated above. These were: D(0, 0) = 2.350402387287603, D(1, 0) = 1.302738263734369, D(2, 0) = 1.073602346434715, D(3, 0) = 1.018271923834063. March 8, 2015 Page 5 of 10

In the next step, we compute the next column by the formula for k 1. We obtain Next we compute for k 2, obtaining D(k, 1) = D(k, 0) + 1 3 ( D(k, 0) D(k 1, 0) ) D(1, 1) = 0.953516889216624, D(2, 1) = 0.997223707334830, D(3, 1) = 0.999828449633845. D(k, 2) = D(k, 1) + 1 ( ) D(k, 1) D(k 1, 1) 15 D(2, 2) = 1.000137495209377, D(3, 2) = 1.000002099120446. Finally, we have and thus D(3, 3) = D(3, 2) + 1 ( ) D(3, 2) D(2, 2) 63 D(3, 3) = 0.999999949976177. (Note that the actual solution is f (0) = 1. significantly better.) Thus the approximations indeed become 6 Derive an O(h 4 ) five point formula to approximate f (x 0 ) that uses f(x 0 h), f(x 0 ), f(x 0 + h), f(x 0 + 2h) and f(x 0 + 3h). Test the formula on f(x) = sin x at x = 1 and try to verify numerically that is has the stated order. Hint: Consider the expression Af(x 0 h)+bf(x 0 +h)+cf(x 0 +2h)+Df(x 0 +3h). Expand in fourth Taylor polynomials, and choose A, B, C and D appropriately. March 8, 2015 Page 6 of 10

We follow the hint and expand the terms in Af(x 0 h) + Bf(x 0 + h) + Cf(x 0 + 2h) + Df(x 0 + 3h) in fourth order Taylor polynomials and reorganize. Af(x 0 h) + Bf(x 0 + h) + Cf(x 0 + 2h) + Df(x 0 + 3h) = ] = A [f(x 0 ) hf (x 0 ) + h2 2! f (x 0 ) h3 3! f (3) (x 0 ) + h4 4! f (4) (x 0 ) h5 5! f (5) (ξ A ) ] + B [f(x 0 ) + hf (x 0 ) + h2 2! f (x 0 ) + h3 3! f (3) (x 0 ) + h4 4! f (4) (x 0 ) + h5 5! f (5) (ξ B ) + C [f(x 0 ) + 2hf (x 0 ) + (2h)2 2! + D [f(x 0 ) + 3hf (x 0 ) + (3h)2 2! f (x 0 ) + (2h)3 3! f (x 0 ) + (3h)3 3! f (3) (x 0 ) + (2h)4 4! f (3) (x 0 ) + (3h)4 4! = (A + B + C + D)f(x 0 ) + ( A + B + 2C + 3D)hf (x 0 ) + (A + B + 4C + 9D) h2 2! f (x 0 ) + ( A + B + 8C + 27D) h3 3! f (3) (x 0 ) + (A + B + 16C + 81D) h4 4! f (4) (x 0 ) + ( A + B + 32C + 243D)4 h5 5! f (5) (ξ) f (4) (x 0 ) + (2h)5 f (5) (ξ C ) 5! f (4) (x 0 ) + (3h)5 f (5) (ξ D ) 5! Where f (5) (ξ) = [f (5) (ξ A ) + f (5) (ξ B ) + f (5) (ξ C ) + f (5) (ξ D )]/4 for some ξ (x 0 h, x 0 + 3h). To get a O(h 4 ) five point approximation for f (x 0 ) we require the coefficient in front of f (x 0 ) to equal 1 and the O(h 2 )-, O(h 3 )- and O(h 4 )-term to vanish. This yields the linear system of equations h( A + B + 2C + 3D) = 1, A + B + 4C + 9D = 0, A + B + 8C + 27D = 0, A + B + 16C + 81D = 0, ] ] which is easily solved for A = 3 12h, B = 18 12h, C = 6 12h, D = 1 12h. Inserting this into the expression and moving the term to the left hand side we get (A + B + C + D)f(x 0 ) = 10 12h f(x 0) 3f(x 0 h) 10f(x 0 ) + 18f(x 0 + h) 6f(x 0 + 2h) + f(x 0 + 3h) 12h = f (x 0 ) + 2.4h 4 f (5) (ξ) = f (x 0 ) + O(h 4 ), so the expression on the left hand side is the desired five point formula. To verify that this formula has indeed fourth order, we compute with it an approximation to the derivative of f(x) = sin x at x = 1 for h = 1/2 i for i = 2, 3,..., 9, and plot the March 8, 2015 Page 7 of 10

10 2 E (Aboslute differentiation error) 10 4 10 6 10 8 10 10 E h 4 10 12 10 14 10 3 10 2 h 10 1 10 0 Figure 1: absolute error, i.e. the abolute value of the difference between the computed value and f (1) = cos 1, against h with logarithmic axes in Matlab, see Figure 1. We include a plot of the function h 4 (marked in red) for comparison. If the formula is indeed fourth order, the absolute error should be O(h 4 ). This means that it behaves like the function ch 4 for some positive constant c and sufficiently small values of h. Denote this error by E. If E ch 4 for small h then taking logarithms we see that log E 4 log h + log c. So if the formula is fourth order, our logarithmic error plot should be a straight line with the same slope as the plot of the function h 4. We observe precisely this. Note that reducing h further would cause the line to flatten out, since numerical rounding errors would then start to dominate. 7 Cf. Cheney and Kincaid, Exercises 4.1.16 17 Assume that the function ϕ has the form ϕ(h) = L c 1 h c 2 h 2 c 3 h 3 c 4 h 4... a) Combine the values ϕ(h) and ϕ(h/2) in order to obtain a higher order approximation of L. b) Try to generalize the idea of Richardson extrapolation to the function ϕ. March 8, 2015 Page 8 of 10

a) We have Consequently, ϕ(h) =L c 1 h c 2 h 2 c 3 h 3 c 4 h 4..., h ϕ(h/2) =L c 1 2 c h 2 2 4 c h 3 3 8 c h 4 4 16... h 2 2ϕ(h/2) ϕ(h) = L + c 2 2 + c 3h 3 3 4 + c 7h 4 4 8 +..., which is an approximation of L of order O(h 2 ). b) Assume now that ψ(h) = L a m h m a m+1 h m+1 a m+2 h m+2... is an order O(h m ) approximation of L. Then and therefore Thus ψ(h/2) = L a m h m h m+1 h m+2 2 m a m+1 2 m+1 a m+2 2 m+2..., 2 m ψ(h/2) ψ(h) = (2 m h m+1 3h m+2 1)L + a m+1 + a m+1 +... 2 4 ψ(h/2) + 1 ( ) ψ(h/2) ψ(h) = L + dm+1 2 m h m+1 + d m+1 h m+2 +... 1 is an order O(h m+1 ) approximation of L. Thus is makes sense to compute approximations of L by: Compute approximations D(k, 0) := φ(2 k h). (This is the same as for the usual Richardson extrapolation.) Compute recursively D(k, m) := D(k, m 1) + 1 ( ) D(k, m 1) D(k 1, m 1). 2 m 1 (This is almost the same as the usual Richardson extrapolation, but the factor 1/(4 m 1) is replaced by 1/(2 m 1).) 8 Write a Matlab-program for the approximation of derivatives using central differences and Richardson extrapolation. Your program should take as an input a function f, a point x where you want to approximate f, a basic step size h, and the desired approximation order. Test your program on the function sin with x = π/3 and on the function from exercise 5. For which parameters do you obtain the best results? March 8, 2015 Page 9 of 10

There is an example program on the webpage of the course. Applying the program to the function sin with x = π/3 and h = 1, it turns out that the error becomes smallest for k = 5 (that is, D(5, 5) is the best approximation of the true value sin (π/3) = 1/2). In the case of the function f(x) = (x 2 + 1)e x with x = 0 and h = 1, the best result is obtained for k = 7. In Figure 2 you find a plot of D(k, k) f (0) versus k for different initial step sizes h. In all cases one observes first a rapid decrease of the error as k increases, followed by a slow (and unsteady) increase of the error. The initial decrease is due to the higher approximation order. At some point, however, rounding errors due to the limited precision of the computations become notable, and the error starts to increase again. 10 5 10 0 D(k,k)-f'(0) 10-5 10-10 10-15 10-20 0 1 2 3 4 5 6 7 8 9 10 h Figure 2: h = 1 blue, h = 1/8 red, h = 1/64 yellow. March 8, 2015 Page 10 of 10