Integration of Ordinary Differential Equations

Similar documents
Fourth Order RK-Method

Numerical Differential Equations: IVP

Chapter 8. Numerical Solution of Ordinary Differential Equations. Module No. 1. Runge-Kutta Methods

ECE257 Numerical Methods and Scientific Computing. Ordinary Differential Equations

Second Order ODEs. CSCC51H- Numerical Approx, Int and ODEs p.130/177

PowerPoints organized by Dr. Michael R. Gustafson II, Duke University

Chapter 11 ORDINARY DIFFERENTIAL EQUATIONS

Polynomial Interpolation

PowerPoints organized by Dr. Michael R. Gustafson II, Duke University

THE TRAPEZOIDAL QUADRATURE RULE

Exercises, module A (ODEs, numerical integration etc)

Applied Math for Engineers

16.7 Multistep, Multivalue, and Predictor-Corrector Methods

Numerical Methods - Initial Value Problems for ODEs

Review Higher Order methods Multistep methods Summary HIGHER ORDER METHODS. P.V. Johnson. School of Mathematics. Semester

Scientific Computing with Case Studies SIAM Press, Lecture Notes for Unit V Solution of

Euler s Method, cont d

A Brief Introduction to Numerical Methods for Differential Equations

The Definition and Numerical Method of Final Value Problem and Arbitrary Value Problem Shixiong Wang 1*, Jianhua He 1, Chen Wang 2, Xitong Li 1

Initial value problems for ordinary differential equations

An Overly Simplified and Brief Review of Differential Equation Solution Methods. 1. Some Common Exact Solution Methods for Differential Equations

Lecture Notes on Numerical Differential Equations: IVP

2.29 Numerical Fluid Mechanics Fall 2011 Lecture 20

Ordinary Differential Equations

TS Method Summary. T k (x,y j 1 ) f(x j 1,y j 1 )+ 2 f (x j 1,y j 1 ) + k 1

PowerPoints organized by Dr. Michael R. Gustafson II, Duke University

Chap. 20: Initial-Value Problems

CS 450 Numerical Analysis. Chapter 9: Initial Value Problems for Ordinary Differential Equations

ODE Runge-Kutta methods

Lecture Notes to Accompany. Scientific Computing An Introductory Survey. by Michael T. Heath. Chapter 9

Scientific Computing: An Introductory Survey

Initial value problems for ordinary differential equations

Differential Equations

Mathematics for chemical engineers. Numerical solution of ordinary differential equations

Jim Lambers MAT 772 Fall Semester Lecture 21 Notes

16.1 Runge-Kutta Method

Initial Value Problems

Consistency and Convergence

Differential Equations

multistep methods Last modified: November 28, 2017 Recall that we are interested in the numerical solution of the initial value problem (IVP):

16.7 Multistep, Multivalue, and Predictor-Corrector Methods

AN OVERVIEW. Numerical Methods for ODE Initial Value Problems. 1. One-step methods (Taylor series, Runge-Kutta)

HIGHER ORDER METHODS. There are two principal means to derive higher order methods. b j f(x n j,y n j )

Solving Ordinary Differential equations

Solving PDEs with PGI CUDA Fortran Part 4: Initial value problems for ordinary differential equations

8.1 Introduction. Consider the initial value problem (IVP):

Ordinary differential equations - Initial value problems

COSC 3361 Numerical Analysis I Ordinary Differential Equations (II) - Multistep methods

Solving Ordinary Differential Equations

EXAMPLE OF ONE-STEP METHOD

Numerical Methods. King Saud University

Lecture 4: Numerical solution of ordinary differential equations

KINGS COLLEGE OF ENGINEERING DEPARTMENT OF MATHEMATICS ACADEMIC YEAR / EVEN SEMESTER QUESTION BANK

Solution of Nonlinear Equations

Chapter 8. Numerical Solution of Ordinary Differential Equations. Module No. 2. Predictor-Corrector Methods

NUMERICAL SOLUTION OF ODE IVPs. Overview

9.6 Predictor-Corrector Methods

Ordinary Differential Equations (ODEs)

1 Ordinary differential equations

Ordinary Differential Equations. Monday, October 10, 11

Computational Methods CMSC/AMSC/MAPL 460. Ordinary differential equations

Physically Based Modeling: Principles and Practice Differential Equation Basics

Introduction to the Numerical Solution of IVP for ODE

(f(x) P 3 (x)) dx. (a) The Lagrange formula for the error is given by

2.1 NUMERICAL SOLUTION OF SIMULTANEOUS FIRST ORDER ORDINARY DIFFERENTIAL EQUATIONS. differential equations with the initial values y(x 0. ; l.

Ph 22.1 Return of the ODEs: higher-order methods

Convergence Rates on Root Finding

Computational Methods CMSC/AMSC/MAPL 460. Ordinary differential equations

CHAPTER 5: Linear Multistep Methods

Research Computing with Python, Lecture 7, Numerical Integration and Solving Ordinary Differential Equations

Ordinary Differential Equations

Chapter 6 - Ordinary Differential Equations

Mini project ODE, TANA22

Virtual University of Pakistan

Solution of First Order Initial Value Problem by Sixth Order Predictor Corrector Method

2tdt 1 y = t2 + C y = which implies C = 1 and the solution is y = 1

Bindel, Spring 2012 Intro to Scientific Computing (CS 3220) Week 12: Monday, Apr 16. f(x) dx,

The family of Runge Kutta methods with two intermediate evaluations is defined by

Numerical solution of ODEs

Multistep Methods for IVPs. t 0 < t < T

1 Error Analysis for Solving IVP

Introduction to standard and non-standard Numerical Methods

8 Numerical Integration of Ordinary Differential

Numerical Methods for Differential Equations

4 Stability analysis of finite-difference methods for ODEs

Numerical Methods for the Solution of Differential Equations

Chapter 1 Error Analysis

Ordinary Differential Equations

Ordinary differential equation II

Bindel, Fall 2011 Intro to Scientific Computing (CS 3220) Week 12: Monday, Apr 18. HW 7 is posted, and will be due in class on 4/25.

Physically Based Modeling Differential Equation Basics

One-Step Hybrid Block Method with One Generalized Off-Step Points for Direct Solution of Second Order Ordinary Differential Equations

Explicit One-Step Methods

Chapter 8: Taylor s theorem and L Hospital s rule

Computation Fluid Dynamics

You may not use your books, notes; calculators are highly recommended.

Backward error analysis

Power series solutions for 2nd order linear ODE s (not necessarily with constant coefficients) a n z n. n=0

10 Numerical Solutions of PDEs

Transcription:

Integration of Ordinary Differential Equations Com S 477/577 Nov 7, 00 1 Introduction The solution of differential equations is an important problem that arises in a host of areas. Many differential equations are too difficult to solve in closed form. Instead, it becomes necessary to employ numerical techniques. Differential equations have a major application in understanding physical systems that involve aerodynamics, fluid dynamics, thermodynamics, heat diffusion, mechanical oscillations, etc. They are used for developing control algorithms for dynamic simulations. Other applications also include optimization and stochastics. We will consider ordinary differential equations ODEs and focus on two classes of problems: a first-order initial-value problems and b linear higher-order boundary-value problems. The basic principles we will see in these two classes of problems carry over to more general problems. For instance, higher-order initial-value problems can be rewritten in vector form to yield a set of simultaneous first-order equations. First-order techniques may then be used to solve this system. Consider a general nth order differential equation in the form y n x = f x,yx,y x,...,y n 1 x, 1 where f is a function from R n to R. A general solution of this equation will usually contain n arbitrary constants. A particular solution is obtained by specifying n auxiliary constraints. For instance, one might specify the values of y and its derivatives at some point x = x 0 as yx 0,y x 0,...,y n 1 x 0. Such a problem is called an initial-value problem. In effect, the auxiliary conditions specify all the relevant information at some starting point x 0, and the differential equation tells us how to proceed from that point. If the n auxiliary conditions are specified at different points, then the problem is called a boundary-value problem. In this case, we tie the function or its derivatives down at several points, and the differential equation tells us the shape of yx between those points. For example, the following n conditions on y are specified at two points x 0 and x 1 : f 1 yx 0,...,y n 1 x 0,yx 1,...,y n 1 x 1 = 0, f n yx 0,...,y n 1 x 0,yx 1,...,y n 1 x 1. = 0. 1

The nth order equation 1 can be transformed into a system of first order equations if we introduce n 1 variables z i = y i, i = 1,...,n 1. This system consists of n equations: y = z 1, z 1 = z,. z n 1 = fx,y,z 1,...,z n 1. We will focus on numerical techniques of solving first-order initial-value problems, assuming basic knowledge of analytics techniques. Numerical techniques are useful both for non-linear differential equations and for equations whose analytic solutions consist of complicated integrals and exponentials, etc. Stated formally, the problem is to recover yx from the differential equation dy dx = fx,y with initial condition yx 0 = y 0. We will introduce three types of numerical methods to solve the above equation. Taylor s Algorithm If the exact solution yx has a Taylor series expansion about x 0, then we can write yx = y 0 +x x 0 y x 0 + x x 0 y x 0 +! Of course, if we do not know yx, then we do not explicitly know the values of its derivatives y x 0,y x 0,... However, if f is sufficiently differentiable, then we can obtain the derivatives of y from f: y x = f x, yx, y = f x +f y y = f x +f y f, y = f xx +f xy f +f yx f +f yy f +f y f x +f yf = f xx +f xy f +f yy f +f x f y +f yf,. From now on we will denote by yx n the true value of yx at the point x = x n, and by y n an approximate value. We will generate a series of points x 0,x 1,... by marching out from x 0. We will obtain y n+1 from y n by using a truncated Taylor series, with approximate derivative values obtained from equations. Suppose we choose a step size h > 0. Our points of interest are x 0,x 1 = x 0 +h,x = x 0 +h,...

Let us define a truncated Taylor series operator: T k x,y = fx,y+ h! f x,y+ + hk 1 f k 1 x,y, for each k = 1,,..., k! where f j x,y = dj dx jf x, yx = y j+1 x. Taylor s series tells us that yx+h = yx+ht k x,y+r k, where r k is a remainder that depends on y k+1 for each k. This formula provides us with a numerical method for obtaining y n+1 from y n. Taylor s algorithm of order k finds an approximate solution to the differential equation y x = fx,y, ya = y 0, over the interval [a,b]. It starts with discretizing [a,b] into N subintervals of equal length h = b a N and set x n = a+nh,n = 0,1,...,N. Then it carries out the following iteration: y n+1 y n +ht k x n,y n, for n = 0,1,...,N 1. The calculation of y n+1 uses information about y and its derivatives that comes from a single point, namely from x n. For this reason, Taylor s algorithm is a one-step method. Taylor s algorithm of order 1, also known as Euler s method, has the basic update rule: y n+1 y n +hfx n,y n. Unfortunately, this method is not very accurate. It requires very small step sizes to obtain good accuracy. And there are stability problems such as error accumulation. Nevertheless, the basic idea of adding small increments to previous estimates to obtain new estimates leads to more advanced methods. 3 Error Estimation We would like to understand the quality of our differential equation solvers by estimating the error between yx n and y n. There are three types of errors: 1. Local discretization error This is the error introduced in a single step of the equation solver as it moves from x n to x n+1. In other words, it is the error in the estimate y n+1 that would result if yx n were known perfectly.. Full discretization error This is the net error between yx n and y n at step n. This error is the sum of the local discretization errors, plus any numerical roundoff errors. 3. Numerical roundoff error Limited machine precision can introduce errors. This type of errors will be ignored by us. In general, an algorithm is said to be of order k if its local discretization error is Oh k+1, where h is the step size. 3

3.1 Local Discretization Error Taylor s theorem tells us that the local error for Taylor s algorithm of order k is simply h k+1 f ξ,yξ k h k+1 y k+1 ξ, or equivalently,, k +1! k +1! where ξ is some point in the interval x n,x n +h. Thus Taylor s algorithm of order k is indeed of order k. And Euler s algorithm is of order 1. 3. Full Discretization Error This type of error can be very difficult to estimate. Let us therefore illustrate the approach with Euler s method. First, we define e n = yx n y n. Taylor s theorem says that Recall that yx n+1 = yx n +hy x n + h y ξ n, for some ξ n x n,x n+1. Subtract this equation from the previous one: y n+1 = y n +hfx n,y n. e n+1 = e n +h f x n,yx n fx n,y n + h y ξ n. Now, we apply the mean-value theorem to f and obtain e n+1 = e n +hf y x n,ȳ n yx n y n + h y ξ n = e n 1+hf y x n,ȳ n + h y ξ n, 3 for some ȳ n between yx n and y n. To solve the recurrence relation 3, we assume that f y and y are both bounded on [a,b], say, x,yx f y L, for x [a,b]. Then it follows that y x Y, e n+1 1+hL e n + h Y. Since e 0 = 0 and 1 + hl > 1, we can prove by induction that e n δ n, where δ n satisfies the recurrence δ n+1 = 1+hLδ n + h Y. 4

This recurrence has the solution Consequently, δ n = hy L 1+hLn 1. e n δ n hy L = hy L e nhl 1 e xn x0l 1. 4 From the above analysis we see that the full discretizaton error approaches zero as the step size h is made smaller. That the error is Oh implies that the convergence may be slow. Example 1. Consider the differential equation y = y, with y1 = 1. The exact solution is yx = 1 x. Suppose we use Euler s method to obtain a numerical solution over the interval [1,]. The step size we choose is h = 0.1. Euler s method says In this problem y 0 = 1 and fx n,y n = y n. First, we obtain the local error as y n+1 = y n +hfx n,y n, n = 0,1,... h E max 1 x y x h = max 1 x x 3 = 0.01. To obtain full error, note that y x = x and f 3 y x,yx = y = x. So we use Y = and L =. Therefore it follows from 4 that e n h e xn 1 1, and max n e n h e 1 0.3195. So we should expect about 1 decimal digit of accuracy. 5

The following table compares the actual results obtained against those yielded from the exact solution. x n y n fx n,y n yx n 1.0 1.0 1.0 1.0 1.1 0.9 0.810 0.90909 1. 0.819 0.67076 0.83333 1.3 0.7519 0.56539 0.7693 1.4 0.69539 0.48356 0.7149 1.5 0.64703 0.41865 0.66667 1.6 0.60516 0.366 0.65 1.7 0.56854 0.334 0.5884 1.8 0.536 0.8753 0.55556 1.9 0.50746 0.575 0.563.0 0.48171 0.305 0.5 The maximum error is just under 0.0, well within our estimates. But it is still not small enough. 4 Runge-Kutta Methods Euler s method is unacceptable because it requires mall step sizes. Higher-order Taylor algorithms are unacceptable because they require higher-order derivatives. Runge-Kutta methods attempt to obtain greater accuracy than, say, Euler s method, without explicitly evaluating higher order derivatives. These methods evaluate fx, y at selected points in the interval [x n,x n+1 ]. The idea is to combine the estimates of y resulting from these selected evaluations in such a way that error terms of order h,h,..., etc., are canceled out, to the desired accuracy. For instance, the Runge-Kutta method of order tries to cancel terms of order h and h, leaving a local discretization error of order h 3. Similarly, the Runge-Kutta method of order 4 tries to cancel terms of order h,h,h 3,h 4, leaving an Oh 5 local discretization error. 5 Runge-Kutta Method of Order We wish to evaluate fx,y at two points in the interval [x n,x n+1 ], then combine the results to obtain an error of order h 3. The basic step of the algorithm is where y n+1 = y n +ak 1 +bk, k 1 = hfx n,y n, k = hfx n +αh,y n +βk 1. In the above, a,b,α, and β are fixed constants. They will be chosen in such a way as to obtain the Oh 3 local discretization error. Intuitively, the basic step first evaluates y at x n,y n by computing y n = fx n,y n. The algorithm then tentatively usesthis approximation toy tostep tothetrial pointx n +αh,y n +βk 1. At this trial point, the algorithm reevaluates the derivative y, then uses both derivative estimates 6

to compute y n+1. By using two points to compute y, the algorithm should be obtaining second derivative information, hence be able to reduce the size of the discretization error. Let us derive the constants a,b,α and β. First, we use Taylor s expansion in x: yx n+1 = yx n +hy x n + h! y x n + h3 3! y x n + = yx n +hfx n,y n + h f x +ff y xn,y n + h3 6 f xx +f xy f +f yy f +f x f y +f y f x n,y n +Oh 4. 5 Meanwhile, Taylor s expansion in two variables x and y says that fx n +αh,y n +βk 1 = fx n,y n +αhf x +βk 1 f y + α h Subsequently, we have y n+1 = y n +ak 1 +bk = y n +a+bhf +bh αf x +βff y +bh 3 α f xx + β k 1 f yy +αhβk 1 f xy +Oh 3. f xx +β f f yy +αβff xy +Oh 4. 6 We want y n+1 to estimate yx n+1 with Oh 3 local error. Let us equate the formulas 5 with 6 and match the terms in h and h, respectively. This yields the following three constraint equations for a,b,α, and β: a+b = 1, αb = 1, βb = 1. Note that for the purpose of local discretization error, yx n matches y n. The three constraint equations on four constants gives us a family of second-order Runge-Kutta methods. One popular method is to choose a = b = 1 and α = β = 1. This effectively places the trial point at x n+1, then moves from y n to y n+1 by averaging the derivatives computed at x n and x n+1. Thus y n+1 = y n + 1 k 1 +k with k 1 = hfx n,y n, k = hfx n +h,y n +k 1. The Runge-Kutta method of order has a local discretization error Oh 3. This is superior than Euler s method, which has local error Oh. In order for a Taylor-based method to obtain Oh 3 error, it must compute second derivatives. The Runge-Kutta method, however, does not need to compute second derivatives. Instead it performs two evaluations of the first derivative. 7

5.1 Runge-Kutta of Order 4 A widely used method is the following order 4 Runge-Kutta method. It produces Oh 5 local discretization error, and thus leads to solutions quickly and accurately. The integration formula has the form y n+1 = y n + 1 6 k 1 +k +k 3 +k 4, where k 1 = hfx n,y n, k = hf x n + h,y n + k 1, k 3 = hf x n + h,y n + k, k 4 = hfx n +h,y n +k 3. The term k 1 estimates the derivative y = fx,y at the start point. The term k estimates y at the midpoint between x n and x n+1, using a y value obtained with k 1. The term k 3 estimates y at the midpoint xn+x n+1, but with an updated y value obtained by using k. The term k 4 estimates y at the point x n+1, using a y value obtained by advancing from y n by k 3. yx n+1 y n+1 3 true value x n+1,yx n+1 5 estimate x n+1,y n+1 4 yx n = y n 1 x n x n +x n+1 x n+1 In the above figure, the curve represents the actual solution to the differential equation y x = fx,y. If yx n = y n as in the picture, then we desire that y n+1 = yx n+1. The picture shows the process whereby y n+1 is computed. Points are labelled by numbers in the order that they are computed. We assume that the vector field given by y = fx,y is continuous, and that other integral curves locally look like the one shown. Example. Reconsider the differential equation in Example 1: y = y with y1 = 1. The table below shows the solution of this problem over the interval [1, ] using the fourth order Runge-Kutta 8

method with a step size h = 0.1. x n y n fx n,y n yx n 1.0 1.0 1.0 1.0 1.1 0.90909 0.8645 0.90909 1. 0.83333 0.69445 0.83333 1.3 0.7693 0.5917 0.7693 1.4 0.7149 0.5100 0.7149 1.5 0.66667 0.44445 0.66667 1.6 0.6500 0.39063 0.65 1.7 0.5884 0.3460 0.5884 1.8 0.55556 0.30864 0.55556 1.9 0.563 0.7701 0.563.0 0.50000 0.5000 0.5 Compare this with the results of Euler s method shown in Example 1. There is no error to the fifth decimal digits. In fact, the error was less than 5 10 7. 6 Adams-Bashforth Method Runge-Kutta propagates a solution over an interval by combining the information from several Euler-style steps each involving one evaluation of the right-hand f s, and then using the information obtained to match a Taylor series expansion up to some higher order. It essentially makes several small steps for each major step from x n to x n+1. An alternative of moving from x n to x n+1 is to make use of prior information at x n 1,x n,... In other words, rather than evaluating fx, y at several intermediate points, make use of known values of fx,y at several past points. Methods that make use of information at several of the x i s are called multi-step methods. We will here look at one such method, known as the Adams- Bashforth method. Suppose we have approximations to yx and y x at the points x 0,...,x n. If we integrate the differential equation y x = f x, yx from x n to x n+1, we obtain xn+1 x n y xdx = xn+1 x n f x, yx dx. Hence xn+1 y n+1 = y n + f x, yx dx. x n How do we evaluate the integral? The trick is to approximate fx,yx with an interpolating polynomial. Of course, we do not know exact values of fx,yx anywhere except at x 0. However, we do have approximate values fx k,y k for k = 0,1,...,n. So we will construct an interpolating polynomial that passes through some of those values. Specifically, if we are interested in an order m + 1 method, we will approximate the function fx,yx with an interpolating polynomial that passes through the m + 1 points x n m,f n m,..., x n,f n. Here f i = fx i,y i. We then integrate this polynomial in order to approximate the integral x n+1 x n fx,yxdx. 9

We have seen interpolating polynomials written in terms of divided differences. It is also possible to write interpolating polynomials in terms of forward differences. Forward differences are defined recursively as { i fk if i = 0; f k = i 1 f k+1 i 1 f k if i > 0. A forward difference table looks like this: x i 0 1 3 4 5 x 0 f 0 f 0 x 1 f 1 f 0 f 1 3 f 0 x f f 1 4 f 0 f 3 f 1 5 f 0 x 3 f 3 f 4 f 1 f 3 3 f x 4 f 4 f 3 f 4 x 5 f 5 One advantage of forward differences is that no division needs to be performed. It turns out that one can write the order m interpolating polynomial that passes through the points x i,f i, i = n m,...,n, as m s p m x = 1 k k f n k, k where k=0 s k s = x x n, h = s s 1 s k +1, if k 1. k Since dx = hds, we obtain that 1 m s y n+1 y n +h 1 k k f n k ds 0 k k=0 } = y n +h {γ 0 f n +γ 1 f n 1 + +γ m m f n m 7 where γ k = 1 k 1 0 s ds. k The sequence {γ k } are precomputable numbers. The simplest case, obtained by setting m = 0 in 7, leads to Euler s method. 10

One popular Adams-Bashforth method is the one of order 4. To derive it, we need to compute γ i, i = 0,1,,3: So γ 0 = 1, γ 1 = γ = γ 3 = 1 0 1 0 1 sds = 1, s s 1 ds = 5 1, 0 s s 1 s 6 ds = 3 8. y n+1 = y n +h f n + 1 f n 1 + 5 1 f n + 3 8 3 f n 3. 8 From the definition of the forward-difference operator we derive that f n 1 = f n f n 1, f n = f n f n 1 +f n, 3 f n 3 = f n 3f n 1 +3f n f n 3. Substituting the above in 8 and regrouping, we obtain the update formula: y n+1 = y n + h 4 The local error can be determined as 55f n 59f n 1 +37f n 9f n 3. E = 51 70 h5 y 5 ξ, for some ξ x n 3,x n+1. Example 3. Let us return to the example y = y with y1 = 1 to be solved over the interval [1,]. In order to start off the order 4 Adams-Bashforth method, we need to know the values of fx,yx at the four points x 0,x 1,x,x 3. In general, one obtains these either by knowing them, or by running a highly accurate method that computes approximations to yx 1,yx, and yx 3. In our case, we will take the values obtained by the Runge-Kutta method in Example. x n y n fx n,y n yx n 1.0 1.0 1.0 1.0 1.1 0.90909 0.8645 0.90909 1. 0.83333 0.69445 0.83333 1.3 0.7693 0.5917 0.7693 1.4 0.71444 0.5104 0.7149 1.5 0.66686 0.44470 0.66667 1.6 0.655 0.39093 0.65 1.7 0.58848 0.34631 0.5884 1.8 0.55580 0.3089 0.55556 1.9 0.5655 0.776 0.563.0 0.5003 0.503 0.5 11

Notice that the maximum error is approximately.5 10 4. This is indeed within the bound we obtained for the local error, namely 51 y E max 1 x 750 h5 5 x 51 = max 1 x 70.1510 x 6 4. 10 4. Both the Runge-Kutta method and the Adams-Bashforth method are methods of order 4, meaning that their local discretization error is Oh 5. However, the constant coefficient hidden in the error term tends to be higher for the Adams-Bashforth method than for the Runge-Kutta method. Consequently, the Runge-Kutta method tends to exhibit greater accuracy. This is shown in Example 3. Adams-Bashforth also has the drawback that it is not self-starting one must supply four initial data points rather than just one. Both methods require evaluation of fx,yx at four points in order to move from x n to x n+1. However, while Runge-kutta must generate three intermediate points at which to evaluate fx, yx, Adams-BAshforth uses information already available. Consequently, Adams-Bashforth requires less computation and is therefore faster. Multi-step methods use the basic formula: 1 y n+1 = y n p +h p k=0 m s 1 k k f n k ds. k The integration is from x n p to x n+1 using interpolation at the points x n m,...,x n. The case p = 0 yields Adams-Bashford. Some especially interesting formulas of this type are and y n+1 = y n 1 +hf n, E = h3 6 y ξ, when m = 1 and p = 1. y n+1 = y n 3 + 4h 3 f n f n 1 +f n, E = 14 45 h5 y 5 ξ, when m = 3 and p = 3. Other important techniques and topics of the initial-value problems include Adaptive control of the step size. Multi-step formulas that integrate over several intervals. Predictor-corrector methods implicit methods. Extrapolation to the limit. Numerical instability. Stiff equations higher order equations or systems of equations whose solutions have different time scales. 1

A Two Theorems about ODEs It can be shown that the initial value problem involving the differential equation y = fx,y has exactly one solution provided that the function f satisfies a few simple regular conditions. Theorem 1 Let f be defined and continuous on the strip S = {x,y a x b,y R n }. Furthermore, let there be a constant L such that fx,y 1 fx,y L y 1 y Lipschitz condition for all x [a,b] and all y 1,y R n. Then for every x 0 [a,b] and every y 0 R n there exists exactly one function yx such that a yx is continuously differentiable on [a, b]; b y x = fx,yx for x [a,b]; c yx 0 = y 0. The second theorem states that the solution of an initial value problem depends continuously on the initial value. Theorem Let f be defined and continuous on the strip S = {x,y a x b,y R n } and satisfy the Lipschitz condition fx,y 1 fx,y L y 1 y, some constant L, for all x,y i S, i = 1,. Let a x 0 b. Then for the solution of yx;s of the initial value problem y = fx,y, yx 0 ;s = s there holds the estimate References yx;s 1 yx;s e L x x 0 s 1 s, for a x b. [1] M. Erdmann. Lecture notes for 16-811 Mathematical Fundamentals for Robotics. The Robotics Institute, Carnegie Mellon University, 1998. [] S. D. Conte and Carl de Boor. Elementary Numerical Analysis. McGraw-Hill, Inc., nd edition, 197. [3] J. Stoer and R. Bulirsch. Introduction to Numerical Analysis. Springer-Verlag New York, Inc., nd edition, 1993. 13