Simple Iteration, cont d

Similar documents
Jim Lambers MAT 460/560 Fall Semester Practice Final Exam

Numerical Methods in Informatics

Jim Lambers MAT 460 Fall Semester Lecture 2 Notes

Jim Lambers MAT 169 Fall Semester Lecture 6 Notes. a n. n=1. S = lim s k = lim. n=1. n=1

x 2 x n r n J(x + t(x x ))(x x )dt. For warming-up we start with methods for solving a single equation of one variable.

STOP, a i+ 1 is the desired root. )f(a i) > 0. Else If f(a i+ 1. Set a i+1 = a i+ 1 and b i+1 = b Else Set a i+1 = a i and b i+1 = a i+ 1

PART I Lecture Notes on Numerical Solution of Root Finding Problems MATH 435

Numerical Methods I Solving Nonlinear Equations

Numerical Methods in Physics and Astrophysics

Hermite Interpolation

Numerical Methods in Physics and Astrophysics

Jim Lambers MAT 419/519 Summer Session Lecture 13 Notes

Chapter 1. Root Finding Methods. 1.1 Bisection method

FIXED POINT ITERATION

FIXED POINT ITERATIONS

Lecture Notes to Accompany. Scientific Computing An Introductory Survey. by Michael T. Heath. Chapter 5. Nonlinear Equations

Chapter 2 Solutions of Equations of One Variable

SOLUTION OF ALGEBRAIC AND TRANSCENDENTAL EQUATIONS BISECTION METHOD

University of Houston, Department of Mathematics Numerical Analysis, Fall 2005

Outline. Math Numerical Analysis. Intermediate Value Theorem. Lecture Notes Zeros and Roots. Joseph M. Mahaffy,

Math Numerical Analysis

Nonlinearity Root-finding Bisection Fixed Point Iteration Newton s Method Secant Method Conclusion. Nonlinear Systems

Numerical solutions of nonlinear systems of equations

ROOT FINDING REVIEW MICHELLE FENG

Numerical Analysis: Solving Nonlinear Equations

CS 450 Numerical Analysis. Chapter 5: Nonlinear Equations

Outline. Scientific Computing: An Introductory Survey. Nonlinear Equations. Nonlinear Equations. Examples: Nonlinear Equations

Math 273a: Optimization Netwon s methods

Convergence Rates on Root Finding

Polynomial Approximations and Power Series

Root Finding For NonLinear Equations Bisection Method

Section 6.3 Richardson s Extrapolation. Extrapolation (To infer or estimate by extending or projecting known information.)

Vectors in Function Spaces

Economics 204 Summer/Fall 2011 Lecture 5 Friday July 29, 2011

Unconstrained optimization I Gradient-type methods

Nonlinearity Root-finding Bisection Fixed Point Iteration Newton s Method Secant Method Conclusion. Nonlinear Systems

a factors The exponential 0 is a special case. If b is any nonzero real number, then

Lecture 39: Root Finding via Newton s Method

Mathematics 1a, Section 4.3 Solutions

Goals for This Lecture:

Homework and Computer Problems for Math*2130 (W17).

3.1 Introduction. Solve non-linear real equation f(x) = 0 for real root or zero x. E.g. x x 1.5 =0, tan x x =0.

you expect to encounter difficulties when trying to solve A x = b? 4. A composite quadrature rule has error associated with it in the following form

Consistency and Convergence

E5295/5B5749 Convex optimization with engineering applications. Lecture 8. Smooth convex unconstrained and equality-constrained minimization

ECE133A Applied Numerical Computing Additional Lecture Notes

MATH 614 Dynamical Systems and Chaos Lecture 3: Classification of fixed points.

Lyapunov Stability Theory

4 Stability analysis of finite-difference methods for ODEs

NONLINEAR DC ANALYSIS

Queens College, CUNY, Department of Computer Science Numerical Methods CSCI 361 / 761 Spring 2018 Instructor: Dr. Sateesh Mane.

Taylor and Maclaurin Series. Copyright Cengage Learning. All rights reserved.

Scientific Computing: An Introductory Survey

Infinite series, improper integrals, and Taylor series

THE SECANT METHOD. q(x) = a 0 + a 1 x. with

MAT 460: Numerical Analysis I. James V. Lambers

Solution of Nonlinear Equations

c 2007 Society for Industrial and Applied Mathematics

Introduction to Numerical Analysis

1. Nonlinear Equations. This lecture note excerpted parts from Michael Heath and Max Gunzburger. f(x) = 0

Solving nonlinear equations

Numerical Methods for Large-Scale Nonlinear Systems

Chapter 3: Root Finding. September 26, 2005

Unit 2: Solving Scalar Equations. Notes prepared by: Amos Ron, Yunpeng Li, Mark Cowlishaw, Steve Wright Instructor: Steve Wright

Numerical Methods. King Saud University

Math 471. Numerical methods Root-finding algorithms for nonlinear equations

SYSTEMS OF NONLINEAR EQUATIONS

5.6 Logarithmic and Exponential Equations

MA 8019: Numerical Analysis I Solution of Nonlinear Equations

MATH 2053 Calculus I Review for the Final Exam

Linear Algebra and Matrix Inversion

Autonomous Systems and Stability

C.7. Numerical series. Pag. 147 Proof of the converging criteria for series. Theorem 5.29 (Comparison test) Let a k and b k be positive-term series

1 Solutions to selected problems

COURSE Iterative methods for solving linear systems

Jim Lambers MAT 772 Fall Semester Lecture 21 Notes

Nonlinear equations and optimization

1. Method 1: bisection. The bisection methods starts from two points a 0 and b 0 such that

INTRODUCTION, FOUNDATIONS

Continuous functions that are nowhere differentiable

Series Solutions. 8.1 Taylor Polynomials

MATH ASSIGNMENT 03 SOLUTIONS

Completion Date: Monday February 11, 2008

Introductory Numerical Analysis

On sequential optimality conditions for constrained optimization. José Mario Martínez martinez

Roundoff Analysis of Gaussian Elimination

Numerical Methods for Differential Equations Mathematical and Computational Tools

Chapter 4. Solution of Non-linear Equation. Module No. 1. Newton s Method to Solve Transcendental Equation

Analysis II: The Implicit and Inverse Function Theorems

11.10a Taylor and Maclaurin Series

Optimization. Charles J. Geyer School of Statistics University of Minnesota. Stat 8054 Lecture Notes

Numerical Methods

Variable. Peter W. White Fall 2018 / Numerical Analysis. Department of Mathematics Tarleton State University

Topic 5: The Difference Equation

Euler s Method, cont d

Applications of the Maximum Principle

CS 323: Numerical Analysis and Computing

CLASS NOTES Models, Algorithms and Data: Introduction to computing 2018

Numerical Methods. Root Finding

min f(x). (2.1) Objectives consisting of a smooth convex term plus a nonconvex regularization term;

Transcription:

Jim Lambers MAT 772 Fall Semester 2010-11 Lecture 2 Notes These notes correspond to Section 1.2 in the text. Simple Iteration, cont d In general, nonlinear equations cannot be solved in a finite sequence of steps. As linear equations can be solved using direct methods such as Gaussian elimination, nonlinear equations usually require iterative methods. In iterative methods, an approximate solution is refined with each iteration until it is determined to be sufficiently accurate, at which time the iteration terminates. Since it is desirable for iterative methods to converge to the solution as rapidly as possible, it is necessary to be able to measure the speed with which an iterative method converges. To that end, we assume that an iterative method generates a sequence of iterates x 0, x 1, x 2,... that converges to the exact solution x. Ideally, we would like the error in a given iterate x k+1 to be much smaller than the error in the previous iterate x k. For example, if the error is raised to a power greater than 1 from iteration to iteration, then, because the error is typically less than 1, it will approach zero very rapidly. This leads to the following definition. Definition (Order and Rate of Convergence) Let {x k } k=0 be a sequence in Rn that converges to x R n and assume that x k = x for each k. We say that the order of convergence of {x k } to x is order r, with asymptotic error constant C, if x k+1 x lim k x k x r = C, where r 1. If r = 1, then the number ρ = log 10 C is called the asymptotic rate of convergence. If r = 1, and 0 < C < 1, we say that convergence is linear. If r = 1 and C = 0, or if 1 < r < 2 for any positive C, then we say that convergence is superlinear. If r = 2, then the method converges quadratically, and if r = 3, we say it converges cubically, and so on. Note that the value of C need only be bounded above in the case of linear convergence. When convergence is linear, the asymptotic rate of convergence ρ indicates the number of correct decimal digits obtained in a single iteration. In other words, 1/ρ + 1 iterations are required to obtain an additional correct decimal digit, where x is the floor of x, which is the largest integer that is less than or equal to x. If g satisfies the conditions of the Contraction Mapping Theorem with Lipschitz constant L, then Fixed-point Iteration achieves at least linear convergence, with an asymptotic error constant 1

that is bounded above by L. This value can be used to estimate the number of iterations needed to obtain an additional correct decimal digit, but it can also be used to estimate the total number of iterations needed for a specified degree of accuracy. From the Lipschitz condition, we have, for k 1, From we obtain x k x L x k 1 x L k x 0 x. x 0 x x 0 x 1 + x 1 x x 0 x 1 + x 1 x x 0 x 1 + L x 0 x x k x Lk 1 L x 1 x 0. Therefore, in order to satisfy x k x ε, the number of iterations, k, must satisfy k ln x 1 x 0 ln(ε(1 L)). ln(1/l) That is, we can bound the number of iterations after performing a single iteration, as long as the Lipschitz constant L is known. We know that Fixed-point Iteration will converge to the unique fixed point in [a, b] if g satisfies the conditions of the Contraction Mapping Theorem. However, if g is differentiable on [a, b], its derivative can be used to obtain an alternative criterion for convergence that can be more practical than computing the Lipschitz constant L. If we denote the error in x k by e k = x k x, we can see from Taylor s Theorem and the fact that g(x ) = x that e k+1 = x k+1 x = g(x k ) g(x ) = g (x )(x k x ) + 1 2 g (ξ k )(x k x ) 2 = g (x )e k + O(e 2 k ), where ξ k lies between x k and x. Therefore, if g (x ) k, where k < 1, then Fixed-point Iteration is locally convergent; that is, it converges if x 0 is chosen sufficiently close to x. This leads to the following result. Theorem (Fixed-point Theorem) Let g be a continuous function on the interval [a, b], and let g be differentiable on [a, b]. If g(x) [a, b] for each x [a, b], and if there exists a constant k < 1 such that g (x) k, x (a, b), then the sequence of iterates {x k } k=0 converges to the unique fixed point x of g in [a, b], for any initial guess x 0 [a, b]. It can be seen from the preceding discussion why g (x) must be bounded away from 1 on (a, b), as opposed to the weaker condition g (x) < 1 on (a, b). If g (x) is allowed to approach 1 as x 2

approaches a point c (a, b), then it is possible that the error e k might not approach zero as k increases, in which case Fixed-point Iteration would not converge. Suppose that g satisfies the conditions of the Fixed-Point Theorem, and that g is also continuously differentiable on [a, b]. We can use the Mean Value Theorem to obtain e k+1 = x k+1 x = g(x k ) g(x ) = g (ξ k )(x k x ) = g (ξ k )e k, where ξ k lies between x k and x. It follows from the continuity of g at x that for any initial iterate x 0 [a, b], Fixed-point Iteration converges linearly with asymptotic error constant g (x ), since, by the definition of ξ k and the continuity of g, e k+1 lim = lim k e k k g (ξ k ) = g (x ). Recall that the conditions we have stated for linear convergence are nearly identical to the conditions for g to have a unique fixed point in [a, b]. The only difference is that now, we also require g to be continuous on [a, b]. The derivative can also be used to indicate why Fixed-point Iteration might not converge. Example The function g(x) = x 2 + 3 16 has two fixed points, x 1 = 1/4 and x 2 = 3/4, as can be determined by solving the quadratic equation x 2 + 3 16 = x. If we consider the interval [0, 3/8], then g satisfies the conditions of the Fixed-point Theorem, as g (x) = 2x < 1 on this interval, and therefore Fixed-point Iteration will converge to x 1 for any x 0 [0, 3/8]. On the other hand, g (3/4) = 2(3/4) = 3/2 > 1. Therefore, it is not possible for g to satisfy the conditions of the Fixed-point Theorem. Furthemore, if x 0 is chosen so that 1/4 < x 0 < 3/4, then Fixed-point Iteration will converge to x 1 = 1/4, whereas if x 0 > 3/4, then Fixed-point Iteration diverges. The fixed point x 2 = 3/4 in the preceding example is an unstable fixed point of g, meaning that no choice of x 0 yields a sequence of iterates that converges to x 2. The fixed point x 1 = 1/4 is a stable fixed point of g, meaning that any choice of x 0 that is sufficiently close to x 1 yields a sequence of iterates that converges to x 1. The preceding example shows that Fixed-point Iteration applied to an equation of the form x = g(x) can fail to converge to a fixed point x if g (x ) > 1. We wish to determine whether this condition indicates non-convergence in general. If g (x ) > 1, and g is continuous in a neighborhood of x, then there exists an interval x x δ such that g (x) > 1 on the interval. If x k lies within this interval, it follows from the Mean Value Theorem that x k+1 x = g(x k ) g(x ) = g (η) x k x, where η lies between x k and x. Because η is also in this interval, we have x k+1 x > x k x. 3

In other words, the error in the iterates increases whenever they fall within a sufficiently small interval containing the fixed point. Because of this increase, the iterates must eventually fall outside of the interval. Therefore, it is not possible to find a k 0, for any given δ, such that x k x δ for all k k 0. We have thus proven the following result. Theorem Let g have a fixed point at x, and let g be continuous in a neighborhood of x. If g (x ) > 1, then Fixed-point Iteration does not converge to x for any initial guess x 0 except in a finite number of iterations. Now, suppose that in addition to the conditions of the Fixed-point Theorem, we assume that g (x ) = 0, and that g is twice continuously differentiable on [a, b]. Then, using Taylor s Theorem, we obtain e k+1 = g(x k ) g(x ) = g (x )(x k x ) + 1 2 g (ξ k )(x k x ) 2 = 1 2 g (ξ k )e 2 k, where ξ k lies between x k and x. It follows that for any initial iterate x 0 [a, b], Fixed-point Iteration converges at least quadratically, with asymptotic error constant g (x )/2. Later, this will be exploited to obtain a quadratically convergent method for solving nonlinear equations of the form f(x) = 0. Iterative Solution of Equations Now that we understand the convergence behavior of Fixed-point Iteration, we consider the application of Fixed-point Iteration to the solution of an equation of the form f(x) = 0. When rewriting this equation in the form x = g(x), it is essential to choose the function g wisely. One guideline is to choose g(x) = x + φ(x)f(x), where the function φ(x) is, ideally, nonzero except possibly at a solution of f(x) = 0. This can be satisfied by choosing φ(x) to be constant, but this can fail, as the following example illustrates. Example Consider the equation x + ln x = 0. By the Intermediate Value Theorem, this equation has a solution in the interval [0.5, 0.6]. Furthermore, this solution is unique. To see this, let f(x) = x + ln x. Then f (x) = x + 1/x > 0 on the domain of f, which means that f is increasing on its entire domain. Therefore, it is not possible for f(x) = 0 to have more than one solution. We consider using Fixed-point Iteration to solve the equivalent equation x = x + ( 1)(x + ln x) = ln x. However, with g(x) = ln x, we have g (x) = 1/x > 1 on the interval [0.5, 0.6]. Therefore, by the preceding theorem, Fixed-point Iteration will fail to converge for any initial guess in this 4

interval. We therefore apply g 1 (x) = e x to both sides of the equation x = g(x) to obtain which simplifies to g 1 (x) = g 1 (g(x)) = x, x = e x. The function g(x) = e x satisfies g (x) < 1 on [0.5, 0.6], as g (x) = e x, and e x < 1 when the argument x is positive. By narrowing this interval to [0.52, 0.6], which is mapped into itself by this choice of g, we can apply the Fixed-point Theorem to conclude that Fixed-point Iteration will converge to the unique fixed point of g for any choice of x 0 in the interval. 5