STOP, a i+ 1 is the desired root. )f(a i) > 0. Else If f(a i+ 1. Set a i+1 = a i+ 1 and b i+1 = b Else Set a i+1 = a i and b i+1 = a i+ 1

Similar documents
Computational Methods CMSC/AMSC/MAPL 460. Solving nonlinear equations and zero finding. Finding zeroes of functions

Nonlinearity Root-finding Bisection Fixed Point Iteration Newton s Method Secant Method Conclusion. Nonlinear Systems

Lecture Notes to Accompany. Scientific Computing An Introductory Survey. by Michael T. Heath. Chapter 5. Nonlinear Equations

1. Method 1: bisection. The bisection methods starts from two points a 0 and b 0 such that

Math 471. Numerical methods Root-finding algorithms for nonlinear equations

Nonlinearity Root-finding Bisection Fixed Point Iteration Newton s Method Secant Method Conclusion. Nonlinear Systems

Numerical Methods in Informatics

Solution of Nonlinear Equations

Simple Iteration, cont d

MA 8019: Numerical Analysis I Solution of Nonlinear Equations

Motivation: We have already seen an example of a system of nonlinear equations when we studied Gaussian integration (p.8 of integration notes)

CS 323: Numerical Analysis and Computing

Outline. Scientific Computing: An Introductory Survey. Nonlinear Equations. Nonlinear Equations. Examples: Nonlinear Equations

Midterm Review. Igor Yanovsky (Math 151A TA)

Nonlinear equations. Norms for R n. Convergence orders for iterative methods

Outline. Math Numerical Analysis. Intermediate Value Theorem. Lecture Notes Zeros and Roots. Joseph M. Mahaffy,

Scientific Computing: An Introductory Survey

Math Numerical Analysis

1. Nonlinear Equations. This lecture note excerpted parts from Michael Heath and Max Gunzburger. f(x) = 0

Numerical Methods I Solving Nonlinear Equations

Numerical Solution of f(x) = 0

CS 450 Numerical Analysis. Chapter 5: Nonlinear Equations

ROOT FINDING REVIEW MICHELLE FENG

x 2 x n r n J(x + t(x x ))(x x )dt. For warming-up we start with methods for solving a single equation of one variable.

Roots of Equations. ITCS 4133/5133: Introduction to Numerical Methods 1 Roots of Equations

Chapter 3: Root Finding. September 26, 2005

A nonlinear equation is any equation of the form. f(x) = 0. A nonlinear equation can have any number of solutions (finite, countable, uncountable)

CHAPTER-II ROOTS OF EQUATIONS

University of Houston, Department of Mathematics Numerical Analysis, Fall 2005

Finding the Roots of f(x) = 0. Gerald W. Recktenwald Department of Mechanical Engineering Portland State University

Finding the Roots of f(x) = 0

Numerical Analysis: Solving Nonlinear Equations

SOLUTION OF ALGEBRAIC AND TRANSCENDENTAL EQUATIONS BISECTION METHOD

Line Search Methods. Shefali Kulkarni-Thaker

FIXED POINT ITERATIONS

CS 323: Numerical Analysis and Computing

Root Finding (and Optimisation)

Solution of Algebric & Transcendental Equations

Numerical Methods. Root Finding

5 Finding roots of equations

Chapter 8: Taylor s theorem and L Hospital s rule

Unit 2: Solving Scalar Equations. Notes prepared by: Amos Ron, Yunpeng Li, Mark Cowlishaw, Steve Wright Instructor: Steve Wright

Exam in TMA4215 December 7th 2012

Lecture 5: Random numbers and Monte Carlo (Numerical Recipes, Chapter 7) Motivations for generating random numbers

MATH 3795 Lecture 12. Numerical Solution of Nonlinear Equations.

Numerical Methods in Physics and Astrophysics

Appendix 8 Numerical Methods for Solving Nonlinear Equations 1

Optimization and Calculus

Math 273a: Optimization Netwon s methods

COMPLETE METRIC SPACES AND THE CONTRACTION MAPPING THEOREM

Math Numerical Analysis Mid-Term Test Solutions

15 Nonlinear Equations and Zero-Finders

Intro to Scientific Computing: How long does it take to find a needle in a haystack?

3.1 Introduction. Solve non-linear real equation f(x) = 0 for real root or zero x. E.g. x x 1.5 =0, tan x x =0.

Notes for Numerical Analysis Math 5465 by S. Adjerid Virginia Polytechnic Institute and State University. (A Rough Draft)

Continuity. Chapter 4

x y More precisely, this equation means that given any ε > 0, there exists some δ > 0 such that

17 Solution of Nonlinear Systems

Scientific Computing. Roots of Equations

LECTURE NOTES ELEMENTARY NUMERICAL METHODS. Eusebius Doedel

Two hours. To be provided by Examinations Office: Mathematical Formula Tables. THE UNIVERSITY OF MANCHESTER. 29 May :45 11:45

Numerical solutions of nonlinear systems of equations

Numerical Methods in Physics and Astrophysics

Shifted QR algorithm

CLASS NOTES Models, Algorithms and Data: Introduction to computing 2018

Jim Lambers MAT 460/560 Fall Semester Practice Final Exam

PART I Lecture Notes on Numerical Solution of Root Finding Problems MATH 435

Solving nonlinear equations (See online notes and lecture notes for full details) 1.3: Newton s Method

you expect to encounter difficulties when trying to solve A x = b? 4. A composite quadrature rule has error associated with it in the following form

Richard S. Palais Department of Mathematics Brandeis University Waltham, MA The Magic of Iteration

Exam 2 extra practice problems

Chapter 4. Solution of Non-linear Equation. Module No. 1. Newton s Method to Solve Transcendental Equation

Numerical Study of Some Iterative Methods for Solving Nonlinear Equations

Eigenvalues and Eigenvectors

1 Newton s Method. Suppose we want to solve: x R. At x = x, f (x) can be approximated by:

Chapter 3. Differentiable Mappings. 1. Differentiable Mappings

Figure 1: Graph of y = x cos(x)

MATH 409 Advanced Calculus I Lecture 12: Uniform continuity. Exponential functions.

MATH 3795 Lecture 13. Numerical Solution of Nonlinear Equations in R N.

x n+1 = x n f(x n) f (x n ), n 0.

Introductory Numerical Analysis

1.1: The bisection method. September 2017

Root Finding Convergence Analysis

FIXED POINT ITERATION

REVIEW OF DIFFERENTIAL CALCULUS

Numerical Methods for Differential Equations Mathematical and Computational Tools

Lecture 10 Polynomial interpolation

Chapter 1. Root Finding Methods. 1.1 Bisection method

2tdt 1 y = t2 + C y = which implies C = 1 and the solution is y = 1

Mean Value Theorem. MATH 161 Calculus I. J. Robert Buchanan. Summer Department of Mathematics

Calculating determinants for larger matrices

SYSTEMS OF NONLINEAR EQUATIONS

Lecture 7. Root finding I. 1 Introduction. 2 Graphical solution

Numerical Methods for Large-Scale Nonlinear Systems

Math 4329: Numerical Analysis Chapter 03: Fixed Point Iteration and Ill behaving problems. Natasha S. Sharma, PhD

BSM510 Numerical Analysis

Mean Value Theorem. MATH 161 Calculus I. J. Robert Buchanan. Summer Department of Mathematics

Suppose that the approximate solutions of Eq. (1) satisfy the condition (3). Then (1) if η = 0 in the algorithm Trust Region, then lim inf.

Variable. Peter W. White Fall 2018 / Numerical Analysis. Department of Mathematics Tarleton State University

Economics 204 Summer/Fall 2011 Lecture 5 Friday July 29, 2011

Transcription:

53 17. Lecture 17 Nonlinear Equations Essentially, the only way that one can solve nonlinear equations is by iteration. The quadratic formula enables one to compute the roots of p(x) = 0 when p P. Formulas were derived for finding the roots of p P 3 and p P 4 by expressions involving radicals ( 1545). The case of quintics was studied unsuccessfully for almost 300 years until Abel (184) proved no such formula existed. As we shall see, the roots of quintics (and other nonlinear equations) can be solved by iteration! Example 17.1 (Eigenvalues). Let A be a n n matrix with n 5. We propose to compute the eigenvalues of A. The eigenvalues of A are the roots of the characteristic polynomial, i.e. p(λ) = det(a λi). The characteristic polynomial p has degree n and leading term is ( 1) n λ n. This problem can only be solved by iteration when n 5. Bisection method. Suppose f is continuous on [a, b] and satisfies f(a)f(b) < 0 (not the same sign). The intermediate value theorem tells us there is at least 1 root f(x ) = 0, x (a, b). The bisection method is used to finding such root. Bisection Algorithm (mathematical) Set a 0 = a and b 0 = b For i = 0, 1,,... Set a i+ 1 = ai+bi If f(a i+ 1 ) = 0 STOP, a i+ 1 is the desired root Else If f(a i+ 1 )f(a i) > 0 Set a i+1 = a i+ 1 and b i+1 = b i Else Set a i+1 = a i and b i+1 = a i+ 1 End For It is obvious that after j steps, either we have found a root or there is a root in (a j, b j ). Note that in that case the root x is at most half of the interval length away from either a j or b j, i.e. where x = a j + b j ε < b j a j + b j = b j a j = b a j. We now describe the matlab version of the mathematical bisection algorithm using minimal memory usage. + ε, 1 %%% R i s the root approximation a f t e r N s t e p s %%% A,B are r e a l numbers 3 %%% F i s a continuous f u n c t i o n 4 %%% F(A)F(B)<=0 5 Function R=BISECTION(A, B, N, F) 6

54 7 %% P r e l i m i n a r i e s : A or B are a root, F(A)F(B)>0 8 SA = s i g n (F(A) ) ; 9 i f (SA == 0) 10 R=A; 11 r eturn ; 1 end 13 14 SB = s i g n (F(B) ) ; 15 i f (SB == 0) 16 R=B; 17 r eturn ; 18 end 19 0 i f (SA == SB) 1 f p r i n t f ( Input e r r o r to b i s e c t i o n \n ) ; R=NaN( 1 ) ; %return not a number 3 r eturn ; 4 end 5 6 %% a l l the p r e l i m i n a r i e s are done 7 f o r I =1:N 8 AV = (A B) / ; 9 FAV = F(AV) ; 30 S=s i g n (FAV) ; 31 i f ( S==0) 3 R=AV; 33 r eturn ; 34 end 35 i f ( S==SA) 36 A=AV; 37 e l s e 38 B=AV; 39 end 40 end 41 end This algorithm will only get the real roots of a real valued function f : R R. Example 17. (Cubic). Let f(x) = x 3 +x = x(x +1). The roots of f are (0, i, i) and bisection will only get the real roots, i.e. x = 0 To find the complex roots, we need to treat f as a complex valued functions. We define F : R C C R by F ( ) R I f(r + ii) = (R + ii) 3 + (R + ii) = R 3 + 3iR I 3RI ii 3 + (R + ii) ( ) = (R 3 3RI + R) + (3R I I 3 R + I)i 3 3RI + R 3R I I 3. + I

55 for real number R, I. Find the complex roots correspond to a system ( ) ( ) ( R R F := 3 3RI + R 0 I 3R I I 3 =. + I 0) The previous illustrates the need of iterative techniques for systems (but is far from the only reason). Let F : R n R n be a vector values function on R n. We want to find roots x R n satisfying F (x ) = 0 R n. This gives n equations and x 1,..., x n are the n unknown. Example 17.3 (n = ). F ( x1 ) := x ( ) ( x 1 + 4x 1 0 4x 1 + x =. 1 0) Refer to Figure 6 for an illustration of the situation. 1 0.5-1 -0.75-0.5-0.5 0 0.5 0.5 0.75 1-0.5-1 Figure 6. There are 4 roots of the system x 1 + 4x 1 = 0 and 4x 1 + x 1 = 0. Example 17.4 (n = 1). Consider f(x) = sin(x) = 0. The roots are x = jπ, j Z = {...,, 1, 0, 1,,...}. Definition 17.1 (Fixed Point Equation). A fixed point equation is one of the form x = G(x), with x R n and G : R n R n.

56 A solution to a fixed point equations is x R n satisfies x = G(x ). We can turn F (x) = 0 into a fixed point problem, i.e. x = x F (x), i.e. G(x) := x F (x). This means that x solves F (x ) = 0 if and only if x = G(x ). Obviously fixed point problems can be turned into F (x) = 0 by setting F (x) = x G(x). We could have also used x = x BF (x) =: G(x) for any non singular n n matrix B. We will see the importance of the matrix B soon.

57 18. Lecture 18 We considered in the last lecture the fixed point formulation: Find x R n satisfying x = G(x ), where G : R n R n. We now discuss an algorithm approximating such x. Fixed Point Iteration or Picard Iteration. Start with an initial iterate (guess) x 0 R n. Then, for i = 0, 1,,... set x i+1 = G(x i ). We now want to understand when x i x. Definition 18.1 (Lipschitz). Let Ω R n. The vector-valued function G : Ω R n is called Lipschitz conitnuous if there is a M 0 with for all x, y Ω. Here for w R n G(x) G(y) M x y w := max w i. i=1,...,n Definition 18. (Contraction Mapping). Let Ω R n. The vector-valued function G : Ω Ω is a contraction mapping if G is Lipschitz continuous with constant M = ρ < 1. Theorem 18.1 (Contraction Mapping Theorem). Let Ω be a closed subset of R n and G be a contraction mapping of Ω into Ω with constant ρ. Then, there is a unique fixed point x Ω (satisfying x = G(x )) and the Picard iteration {x j } j=0, starting with any x 0 Ω, converges to x and satisfies x j x ρj 1 ρ x0 x. This is often called linear convergence. We postpone the proof for later. To understand the contraction mapping hypothesis, we consider the equation F (x) = 0, where F : R R, and assume that it has as solution x R, i.e. F (x ) = 0. Furthermore, we assume that F C in a neighborhood B δ1 (x ) := [x δ 1, x + δ 1 ] and that F (x ) 0. Notice that the latter condition, guarantees that there is a neighborhood B δ (x ) (δ δ 1 ) such that for every ξ B δ (x ). This implies that F (ξ) 1 F (x ) F (ξ) 1 F (x ) 1 ξ B δ (x ). Now for δ δ (to be determined), we pick w B δ (x ) and set (9) G(x) = x (F (w)) 1 F (x). Clearly x is a fixed point of G and for x, y B δ (x ) G(x) G(y) = G (y)(x y) + G (ξ) (x y), for some ξ between x and y. Therefore, G(x) G(y) G (y) x y + G (ξ) (x y).

58 However, G (y) = 1 (F (w)) 1 F (y) = (F (w) F (y))(f (w)) 1 = F (ξ 1 )(w y)(f (w)) 1 for ξ 1 between w and y. Now by the C assumption, there exists a constant M 0 such that F (θ) M for every θ B δ1 (x ). Thus, so G (y) 4M F (x ) 1 δ (10) G(x) G(y) { 4M F (x ) 1 δ + Mδ } x y, using the fact that w y δ and x y δ. Given 0 < ρ < 1, we chose δ so that { 4M F (x ) 1 + M } δ < ρ. This implies that G is a contraction mapping and so the Picard iterates x i+1 = x i (F (w)) 1 F (x i ) converges to x provided x 0 x δ and w x δ. The next theorem generalize this argumentation to R n. Theorem 18. (Secant Algorithm). Assume F : R n R n and F (x ) = 0 for some x R n with (1) F C in a neighborhood of x ; () DF (x ) = the derivative matrix at x given by (DF (x )) ij = x j F i (x ) is nonsingular. Then there is a δ > 0 such that if w B δ (x ) and x 0 B δ (x ), the iteration converges to x. x i+1 = x i (DF (x )) 1 F (x i ) The proof is as in the case of scalar functions discussed before the theorem but more complicated due to matrix notations. The major obstacle to apply this algorithm is getting close enough to the root to come up with w (then we can always take x 0 = w).

59 19. Lecture 19 We start with the proof of the contraction mapping theorem (Theorem 18.1). Proof of Theorem 18.1. Let j k and l 0, x j+1 = G(x j ) with x 0 Ω R n (closed) and G a contraction (Lipschitz constant ρ < 1) on Ω. Note that x k+l+1 x k = x k+l+1 x k+l + x k+l x k+l 1 +... + x k+1 x k. Now, and so x m+1 x m = G(x m ) G(x m 1 ) x m+1 x m = G(x m ) G(x m 1 ) ρ x m x m 1. Repeating x m+k+1 x m+k ρ k x m+1 x m and thus x k+l+1 x k x k+l+1 x k+l + x k+l x k+l 1 +... + x k+1 x k This means that if m, l > j (ρ l + ρ l 1 +... + 1) x k+1 x k (ρ l + ρ l 1 +... + 1)ρ k x 1 x 0 ρ k j ρ j }{{} 1 ρ x1 x 0 ρj 1 ρ x1 x 0. 1 x m x l ρj 1 ρ x1 x 0. The quantity on the right side can be made as small as we want by taking j large. This implies that the sequence {x j } is a Cauchy sequence and so converges to some x Ω. (Recall that R n is complete and Ω closed implies that Ω is complete). Moreover, x G(x ) x x j + G(x j 1 ) G(x ) x x j + ρ x x j 1. As x j converges to x, the quantity on the right can be made as small as desired buy taking j large, i.e. x = G(x ). This shows that every Picard iteration converges to a fixed point. These fixed points are unique. Indeed, if x 1 = G(x 1) is another fixed point, then x 1 x = G(x 1) G(x ) ρ x 1 x, i.e. and therefore x 1 x or x 1 = x. (1 ρ) x 1 x 0

60 Newton s Method. We start with a motivation. We look for zero of F (x) = 0, where F : R n R n. Assume F C and that there is a zero x of F such that DF (x ) is nonsingular. As in the previous lecture, there is a neighborhood B δ (x ) such that DF (x) is nonsingular when x B δ (x ). Assume that x j B δ (x ). The Taylor theorem for vector fields guarantees that F (x) = F (x i ) + DF (x j )(x x j ) + ε(x), where ε(x) = O( x x j ). Ignoring the error term and recalling that we want F (x ) = 0, we chose x j+1 by 0 = F (x j ) + DF (x j )(x j+1 x j ), i.e. x j+1 = x j DF (x j ) 1 F (x j ). This is the Newton s iterative method. Note that this iteration is of the form x j+1 = G j (x j ), G j (x) = x DF (x j ) 1 F (x) so it is not quite a fixed point iteration because G changes at each step! The mathematical analysis of this iterative scheme follows that provided before Theorem 18.. For simplicity, we again consider the case Ω R. The main difference is that w in (9) is replaced by x j. As in the previous analysis, for (1) x satisfying F (x ) = 0; () F (x ) 0; (3) F C in a neighborhood of x we assume that x j B δ (x ) with δ δ (δ to be determined). Now x x j+1 = G j (x ) G j (x j ) so that x x j+1 G j (x ) G j (x j ). Since x j B δ (x ), we have as in (10) G j (x ) G j (x j ) { 4M F (x ) 1 + M } δ x x j. As long as x x j δ, we may take δ = x x j to conclude x x j+1 M{4 F (x ) 1 + 1} x x j. This is called quadratic convergence. Once Newton s iterates get close to the solution so that quadratic convergence kicks in, the convergence is extremely fast. For a geometric interpretation of the Newton s method, we refer to Figure 7. Example 19.1 (Newton). Consider the function f(x) = x 3 e x, which has only one root (at x = 0), see Figure 8. We compute f (x) = (3x x 4 )e x so f (x) > 0 for x < 3/ and f (x) < 0 for x > 3/ The geometric interpretation of the Newton method implies that (1) if x 0 > 3/, Newton s method diverge to

61 F (x j ) + (x x j )F (x j ) F (x) x j+1 x j Figure 7. Geometric Interpretation of the Newton s method. 0.5 0.5 -.4 - -1.6-1. -0.8-0.4 0 0.4 0.8 1. 1.6.4-0.5-0.5 Figure 8. f(x) = x 3 e x. () if x 0 < 3/, Newton s method diverge to (3) it diverge in a neighborhood of both 3/ and 3/ (4) it converges to 0 otherwise.