A nonlinear equation is any equation of the form. f(x) = 0. A nonlinear equation can have any number of solutions (finite, countable, uncountable)

Similar documents
Nonlinear equations. Norms for R n. Convergence orders for iterative methods

d(x n, x) d(x n, x nk ) + d(x nk, x) where we chose any fixed k > N

University of Houston, Department of Mathematics Numerical Analysis, Fall 2005

NATIONAL BOARD FOR HIGHER MATHEMATICS. Research Scholarships Screening Test. Saturday, February 2, Time Allowed: Two Hours Maximum Marks: 40

1. Nonlinear Equations. This lecture note excerpted parts from Michael Heath and Max Gunzburger. f(x) = 0

Appendix 8 Numerical Methods for Solving Nonlinear Equations 1

COURSE Iterative methods for solving linear systems

STOP, a i+ 1 is the desired root. )f(a i) > 0. Else If f(a i+ 1. Set a i+1 = a i+ 1 and b i+1 = b Else Set a i+1 = a i and b i+1 = a i+ 1

Analysis II: Basic knowledge of real analysis: Part V, Power Series, Differentiation, and Taylor Series

PART I Lecture Notes on Numerical Solution of Root Finding Problems MATH 435

Scientific Computing: An Introductory Survey

Solution of Algebric & Transcendental Equations

Solutions of Equations in One Variable. Newton s Method

Lecture Notes to Accompany. Scientific Computing An Introductory Survey. by Michael T. Heath. Chapter 5. Nonlinear Equations

1. Method 1: bisection. The bisection methods starts from two points a 0 and b 0 such that

Outline. Scientific Computing: An Introductory Survey. Nonlinear Equations. Nonlinear Equations. Examples: Nonlinear Equations

THE INVERSE FUNCTION THEOREM

LECTURE NOTES ELEMENTARY NUMERICAL METHODS. Eusebius Doedel

Numerical Methods for Differential Equations Mathematical and Computational Tools

Economics 204 Summer/Fall 2011 Lecture 5 Friday July 29, 2011

1 + lim. n n+1. f(x) = x + 1, x 1. and we check that f is increasing, instead. Using the quotient rule, we easily find that. 1 (x + 1) 1 x (x + 1) 2 =

Math Refresher - Calculus

A New Modification of Newton s Method

Numerical solutions of nonlinear systems of equations

MA 8019: Numerical Analysis I Solution of Nonlinear Equations

Lecture 4 - The Gradient Method Objective: find an optimal solution of the problem

CS 450 Numerical Analysis. Chapter 5: Nonlinear Equations

Lecture 4 - The Gradient Method Objective: find an optimal solution of the problem

Math 471. Numerical methods Root-finding algorithms for nonlinear equations

Measure and Integration: Solutions of CW2

Numerical Analysis: Solving Nonlinear Equations

Analysis Finite and Infinite Sets The Real Numbers The Cantor Set

Review of Power Series

Chapter 6. Nonlinear Equations. 6.1 The Problem of Nonlinear Root-finding. 6.2 Rate of Convergence

NONLINEAR DC ANALYSIS

h(x) lim H(x) = lim Since h is nondecreasing then h(x) 0 for all x, and if h is discontinuous at a point x then H(x) > 0. Denote

Numerical Methods. Root Finding

Chapter 3: Root Finding. September 26, 2005

Math 473: Practice Problems for Test 1, Fall 2011, SOLUTIONS

Economics 204 Fall 2013 Problem Set 5 Suggested Solutions

International Competition in Mathematics for Universtiy Students in Plovdiv, Bulgaria 1994

Definitions & Theorems

Optimization. Yuh-Jye Lee. March 21, Data Science and Machine Intelligence Lab National Chiao Tung University 1 / 29

Section Taylor and Maclaurin Series

8.7 Taylor s Inequality Math 2300 Section 005 Calculus II. f(x) = ln(1 + x) f(0) = 0

Selçuk Demir WS 2017 Functional Analysis Homework Sheet

Unit 2: Solving Scalar Equations. Notes prepared by: Amos Ron, Yunpeng Li, Mark Cowlishaw, Steve Wright Instructor: Steve Wright

Ordinary Differential Equations II

Notes for Numerical Analysis Math 5465 by S. Adjerid Virginia Polytechnic Institute and State University. (A Rough Draft)

Math 421, Homework #9 Solutions

CALCULUS JIA-MING (FRANK) LIOU

PHYS 410/555 Computational Physics Solution of Non Linear Equations (a.k.a. Root Finding) (Reference Numerical Recipes, 9.0, 9.1, 9.

SMSTC (2017/18) Geometry and Topology 2.

Course 224: Geometry - Continuity and Differentiability

FIXED POINT ITERATIONS

1 Markov decision processes

C.7. Numerical series. Pag. 147 Proof of the converging criteria for series. Theorem 5.29 (Comparison test) Let a k and b k be positive-term series

Functions of Several Variables (Rudin)

MATH 2053 Calculus I Review for the Final Exam

Numerical Methods in Informatics

Lecture 2: Review of Prerequisites. Table of contents

Econ Slides from Lecture 10

MA4001 Engineering Mathematics 1 Lecture 15 Mean Value Theorem Increasing and Decreasing Functions Higher Order Derivatives Implicit Differentiation

CHAPTER-II ROOTS OF EQUATIONS

x y More precisely, this equation means that given any ε > 0, there exists some δ > 0 such that

Math 328 Course Notes

SOLUTION OF ALGEBRAIC AND TRANSCENDENTAL EQUATIONS BISECTION METHOD

8 Numerical methods for unconstrained problems

COMPLETE METRIC SPACES AND THE CONTRACTION MAPPING THEOREM

TAYLOR AND MACLAURIN SERIES

Solutions Preliminary Examination in Numerical Analysis January, 2017

CLASS NOTES Models, Algorithms and Data: Introduction to computing 2018

f(x) f(z) c x z > 0 1

Notes on uniform convergence

CS 542G: Robustifying Newton, Constraints, Nonlinear Least Squares

MA2AA1 (ODE s): The inverse and implicit function theorem

Caculus 221. Possible questions for Exam II. March 19, 2002

Chapter 8: Taylor s theorem and L Hospital s rule

x 2 x n r n J(x + t(x x ))(x x )dt. For warming-up we start with methods for solving a single equation of one variable.

e x = 1 + x + x2 2! + x3 If the function f(x) can be written as a power series on an interval I, then the power series is of the form

Definable Extension Theorems in O-minimal Structures. Matthias Aschenbrenner University of California, Los Angeles

Chapter 3. Differentiable Mappings. 1. Differentiable Mappings

ROOT FINDING REVIEW MICHELLE FENG

Math 140A - Fall Final Exam

Math 4B Notes. Written by Victoria Kala SH 6432u Office Hours: T 12:45 1:45pm Last updated 7/24/2016

B5.6 Nonlinear Systems

Math 209B Homework 2

Mathematical Analysis Outline. William G. Faris

AP Calculus Testbank (Chapter 9) (Mr. Surowski)

NATIONAL BOARD FOR HIGHER MATHEMATICS. M. A. and M.Sc. Scholarship Test. September 19, Time Allowed: 150 Minutes Maximum Marks: 30

Math 117: Infinite Sequences

Math 216 Final Exam 14 December, 2017

Principle of Mathematical Induction

Math 115 HW #5 Solutions

Taylor and Maclaurin Series. Approximating functions using Polynomials.

Iterative Methods. Splitting Methods

Exercises from other sources REAL NUMBERS 2,...,

Solutions Final Exam May. 14, 2014

Assignment 1: Functions of several variables.

Chapter 1. Root Finding Methods. 1.1 Bisection method

Transcription:

Nonlinear equations Definition A nonlinear equation is any equation of the form where f is a nonlinear function. Nonlinear equations x 2 + x + 1 = 0 (f : R R) f(x) = 0 (x cos y, 2y sin x) = (0, 0) (f : R 2 R 2 ) y (t) = cos(y(t)) (f : C 1 (Ω) C 0 (Ω), f(y) = y cos(y) Remark A nonlinear equation can have any number of solutions (finite, countable, uncountable) J. Janela slides AN - 2014/2015 42 / 41

Example Suppose that in the beginning of each year a bank client makes a deposit of v euros in an investment fund (constant interest rate) and at the end of n years withdraws M euros. What was the interest rate r? Since we have M = v 1 + r [(1 + r) n 1], r the answer to our question is the solution of the equation f(r) = 0, where f(r) = M v 1 + r [(1 + r) n 1] r J. Janela slides AN - 2014/2015 43 / 41

Example (Implied volatility) r(t t) C(S, t) = N(d 1 )S N(d 2 )Ke x 1 N(x) = exp( t 2 /2)dt 2π 1 d 1 = σ [ ln(s/k) + (r + σ 2 /2)(T t) ] T t d 2 =d 1 σ T t C(S, t): option price, r: interest rate free of risk, T t: time to maturity, S: value of the underlying asset, K: strike. The implied volatility is computed considering that every other parameter is known. J. Janela slides AN - 2014/2015 44 / 41

The fixed point method in R The fixed point method applies to the solution of equations in the form g(x) = x Any point z D g that satisfies the equation is called a fixed point of g. The name is due to the fact that the application of g does not change the value of z. The fixed point method 1 Pick an initial guess x 0. 2 Apply function g getting a sequence defined recursively by x n+1 = g(x n ). 3 The solution to the equation is z = lim x n. J. Janela slides AN - 2014/2015 45 / 41

Remark Does the limit exist? Is it really the solution of the equation? Example cos x = x J. Janela slides AN - 2014/2015 46 / 41

Example 2x = x J. Janela slides AN - 2014/2015 47 / 41

But does this always work?? Example x2 4 x 2 + 4 = x, x 0 = 1.97 J. Janela slides AN - 2014/2015 48 / 41

Theorem (Fixed point theorem) Let g : [a, b] R be a continuous function. Then 1 If g([a, b]) [a, b] then g has at least one fixed point in [a, b]. 2 Additionally, if g is contractive in [a, b], it has one and only one fixed point in [a, b], given by the limit of the sequence x n+1 = g(x n ), for all x 0 [a, b]. Contractivity A function g : [a, b] R is contractive if there is a constant 0 L < 1 such that g(x) g(y) L x y, x, y [a, b] If g is differentiable this condition is equivalente to existing a constant 0 L < 1 such that g (x) < 1, x [a, b]. J. Janela slides AN - 2014/2015 49 / 41

Error bounds Theorem (FPT: Error estimates) Under the assumptions of the fixed point theorem, we have 1 z x n L z x n 1, n 1 2 z x n L n z x 0, n 1 3 z x n L 1 L x n x n 1, n 1 4 z x n Ln 1 L x 1 x 0, n 1 Estimates (1), (2) are called a priori estimates whereas estimates (3),(4) are called a posteriori estimates. Remark If g (z) < 1 there is a neighbourhood of z where the assumptions of the FPT hold. If g (z) > 1 the FPM diverges, unless x k = z, for some k. J. Janela slides AN - 2014/2015 50 / 41

Monotonic convergence Under the same conditions of the fixed point method: If 1 < g (x) < 0, x [a, b] the convergence is alternate and the fixed point is always between consecutive iterations. If 0 < g (x) < 1 the convergence is monotonous. J. Janela slides AN - 2014/2015 51 / 41

Rate of convergence A sequence x n z is said to converge with rate p if lim z x n = K( 0) z x n 1 p The constant K is denoted as the asymptotic convergence factor. Theorem Let g be a function with p continuous derivatives verifying the FPT assumptions and z its fixed point. If g (z) = g (p 1) (z) = 0, g (p) (z) 0 then the FPM has rate of convergence p. J. Janela slides AN - 2014/2015 52 / 41

Using Taylor s polynomial we can check that if a fixed point method has order p then which means that Remark lim z x n z x n 1 p = lim g (p+1) (η n ) = g(p) (z) n (n + 1)! (p + 1)! z x n g(p+1) (z) z x n 1 p (p + 1)! If p = 1 we have e n C e n 1. If p = 2 we have e n C e n 1 2. J. Janela slides AN - 2014/2015 53 / 41

Newton s method Newton s method is a particular case of the FPM to solve equations of the type f(x) = 0. The method is based on the observation that if f (x) 0 then f(x) = 0 f(x) f (x) = 0 x = x f(x) f (x) Remark If f 0, every zero of f(x) is a fixed point of g(x) = x f(x) f (x) Newton s Method 1 Choose an initial approximation x 0 2 For n 0 compute x n+1 = x n f(xn) f (x n) 3 Repeat until convergence (up to a prescribed accuracy). J. Janela slides AN - 2014/2015 54 / 41

Geometric interpretation Choose the initial approximation J. Janela slides AN - 2014/2015 55 / 41

Geometric interpretation Choose the initial approximation Take the tangent line to the graphic at x = x 0. The equation of the tangent is given by y f(x 0 ) = f (x 0 )(x x 0 ) J. Janela slides AN - 2014/2015 55 / 41

Geometric interpretation Choose the initial approximation Take as new approximation the point of the tangent line crosses de x-axis x 1 = x 0 f(x 0) f (x 0 ) J. Janela slides AN - 2014/2015 55 / 41

Convergence of Newton s method Local convergence If f C 2 (V ε (z)) and f (z) 0 Newton s method converges to z at least with order 2. More precisely, z x n+1 = f (η) 2f (x n ) z x n 2 K z x n 2, where K = max f 2 min f. Applying this estimate recursively, we get z x n 1 K (K z x 0 ) 2n. J. Janela slides AN - 2014/2015 56 / 41

Global convergence Suppose that f : [a, b] R is C 2 ([a, b]) and 1 f(a) f(b) < 0 2 f (x) 0, x [a, b] 3 f (x) 0 or f (x) 0, x [a, b] Then, 1 Choosing x 0 [a, b] such that f(x 0 )f (x) 0, Newton s method is convergent. 2 If f(a) f (a) f(b) < b a and f (b) < b a the method converges for every initial approximation x 0 [a, b]. J. Janela slides AN - 2014/2015 57 / 41

Systems of equations Determine x = (x 1, x 2,, x n ) R n satisfying f 1 (x 1,, x n ) = 0 f F (x) = 0 2 (x 1,, x n ) = 0 f n (x 1,, x n ) = 0 or x 1 = g 1 (x 1, x 2,, x n ) x x = G(x) 2 = g 2 (x 1, x 2,, x n ) x n = g n (x 1, x 2,, x n ) J. Janela slides AN - 2014/2015 58 / 41

Theorem (Fixed point in R n ) Let Ω R a nonempty closed set and G : Ω Ω a contractive function for some norm. Then, the system x = G(x) has one and only one solution z Ω, the fixed point iteration x x+1 = G(x n ) converges to z, for every x 0 Ω and the error bounds are similar to the ones obtained in R. Remark A function G : Ω Ω is contractive in a norm if there exists a constant 0 L < 1 such that G(x) G(y) L x y Remark If G is differentiable, L = sup J G (x). x Ω J. Janela slides AN - 2014/2015 59 / 41

Definition (Norm) Let X be a real vector space. An application : X R + 0 called a norm if: x = 0 if and only if x = 0. αx = α x, x X, α R x + y x + y, x, y X. is Example (Norms in R n.) x 1 = ( n n ) 1/2 ( n ) 1/p x i, x 2 = x 2 i, x p = x i p i=1 i=1 i=1 x = max x i i=1,,n J. Janela slides AN - 2014/2015 60 / 41

Why should we need different norms? d(a, B) = ka Bk = J. Janela slides AN - 2014/2015 61 / 41

Matrix Norms Norms in spaces of linear operators If U and V are normed spaces then L(U, V ) is a normed space considering the norm Ax Y A = sup, A L(X, Y ) x X\0 x X If U = V = R n, the linear operators are represented by matrices Matrix norms Ax 1 A 1 = sup = max x 0 X 1 j=1,,n Ax A = sup = max x 0 x n A ij i=1 i=1,,n j=1 n A ij J. Janela slides AN - 2014/2015 62 / 41

Matrix norms Proposition Let λ be an eigenvalue of A R n n and any matrix norm. Then we have A λ. Proposition If all the eigenvalues λ i of A R n n satisfy λ i < 1 then, for some norm, we have A < 1. Remark If G : Ω R n Ω has enough regularity and all the eigenvalues λ i (x) of J G (x) satisfy λ i (x) < 1, x Ω, the system G(x) = x has one and only one solution in Ω and the fixed point method converges to this solution. J. Janela slides AN - 2014/2015 63 / 41

Example A pharmaceutical company synthesises an antiviral serum based on the active substances S 1, S 2, S 3 and S 4. It has been determined that if an individual is given α milligrams of the serum, after a fixed amount of time the concentration of each substance (mg/l) is given implicitly (for α [0, 5]) by the relations 16x 1 cos (α (x 2 2x 1 )) = 0 16x 2 + 0.75 sin (α ( x 3 3x 1 )) = 0 16x 3 cos (α (x 4 2x 3 )) = 0 16x 4 0.75 sin (2αx 3 ) = 0 Knowing that S 2 should never exceed a concentration of 0.03mg/l what is the highest safe dosaje that can be administered? J. Janela slides AN - 2014/2015 64 / 41

The system can be written as x 1 = 1 16 cos (α (x 2 2x 1 )) x 2 = 3 64 sin (α (x 3 + 3x 1 )) x 3 = 1 16 cos (α (x 4 2x 3 )) x 4 = 3 64 sin (2αx 3) Observing the system we can immediately see that, if there is a solution z R 4, then z [ 1 16, 1 16 ] [ 3 64, 3 64 ] [ 1 16, 1 16 ] [ 3 64, 3 64 ] In fact, we can be more precise... For instance, (x 1, x 2 ) [ 1 16, 1 16 ] [ 3 64, 3 64 ] α(x 2 2x 1 ) [ 55 64, 55 64 ] cos α(x 2 2x 1 ) [0.6; 1] x 1 [0.0375; 0.0625] J. Janela slides AN - 2014/2015 65 / 41

Making similar calculations for the other variables we can choose to search a solution in the compact set Ω = [0.0375; 0.0625] [0.044; 0.046875] [0.0375; 0.0625] [0.0172; 0.0275 The jacobian matrix J G (x) is given by α 8 sin α 16 sin 0 0 9α 64 cos 0 3α 64 cos 0 0 0 α 8 sin α 16 sin 0 0 6α 64 sin 0 and, therefore, J G (x) = α max{ 3 16, 3 16, 3 16, 3 32 } = 3α 16 15 16 < 1. J. Janela slides AN - 2014/2015 66 / 41

What can we conclude? Since G : Ω Ω is continuous and differentiable in the compact set Ω, and J G (x) 1 in Ω, the system has only and only one solution in this set and the fixed point method converges to this solution, for all initial approximation x 0 Ω. If we take any x 0 R n then G(x 0 ) Ω, and so the convergence is guaranteed for any x 0 R n. We have the error bounds z x (n) (15/16) n z x (0) 1 16 (15/16)n z x (n) 15/16 1 15/16 x(n) x (n 1) z x (n) (15/16)n 1 15 x (1) x (0) 16 J. Janela slides AN - 2014/2015 67 / 41

i x (i) 1 x (i) 2 x (i) 3 x (i) 4 z x (i) 1 0.0625 0.0625 0.0625 0.046875 0.315 10 1 2 0.0614046 0.0319518 0.0607912 0.017169 0.123 10 2 3 0.0601926 0.0314343 0.0594588 0.0167209 0.491 10 3 4 0.0602878 0.0309125 0.0595855 0.0163703 0.308 10 4 5 0.0602525 0.0309561 0.0595512 0.0164037 0.129 10 4 6 0.0602582 0.0309413 0.059557 0.0163947 1.964 10 6 7 0.0602569 0.0309437 0.0595559 0.0163962 4.522 10 7 8 0.0602571 0.0309432 0.0595561 0.0163959 8.510 10 8 21 0.062571 0.0309432 0.059556 0.0163959 6.592 10 17 Table: Evolution of the computational error. J. Janela slides AN - 2014/2015 68 / 41

Newton s method for systems If the Jacobian matrix of F is invertible, we have that F (x) = 0 J 1 1 (x) F (x) = 0 x = x J (x) F (x) F F Newton s method for systems of equations 1 Take an initial guess x 0 2 For each n > 0 compute the solution y n of J F (x n )y n = F (x n ) and update x n+1 = x n + y n. 3 Stop when some criteria is satisfied. J. Janela slides AN - 2014/2015 69 / 41

Performance of Newton s method (same example) i x (i) 1 x (i) 2 x (i) 3 x (i) 4 z x (i) 1 0.0625 0.0625 0.0625 0.046875 0.315 10 1 2 0.0604482 0.0310313 0.059709 0.0164386 0.191 10 3 3 0.0602571 0.0309433 0.0595561 0.016396 0.8329 10 7 4 0.0602571 0.0309432 0.059556 0.0163959 0.1304 10 14 5 0.0602571 0.0309432 0.059556 0.0163959 0.6938 10 17 Table: Newton s method, evolution of the error J. Janela slides AN - 2014/2015 70 / 41