2.4 - Convergence of the Newton Method and Modified Newton Method
|
|
- Brittany Bell
- 6 years ago
- Views:
Transcription
1 2.4 - Convergence of the Newton Method and Modified Newton Method Consider the problem of finding the solution of the equation: for in Assume that is continuous and for in 1. Newton's Method: Suppose that is a simple zero of Then we know where In Section 2.2, the Newton Method is introduced for solving Algorithm Newton Method: newton.m Recall Given in and for (1) compute (2) if or then the algorithm is terminated and otherwise goto Step (1). Convergence and Rate of Convergence: In this section, we study the convergence of Newton's Method and will answer the following questions: (i) Does the Newton Method always converge? (ii) If Newton Method converges, what is the rate of convergence and what is the order of convergence? (iii) Does the Newton Method still converge when has multiple zeros at From Section 2.3, we know that the sequence of iterations Algorithm with function generated by Newton's Method can also be considered as iterations generated by the Fixed-Point By the Fixed-Point Theorem, the sequence of iterations converges to if where in and the rate of convergence is We now investigate under what condition the function has the property that for. We know that If for all in and for all in then there exists an interval such that Notes: Let and be continuous in If is in and then there exists a such that generated by Newton's method converges to for any in Note that: 1. Since the convergence of Newton method deps on locally. 2. where for all in Determine if the iterations generated by Newton's Method for solving Consider converges.
2 (1) and (2) Approximate by finding the solution of for within 10 Let Then. By the Fixed-Point Theorem, we can find an upper bound for the number of iterations needed: Let Use the MatLab program newton.m to find an approximation. x.^2-2; 2*x; >>[xsol, flag, ct]=newton(fun, fpfun,1,10^(-8),100) flag = 1 ct = 5 Find an approximation of the solution of the equation for in within Compare the convergence with the Bisection method and the Fixed point Iteration with Newton's Method: >>fun=@(x) x.^3+x+1; >>fpfun=@(x) 3*x.^2+1; >>[xsol,flag,ct]=newton(fun,fpfun,-1,10^(-5),100) flag =1 ct =5 Bisection method: >>bisect the left point a = -1 the right point b = 0 epsilon = 10^(-5) the approximation to the solution - cn = function value at the solution - fcn = e-005 number of iterations - Nit =17 Fixed-point Algorithm: >>fixpt input initial point p0 = -1
3 input the tolerance for stopping criterion, ex:.0001,10^(-8) 10^(-5) input the maximum number of iterations 100 Algorithm converges with number of iterations = 25 fixed point p = Order of convergence: Theorem Modified Fixed Point Theorem: If and Then quadratically with an asymptotic error constant Assume that but By the Taylor Theorem, we have for Then So, the order of convergence is 2 and the asymptotic error constant Theorem Let be generated by the Newton Method. If is a simple zero of then quadratically with an asymptotic error constant By the Modified Fixed-Point Theorem, we have Then the asymptotic error constant is 2. Zero of Multiplicity : Definition A solution of is a zeros of multiplicity of if for can be expressed as where Theorem Proof Let be continuously differentiable on. has a simple zero at in if and only if but Let have a simple zero at in By the definition, where Since Now let Since exists on for by the Taylor Theorem Since So, where Theorem Let have continuous derivatives on has a zero of multiplicity at if and only if
4 but Let has a zero at Determine the multiplicity of this zero. Three ways to determine 1. Use Theorem 2.4.4: find such that but Check: So, 2. Use the definition: find such that where That is So, 3. Use the definition and Taylor series of So, Modified Newton's Methods: Let be a zero of multiplicity of on Then we know Though as So, for large In this case, does still converge to (theoretically)? If so, what is the order of convergence? Let To answer the first question, we need to find out if So, still converges to However, since linearly instead of quadratically. If but then
5 Consider solving the equation: using Newton's Method with. log(x+1)-x; 1./(x+1)-1; >>[xsol,flag,ct]=newton(fun,fpfun,1,10^(-5),100) flag =1 ct =9 >> orderconv(xsol,0,9,1) Clearly, linearly. Since when and when is a zero of multiplicity 2 of Can we improve the convergence of Newton's Method for the problems where 1. Modified Newton's Methods: newtonm.m 2. where Let be a solution of For each method, we will check if a. and b. quadratically. Let us just check out the case where By the definition, but 1. Let and Then So, is a fixed point of Since converges to and converges to quadratically. 2. This method can also be considered as solving using Newton's Method where has a simple zero at
6 Since has a simple zero at and quadratically. Consider solving the equation: using both modified Newton's Methods with >>fun=@(x) log(x+1)-x; >>fpfun=@(x) 1./(x+1)-1; >>[xsol,flag,ct]=newtonm(fun,fpfun,1,10^(-5),100,2) flag = 1 ct = 4 >>orderconv(xsol,0,4,2) and Define and >>fun=@(x) -(log(x+1)-x).*(x+1)./x; >>fpfun=@(x) 1+(log(x+1)-x).*(x+1)./x.^2; >>[xsol,flag,ct]=newton(fun,fpfun,1,10^(-5),100) flag =1 ct =4 >> orderconv(xsol,0,4,2) Use the Newton Method to solve for in Determine the multiplicity of the zero of >>fun=@(x) exp(3*x)-27*x.^6+27*x.^4.*exp(x)-9*x.^2.*exp(2*x); >>fpfun=@(x) 3*exp(3*x)-162*x.^5+27*x.^3*(4+x).*exp(x)-18*x.*(1+x).*exp(2*x); >>[xsol,flag,ct]=newton(fun,fpfun,3,10^(-8),200) ct=24 xsol(ct)=
7 r1(k)=abs(xsol(k+1)-xsol(ct))/abs(xsol(k)-xsol(ct)); >>plot(r1) >>title('\x_{n+1}-x*\/x_n-x*\') r2(k)=abs(xsol(k+1)-xsol(ct))/(abs(xsol(k)-xsol(ct)))^2; >>plot(r2) >>title('\x_{n+1}-x*\/x_n-x*\^2') So, the Newton Method converges linearly and Now let and use the Modified Newton Method: >> [xsol,flag,ct]=newtonm(fun,fpfun,3,10^(-8),200,2); ct=14 xsol(14)= r3(k)=abs(xsol(k+1)-xsol(ct))/abs(xsol(k)-xsol(ct)); >>subplot(1,2,1),plot(r3) >>title('\x_{n+1}-x*\/x_n-x*\') r4(k)=abs(xsol(k+1)-xsol(ct))/(abs(xsol(k)-xsol(ct)))^2; >>subplot(1,2,2),plot(r4) >>title('\x_{n+1}-x*\/x_n-x*\^2') So, the Modified Newton Method converges linearly with Now let and use the Modified Newton Method again: >> [xsol,flag,ct]=newtonm(fun,fpfun,3,10^(-8),200,3); ct=9 xsol(9)= r5(k)=abs(xsol(k+1)-xsol(ct))/abs(xsol(k)-xsol(ct)); >>subplot(1,2,1),plot(r5) >>title('\x_{n+1}-x*\/x_n-x*\') r6(k)=abs(xsol(k+1)-xsol(ct))/(abs(xsol(k)-xsol(ct)))^2; >>subplot(1,2,2),plot(r6) >>title('\x_{n+1}-x*\/x_n-x*\^2')
8 Hence, the Modified Newton Method convergence quadratically and Exercises: 1. Let 1. Show first that has a zero of multiplicity 3 at 2. Use Newton Method to solve this equation with within 3. Verify using the graphs of the sequences and that the iterations obtained in (2) converge linearly. Estimate the asymptotic error constant. 4. Apply one modified Newton Method to solve this equation again with within 5. Verify using the graphs of the sequences and that the iterations obtained in (4) converge quadratically. 2. Let 1. Show first that has a zero of multiplicity 3 at 2. Use Newton Method to solve this equation with within 3. Verify using the graphs of the sequences and that the iterations obtained in (2) converge linearly. Estimate the asymptotic error constant. 4. Apply one modified Newton Method to solve this equation again with within 5. Verify using the graphs of the sequences and that the iterations obtained in (4) converge quadratically. 3. Let 1. Verify that is a zero of Find the multiplicity of the zero 2. Use Newton Method to find the zero of within with Use a proper Modified Newton Method to find the zero (quadratically) within with. 3. Use Newton Method to find another zero of within with What is the zero? Estimate the order of convergence and asymptotic error constant Is this zero simple? If not, find a proper Modified Newton Method to find this zero (quadratically) within with What is the multiplicity of this zero? 4. The sum of two real numbers is 20. If each number is added to its square root, the product of the two sums is Determine the two numbers to within by the Newton Method. (Ref [1]) 5. Consider the population equation: Approximate within using Newton Method. Report your result including and (Ref [1])
Zeros of functions with Matlab: Bisection method and fixed point iteration
Zeros of functions with Matlab: Bisection method and fixed point iteration Emma Perracchione Corso di Calcolo Numerico per Ingegneria Meccanica - Matr. PARI (Univ. PD) A.A. 2017/2018, secondo semestre
More informationAccelerating Convergence
Accelerating Convergence MATH 375 Numerical Analysis J. Robert Buchanan Department of Mathematics Fall 2013 Motivation We have seen that most fixed-point methods for root finding converge only linearly
More informationNUMERICAL METHODS. x n+1 = 2x n x 2 n. In particular: which of them gives faster convergence, and why? [Work to four decimal places.
NUMERICAL METHODS 1. Rearranging the equation x 3 =.5 gives the iterative formula x n+1 = g(x n ), where g(x) = (2x 2 ) 1. (a) Starting with x = 1, compute the x n up to n = 6, and describe what is happening.
More informationMATH 3795 Lecture 12. Numerical Solution of Nonlinear Equations.
MATH 3795 Lecture 12. Numerical Solution of Nonlinear Equations. Dmitriy Leykekhman Fall 2008 Goals Learn about different methods for the solution of f(x) = 0, their advantages and disadvantages. Convergence
More informationNON-LINEAR ALGEBRAIC EQUATIONS Lec. 5.1: Nonlinear Equation in Single Variable
NON-LINEAR ALGEBRAIC EQUATIONS Lec. 5.1: Nonlinear Equation in Single Variable Dr. Niket Kaisare Department of Chemical Engineering IIT Madras NPTEL Course: MATLAB Programming for Numerical Computations
More informationNumerical Methods Lecture 3
Numerical Methods Lecture 3 Nonlinear Equations by Pavel Ludvík Introduction Definition (Root or zero of a function) A root (or a zero) of a function f is a solution of an equation f (x) = 0. We learn
More information10.3 Steepest Descent Techniques
The advantage of the Newton and quasi-newton methods for solving systems of nonlinear equations is their speed of convergence once a sufficiently accurate approximation is known. A weakness of these methods
More informationVariable. Peter W. White Fall 2018 / Numerical Analysis. Department of Mathematics Tarleton State University
Newton s Iterative s Peter W. White white@tarleton.edu Department of Mathematics Tarleton State University Fall 2018 / Numerical Analysis Overview Newton s Iterative s Newton s Iterative s Newton s Iterative
More information2.3 The Fixed-Point Algorithm
.3 The Fied-Point Algorithm 1. Mean Value Theorem: Theorem Rolle stheorem: Suppose that f is continuous on a, b and is differentiable on a, b. If f a f b, then there eists a number c in a, b such that
More informationNumerical Methods. Root Finding
Numerical Methods Solving Non Linear 1-Dimensional Equations Root Finding Given a real valued function f of one variable (say ), the idea is to find an such that: f() 0 1 Root Finding Eamples Find real
More informationx 2 x n r n J(x + t(x x ))(x x )dt. For warming-up we start with methods for solving a single equation of one variable.
Maria Cameron 1. Fixed point methods for solving nonlinear equations We address the problem of solving an equation of the form (1) r(x) = 0, where F (x) : R n R n is a vector-function. Eq. (1) can be written
More informationMath 551 Homework Assignment 3 Page 1 of 6
Math 551 Homework Assignment 3 Page 1 of 6 Name and section: ID number: E-mail: 1. Consider Newton s method for finding + α with α > 0 by finding the positive root of f(x) = x 2 α = 0. Assuming that x
More informationHomework and Computer Problems for Math*2130 (W17).
Homework and Computer Problems for Math*2130 (W17). MARCUS R. GARVIE 1 December 21, 2016 1 Department of Mathematics & Statistics, University of Guelph NOTES: These questions are a bare minimum. You should
More informationNUMERICAL AND STATISTICAL COMPUTING (MCA-202-CR)
NUMERICAL AND STATISTICAL COMPUTING (MCA-202-CR) Autumn Session UNIT 1 Numerical analysis is the study of algorithms that uses, creates and implements algorithms for obtaining numerical solutions to problems
More information(6.2) -The Shooting Method for Nonlinear Problems
(6.2) -The Shooting Method for Nonlinear Problems Consider the boundary value problems (BVPs) for the second order differential equation of the form (*) y f x,y,y, a x b, y a and y b. When f x,y,y is linear
More informationFIXED POINT ITERATION
FIXED POINT ITERATION The idea of the fixed point iteration methods is to first reformulate a equation to an equivalent fixed point problem: f (x) = 0 x = g(x) and then to use the iteration: with an initial
More informationLine Search Methods. Shefali Kulkarni-Thaker
1 BISECTION METHOD Line Search Methods Shefali Kulkarni-Thaker Consider the following unconstrained optimization problem min f(x) x R Any optimization algorithm starts by an initial point x 0 and performs
More informationSolution Methods. Richard Lusby. Department of Management Engineering Technical University of Denmark
Solution Methods Richard Lusby Department of Management Engineering Technical University of Denmark Lecture Overview (jg Unconstrained Several Variables Quadratic Programming Separable Programming SUMT
More informationChapter 1. Root Finding Methods. 1.1 Bisection method
Chapter 1 Root Finding Methods We begin by considering numerical solutions to the problem f(x) = 0 (1.1) Although the problem above is simple to state it is not always easy to solve analytically. This
More informationGoals for This Lecture:
Goals for This Lecture: Learn the Newton-Raphson method for finding real roots of real functions Learn the Bisection method for finding real roots of a real function Look at efficient implementations of
More informationECONOMICS 207 SPRING 2006 LABORATORY EXERCISE 5 KEY. 8 = 10(5x 2) = 9(3x + 8), x 50x 20 = 27x x = 92 x = 4. 8x 2 22x + 15 = 0 (2x 3)(4x 5) = 0
ECONOMICS 07 SPRING 006 LABORATORY EXERCISE 5 KEY Problem. Solve the following equations for x. a 5x 3x + 8 = 9 0 5x 3x + 8 9 8 = 0(5x ) = 9(3x + 8), x 0 3 50x 0 = 7x + 7 3x = 9 x = 4 b 8x x + 5 = 0 8x
More informationPanHomc'r I'rui;* :".>r '.a'' W"»' I'fltolt. 'j'l :. r... Jnfii<on. Kslaiaaac. <.T i.. %.. 1 >
5 28 (x / &» )»(»»» Q ( 3 Q» (» ( (3 5» ( q 2 5 q 2 5 5 8) 5 2 2 ) ~ ( / x {» /»»»»» (»»» ( 3 ) / & Q ) X ] Q & X X X x» 8 ( &» 2 & % X ) 8 x & X ( #»»q 3 ( ) & X 3 / Q X»»» %» ( z 22 (»» 2» }» / & 2 X
More informationSolution of Algebric & Transcendental Equations
Page15 Solution of Algebric & Transcendental Equations Contents: o Introduction o Evaluation of Polynomials by Horner s Method o Methods of solving non linear equations o Bracketing Methods o Bisection
More informationSOLUTIONS to Exercises from Optimization
SOLUTIONS to Exercises from Optimization. Use the bisection method to find the root correct to 6 decimal places: 3x 3 + x 2 = x + 5 SOLUTION: For the root finding algorithm, we need to rewrite the equation
More informationMANY BILLS OF CONCERN TO PUBLIC
- 6 8 9-6 8 9 6 9 XXX 4 > -? - 8 9 x 4 z ) - -! x - x - - X - - - - - x 00 - - - - - x z - - - x x - x - - - - - ) x - - - - - - 0 > - 000-90 - - 4 0 x 00 - -? z 8 & x - - 8? > 9 - - - - 64 49 9 x - -
More informationNumerical Optimization
Unconstrained Optimization Computer Science and Automation Indian Institute of Science Bangalore 560 01, India. NPTEL Course on Unconstrained Minimization Let f : R n R. Consider the optimization problem:
More informationApplied Computational Economics Workshop. Part 3: Nonlinear Equations
Applied Computational Economics Workshop Part 3: Nonlinear Equations 1 Overview Introduction Function iteration Newton s method Quasi-Newton methods Practical example Practical issues 2 Introduction Nonlinear
More information446 CHAP. 8 NUMERICAL OPTIMIZATION. Newton's Search for a Minimum of f(x,y) Newton s Method
446 CHAP. 8 NUMERICAL OPTIMIZATION Newton's Search for a Minimum of f(xy) Newton s Method The quadratic approximation method of Section 8.1 generated a sequence of seconddegree Lagrange polynomials. It
More informationMath Numerical Analysis
Math 541 - Numerical Analysis Lecture Notes Zeros and Roots Joseph M. Mahaffy, jmahaffy@mail.sdsu.edu Department of Mathematics and Statistics Dynamical Systems Group Computational Sciences Research Center
More informationComputational Methods CMSC/AMSC/MAPL 460. Solving nonlinear equations and zero finding. Finding zeroes of functions
Computational Methods CMSC/AMSC/MAPL 460 Solving nonlinear equations and zero finding Ramani Duraiswami, Dept. of Computer Science Where does it arise? Finding zeroes of functions Solving functional equations
More informationOutline. Math Numerical Analysis. Intermediate Value Theorem. Lecture Notes Zeros and Roots. Joseph M. Mahaffy,
Outline Math 541 - Numerical Analysis Lecture Notes Zeros and Roots Joseph M. Mahaffy, jmahaffy@mail.sdsu.edu Department of Mathematics and Statistics Dynamical Systems Group Computational Sciences Research
More informationMath Numerical Analysis Mid-Term Test Solutions
Math 400 - Numerical Analysis Mid-Term Test Solutions. Short Answers (a) A sufficient and necessary condition for the bisection method to find a root of f(x) on the interval [a,b] is f(a)f(b) < 0 or f(a)
More informationCHAPTER 2 EXTRACTION OF THE QUADRATICS FROM REAL ALGEBRAIC POLYNOMIAL
24 CHAPTER 2 EXTRACTION OF THE QUADRATICS FROM REAL ALGEBRAIC POLYNOMIAL 2.1 INTRODUCTION Polynomial factorization is a mathematical problem, which is often encountered in applied sciences and many of
More informationSolutions of Equations in One Variable. Newton s Method
Solutions of Equations in One Variable Newton s Method Numerical Analysis (9th Edition) R L Burden & J D Faires Beamer Presentation Slides prepared by John Carroll Dublin City University c 2011 Brooks/Cole,
More informationi.e., into a monomial, using the Arithmetic-Geometric Mean Inequality, the result will be a posynomial approximation!
Dennis L. Bricker Dept of Mechanical & Industrial Engineering The University of Iowa i.e., 1 1 1 Minimize X X X subject to XX 4 X 1 0.5X 1 Minimize X X X X 1X X s.t. 4 1 1 1 1 4X X 1 1 1 1 0.5X X X 1 1
More informationQR Decomposition. When solving an overdetermined system by projection (or a least squares solution) often the following method is used:
(In practice not Gram-Schmidt, but another process Householder Transformations are used.) QR Decomposition When solving an overdetermined system by projection (or a least squares solution) often the following
More information1 The best of all possible worlds
Notes for 2017-03-18 1 The best of all possible worlds Last time, we discussed three methods of solving f(x) = 0: Newton, modified Newton, and bisection. Newton is potentially faster than bisection; bisection
More informationThe Steepest Descent Algorithm for Unconstrained Optimization
The Steepest Descent Algorithm for Unconstrained Optimization Robert M. Freund February, 2014 c 2014 Massachusetts Institute of Technology. All rights reserved. 1 1 Steepest Descent Algorithm The problem
More informationNumerical Optimization
Unconstrained Optimization (II) Computer Science and Automation Indian Institute of Science Bangalore 560 012, India. NPTEL Course on Unconstrained Optimization Let f : R R Unconstrained problem min x
More informationChapter 2 Solutions of Equations of One Variable
Chapter 2 Solutions of Equations of One Variable 2.1 Bisection Method In this chapter we consider one of the most basic problems of numerical approximation, the root-finding problem. This process involves
More informationSOLUTION OF ALGEBRAIC AND TRANSCENDENTAL EQUATIONS BISECTION METHOD
BISECTION METHOD If a function f(x) is continuous between a and b, and f(a) and f(b) are of opposite signs, then there exists at least one root between a and b. It is shown graphically as, Let f a be negative
More informationNumerical Methods in Informatics
Numerical Methods in Informatics Lecture 2, 30.09.2016: Nonlinear Equations in One Variable http://www.math.uzh.ch/binf4232 Tulin Kaman Institute of Mathematics, University of Zurich E-mail: tulin.kaman@math.uzh.ch
More informationSimple Iteration, cont d
Jim Lambers MAT 772 Fall Semester 2010-11 Lecture 2 Notes These notes correspond to Section 1.2 in the text. Simple Iteration, cont d In general, nonlinear equations cannot be solved in a finite sequence
More informationMATH 3795 Lecture 13. Numerical Solution of Nonlinear Equations in R N.
MATH 3795 Lecture 13. Numerical Solution of Nonlinear Equations in R N. Dmitriy Leykekhman Fall 2008 Goals Learn about different methods for the solution of F (x) = 0, their advantages and disadvantages.
More informationLecture 5. September 4, 2018 Math/CS 471: Introduction to Scientific Computing University of New Mexico
Lecture 5 September 4, 2018 Math/CS 471: Introduction to Scientific Computing University of New Mexico 1 Review: Office hours at regularly scheduled times this week Tuesday: 9:30am-11am Wed: 2:30pm-4:00pm
More informationMidterm Review. Igor Yanovsky (Math 151A TA)
Midterm Review Igor Yanovsky (Math 5A TA) Root-Finding Methods Rootfinding methods are designed to find a zero of a function f, that is, to find a value of x such that f(x) =0 Bisection Method To apply
More informationRoot Finding with Newton s Method
Root Finding with Newton s Method 1 Newton s Method derivation of the method an implementation with SymPy and Julia 2 Convergence of Newton s Method linear and quadratic convergence geometric convergence
More informationMath Introduction to Numerical Methods - Winter 2011 Homework 2 Assigned: Friday, January 14, Due: Thursday, January 27,
Math 371 - Introduction to Numerical Methods - Winter 2011 Homework 2 Assigned: Friday, January 14, 2011. Due: Thursday, January 27, 2011.. Include a cover page. You do not need to hand in a problem sheet.
More informationPART I Lecture Notes on Numerical Solution of Root Finding Problems MATH 435
PART I Lecture Notes on Numerical Solution of Root Finding Problems MATH 435 Professor Biswa Nath Datta Department of Mathematical Sciences Northern Illinois University DeKalb, IL. 60115 USA E mail: dattab@math.niu.edu
More information5.6 Logarithmic and Exponential Equations
SECTION 5.6 Logarithmic and Exponential Equations 305 5.6 Logarithmic and Exponential Equations PREPARING FOR THIS SECTION Before getting started, review the following: Solving Equations Using a Graphing
More informationMA 8019: Numerical Analysis I Solution of Nonlinear Equations
MA 8019: Numerical Analysis I Solution of Nonlinear Equations Suh-Yuh Yang ( 楊肅煜 ) Department of Mathematics, National Central University Jhongli District, Taoyuan City 32001, Taiwan syyang@math.ncu.edu.tw
More informationSolving nonlinear equations (See online notes and lecture notes for full details) 1.3: Newton s Method
Solving nonlinear equations (See online notes and lecture notes for full details) 1.3: Newton s Method MA385 Numerical Analysis September 2018 (1/16) Sir Isaac Newton, 1643-1727, England. Easily one of
More informationLecture 7. Root finding I. 1 Introduction. 2 Graphical solution
1 Introduction Lecture 7 Root finding I For our present purposes, root finding is the process of finding a real value of x which solves the equation f (x)=0. Since the equation g x =h x can be rewritten
More informationNonlinear Equations. Not guaranteed to have any real solutions, but generally do for astrophysical problems.
Nonlinear Equations Often (most of the time??) the relevant system of equations is not linear in the unknowns. Then, cannot decompose as Ax = b. Oh well. Instead write as: (1) f(x) = 0 function of one
More informationOptimization Methods. Lecture 19: Line Searches and Newton s Method
15.93 Optimization Methods Lecture 19: Line Searches and Newton s Method 1 Last Lecture Necessary Conditions for Optimality (identifies candidates) x local min f(x ) =, f(x ) PSD Slide 1 Sufficient Conditions
More informationPractical Numerical Analysis: Sheet 3 Solutions
Practical Numerical Analysis: Sheet 3 Solutions 1. We need to compute the roots of the function defined by f(x) = sin(x) + sin(x 2 ) on the interval [0, 3] using different numerical methods. First we consider
More informationNumerical Methods. King Saud University
Numerical Methods King Saud University Aims In this lecture, we will... Introduce the topic of numerical methods Consider the Error analysis and sources of errors Introduction A numerical method which
More informationNumerical Solution of f(x) = 0
Numerical Solution of f(x) = 0 Gerald W. Recktenwald Department of Mechanical Engineering Portland State University gerry@pdx.edu ME 350: Finding roots of f(x) = 0 Overview Topics covered in these slides
More informationNumerical solutions of nonlinear systems of equations
Numerical solutions of nonlinear systems of equations Tsung-Ming Huang Department of Mathematics National Taiwan Normal University, Taiwan E-mail: min@math.ntnu.edu.tw August 28, 2011 Outline 1 Fixed points
More informationComputational Methods. Solving Equations
Computational Methods Solving Equations Manfred Huber 2010 1 Solving Equations Solving scalar equations is an elemental task that arises in a wide range of applications Corresponds to finding parameters
More informationHalley s Method: A Cubically Converging. Method of Root Approximation
Halley s Method: A Cubically Converging Method of Root Approximation Gabriel Kramer Brittany Sypin December 3, 2011 Abstract This paper will discuss numerical ways of approximating the solutions to the
More informationBEKG 2452 NUMERICAL METHODS Solution of Nonlinear Equations
BEKG 2452 NUMERICAL METHODS Solution of Nonlinear Equations Ser Lee Loh a, Wei Sen Loi a a Fakulti Kejuruteraan Elektrik Universiti Teknikal Malaysia Melaka Lesson Outcome Upon completion of this lesson,
More informationUniversity of Delaware Department of Mathematical Sciences Math 353 Engineering Mathematics III 06S C. Bacuta
University of Delaware Department of Mathematical Sciences Math 353 Engineering Mathematics III 06S C. Bacuta Homework 4: Due Friday, 0/4/06, 10:00am Part I) (1) Section., Problems #3(a,d), 9, 11. () Apply
More informationMath 56 Homework 1 Michael Downs. ne n 10 + ne n (1)
. Problem (a) Yes. The following equation: ne n + ne n () holds for all n R but, since we re only concerned with the asymptotic behavior as n, let us only consider n >. Dividing both sides by n( + ne n
More informationNumerical Methods I Solving Nonlinear Equations
Numerical Methods I Solving Nonlinear Equations Aleksandar Donev Courant Institute, NYU 1 donev@courant.nyu.edu 1 MATH-GA 2011.003 / CSCI-GA 2945.003, Fall 2014 October 16th, 2014 A. Donev (Courant Institute)
More informationSolving nonlinear equations
Chapter Solving nonlinear equations Bisection Introduction Linear equations are of the form: find x such that ax + b = 0 Proposition Let f be a real-valued function that is defined and continuous on a
More informationMATH ASSIGNMENT 07 SOLUTIONS. 8.1 Following is census data showing the population of the US between 1900 and 2000:
MATH4414.01 ASSIGNMENT 07 SOLUTIONS 8.1 Following is census data showing the population of the US between 1900 and 2000: Years after 1900 Population in millions 0 76.0 20 105.7 40 131.7 60 179.3 80 226.5
More informationDetermining the Roots of Non-Linear Equations Part I
Determining the Roots of Non-Linear Equations Part I Prof. Dr. Florian Rupp German University of Technology in Oman (GUtech) Introduction to Numerical Methods for ENG & CS (Mathematics IV) Spring Term
More informationUnconstrained optimization I Gradient-type methods
Unconstrained optimization I Gradient-type methods Antonio Frangioni Department of Computer Science University of Pisa www.di.unipi.it/~frangio frangio@di.unipi.it Computational Mathematics for Learning
More informationExample Graphs of f x are given below. (c) no zero in 2,2
. - Bisection Method The idea of the Bisection Method is based on the Intermediate Value Theorem that ou studied in Calculus I. Recall that: Intermediate Value Theorem: If f is continuous on a,b and K
More informationNonlinearity Root-finding Bisection Fixed Point Iteration Newton s Method Secant Method Conclusion. Nonlinear Systems
Nonlinear Systems CS 205A: Mathematical Methods for Robotics, Vision, and Graphics Justin Solomon CS 205A: Mathematical Methods Nonlinear Systems 1 / 24 Part III: Nonlinear Problems Not all numerical problems
More informationSYSTEMS OF NONLINEAR EQUATIONS
SYSTEMS OF NONLINEAR EQUATIONS Widely used in the mathematical modeling of real world phenomena. We introduce some numerical methods for their solution. For better intuition, we examine systems of two
More informationR x n. 2 R We simplify this algebraically, obtaining 2x n x n 1 x n x n
Math 42 Homework 4. page 3, #9 This is a modification of the bisection method. Write a MATLAB function similar to bisect.m. Here, given the points P a a,f a and P b b,f b with f a f b,we compute the point
More informationINTRODUCTION TO COMPUTATIONAL MATHEMATICS
INTRODUCTION TO COMPUTATIONAL MATHEMATICS Course Notes for CM 271 / AMATH 341 / CS 371 Fall 2007 Instructor: Prof. Justin Wan School of Computer Science University of Waterloo Course notes by Prof. Hans
More informationNonlinear Programming
Nonlinear Programming Kees Roos e-mail: C.Roos@ewi.tudelft.nl URL: http://www.isa.ewi.tudelft.nl/ roos LNMB Course De Uithof, Utrecht February 6 - May 8, A.D. 2006 Optimization Group 1 Outline for week
More informationROOTFINDING. We assume the interest rate r holds over all N in +N out periods. h P in (1 + r) N out. N in 1 i h i
ROOTFINDING We want to find the numbers x for which f(x) = 0, with f a given function. Here, we denote such roots or zeroes by the Greek letter α. Rootfinding problems occur in many contexts. Sometimes
More informationSTOP, a i+ 1 is the desired root. )f(a i) > 0. Else If f(a i+ 1. Set a i+1 = a i+ 1 and b i+1 = b Else Set a i+1 = a i and b i+1 = a i+ 1
53 17. Lecture 17 Nonlinear Equations Essentially, the only way that one can solve nonlinear equations is by iteration. The quadratic formula enables one to compute the roots of p(x) = 0 when p P. Formulas
More informationMath 473: Practice Problems for Test 1, Fall 2011, SOLUTIONS
Math 473: Practice Problems for Test 1, Fall 011, SOLUTIONS Show your work: 1. (a) Compute the Taylor polynomials P n (x) for f(x) = sin x and x 0 = 0. Solution: Compute f(x) = sin x, f (x) = cos x, f
More informationLecture 8. Root finding II
1 Introduction Lecture 8 Root finding II In the previous lecture we considered the bisection root-bracketing algorithm. It requires only that the function be continuous and that we have a root bracketed
More informationRoot Finding For NonLinear Equations Bisection Method
Root Finding For NonLinear Equations Bisection Method P. Sam Johnson November 17, 2014 P. Sam Johnson (NITK) Root Finding For NonLinear Equations Bisection MethodNovember 17, 2014 1 / 26 Introduction The
More information1 Newton s Method. Suppose we want to solve: x R. At x = x, f (x) can be approximated by:
Newton s Method Suppose we want to solve: (P:) min f (x) At x = x, f (x) can be approximated by: n x R. f (x) h(x) := f ( x)+ f ( x) T (x x)+ (x x) t H ( x)(x x), 2 which is the quadratic Taylor expansion
More informationOptimization and Calculus
Optimization and Calculus To begin, there is a close relationship between finding the roots to a function and optimizing a function. In the former case, we solve for x. In the latter, we solve: g(x) =
More informationUnit 2: Solving Scalar Equations. Notes prepared by: Amos Ron, Yunpeng Li, Mark Cowlishaw, Steve Wright Instructor: Steve Wright
cs416: introduction to scientific computing 01/9/07 Unit : Solving Scalar Equations Notes prepared by: Amos Ron, Yunpeng Li, Mark Cowlishaw, Steve Wright Instructor: Steve Wright 1 Introduction We now
More information1.1: The bisection method. September 2017
(1/11) 1.1: The bisection method Solving nonlinear equations MA385/530 Numerical Analysis September 2017 3 2 f(x)= x 2 2 x axis 1 0 1 x [0] =a x [2] =1 x [3] =1.5 x [1] =b 2 0.5 0 0.5 1 1.5 2 2.5 1 Solving
More informationNumerical Study of Some Iterative Methods for Solving Nonlinear Equations
International Journal of Engineering Science Invention ISSN (Online): 2319 6734, ISSN (Print): 2319 6726 Volume 5 Issue 2 February 2016 PP.0110 Numerical Study of Some Iterative Methods for Solving Nonlinear
More informationTHE SECANT METHOD. q(x) = a 0 + a 1 x. with
THE SECANT METHOD Newton s method was based on using the line tangent to the curve of y = f (x), with the point of tangency (x 0, f (x 0 )). When x 0 α, the graph of the tangent line is approximately the
More informationCLASS NOTES Models, Algorithms and Data: Introduction to computing 2018
CLASS NOTES Models, Algorithms and Data: Introduction to computing 2018 Petros Koumoutsakos, Jens Honore Walther (Last update: April 16, 2018) IMPORTANT DISCLAIMERS 1. REFERENCES: Much of the material
More informationMath 2Z03 - Tutorial # 4. Oct. 5th, 6th, 7th, 2015
Math 2Z03 - Tutorial # 4 Oct. 5th, 6th, 7th, 2015 Tutorial Info: Tutorial Website: http://ms.mcmaster.ca/ dedieula/2z03.html Office Hours: Mondays 3pm - 5pm (in the Math Help Centre) Tutorial #4: 3.1 Theory
More informationShiqian Ma, MAT-258A: Numerical Optimization 1. Chapter 3. Gradient Method
Shiqian Ma, MAT-258A: Numerical Optimization 1 Chapter 3 Gradient Method Shiqian Ma, MAT-258A: Numerical Optimization 2 3.1. Gradient method Classical gradient method: to minimize a differentiable convex
More informationRegister machines L2 18
Register machines L2 18 Algorithms, informally L2 19 No precise definition of algorithm at the time Hilbert posed the Entscheidungsproblem, just examples. Common features of the examples: finite description
More informationOrder of convergence
Order of convergence Linear and Quadratic Order of convergence Computing square root with Newton s Method Given a > 0, p def = a is positive root of equation Newton s Method p k+1 = p k p2 k a 2p k = 1
More informationToday s class. Numerical differentiation Roots of equation Bracketing methods. Numerical Methods, Fall 2011 Lecture 4. Prof. Jinbo Bi CSE, UConn
Today s class Numerical differentiation Roots of equation Bracketing methods 1 Numerical Differentiation Finite divided difference First forward difference First backward difference Lecture 3 2 Numerical
More informationNonlinear Equations. Chapter The Bisection Method
Chapter 6 Nonlinear Equations Given a nonlinear function f(), a value r such that f(r) = 0, is called a root or a zero of f() For eample, for f() = e 016064, Fig?? gives the set of points satisfying y
More informationIf you need the source code from the sample solutions, copy and paste from the PDF should work. If you re having trouble, send me an .
AM117 Sample Solutions, HW#1 Hi, I m Andreas Kloeckner, your grader for this course. Feel free to email me at kloeckner@dam.brown.edu. The code for the sample solutions is written so that it will run in
More informationPreliminary Examination in Numerical Analysis
Department of Applied Mathematics Preliminary Examination in Numerical Analysis August 7, 06, 0 am pm. Submit solutions to four (and no more) of the following six problems. Show all your work, and justify
More informationChapter 4. Unconstrained optimization
Chapter 4. Unconstrained optimization Version: 28-10-2012 Material: (for details see) Chapter 11 in [FKS] (pp.251-276) A reference e.g. L.11.2 refers to the corresponding Lemma in the book [FKS] PDF-file
More information3.1 Introduction. Solve non-linear real equation f(x) = 0 for real root or zero x. E.g. x x 1.5 =0, tan x x =0.
3.1 Introduction Solve non-linear real equation f(x) = 0 for real root or zero x. E.g. x 3 +1.5x 1.5 =0, tan x x =0. Practical existence test for roots: by intermediate value theorem, f C[a, b] & f(a)f(b)
More informationSkill 6 Exponential and Logarithmic Functions
Skill 6 Exponential and Logarithmic Functions Skill 6a: Graphs of Exponential Functions Skill 6b: Solving Exponential Equations (not requiring logarithms) Skill 6c: Definition of Logarithms Skill 6d: Graphs
More informationNon-linear least squares
Non-linear least squares Concept of non-linear least squares We have extensively studied linear least squares or linear regression. We see that there is a unique regression line that can be determined
More informationHomework No. 2 Solutions TA: Theo Drivas
550.386 Homework No. Solutions TA: Theo Drivas Problem 1. (a) Consider the initial-value problems for the following two differential equations: (i) dy/dt = y, y(0) = y 0, (ii) dy/dt = y, y(0) = y 0. Use
More information