Line Search Methods. Shefali Kulkarni-Thaker

Similar documents
1. Method 1: bisection. The bisection methods starts from two points a 0 and b 0 such that

McMaster University CS-4-6TE3/CES Assignment-1 Solution Set

x 2 x n r n J(x + t(x x ))(x x )dt. For warming-up we start with methods for solving a single equation of one variable.

15 Nonlinear Equations and Zero-Finders

Nonlinear Equations. Chapter The Bisection Method

MATH 4211/6211 Optimization Basics of Optimization Problems

Motivation: We have already seen an example of a system of nonlinear equations when we studied Gaussian integration (p.8 of integration notes)

MATH 3795 Lecture 12. Numerical Solution of Nonlinear Equations.

Computational Methods CMSC/AMSC/MAPL 460. Solving nonlinear equations and zero finding. Finding zeroes of functions

Numerical Methods in Informatics

Chapter 3: Root Finding. September 26, 2005

1 Newton s Method. Suppose we want to solve: x R. At x = x, f (x) can be approximated by:

ECS550NFB Introduction to Numerical Methods using Matlab Day 2

Nonlinearity Root-finding Bisection Fixed Point Iteration Newton s Method Secant Method Conclusion. Nonlinear Systems

Outline. Math Numerical Analysis. Intermediate Value Theorem. Lecture Notes Zeros and Roots. Joseph M. Mahaffy,

Math Numerical Analysis

3.1 Introduction. Solve non-linear real equation f(x) = 0 for real root or zero x. E.g. x x 1.5 =0, tan x x =0.

Numerical Optimization

UNCONSTRAINED OPTIMIZATION

IE 5531: Engineering Optimization I

STOP, a i+ 1 is the desired root. )f(a i) > 0. Else If f(a i+ 1. Set a i+1 = a i+ 1 and b i+1 = b Else Set a i+1 = a i and b i+1 = a i+ 1

ROOT FINDING REVIEW MICHELLE FENG

Midterm Review. Igor Yanovsky (Math 151A TA)

Numerical Methods. Root Finding

Solution of Nonlinear Equations

Scientific Computing: An Introductory Survey

IE 5531: Engineering Optimization I

Numerical Solution of f(x) = 0

Numerical solutions of nonlinear systems of equations

Numerical Methods Dr. Sanjeev Kumar Department of Mathematics Indian Institute of Technology Roorkee Lecture No 7 Regula Falsi and Secant Methods

PART I Lecture Notes on Numerical Solution of Root Finding Problems MATH 435

Lecture Notes to Accompany. Scientific Computing An Introductory Survey. by Michael T. Heath. Chapter 5. Nonlinear Equations

Finding the Roots of f(x) = 0. Gerald W. Recktenwald Department of Mechanical Engineering Portland State University

Finding the Roots of f(x) = 0

Roots of Equations. ITCS 4133/5133: Introduction to Numerical Methods 1 Roots of Equations

17 Solution of Nonlinear Systems

1 Numerical optimization

Today s class. Numerical differentiation Roots of equation Bracketing methods. Numerical Methods, Fall 2011 Lecture 4. Prof. Jinbo Bi CSE, UConn

Outline. Scientific Computing: An Introductory Survey. Nonlinear Equations. Nonlinear Equations. Examples: Nonlinear Equations

Variable. Peter W. White Fall 2018 / Numerical Analysis. Department of Mathematics Tarleton State University

1-D Optimization. Lab 16. Overview of Line Search Algorithms. Derivative versus Derivative-Free Methods

CHAPTER-II ROOTS OF EQUATIONS

1 Numerical optimization

Solution of Algebric & Transcendental Equations

Introduction to Nonlinear Optimization Paul J. Atzberger

Determination of Feasible Directions by Successive Quadratic Programming and Zoutendijk Algorithms: A Comparative Study

CS 323: Numerical Analysis and Computing

Nonlinear Programming

Numerical Methods I Solving Nonlinear Equations

GENG2140, S2, 2012 Week 7: Curve fitting

Numerical Methods. King Saud University

Solving Non-Linear Equations (Root Finding)

Optimization. Totally not complete this is...don't use it yet...

Root Finding: Close Methods. Bisection and False Position Dr. Marco A. Arocha Aug, 2014

Chapter 6. Nonlinear Equations. 6.1 The Problem of Nonlinear Root-finding. 6.2 Rate of Convergence

Chapter 4. Solution of Non-linear Equation. Module No. 1. Newton s Method to Solve Transcendental Equation

Chapter 2 Solutions of Equations of One Variable

Single Variable Minimization

8 Numerical methods for unconstrained problems

Unconstrained minimization of smooth functions

NUMERICAL AND STATISTICAL COMPUTING (MCA-202-CR)

Optimization II: Unconstrained Multivariable

Computational Methods. Solving Equations

Maria Cameron. f(x) = 1 n

Math 471. Numerical methods Root-finding algorithms for nonlinear equations

Bisection and False Position Dr. Marco A. Arocha Aug, 2014

MATH 3795 Lecture 13. Numerical Solution of Nonlinear Equations in R N.

Goals for This Lecture:

SOLUTION OF ALGEBRAIC AND TRANSCENDENTAL EQUATIONS BISECTION METHOD

CLASS NOTES Models, Algorithms and Data: Introduction to computing 2018

Applied Computational Economics Workshop. Part 3: Nonlinear Equations

Optimization Tutorial 1. Basic Gradient Descent

MVE165/MMG631 Linear and integer optimization with applications Lecture 13 Overview of nonlinear programming. Ann-Brith Strömberg

CS 323: Numerical Analysis and Computing

Chapter 4. Unconstrained optimization

Nonlinearity Root-finding Bisection Fixed Point Iteration Newton s Method Secant Method Conclusion. Nonlinear Systems

Chapter 1. Root Finding Methods. 1.1 Bisection method

Determining the Roots of Non-Linear Equations Part I

6.252 NONLINEAR PROGRAMMING LECTURE 10 ALTERNATIVES TO GRADIENT PROJECTION LECTURE OUTLINE. Three Alternatives/Remedies for Gradient Projection

Root Finding (and Optimisation)

E5295/5B5749 Convex optimization with engineering applications. Lecture 8. Smooth convex unconstrained and equality-constrained minimization

Unit 2: Solving Scalar Equations. Notes prepared by: Amos Ron, Yunpeng Li, Mark Cowlishaw, Steve Wright Instructor: Steve Wright

Review of Classical Optimization

Numerical Analysis Fall. Roots: Open Methods

Lecture V. Numerical Optimization

A Review of Bracketing Methods for Finding Zeros of Nonlinear Functions

Optimization. Next: Curve Fitting Up: Numerical Analysis for Chemical Previous: Linear Algebraic and Equations. Subsections

Numerisches Rechnen. (für Informatiker) M. Grepl P. Esser & G. Welper & L. Zhang. Institut für Geometrie und Praktische Mathematik RWTH Aachen

Methods that avoid calculating the Hessian. Nonlinear Optimization; Steepest Descent, Quasi-Newton. Steepest Descent

Higher-Order Methods

Contraction Mappings Consider the equation

Solving Nonlinear Equations

Notes for Numerical Analysis Math 5465 by S. Adjerid Virginia Polytechnic Institute and State University. (A Rough Draft)

BEKG 2452 NUMERICAL METHODS Solution of Nonlinear Equations

INTRODUCTION TO NUMERICAL ANALYSIS

Scientific Computing. Roots of Equations

CS 450 Numerical Analysis. Chapter 5: Nonlinear Equations

Solving nonlinear equations (See online notes and lecture notes for full details) 1.3: Newton s Method

Figure 1: Graph of y = x cos(x)

Line Search Techniques

Transcription:

1 BISECTION METHOD Line Search Methods Shefali Kulkarni-Thaker Consider the following unconstrained optimization problem min f(x) x R Any optimization algorithm starts by an initial point x 0 and performs series of iterations to reach the optimal point x. At any k th iteration the next point is given by the sum of old point and the direction in which to search of a next point multiplied by how far to go in that direction. Thus x k+1 = x k + α k d k here d k is a search direction and α k is a positive scalar determining how far to go in that direction. It is called as the step length. Additionally, the new point must be such that the function value (the function which we are optimizing) at that point should be less than or equal to the previous point. This is quite obvious because if we are moving in a different direction where the function value is increasing we are not really moving towards the minimum. Thus, f(x k+1 ) f(x k ) f(x k + α k d k ) f(x k ) The general optimization algorithm begins with an initial point, finds a descent search direction, determines the step length and checks the termination criteria. So, when we are trying to find the step length, we already know that the direction in which we are going is descent. So, ideally we want to go far enough so that the function reaches its minimum. Thus, given the previous point and a descent search direction, we are trying to find a scalar step length such that the value of the function is minimum in that direction. In its mathematical form we may write, min f(x + αd) α 0 Since, x and d are known, this problem reduces to a univariate, 1 D, minimization problem.assuming that f is smooth and continuous, we find its optimum where its first-derivative is zero, i.e. f(x + αd) = 0. All we are doing is trying to find zero of a function (i.e. the point where the curves intersects the x axis). We will visit some known and some newer zero-finding (or root-finding) techniques. When, the dimensionality of the problem or the degree of equation increases finding an exact zero is difficult. Usually, we are looking for an interval 0 [a, b] so that a b < ɛ, where ɛ is an acceptable tolerance. 1 Bisection Method In bisection method we reduce begin with an interval so that 0 [a, b] and divide the interval in two halves,i.e. [a, a+b a+b ] and [, b]. A next search interval is chosen by comparing and finding which one 2 2 has zero. This is done by evaluating the sign. The algorithm for this is given as follows: Choose a, b so that f(a)f(b) < 0 1. m = a+b 2 1

1.1 Features of bisection 1 BISECTION METHOD 2. f(m) = 0 Stop and return. 3. f(m)f(a) < 0; b = m 4. f(m)f(b) < 0; a = m 5. a b < ɛ stop and return. 1.1 Features of bisection Guaranteed Convergence. Slow as it has linear convergence with β = 0.5. At the k th iteration b a 2 k = ɛ giving total evaluations as k = log( b a 2 ) 1.2 Example Converging f(x) = x with [a, b] = [ 0.6, 0.75] and ɛ = 10 4. Number a f(a) b f(b) m f(m) 1-0.6-0.441176 0.75 0.48 0.075 0.07458 2-0.6-0.441176 0.075 0.07458-0.2625-0.245578 3-0.2625-0.245578 0.075 0.07458-0.09375-0.092933 4-0.09375-0.092933 0.075 0.07458-0.009375-0.009374 5-0.009375-0.009374 0.075 0.07458 0.032813 0.032777 6-0.009375-0.009374 0.032813 0.032777 0.011719 0.011717 7-0.009375-0.009374 0.011719 0.011717 0.001172 0.001172 8-0.009375-0.009374 0.001172 0.001172-0.004102-0.004101 9-0.004102-0.004101 0.001172 0.001172-0.001465-0.001465 10-0.001465-0.001465 0.001172 0.001172-0.000146-0.000146 11-0.000146-0.000146 0.001172 0.001172 0.000513 0.000513 12-0.000146-0.000146 0.000513 0.000513 0.000183 0.000183 13-0.000146-0.000146 0.000183 0.000183 0.000018 0.000018 14-0.000146-0.000146 0.000018 0.000018-0.000064-0.000064 Figure 1: The bisection method If the function is smooth and continuous, the speed of convergence can be improved by using the derivative information 2

2 NEWTON S METHOD 2 Newton s Method In Newton s method does a linear approximation of the function and finding the x-intercept of that approximation, thereby improving the performance of the bisection method. Linear approximation can be done by using Taylor s series. 2.1 Features of Newton s method f (x k+1 ) = f(x k ) + f (x k )(x k+1 x k ) f (x k+1 ) = 0 x k+1 = x k f(x k) f (x k ) Newton s method has a quadratic convergence when the chosen point is close enough to zero. If the derivative is zero at the root, it has only local quadratic convergence. Numerical difficulties occur when the first-derivative is zero. If a poor starting point is chosen the method may fail to converge or diverge. If it is difficult to find analytical derivation, secant method may be used. For a smooth, continuous function when proper starting point is chosen, Newton s method can be real fast. The convergence of f(x) = x with x 0 = 0.5 and ɛ = 10 4. X f(x) 1 0.5 2-0.333333 3 0.083333 4-0.001166 5 0 Figure 2: Newton s method converging with x 0 = 0.5 The divergence of f(x) = x with x 0 = 0.75 and ɛ = 10 4. 3

3 SECANT METHOD 1 0.75 2-1.928571 3-5.275529 4-10.944297 5-22.072876 6-44.236547.. 20-725265.55757 Figure 3: Newton s method diverging with x 0 = 0.75 Newton s method to find the minimum uses the first derivative of the approximation an equalling it to zero. 3 Secant Method f (x k+1 ) = f(x k ) + f (x k )(x k+1 x k ) f (x k+1 ) = f (x k ) + f (x k )(x k+1 x k ) f (x k+1 ) = 0 x k+1 = x k f (x k ) f (x k ) When f is expensive or cumbersome to calculate, one can use secant s method to approximate the derivative. The derivation of this method comes by replacing first derivative in the newton s method by its approximation (finite differentiation), i.e f (x k ) = f k f k 1 x k x k 1, where f k = f(x k ) x k+1 = x k x k x k 1 f k f k 1 f k Just like Newton s method the secant s method to find the minimum is given by: x k+1 = x k x k x k 1 f k f f k 1 k Convergence of secant method is super-linear with β = 1.618. The table below shows the secant method convergence for f = x with -0.6 and 0.75 as initial points 4

4 GOLDEN SECTION METHOD Number x k 1 x k f(x k 1 ) f(x k ) 1-0.6 0.75-0.441176 0.48 2 0.75 0.046552 0.48 0.046451 3 0.046552-0.028817 0.046451-0.028793 4-0.028817 0.000024-0.028793 0.000024 Figure 4: Secant s method converging with x 0 = 0.6; x 1 = 0.75 4 Golden Section Method The table below shows the Golden Section method convergence for f = initial points Number a b f(a) f(b) 0-0.6 0.75-0.08375 0.222146 1-0.6 0.234346-0.441176 0.222146 2-0.6-0.084346-0.441176-0.08375 3-0.6-0.281308-0.441176-0.26068..... 20-0.6-0.599911-0.441176-0.441146 x min = 0.599966 f(x min ) = 0.441165 x with -0.6 and 0.75 as 4.1 Another example for golden-section Consider the function: f(x 1, x 2 ) = (x 2 x 2 1) 2 + e x2 1 You are at the point (0,1).Find the minimum of the function in the direction (line) (1, 2) T using the Golden-Section line-search algorithm on the step-length interval [0, 1]. Stop when the length of the interval is less than 0.2. Note: step-length interval could be described by the parameter t, and, so, all the points along the direction (1, 2) T can be expressed as (0, 1) + t (1, 2). 5

4.1 Another example for golden-section 4 GOLDEN SECTION METHOD Figure 5: Golden Section method converging for f = x with a = 0.6; b = 0.75 The equation to find new point is defined as: x k+1 = x k + α k d k Here, d k = (1, 2) and x 0 = (0, 1) In the initial step length interval [0, 1], let α 1 = 0.618 and α 2 = 0.382. Thus, x 1 = (0, 1) T + 0.382(1, 2) T = (0.382, 1.764) T ; f(x 1 ) = 3.7753 x 2 = (0, 1) T + 0.618(1, 2) T = (0.618, 2.236) T ; f(x 2 ) = 4.9027 Since f(x 1 ) < f(x 2 ), the new step-length interval is [0, 0.618] α 3 = 0.618 0.382 = 0.2361 x 3 = (0, 1) T + 0.2361(1, 2) T = (0.2361, 1.4722) T ; f(x 3 ) = 3.0637 The new step-length interval is [0, 0.2361] Using the Golden Section search method, the new step length is α 4 = 0.2361 0.618 = 0.1459 x 4 = (0, 1) T + 0.1459(1, 2) T = (0.1459, 1.2918) T ; f(x 4 ) = 2.6357 The new step-length interval is [0, 0.1459] α 5 = 0.1459 0.618 = 0.0902 x 5 = (0, 1) T + 0.0902(1, 2) T = (0.0902, 1.1804) T ; f(x 5 ) = 2.3824 The new step-length interval is [0, 0.0902] α 6 = 0.0902 0.618 = 0.0557 x 6 = (0, 1) T + 0.0557(1, 2) T = (0.0557, 1.1114) T ; f(x 6 ) = 2.2314 6

5 REFERENCES Step-length Interval α i = α i 1 0.618 x 1 x 2 f(x 1, x 2 ) = (x 2 x 2 1) 2 + e x2 1 [0,.0557] 0.0344 0.0344 1.0688 2.1410 [0,.0344] 0.0213 0.0213 1.0426 2.0865 [0,.0213] 0.0132 0.0132 1.0264 2.0533 [0,.0132] 0.0081 0.0081 1.0162 2.0326 [0,.0081] 0.0050 0.0050 1.0100 2.0201 [0,.0050] 0.0031 0.0031 1.0062 2.0124 [0,.00031] 0.0002 0.0002 1.0004 2.0008 [0,.0002] 0.0001 0.0001 1.0002 2.0004 [0,.0001] 0.00006 0.00006 1.00012 2.0002 [0,.00006] 0.000034 0.00004 1.00008 2.0002 5 References 1. Jorge Nocedal, Stephen J.Wright: Numerical Optimization, Springer Series in Operations Research (1999). 2. E. de Klerk, C. Roos, and T. Terlaky, Nonlinear Optimization (pdf) (ps), 1999-2004, Delft. 3. http://de2de.synechism.org/c3/sec36.pdf 7