Optimal Starting Approximations. Newton's Method. By P. H. Sterbenz and C. T. Fike

Similar documents
MTH310 EXAM 2 REVIEW

Section Properties of Rational Expressions

Math 2: Algebra 2, Geometry and Statistics Ms. Sheppard-Brick Chapter 4 Test Review

Homework 8 Solutions to Selected Problems

Simplifying Rational Expressions and Functions

Mathematical Olympiad Training Polynomials

March Algebra 2 Question 1. March Algebra 2 Question 1

be any ring homomorphism and let s S be any element of S. Then there is a unique ring homomorphism

Mathematics Notes for Class 12 chapter 7. Integrals

Function: exactly Functions as equations:

where c R and the content of f is one. 1

8. Limit Laws. lim(f g)(x) = lim f(x) lim g(x), (x) = lim x a f(x) g lim x a g(x)

Chapter 6: Rational Expr., Eq., and Functions Lecture notes Math 1010

Aim: How do we prepare for AP Problems on limits, continuity and differentiability? f (x)

LIMITS AT INFINITY MR. VELAZQUEZ AP CALCULUS

Section IV.23. Factorizations of Polynomials over a Field

Linear DifferentiaL Equation

Numerical Methods in Informatics

Teacher: Mr. Chafayay. Name: Class & Block : Date: ID: A. 3 Which function is represented by the graph?

Math123 Lecture 1. Dr. Robert C. Busby. Lecturer: Office: Korman 266 Phone :

Chapter 8. Exploring Polynomial Functions. Jennifer Huss

1 Solutions to selected problems

PARTIAL FRACTIONS AND POLYNOMIAL LONG DIVISION. The basic aim of this note is to describe how to break rational functions into pieces.

Math 261 Exercise sheet 5

Computation of Eigenvalues of Singular Sturm-Liouville Systems

Numerical Methods in Physics and Astrophysics

50 Algebraic Extensions

5-8 Study Guide and Intervention

GROUP OF TRANSFORMATIONS*

Limits at Infinity. Horizontal Asymptotes. Definition (Limits at Infinity) Horizontal Asymptotes

Math 141: Lecture 11

Approximations for the Bessel and Struve Functions

MSM120 1M1 First year mathematics for civil engineers Revision notes 3

Numerical Methods in Physics and Astrophysics

MATH 431 PART 2: POLYNOMIAL RINGS AND FACTORIZATION

Minimum Polynomials of Linear Transformations

MATH 614 Dynamical Systems and Chaos Lecture 3: Classification of fixed points.

6.3 Partial Fractions

Polynomials. Chapter 4

Section III.6. Factorization in Polynomial Rings

Partial Fraction Decomposition Honors Precalculus Mr. Velazquez Rm. 254

CS 323: Numerical Analysis and Computing

Exam 1. (2x + 1) 2 9. lim. (rearranging) (x 1 implies x 1, thus x 1 0

Warm Up Lesson Presentation Lesson Quiz. Holt Algebra 2 2

Horizontal and Vertical Asymptotes from section 2.6

Algebra Review 2. 1 Fields. A field is an extension of the concept of a group.

For more information visit

x 2 + y 2 = 1. dx 2y and we conclude that, whichever function we chose, dx y2 = 2x + dy dx x2 + d dy dx = 2x = x y sinh(x) = ex e x 2

171S4.3 Polynomial Division; The Remainder and Factor Theorems. October 26, Polynomial Division; The Remainder and Factor Theorems

171S4.3 Polynomial Division; The Remainder and Factor Theorems. March 24, Polynomial Division; The Remainder and Factor Theorems

Math 547, Exam 2 Information.

Analyzing Power Series using Computer Algebra and Precalculus Techniques

Polynomial Rings. i=0. i=0. n+m. i=0. k=0

Review all the activities leading to Midterm 3. Review all the problems in the previous online homework sets (8+9+10).

BOUNDS ON THE EXPECTATION OF FUNCTIONS OF MARTINGALES AND SUMS OF POSITIVE RV'S IN TERMS OF NORMS OF SUMS OF INDEPENDENT RANDOM VARIABLES

Implementation of Theorems in D.S. Dummit s Solving Solvable Quintics(1991)

Advanced Mathematics Unit 2 Limits and Continuity

Advanced Mathematics Unit 2 Limits and Continuity

Chapter 4. Remember: F will always stand for a field.

A General Class of Optimal Polynomials Related to Electric Filters

Polynomial Rings. (Last Updated: December 8, 2017)

The function graphed below is continuous everywhere. The function graphed below is NOT continuous everywhere, it is discontinuous at x 2 and

ANALYTIC TRANSFORMATIONS OF EVERYWHERE DENSE POINT SETS*

A numerical method for solving the equations of stability of general slip surfaces

Chapter 3. Rings. The basic commutative rings in mathematics are the integers Z, the. Examples

Equations and Inequalities

MTH4100 Calculus I. Lecture notes for Week 4. Thomas Calculus, Sections 2.4 to 2.6. Rainer Klages

Final Exam A Name. 20 i C) Solve the equation by factoring. 4) x2 = x + 30 A) {-5, 6} B) {5, 6} C) {1, 30} D) {-5, -6} -9 ± i 3 14

6x 3 12x 2 7x 2 +16x 7x 2 +14x 2x 4

Lecture 9: Krylov Subspace Methods. 2 Derivation of the Conjugate Gradient Algorithm

PARTIAL FRACTIONS AND POLYNOMIAL LONG DIVISION. The basic aim of this note is to describe how to break rational functions into pieces.

Quadratic Model To Solve Equations Of Degree n In One Variable

+ 1 3 x2 2x x3 + 3x 2 + 0x x x2 2x + 3 4

2009 A-level Maths Tutor All Rights Reserved

Math /Foundations of Algebra/Fall 2017 Numbers at the Foundations: Real Numbers In calculus, the derivative of a function f(x) is defined

The most factored form is usually accomplished by common factoring the expression. But, any type of factoring may come into play.

Great Theoretical Ideas in Computer Science

ON DIVISION ALGEBRAS*

7.5 Partial Fractions and Integration

Midterm Review. Name: Class: Date: ID: A. Short Answer. 1. For each graph, write the equation of a radical function of the form y = a b(x h) + k.

Chapter 5B - Rational Functions

b n x n + b n 1 x n b 1 x + b 0

3 Polynomial and Rational Functions

PUTNAM TRAINING POLYNOMIALS. Exercises 1. Find a polynomial with integral coefficients whose zeros include

Solution of Nonlinear Equations: Graphical and Incremental Sea

Section 0.2 & 0.3 Worksheet. Types of Functions

Math 320: Real Analysis MWF 1pm, Campion Hall 302 Homework 7 Solutions Please write neatly, and in complete sentences when possible.

2.8 Endomorphisms. Rong-Jaye Chen ECC Department of Computer Science, National Chiao Tung University. Rong-Jaye Chen 2.

TIGHT BOUNDS FOR THE FIRST ORDER MARCUM Q-FUNCTION

15. Polynomial rings Definition-Lemma Let R be a ring and let x be an indeterminate.

Abstract Algebra: Chapters 16 and 17

MATH 409 Advanced Calculus I Lecture 11: More on continuous functions.

Math 115 Spring 11 Written Homework 10 Solutions

:i.( c -t.t) -?>x ( -\- ) - a.;-b 1 (o..- b )(a..+al,-+ b:r) x x x -3x 4-192x

Evaluating Limits Analytically. By Tuesday J. Johnson

Chapter 7 Polynomial Functions. Factoring Review. We will talk about 3 Types: ALWAYS FACTOR OUT FIRST! Ex 2: Factor x x + 64

CHAPTER 2 POLYNOMIALS KEY POINTS

12d. Regular Singular Points

Section 3.4. Second Order Nonhomogeneous. The corresponding homogeneous equation. is called the reduced equation of (N).

1 Solutions to selected problems

Transcription:

Optimal Starting Approximations Newton's Method for By P. H. Sterbenz and C. T. Fike Abstract. Various writers have dealt with the subject of optimal starting approximations for square-root calculation by Newton's method. Three optimality criteria that have been used can be shown to lead to closely related approximations. This fact makes it surprisingly easy to choose a starting approximation of some prescribed form so that the maximum relative error after any number of Newton iterations is as small as possible. 1. Introduction. The choice of polynomial and rational starting approximations for square-root calculation by Newton's method has been the subject of various investigations (e.g., [l]-[5]). These approach the problem from several different points of view. In this paper we will show how approximations obtained from these different viewpoints are related and how some of them can be derived from others. The problem of evaluating for any x > 0 is easily reduced to the problem of evaluating for x in some closed interval [a, b] such that 0 < a < b. Here [a, b] depends on the radix of the floating-point number system of the computer to be used; typical possibilities are [1/16, 1] and [\, 2]. The following procedure is used to compute an approximate value for. Using a polynomial or rational approximation f(x) to, valid in [a, b], compute a starting value y o = f(x) and then obtain y h 2/2) ' ' '> V«by means of the relation Vk+i = h(yt + x/yk), k = 0,1, -, n 1. Then y ~. It is customary not to test for convergence, since the number of iterations required in practice is quite small. Instead, the number of iterations to is chosen so that yn is a sufficiently accurate approximation to for all x in [a, b]. Let the relative error in the fcth Newton iterate be denoted by Rk(x) ; that is, let Rk(x) = ^^. Then Ro(x) is the relative error of the starting approximation ; that is, It is well known that Ro(x) = f(x) - Rk(x) (x) Rk+lix) = 2[l+Rk(x)]- The better the approximation f(x) is, the fewer iterations will be required. Thus we wish to select the coefficients of the approximation f(x) in order to obtain a best fit Received October 2, 1967, revised September 18, 1968. 313

314 P. H. STERBENZ AND C. T. FIKE in some sense. We will look for a best-fit approximation in a set V of admissible functions. Here either V is the set of all polynomials of degree ^ M or else V is the set of all rational functions p(x)/q(x) where p(x) and q(x) are relatively prime polynomials of degree ^ M and N respectively and q(x) does not vanish for a ^ x b. Let W be the set of all f(x) in V for which f(x) > 0 for a g x ^ b. Now/(a.) = Va is in TF, and we find that for this function R0(x) \ < lfora ^ x ^ b. For any f(x) in F but not in W we would have Äo(a;) =ï 1. Therefore, in searching for a best-fit approximation to we restrict our attention to W. 2. Optimality Criteria. In trying to select the best approximation from W, three criteria have been used. First, we can try to minimize the maximum of /2o(a;)[ for a á x è b. This leads to the best-fit approximation to in the relative-error sense for the interval [a, b], a type of approximation treated extensively in the literature (e.g., [2] and [4]). Another criterion, for which it may be easier to find the optimal approximation by analytical methods, is to minimize the maximum of [5(x), where (2) 5(a0 = log^. Here (3) a(a;)=log[l+äo(ao]. Several approximations to that are optimal in this sense also appear in the literature (e.g., [1] and [3]). A third approach is to note that we are going to use a fixed number of Newton iterations and always going to take yn as an approximation to. Therefore, we can try to minimize the maximum of Ä (a;) for the last iteration. It is easy to show (see [5]) that the f(x) that minimizes the maximum of Äi(a;) also minimizes the maximum of Ä (a;) for every to ^ 1. The three alternatives are thus to select the f(x) in W that minimizes the absolute value of either Ro(x), R\(x), or h(x). Minimizing Äi(a;) (and consequently also Ä (a;) ) would appear to be what we would like to do in practice, but this would also appear to be the most difficult of the three alternatives to deal with analytically. Thus Moursund [5] in his results concerning the existence of an f(x) optimal by this criterion resorted to a generalization of the classical Chebyshev approximation theory. Fortunately, as we will show below, there are surprisingly simple relationships among the three optimization criteria which make it possible to avoid such difficulties. 3. Relationships Among Optimization Criteria. It follows from (3) that Ro(x) = e5(x) 1. Therefore, we have which simplifies to R1(x) = [es(x) - l]2/2ehx), (4) Rt(x) = cosh 5(a;) - 1. Since cosh x is an even function and is monotone for all x > 0, this yields max Äi(a:) = cosh (max 5(a;) ) la,6] \[a,i>] /

OPTIMAL STARTING APPROXIMATIONS 315 Therefore, if we minimize the maximum of \ô(x)\, we have also minimized the maximum of ßi(x). Thus, we have Theorem 1. * A function f(x) in W min imizes the maximum of \8(x)\ if and only if it minimizes the maximum of [72i(a;). We now turn to the relationship between 8(x) and R0(x). We note that for any a and any f(x) in W, (5) af(x)v- Vx = aeh* - 1, and, if a > 0, (6) log'5^ = log [a(l+ «.(*))]. Lemma 1. For f(x) in W, let the minimum and maximum of ô(x) for a ^ x ^ b be denoted by Xi and X2 respectively. Let X be the larger of Xi, IX2I ; and let Then u = tanh X. Proof. From (5), ß = max [a, b] f(x) /cosh X /(a:)/cosh X _ 1 ä(l) _ cosh X so the minimum and maximum of this expression are ß\ and ßi respectively, where Now either X = X2 or X = Xi, so Since 1 \i, ßi = 7-T-e 1. coshx ß = max la.t] e -1 1 cosh X e± /cosh X 1 = ±tanh X, the lemma follows. Lemma 2. For f(x) in W let the minimum and maximum cf Ro(x) for a ^ x ^ b be denoted by ß\ and ßi. respectively. Let ß be the larger of \m\, \ßi\ ; and suppose that M < 1. Let g(x) = log /1 m_ (1 ß 2x1/2 ) ' / ' and let the minimum and maximum of g(x) for a ^ x ^ b be denoted by Xi and X2 respectively. Let X be the larger of Xi, X2. Then X = arctanh u. Moreover, if ßi = / 2, then Xi = X-2- Proof. From (6), * We are indebted to the referee for pointing out that this result was also discovered by King and Phillips and will appear in [6].

316 P. H. STERBENZ AND C. T. FIKE and therefore, v. 1 4- Ro(x)»(*) = log7,-2tt72- (1 - M ) Then (7) X; = log Xi è log 1 + Mi 2\l/2 (1 - M ) (l-^)1'2 (8) X2 ^ log 1 + M (1 - M2)1/2 " Now either ß = mi or u = U2, so equality holds in at least one of (7) or (8). Since 1 ± ß arctanh m, a - m r we have X = arctanh ß. If u2 = Mi, then equality holds in both (7) and (8), and Xi = X2. Now let/*(a:) be the best-fit approximation to in TF in the sense of relative error; that is, let f*(x) be that function in IF which minimizes max^,-,] Ä0(a;). From the theory of best-fit approximations (see, e.g., [7]), we know that/*(a;) exists, is unique, and is characterized by the fact that it yields an equal-ripple Ro(x) with a sufficient number of extreme points. Let log 772 = ±lo8 fer-± max [o,i>] /*(») - Vx Since, as we have observed, there is an f(x) in IF for which \Ro(x) < 1 for a ^ x ^ b, it follows that m* < 1- Now let and By Lemma 2, (9) Now, by Lemma 1, max [a, 6] 7(x) = X = max [a, b] f*(x) /. 2s 1/2 (1 - M* ) log 7(x) X = arctanh m* /(a:)/cosh X tanh X = m* If f(x) is any function in W, and if X = max[a,4] 5(a;), then by Lemma 1 max [a.b] /(a:)/cosh X tanh X.

But /(a;)/cosh X is in IF, and therefore OPTIMAL starting approximations 317 (10) tanh X ^ m* Then (11) X ^ X. Therefore f(x) minimizes the maximum of 5(a;). To prove the uniqueness of f(x), we note that if f(x) is any function in IF with X = X, then by the uniqueness of f*(x) we have/(.r)/cosh X = f*(x) and f*(x) f(x) = f*(x) cosh arctanh m* = -~!TTr> (1 M* ) Thus J(x) is the unique polynomial in IF which minimizes the maximum of 5(a;). Moreover, since f*(x) yields an equal-ripple R0(x) and since x yields an extreme value for f*(x) if and only if it yields an extreme value for f(x), it follows from Lemma 2 that f(x) yields an equal-ripple 5(x) with the same number of extreme points that Ro(x) has. The fundamental relationship between f(x) and f*(x) is given by (12) J(x) = f*(x)/(\ -m*2)1/2, (13) f*(x) = 7(x)/cosh X. Thus we have proved Theorem 2. There is a unique function f(x) in W which minimizes the maximum of \8(x)\ for a ^ x g b. It is characterized by the fact that it yields an equal-ripple o(x), and it is related to the function f*(x) in W which minimizes the maximum of \Ro(x)\ by (12) and (13). 4. Conclusion. The significance of the above results can be illustrated in the following way. Consider the problem of approximating in the interval [1/16, 1] by means of a rational function expressible in the form A + B/(x + C). The coefficients of the rational function/i(a;) that makes the maximum of \ô(x) a minimum can easily be obtained with the aid of formulas given by Maehly (posthumously in the appendix of [3]). (It may be noted that Maehly showed by analytical methods how to get exact values for the coefficients, not just approximate numerical values.) Numerical values for the coefficients of the rational function fi(x) that makes the maximum of ño (a:) a minimum were given by Fike [4]; these results were obtained with Remez' method. Numerical values for the coefficients of the rational function fi(x) that makes the maximum of Äi(a;) a minimum were also given by Fike [8]; these results were obtained by an ad hoc method similar to Remez' method. It is now clear from the results stated in this paper that/i(a;) and/s(a;) are the same and that fi(x) is merely a constant multiple of them. IBM Systems Research Institute New York, New York 10017

318 P. H. STERBENZ AND C. T. FIKE 1. H. J. Maehly, Approximations for the CDC 1604, Control Data Corp., 1960. 2. J. Eve, "Starting approximations for the iterative calculation of square roots," Comput. J., v. 6, 1963, pp. 274-276. 3. W. J. Cody, "Double-precision square root for the CDC-3600," Comm. ACM, v. 7, 1964, pp. 715-718. 4. C. T. Fike, "Starting approximations for square-root calculation on IBM System/360," Comm. ACM, v. 9, 1966, pp. 297r299. 5. D. G. Mouesund, "Optimal starting values for Newton-Raphson calculation of vx," Comm. ACM, v. 10, 1967, pp. 430-432. 6. R. F. King & D. L. Phillips, "The logarithmic error and Newton's method for the square root," Comm. ACM, v. 12, 1969, pp. 87-88. 7. N. I. Achiesek, Theory of Approximation, OGIZ, Moscow, 1947; English transi., Ungar, New York, 1956. MR 10, 33; MR 20 #1872. 8. C. T. Fike, "Letter to the editor," Comm. ACM, v. 10, 1967, pp. 683-684.