A BASIC FAMILY OF ITERATION FUNCTIONS FOR POLYNOMIAL ROOT FINDING AND ITS CHARACTERIZATIONS. Bahman Kalantari 3. Department of Computer Science

Similar documents
Linear Algebra (part 1) : Vector Spaces (by Evan Dummit, 2017, v. 1.07) 1.1 The Formal Denition of a Vector Space

ON THE ARITHMETIC-GEOMETRIC MEAN INEQUALITY AND ITS RELATIONSHIP TO LINEAR PROGRAMMING, BAHMAN KALANTARI

LECTURE NOTES IN CRYPTOGRAPHY

4.1 Eigenvalues, Eigenvectors, and The Characteristic Polynomial

Course 311: Michaelmas Term 2005 Part III: Topics in Commutative Algebra

Vector Space Basics. 1 Abstract Vector Spaces. 1. (commutativity of vector addition) u + v = v + u. 2. (associativity of vector addition)

Institute for Advanced Computer Studies. Department of Computer Science. On the Convergence of. Multipoint Iterations. G. W. Stewart y.

EIGENVALUES AND EIGENVECTORS 3

Linear Algebra, 4th day, Thursday 7/1/04 REU Info:

Chapter 2: Linear Independence and Bases

COMPLETELY INVARIANT JULIA SETS OF POLYNOMIAL SEMIGROUPS

SOLUTION TO RUBEL S QUESTION ABOUT DIFFERENTIALLY ALGEBRAIC DEPENDENCE ON INITIAL VALUES GUY KATRIEL PREPRINT NO /4

Lecture 16: 9.2 Geometry of Linear Operators

PROOF OF TWO MATRIX THEOREMS VIA TRIANGULAR FACTORIZATIONS ROY MATHIAS

How many units can a commutative ring have?

TOPOLOGICAL GROUPS MATH 519

Problem Set 9 Due: In class Tuesday, Nov. 27 Late papers will be accepted until 12:00 on Thursday (at the beginning of class).

Contents. 2.1 Vectors in R n. Linear Algebra (part 2) : Vector Spaces (by Evan Dummit, 2017, v. 2.50) 2 Vector Spaces

1 Introduction It will be convenient to use the inx operators a b and a b to stand for maximum (least upper bound) and minimum (greatest lower bound)

VII Selected Topics. 28 Matrix Operations

Chapter 8. P-adic numbers. 8.1 Absolute values

460 HOLGER DETTE AND WILLIAM J STUDDEN order to examine how a given design behaves in the model g` with respect to the D-optimality criterion one uses

STABILITY OF INVARIANT SUBSPACES OF COMMUTING MATRICES We obtain some further results for pairs of commuting matrices. We show that a pair of commutin

Contents. 4 Arithmetic and Unique Factorization in Integral Domains. 4.1 Euclidean Domains and Principal Ideal Domains

Spurious Chaotic Solutions of Dierential. Equations. Sigitas Keras. September Department of Applied Mathematics and Theoretical Physics

Linear Systems and Matrices

Continuity. Chapter 4

Homework #2 Solutions Due: September 5, for all n N n 3 = n2 (n + 1) 2 4

ON SUM OF SQUARES DECOMPOSITION FOR A BIQUADRATIC MATRIX FUNCTION

Richard DiSalvo. Dr. Elmer. Mathematical Foundations of Economics. Fall/Spring,

arxiv: v2 [math.gm] 3 Jan 2018

Garrett: `Bernstein's analytic continuation of complex powers' 2 Let f be a polynomial in x 1 ; : : : ; x n with real coecients. For complex s, let f

Fraction-free Row Reduction of Matrices of Skew Polynomials

SOME NEW EXAMPLES OF INFINITE IMAGE PARTITION REGULAR MATRICES

Novel Approach to Analysis of Nonlinear Recursions. 1 Department of Physics, Bar-Ilan University, Ramat-Gan, ISRAEL

MATH 423 Linear Algebra II Lecture 3: Subspaces of vector spaces. Review of complex numbers. Vector space over a field.

(Inv) Computing Invariant Factors Math 683L (Summer 2003)

LINEAR EQUATIONS WITH UNKNOWNS FROM A MULTIPLICATIVE GROUP IN A FUNCTION FIELD. To Professor Wolfgang Schmidt on his 75th birthday

Factorization in Integral Domains II

SOLUTION OF GENERALIZED LINEAR VECTOR EQUATIONS IN IDEMPOTENT ALGEBRA

5 Eigenvalues and Diagonalization

Automatica, 33(9): , September 1997.

THEODORE VORONOV DIFFERENTIABLE MANIFOLDS. Fall Last updated: November 26, (Under construction.)

On Böttcher s mysterious identity

Generalized eigenvector - Wikipedia, the free encyclopedia

Vector Spaces ปร ภ ม เวกเตอร

MATH 51H Section 4. October 16, Recall what it means for a function between metric spaces to be continuous:

Coins with arbitrary weights. Abstract. Given a set of m coins out of a collection of coins of k unknown distinct weights, we wish to

q-alg/ Mar 96

Continuity. Chapter 4

ADDENDUM B: CONSTRUCTION OF R AND THE COMPLETION OF A METRIC SPACE

6 Lecture 6: More constructions with Huber rings

Roots of Unity, Cyclotomic Polynomials and Applications

Factorization in Polynomial Rings

THE FUNDAMENTAL THEOREM OF ALGEBRA VIA PROPER MAPS

2. Intersection Multiplicities

A NOTE ON RATIONAL OPERATOR MONOTONE FUNCTIONS. Masaru Nagisa. Received May 19, 2014 ; revised April 10, (Ax, x) 0 for all x C n.

RATIONAL POINTS ON CURVES. Contents

6.1 Matrices. Definition: A Matrix A is a rectangular array of the form. A 11 A 12 A 1n A 21. A 2n. A m1 A m2 A mn A 22.

LECTURE NOTES ELEMENTARY NUMERICAL METHODS. Eusebius Doedel

be any ring homomorphism and let s S be any element of S. Then there is a unique ring homomorphism

MATH 326: RINGS AND MODULES STEFAN GILLE

2 Two-Point Boundary Value Problems

Detailed Proof of The PerronFrobenius Theorem

Determinants Chapter 3 of Lay

BP -HOMOLOGY AND AN IMPLICATION FOR SYMMETRIC POLYNOMIALS. 1. Introduction and results

GEOMETRIC CONSTRUCTIONS AND ALGEBRAIC FIELD EXTENSIONS

2 MICHAEL FILASETA rather than y n (x). The polynomials z n (x) are monic polynomials with integer coecients, and y n (x) is irreducible if and only i

Definition 5.1. A vector field v on a manifold M is map M T M such that for all x M, v(x) T x M.

1 Matrices and Systems of Linear Equations

where c R and the content of f is one. 1

Lifting to non-integral idempotents

MINKOWSKI THEORY AND THE CLASS NUMBER

Linear Algebra (part 1) : Matrices and Systems of Linear Equations (by Evan Dummit, 2016, v. 2.02)

290 J.M. Carnicer, J.M. Pe~na basis (u 1 ; : : : ; u n ) consisting of minimally supported elements, yet also has a basis (v 1 ; : : : ; v n ) which f

Math Camp Notes: Linear Algebra I

The WENO Method for Non-Equidistant Meshes

ELEMENTARY LINEAR ALGEBRA

Notes on the Matrix-Tree theorem and Cayley s tree enumerator

METRIC HEIGHTS ON AN ABELIAN GROUP

10. Smooth Varieties. 82 Andreas Gathmann

+ 1 3 x2 2x x3 + 3x 2 + 0x x x2 2x + 3 4

On the decay of elements of inverse triangular Toeplitz matrix

1: Introduction to Lattices

1 Determinants. 1.1 Determinant

Vector Spaces ปร ภ ม เวกเตอร

17 Galois Fields Introduction Primitive Elements Roots of Polynomials... 8

MADHAVA MATHEMATICS COMPETITION, December 2015 Solutions and Scheme of Marking

Chapter 2. Real Numbers. 1. Rational Numbers

Linear Algebra: Linear Systems and Matrices - Quadratic Forms and Deniteness - Eigenvalues and Markov Chains

QUALITATIVE CONTROLLABILITY AND UNCONTROLLABILITY BY A SINGLE ENTRY

PART I Lecture Notes on Numerical Solution of Root Finding Problems MATH 435

Determinants An Introduction

2 EBERHARD BECKER ET AL. has a real root. Thus our problem can be reduced to the problem of deciding whether or not a polynomial in one more variable

Max-Min Problems in R n Matrix

A matrix over a field F is a rectangular array of elements from F. The symbol

0 Sets and Induction. Sets

MATH 117 LECTURE NOTES

Factoring Polynomials with Rational Coecients. Kenneth Giuliani

Properties of Arithmetical Functions

Transcription:

A BASIC FAMILY OF ITERATION FUNCTIONS FOR POLYNOMIAL ROOT FINDING AND ITS CHARACTERIZATIONS Bahman Kalantari 3 Department of Computer Science Rutgers University, New Brunswick, NJ 08903 Iraj Kalantari Department of Mathematics Western Illinois University, Macomb, IL, 6455 Rahim Zaare-Nahandi y Department of Mathematics University of Tehran, Tehran, Iran 3 This research is supported in part by the National Science Foundation under Grant No. CCR-920837 y This research is supported in part by a grant from University of Tehran.

Abstract Let p(x) be a polynomial of degree n 2 with coecients in a subeld K of the complex numbers. For each natural number m 2, let L m (x) be the m 2 m lower triangular matrix whose diagonal entries are p(x) and for each j = ; : : : ; m 0, its j-th subdiagonal entries are p (j) (x)=j!. For i = ; 2, let L (i) m (x) be the matrix obtained from L m (x) by deleting its rst i rows and its last i columns. L () (x). Then, the function B m (x) = x 0 p(x) det(l () m0 (x))=det(l() m (x)) is a member of S(m; m + n 0 2), where for any M m, S(m; M) is the set of all rational iteration functions such that for all roots of p(x), g(x) = + P M i=m i (x)( 0 x) i, with i (x)'s also rational and well-dened at. Given g 2 S(m; M), and a simple root of p(x), g (i) () = 0, i = ; : : :; m 0, and m () = (0) m g (m) ()=m!. For B m (x) we obtain m () = (0) m det(l (2) m+ ())=det(l() m ()). For m = 2 and 3, B m (x) coincides with Newton's and Halley's, respectively. If all roots of p(x) are simple, B m (x) is the unique member of S(m; m + n 0 2). By making use of the identity 0 = P n i=0 [p (i) (x)=i!]( 0 x) i, we arrive at two recursive formulas for constructing iteration functions within the S(m; M) family. In particular the B m 's can be generated using one of these formulas. Moreover, the other formula gives a simple scheme for constructing a family of iteration functions credited to Euler as well as Schroder, whose m-th order member belong to S(m; mn), m > 2. The iteration functions within S(m; M) can be extended to arbitrary smooth functions f, with the automatic replacement of p (j) with f (j) in g as well as m ().

. INTRODUCTION. Let p(x) be a polynomial of degree n 2 with coecients in a subeld K of the complex numbers. For each natural number m 2, let L m (x) be the m 2 m lower triangular matrix whose diagonal entries are p(x) and for each j = ; : : : ; m 0, its j-th subdiagonal entries are p (j) (x)=j!, i.e. L m (x) = 0 B @ p(x) 0 0 : : : 0 p 0 (x) p(x) 0 : : : 0 p 00 (x) p 0 (x) p(x) : : : 0 2!....... p (m0) (x) (m0)! p (m02) (x) (m02)! p (m03) (x) (m03)! : : : p(x) C A : (:) For i = ; 2, let L (i) m (x) be the (m0i)2(m0i) matrix obtained from L m(x) by deleting its rst i rows, and its last i columns, i.e. L () m = 0 B @ p 0 (x) p(x) 0 : : : 0 p 00 (x) p 0 (x) p(x) : : : 0 2!....... p (m0) (x) (m0)! p (m02) (x) (m02)! with L () (x). The function p (m03) (x) (m03)! : : : p 0 (x) C A ; L (2) m = 0 B @ p 00 (x) 2! p 0 (x) : : : 0 p 000 (x) 3!. p (m0) (x) (m0)! p 00 (x) : : : 0 2!..... p (m02) (x) (m02)! : : : B m (x) = x 0 p(x) det(l() m0(x)) det(l () m (x)) ; (:2) denes a xed-point iteration function which has local m-th order convergence for simple roots of p(x) (see Remark. below). Here det is the determinant function. We shall refer to the set of all B m (x)'s, m = 2; 3; : : :, as the Basic F amily of xed-point iteration functions for polynomial root nding. Specic members of the family are B 2 (x) = x 0 p(x) p 0 (x) ; which is the well-known Newton's function, and B 3 (x) = x 0 which is Halley's function. For m = 4, one obtains B 4 (x) = x 0 2p 0 (x)p(x) 2(p 0 (x)) 2 0 p 00 (x)p(x) ; 6p 0 (x) 2 p(x) 0 3p 00 (x)p(x) 2 p 000 (x)p(x) 2 + 6p 0 (x) 3 0 6p 00 (x)p 0 (x)p(x) : p 00 (x) 2! C A ;

The function B 2 (x) is derived in almost all the numerical analysis textbooks with the usual geometric interpretations (see e.g. Atkinson [2]). The Function B 3 (x), credited to the astronomer Halley [2], is perhaps the best-known third-order method, and has been rediscovered and/or derived through various interesting means, see e.g. Bateman [3], Wall [25], Bodewig [4], Hamilton [3], Stewart [22], Frame [7, 8, 9], Traub [23], Hansen and Patrick [4], Popovski [7], Gander [0]. For the interesting history of Halley's function see the recent paper of Scavo and Thoo [8]. p(x)= Halley's function can also be obtained by applying Newton's method to the function q p 0 (x), see Bateman [3], as well as Brown [5], and Alefeld []. Most recently, this was rediscovered by Gerlach [] who also generalized it to get high order methods : Let F (x) = p(x), and dene inductively for i = 2; 3; : : :, F i (x) = F i0 (x)=[f i0 (x)] =i. Then for m = 2; 3; : : :, the function G m (x) = x 0 F m0 (x)=f 0 m0 (x), under some very restrictive conditions, will have m-th order convergence rate. Interestingly, G 4 (x) coincides with B 4 (x) and possibly they are all identical. Gerlach however oers no closed formula for G m (x). Moreover, the domain of G m 's is very restrictive. Indeed the computation of G m (x) is too cumbersome and Gerlach only obtains G 4 (x) for p(x) = x 3 0. In [5], the closed formula for B m (x) corresponding to this cubic polynomial as well as to p(x) = x 2 0 was obtained. Indeed for general functions, B 4 (x) had been obtained previously and through dierent schemes by Snyder [2], Traub [23](page 9), and others. Interestingly, Hamilton [3], made use of determinants in deriving Halley's function. In fact Hamilton does make use of the determinant of the matrix L () m (x), to arrive at high order methods, but he does not obtain B m (x) for any m 4. For instance he obtains a forth-order method which is dierent than B 4 (x) (see page 52, [3]). The algebraic characterization and the existence of the Basic Family was proved in [6]. The present paper gives further characterizations, closed formula for its members, their corresponding asymptotic constant, and thus oers a deeper understanding of each of its individual members, including Newton's and Halley's functions. Denition.. For each pair of nonnegative integers m and M, M m, let S(m; M) 2

be the set of all function g(x) 2 K(x), i.e. a rational function with coecient in K, so that for all roots of p(x) we have g(x) = + M i=m i (x)( 0 x) i ; (:3) where i (x) 2 K(x), for i = m; : : : ; M, and well-dened for any simple root, i.e. lim x! i (x) i () exists. Moreover, m (x) and M (x) are not identically zero. Remark.. Given g 2 S(m; M), from the continuity of i (x)'s at a simple root, and since m 2 it follows that there exists a neighborhood of for which jg(x) 0 j Cjx 0 j m, for some constant C. This implies the existence of a neighborhood I of for which for any x 0 2 I the xed-point iteration converges to, and clearly x k+ = g(x k ); k 0; (:4) ( 0 x k+ ) lim k! ( 0 x k ) = m 0 m(); (:5) i.e. the rate of convergence of the sequence fx k g k=0 is of order m. Also note that if x 0 is in K, then so are all the iterates in (.4). Denoting m (x) corresponding to B m (x) by (m) m (x), we show that for a simple root, the asymptotic constant of convergence for B m (x) is (m) m () = (0) m det(l(2) m+()) det(l () m ()) = In particular for m = 2 and m = 3 we get (0)m p 0 det(l(2) m0 () m+()): (:6) (2) 2 () = p00 () 2p 0 () ; (3) 3 () = 0 2p0 ()p 000 () 0 3p 00 () 2 2p 0 () 2 : We show that if all roots of p are simple, B m (x) is the unique member of S(m; m + n 02). The following denition provides means by which members from two dierent S(m; M)'s can be compared. 3

Denition.2. Given g 2 S(m; M), we dene the order of g to be m. The coef f ient vector of g is the vector 0(x) = ( m (x); : : : ; M (x)). The leading coefficient of g is m. The width of g is M 0m+. We dene the depth of g to be d if the formula for g depends on p (j), j = 0; : : : ; d. Similarly we dene the depth of the leading coecient of g. The simple-root-depth of g is dened to be, if for any simple root of p(x), m () depends on p (j) (), j = 0; : : : ;. Remark.2. If g and h are in S(m; M), so is any ane combination, g + h, where + =. As, B m (x) will be shown to belong to S(m; n + m 0 2), it follows that for any m 2, and M m + n 0 2, the set S(m; M) is nonempty. If p(x) has simple roots, as B m (x) will be shown to be the unique element of S(m; m+n02), it follows that S(m; M) is empty if M < n + m 0 2. Thus, B m (x) is also minimal with respect to width. Traub [23], has shown that any m-th order method must have depth at least equal to m 0. Thus, in the sense of depth too, B m (x) is minimal. Remark.3. Given g 2 S(m; M), from Taylor's Theorem and the continuity of i 's at a simple root, it is easy to conclude that that g (i) () = 0, for all i = ; : : : ; m 0, and that m () = (0) m g(m) () : (:7) m! It should be noted that as functions m (x) and (0) m g (m) (x)=m! are not identical. For example take g(x) = x 0 p(x)=p 0 (x). In other words g (m) (x) and m (x) may have dierent depths, but they have the same simple-root-depth. Denition.3. Let S (m; M) be the set of all g(x) 2 K(x) such that p(y) divides M i (x)(y 0 x) i ; (:8) 0 (y) = 0g(x) + y + in K(x)[y], i.e. polynomials in y with coecients in K(x), such that i (x) is well-dened for any simple root of p(x), and neither m (x) nor M (x) are identically zero. i=m Remark.4. If is a root of p, then 0 () = 0. Thus, it follows that S (m; M) is a subset of S(m; M). Also, if all the roots of p(x) are simple (in particular if it is 4

irreducible), then S (m; M) = S(m; M). To see this pick g 2 S(m; M), and dene a corresponding 0 (y). Then there exists q(y); r(y) 2 K(x)[y] such that 0 (y) = p(y)q(y) + r(y); (:9) where the degree of r(y) is less than n. Since 0 () = 0, for all roots of p, it follows that r(y) 0 (see e.g. [24]). Note that in particular we must have g(x) = x 0 0 (x): (:0) In general however, S(m; M) and S (m; M) are not the same. For example consider p(x) = x 2. Then g (x) = x=2 and g 2 (x) = x 2 both lie in S(2; 2), but g 2 (x) is not in S (2; 2). Denition.4. Let S(m; M) be the set of all g(x) which can be represented as (:3) but with the relaxation of the condition that i ()'s are well-dened. We will give two formulas for generating new members of S(m; M) with higher orders. Both formulas make use of the following relationship which is a consequence of Taylor's Theorem, but can also be derived using purely algebraic means (see [6]) 0 = n i=0 p (i) (x) ( 0 x) i : (:) i! From (:) we have Writing 0p(x) = n i= p (i) (x) ( 0 x) i : (:2) i! x = 0 ( 0 x); (:3) and adding corresponding sides to (.2) we get B (x) x 0 p(x) = + (p 0 (x) 0 )( 0 x) + n i=2 p (i) (x) ( 0 x) i 2 S(; n): (:4) i! From (.) we also immediately get Newton's function as a member of S(2; n) B 2 (x) = x 0 p(x) p 0 (x) = + n 5 i=2 p (i) (x) i!p 0 (x) ( 0 x)i : (:5)

As we shall see from B (x) and B 2 (x) together with the application of the two formulas one can construct a large class of iterative solutions which include the Basic Family, and the Euler-Schroder family. Finally, we show that the iteration functions within S(m; M) can be extended to arbitrary smooth functions f, with the automatic replacement of p (j) with f (j) in g, and the asymptotic constant of convergence m (). In Section 2, we give an algebraic proof of the existence of B m (x). The result will be used in Section 3 for deriving a closed formula for B m (x), and in Section 4, which describes the two formulas for generation of new iteration functions. In Section 5, we consider the extension of iteration function within S(m; M) to arbitrary smooth functions. 2. ALGEBRAIC PROOF OF EISTENCE OF THE BASIC FAMILY In this section we prove algebraically the existence of an iteration function in S(m; m+ n 0 2) as well as its uniqueness under the assumption that p(x) has simple roots. We shall only be concerned with the existence of B m (x) as opposed to its closed form. The theorem of this section will be used in the subsequent sections in deriving the closed form, as well as proving the equivalence of S(m; m + n 0 2), S (m; m + n 0 2), and S(m; m + n 0 2), given that p(x) has simple roots. The proof of the theorem also motivates the denition of S (m; m + n 0 2) which in turn gives rise to the closed formula for B m (x). Moreover the theorem will be used to prove that the Basic Family can also be obtained recursively using one the formulas derived in Section 4. Theorem 2.. For any natural number m 2, S(m; m + n 0 2) is nonempty, and has a unique element if p(x) has simple roots. To prove Theorem 2. we rst need an auxiliary lemma. Given two natural numbers s and t let s t! s(s 0 ) : : : (s 0 t + ) ; (2:) t! the binomial coecient for s t, and zero, for s < t. For a given pair of natural numbers 6

m and r dene the following r 2 r matrix U m;r = 0 B @ m m 2. m r m+ m+ 2. m+ r : : : m+r0 m+r0 : : : 2.... : : : m+r0 r C A : (2:2) Lemma 2.. For any natural numbers m and r, U m;r is invertible (in fact det(u m;r ) = ). m+r0 r Proof. To prove the lemma we will show that by performing elementary operations on U m;r, we can reduce it to the following matrix W m;r = 0 B @ m (m + ) : : : (m + r 0 ) m 2 (m + ) 2 : : : (m + r 0 ) 2...... m r (m + ) r : : : (m + r 0 ) r Since W m;r = V T D, where V is an invertible Vandermonde matrix, and D = diag(m; m+ ; : : : ; m + r 0 ), it follows that W m;r is invertible. To obtain this reduction, we rst multiply the j-th row of U m;r by j!. Let U denote the new matrix. The rst column of U has entries which can be written as the following polynomials in m: f (m) = m, f 2 (m) = m 2 0m; : : :, f r (m) = m r + r0 m r0 + r02 m r02 +: : :+ m, for some coecients i, i = ; : : : ; r 0. The second column of U can be written as the same polynomials evaluated at (m + ), and so on. Thus, U is the r 2 r matrix whose i-th row vector is given by U = 0 B @ f (m) f (m + ) : : : f (m + r 0 ) f 2 (m) f 2 (m + ) : : : f 2 (m + r 0 )...... f r (m) f r (m + ) : : : f r (m + r 0 ) By adding scalar multiples of the rst row of U to other rows, we obtain a new matrix U 2 whose rst and second rows are the corresponding rows of W m;r, and whose remaining rows are polynomials free of the linear terms. Next, by adding scalar multiples of the second row of U 2 to its i-th rows, i = 3; 4; : : : ; r, we obtain U 3 whose rst three rows are those of W m;r and whose remaining rows are polynomials free of the linear and quadratic 7 C A : C A :

terms. Clearly, repeating this process we arrive at W m;r. 2 Proof of Theorem 2.. Let be any root of p(x). Recalling (:3) and replacing k by (0) k k for convenience, we will prove the existence of rational functions g(x) and k (x), k = m; : : : ; (m + n 0 2) with coecient over K so that 0g(x) + + m+n02 k=m Using that p() = 0, for any k n we can write k = n0 l=0 k (x)(x 0 ) k = 0: (2:3) l l ; (2:4) where l 2 K. From (2:4) it is easy to see that for each j = 0; : : : ; (n 0 2), (x 0 ) m+j = n0 i=0 i P m+j i (x); (2:5) where P m+j i (x)'s are polynomials with coecients in K. Furthermore, for i m + j, and for i > m + j, P m+j i P m+j i (x) = m + j i Using (2.5), we can rewrite (2.3) as where h i (x) =! x m+j0i + lower order terms; (2:6) (x) 0, which from (2.) can be written as P m+j i (x) = 0 = m + j i 0g(x) + h (x) = + m+n02 k=m n0 i=0 m+n02 k=m! x m+j0i : (2:7) i h i (x) = 0; (2:8) k (x)p k (x); (2:9) k (x)p k i (x); i = 0; i = 2; : : : ; n: (2:0) Thus, S(m; m+n02) is nonempty if the system of equations h i (x) 0, i = ; : : : ; (n0), is solvable, in which case g(x) = h 0 (x) = m+n02 k=m k (x)p k 0 (x) 2 S(m; m + n 0 2): (2:) 8

Equivalently, S(m; m + n 0 2) is nonempty if the system of nonlinear equations is solvable where 5 m (x)0 m (x) = e; (2:2) 0 m (x) = [ m (x); m+ (x); : : : ; m+n02 (x)] T ; e = [0; 0; : : : ; 0] T ; (2:3) and where 5 m (x) is the following (n 0 ) 2 (n 0 ) matrix 5 m (x) = 0 B @ P m (x) P m+ (x) : : : P m+n02 (x) P m 2 (x) P m+ 2 (x) : : : P m+n02 2 (x)...... Pn0(x) m P m+ n0 (x) : : : P m+n02 (x) n0 C A : (2:4) All that remains to be done is to show that the determinant of 5 m (x) which is a polynomial in the indeterminate x, is not identically zero, in which case 0 m (x) = 5 0 m (x)e: (2:5) To show that det(5 m (x)) is not identically zero, it suces to prove that its highest term is not zero. From (2.6) and (2.7), the highest term of det(5 m (x)) is the determinant of H m (x), the (n 0 ) 2 (n 0 ) given by H m (x) = 0 B @ It is easy to check that m m 2 m n0 x m0 m+ x m02 m+ 2. x m0n+ m+ n0 hence not identically zero by Lemma 2.. x m : : : m+n02 x m+n03 m+n02 x m+n04 2 x m0 : : :..... x m0n+2 : : : m+n02 n0 x m0 C A : (2:6) det(h m (x)) = x (m0)(n0) det(u m;n0 ); (2:7) We must show that m (x) and m+n02 (x) are not identically zero. To prove this rst suppose that p(x) has simple roots and consider the expression 0g(x) + P n0 i=0 i h i (x) as a polynomial over K(x) in the indeterminate. Every root of p(x) is a root of this 9

polynomial. On the other hand its degree is n 0. It follows that 0g(x) + h 0 (x) = 0, and h i (x) = 0, for i = ; : : : ; (n 0 ). In other words (2:5) must be satised implying the uniqueness of g(x). Let us denote this unique solution by B m (x) = + m+n02 i=m (m) i (x)( 0 x) i : (2:8) Next we prove that B m (x) and B m+ (x) are distinct. Otherwise, we must have 0 m (x) = 0 m+ (x). From (2:2) the following must hold 5 m (x)0 m (x) = 5 m+ (x)0 m (x): (2:9) Equivalently, (5 m+ (x) 0 5 m (x))0 m (x) = 0: (2:20) But by the same reasoning as before, the matrix of the leading coecient of (2.20) is H m+ (x), whose determinant is nonzero. This implies 0 m (x) 0, a contradiction. Hence the proof of distinctness m Next we prove that for all m = 2 the leading coecient, (m) (x), and the last coecient, (m) m+n02(x), are nonzero functions. As we have seen for m = 2, the function x 0 p=p 0 (see (:5)) lies in S(2; n). From uniqueness we must have B 2 (x) = x 0 p=p 0. Now assume m > 2 and consider B m0 (x), B m (x), and B m+ (x). If the leading coecient of B m (x) is identically zero, then 2 B m(x) + 2 B m+(x) 2 S(m + ; m + n 0 ); contradicting uniqueness. If the last coecient of B m (x) is identically zero, then again a contradiction. 2 B m(x) + 2 B m0(x) 2 S(m 0 ; m + n 0 3); To complete the proof we must show that the leading coecient of B m (x), as well as the last coecient remain to be nonzero even if p(x) has multiple roots. But in this case too we can arrive at the same set of polynomials P m+j i (x) and hence the same equation (2:). 2 0

3. DERIVING CLOSED FORM FOR MEMBERS OF THE BASIC FAMILY In this section we shall derive the closed form for members of the Basic Family. In doing so, it is more convenient to consider the functions in S(m; M), with m 2, M m+n02. Recall that g 2 S(m; M) implies there exists i (x) 2 K(x), i = m; : : : ; M, such that p(y) divides 0 (y) = 0g(x) + y + M l=m l (x)(y 0 x) l ; (3:) in K(x)[y]. Equivalently, if we let z = y 0 x, then f(z) = p(z + x) must divide h(z) = 0 (z + x). From Taylor's Theorem as applied to the polynomial p(y) we have Since 0 (x) = x 0 g(x), f(z) = n i=0 h(z) = 0 (x) + z + p i (x) z i : (3:2) i! M l=m From the divisibility condition, there exist a polynomial q(z) = M0n j=0 l (x)z l : (3:3) v j z j ; (3:4) with v j 2 K(x) such that h(z) = f(z)q(z): (3:5) Equivalently, while suppressing x for brevity we get the system of equations i+j=l i+j=l pv 0 = 0 ; (3:6) pv 0 + p 0 v = ; (3:7) p (i) i! v j = 0; l = 2; : : : ; m 0 ; (3:8) p (i) i! v j = l (x); l = m; : : : ; M: (3:9) The matrix corresponding to the rst m equations, i.e. (3:6), (3:7), and (3:8) is precisely the matrix L m (see (.)). Letting b to be the m-vector ( 0 ; ; 0; : : : ; 0) T and v the m- vector (v 0 ; : : : ; v m0 ) T, we get L m v = b: (3:0)

Lemma 3.. For any natural number k the matrix 0 n n 0 0 : : : 0 n n n 2 Z n;k = 0 : : : 0 B @....... C A n is invertible. n k n k0 n k02 : : : ; Proof. All that is needed is to reduce it the a matrix of the type U m;r (see (2.2)). n Using the identity i + n i0 = n+ i, starting with the last column of Z = Z n;k we add to each column its previous column to get a new matrix Z 2. We repeat this process to columns of Z 2 except for its 2-nd column and get a new matrix Z 3. Next we repeat the process to columns of Z 3 except for its 3-rd column. This process will eventually result in a matrix of the type U m;r. 2. From Lemma 3., analogous to the proof of invertibility of H m (x), we can immediately conclude Corollary 3.. For each natural number m 2, the matrix L () m is invertible. 2 Theorem 3.. Let B m (x) be the function in S(m; m + n 0 2) whose existence was proved in Theorem 2.. Then B m (x) = x 0 p(x) det(l() m0(x)) det(l () m (x)) : Moreover, it lies in S(m; m + n 0 2), and if is a simple root of p(x) then m (m) det(l(2) () = m+()) (0)m det(l () m ()) : Proof. Letting j to be the determinant of L m with the j-th column replaced with b, from Cramer's rule as applied to the system (3:0) we get v j = j ; j = 0; : : : ; m 0 : (3:) det(l m ) 2

Since p is nonzero, det(l m ) is not identically zero, thus each v j is well-dened and unique. Since for B m (x), M = m + n 0 2, we must have v m0 = 0. In particular by expanding the determinant of m0 along its last column, and then expanding the second resulting determinant along its rst row, we get v m0 = m0 = 0 det(l () m ) 0 p(x)det(l() m0) = 0: (3:2) From Corollary 3., L () m is invertible so that (3:2) can be solved for 0 which results in the closed formula for B m (x). To show B m (x) also lies in S(m; m + n 0 2), we will compute (m) l (), l = m; : : : ; m + n 0 2. First consider l = m. The m + equations corresponding to (3:6), (3:7), (3:8), and the rst equation of (3:9) at x =. Since the rst of these equations is zero, if we let b (m) () denote the m-vector (; 0; : : : ; 0; m (m) ()), we get the following m 2 m system L () m+()v() = b (m) (): (3:3) The above system has a unique solution since the coecient matrix is a lower triangular matrix with diagonal entries p 0 () 6= 0. Using the fact that v m0 () = 0, the Cramer's rule, and by expansion along the last column of the matrix L () m+() with its last column replaced by b (m) (), we get m (m) ()det(l() m ()) 0 (0)m0 det(l (2) m+()) = 0: (3:4) The above can be solved for m (m) () which results in the stated equation of the theorem. For l = m + ; : : : ; m + n 0 2, the corresponding equation in (m) i (m) is of the form l ()det(l () ()) + w(); (3:5) for some w(), hence are well-dened proving that B m (x) lies in S(m; m + n 0 2). 2 m Let us treat 0 (x) in equation (3:6) also as a variable, and consider the the m equation in m + unknowns corresponding to equation (3:6) 0 (3:8), i.e. L m v 0 0 e = e 2 ; (3:6) where e i is the column m-vector with the i-th component equal to, and all other components equal to 0. A basic solution of the above system is a solution with at most m 3

nonzero components. The following is evident Theorem 3.2. The basic solution of the system of equations in (3:6) corresponding to v i = 0, i = 0; : : : ; m 0 are B i+ (x), where B 0 (x) = x. 2 Remark 3.. If g 2 S(m; M) has the property that its corresponding equation in (3:0) satises v m0 = 0, and v m = 0 (a condition that is necessarily satised for M = m + n 0 2), then m (x) depends only on p (i) (x), i = 0; : : : ; m, i.e. the simple-rootdepth of g will be m. Remark 3.2. Since Taylor's Theorem holds in any eld of characteristic zero, the existence of B m (x) extends to the case where K is such a eld. 4. TWO FORMULAS AND THEIR APPLICATIONS. In this section we derive two formulas that will result in the generation of very large class of iteration functions. Both will all rely on (:2). To derive the rst formula, suppose that g 2 S(m; M), with m, i.e. g = + M i=m i ( 0 x) i : (4:); where for the sake of simplicity we have suppressed the argument x. From (:2) we have where (m) i = Multiplying (4:2) by 0 m = (m) m (0p) m = i +:::+i m=i;ij M 0 i=m (m) i ( 0 x) i ; (4:2) p (i ) p (im) i! i m! ; i = m; : : : ; mn: (4:3) = 0 m=p 0m and adding to (4.) we get h = g 0 (0) m m ( p p 0 )m = + M 0 i=m+ i ( 0 x) i ; (4:4) where M 0 = maxfm; mng; 4

with i 0 for i > M, and (m) i (m) i i = i 0 m (p 0 ) m ; i = m + ; : : : ; M 0 ; (4:5) 0 for i > mn. Theorem 4.. Assume that the leading coecient of h is not identically zero. Then, h lies in S(m; M), for some m m +, M M 0. If m = m +, then m+ () = m+ () 0 m () m 2 p 00 () p 0 () : (4:6) Proof. Let be a simple root of p(x). Since g 2 S(m; M) for i = m; : : : ; M, i () is well-dened. Clearly, for i = m + ; : : : ; mn, i () is well-dened. The equation (4.6) follows from (4.5) and that (m) m+ = (m=2)(p 00 p 0m ). 2 Next we derive another formula. Let g be in S(m 0 ; M m0 ), and h in S(m; M m ). Let M m+ = maxfm m ; M m0 ; m + n 0 g. We write where where From (4.7) and (4.8) we get g = + M m+ i=m0 i 0; 8i > M m0 ; h = + M m+ i=m i 0; 8i > M m : g 0 h = m0 ( 0 x) m0 + i ( 0 x) i ; (4:7) i ( 0 x) i ; (4:8) M m+ i=m ( i 0 i )( 0 x) i : (4:9) Also, multiplying (.2) by ( 0 x) m0, and since p j 0, for j > n, we get 0 = p( 0 x) m0 + M m+ i=m p (i+0m) (i + 0 m)! ( 0 x)i : (4:0) Multiplying (4.9) by p, and (4.0) by 0 m0 and adding the resulting equations we get M m+ p p(g 0 h) = p( (i+0m) i 0 i ) 0 m0 ( 0 x) i : (4:) (i + 0 m)! i=m 5

Multiplying (4.) by 0 m =D, where and adding the result to (4.8) we get D = [p( m 0 m ) 0 p 0 m0 ]; (4:2) f = h 0 m p(g 0 h) D = + M m+ i=m+ i ( 0 x) i ; (4:3) with p( i 0 i ) 0 p(i+0m) m0 (i+0m)! i = i 0 m ; 8i = m + ; : : : ; M m+ : (4:4) D Theorem 4.2. Suppose that D is not identically zero. Then f 2 S(m; M), where m m +, M M m+. If m = m +, then m+ () = m+ () 0 m () p00 () 2p 0 () : (4:5) Proof. Let be a simple root of p(x). Clearly f lies in S(m; M) for some m+ m M M m+. Since m0 is not identically zero, and since it is a rational function over K, together with the fact that p 0 () 6= 0 it follows from (4.4) that for i = m + ; : : : ; M m+ we have lim p (i+0m) () i(x) = i () 0 m () x! (i + 0 m)!p 0 () ; i.e. i () is well-dened. In particular we get (4:5). 2 Next we consider two application of these formulas. Theorem 4.3. Suppose that in (4:3), g = B m0 and h = B m, then f = B m+, for all m 2. Proof. It suces to prove this under the assumption that p has only simple roots. In order to prove the theorem we rst need to show that the function D is not identically zero. From (4:) we have p(g 0 h) = D + m+n0 i=m+ p( i 0 i ) 0 m0 p (i+0m) (i + 0 m)! ( 0 x) i : (4:6) 6

If D is identically zero, then B m+ + p(g 0 h) 2 S(m + ; m + n 0 ): (4:7) But this contradicts the uniqueness result proved in Theorem 2.. Thus, D is not identically zero, implying that f 2 S(m + ; m + n 0 ). Again by Theorem 2., we must have that f = B m+. 2 Theorem 4.4. Consider the family E ; E 2 ; E 3 ; : : : ; of iteration functions obtained from the repeated application of (4:4) starting with E = B = x 0 p(x). For m 3, E m 2 S(m; mn). Moreover, the depth of E m is m 0. Proof. The fact that E m lies in S(m; mn) is a trivial consequence of Theorem 4.. The claim on the depth can be proved inductively using that theorem and (4:5). 2 For a given i let us denote the i 's corresponding to E m by (m) i 's. We now obtain m (m) for m = ; 2; 3; 4, and 5. From (4.4) this will give E 2 ; E 3 ; E 4, and E 5. From (.4) we have () = (p 0 0 ). This gives E 2 = x 0 p=p 0. It follows that (2) 2 = p 00 =2p 0, (2) 3 = p 000 =3!p 0, so (3) 3 = p000 6p 0 0 2p00 4p 02 : Next we need (4) 4 = (3) 4 0 (3) 3 (3=2)(p 00 =p 0 ). This in turn requires the computation of (3) 4 = (2) 4 0 (2) 2 (2) 4 =p 02. Note that (2) 4 = p (4) =4!p 0, and (2) 4 = 2p 0 p 000 =3! + p 002 =4. Substituting these and simplifying we get Thus, and E 2 0 E 5 (4) 4 = p(4) 4!p 0 5p00 p 000 0 2p + 5p003 02 8p : 03 E 5 = x 0 ( p p 0 ) + ( p p00 )2 p0 2p 0 0 ( p p 0 )3 ( p000 6p 0 0 2p00 4p ) + ( p 02 p 0 )4 ( p(4) 4!p 0 0 5p00 p 000 2p 02 + 5p003 ); 03 8p coincide with the rst few members of a family credited both to Euler as well as Schroder. For historical comments as well as other approaches for deriving this family see Shub and Smale [9]. 7

5. ETENSIONS TO NON-POLYNOMIAL ROOT FINDING In the previous sections we have shown the existence of many xed point iteration functions for nding roots of a given polynomial p(x), namely functions within S(m; M), any of whose members, say g(x), has m-th order rate of convergence to any simple root of p(x). In this section we will prove that these high order methods derived for polynomials extend to high order methods for arbitrary smooth functions f(x), having the same order of convergence to simple roots, where in the formulas for g(x), one simply replaces p(x) and its higher derivatives with f(x) and its corresponding derivatives. Moreover, for a simple root, its asymptotic constant of convergence m () can be obtained by simply replacing p (j) () with f (j) (). In particular, both depth and simple-root-depth will be unchanged. Theorem 5.. Assume that there exists h : < m! <, m 2, so that for any polynomial p(x) of degree at most (2m 0 ) the following property holds: If is a simple root of p(x), then h is in C m in a neighborhood of the point z p () = (; p(); p 0 (); : : : ; p (m0) ()), and the function g p (x) = h(x; p(x); p 0 (x); : : : ; p (m0) (x)), satises g p () = ; (i) g (i) p () = 0; i = ; : : : ; m 0 : (ii) Then, the above properties can be extended to any function f(x) which is in C 2m0 in a neighborhood of a simple root in the sense that p (j) can be replaced with f (j). Moreover g (m) f () = g p (m) n (); (iii) where p n (x) = P n i=0 f (i) () i! (x 0 ) i, the Taylor polynomial of degree n = 2m 0 at. Proof. Since f is in C n, f (j) (x) = p n (j) (x) + r n (j) (x), j = 0; : : : ; n, where r n (0) (x) is the remainder term. Since f (j) () = p n (j) () for all j = 0; : : : ; n we have r (j) n () = 0; j = ; : : : ; n: (5:) In particular, is a simple root of p n (x), and properties (i) and (ii) apply to p n (x). Also, the fact that h is in C n in a neighborhood of the point z f () follows immediately. From 8

the chain rule we have g 0 f = @h @x + @h @f f 0 + @h @f 0 f 00 + : : : + @h @f f (m) : (5:2) (m0) If we now substitute for f(x) and f (j) (x) in (5:2), using (5.) and the continuity of the partial derivatives of h we conclude that g 0 f(x) is continuous in an interval containing z f (). From the repeated application of the chain rule, the assumption that h is in C m, that f is in C 2m0, and the use of (5.), it follows by induction that for all j = 0; : : : ; m, g (j) f (x) is continuous in an interval containing z f(). Moreover g (j) f () = g(j) p n (), for all j = 0; : : : ; m. 2 Corollary 5.. The depth of B m (x) with respect to f 2 C 2m0 remains to be m 0. Moreover, the simple-root-depth is m. As an application of the above consider the replacement of f in Newton's method with f(x)=f 0 (x) which results in a new iteration function g(x) credited to Schroder (e.g. see Scavo and Thoo [8]). Then, from the corollary g 00 ()=2 can easily be computed in terms of f and its higher derivatives. Remark 5.. The assumption f 2 C 2m0 in the above corollary can be relaxed for B m (x). It suce to have f 2 C m. For example for B 2 (x) = x 0 p(x)=p 0 (x), it is enough to assume that f 2 C 2, despite the fact that B 00 2(x) relies on the third derivative of f. Concluding Remarks. In this paper we have characterized a family of rational iteration functions for polynomial root-nding having an m-th order rate of convergence to simple roots. Although individual members of the family like Newton's and Halley's iteration functions have been known, no closed formula for the general member of the Basic Family, nor their asymptotic constant had previously been known. These formulas hold in the more general case where the underlying eld K is any eld of characteristic zero. We have shown that the members of this family are \optimal" with respect to the notions of depth, width, simple-root-depth (see Denition.2, and Remark.2). 9 Under the assumption of the

simplicity of the roots, the m-th order iteration function was shown to be the unique member of the class S(m; m + n 0 2). We also described two simple recursive formulas for generation of new iteration functions which in particular resulted in the generation of the Euler-Schroder family whose m-th order member belongs to S(m; mn), m > 2. Shub and Smale [9] refer to an \incremental" version of Euler-Schroder family as the \the most appropriate for practically computing zeros of complex polynomials" (see Example 5 k, page 0). In view of the development of the Basic Family, the closed forms, and properties established in this paper, it is fair to venture that their incremental version would be at least as good as the Euler-Schroder family. In particular, the closed form of the Basic Family suggests parallel implementations (see e.g. [5] for a theoretical analysis of parallel implementation of the Basic Family as applied to square-root computation). Another theoretical and practical problem of interest is the study of global convergence of the Basic Family. Acknowledgments. We would like to thank Professor Steve Smale, [20], for suggesting the examination of existence of possible links between the Basic Family (whose algebraic existence was proved in [6]) and the Euler-Schroder family. That suggestion resulted in the discovery of the rst formula (4:4). We also thank Professor Jerry Richter for bringing to our attention a paper by Gerlach []. References [] Alefeld, G.(98). On the convergence of Halley's Method. Amer. Math. Monthly 88, 530-536. [2] Atkinson, K.E.(989). An Introduction to Numerical Analysis. Second Edition, John Wiley and Sons, Inc. [3] Bateman, H.(938). Halley's method for solving equations. Amer. Math. Monthly 45, -7. [4] Bodewig, E.(949). On types of convergence and on the behavior of approximations in the neighborhood of a multiple root of an equation. Quart. App. Math. 7, 325-333. 20

[5] Brown, G.H., Jr.(977). On Halley's variation of Newton's method. Amer. Math. Monthly 84, 726-728. [6] Davies, M and Dawson, B.(975). On the global convergence of Halley's iteration formula. Numer. Math. 24, 33-35. [7] Frame, J.S.(944). A variation of Newton's method. Amer. Math. Monthly 5, 36-38. [8] Frame, J.S.(945). Remarks on a variation of Newton's method. Amer. Math. Monthly 52, 22-24. [9] Frame, J.S.(953). The solution of equations by continued fraction. Amer. Math. Monthly 60, 293-305. [0] Gander, W.(985). On Halley's iteration method. Amer. Math. Monthly 92, 3-34. [] Gerlach, J.(994). Accelerated convergence in Newton's method. SIAM Review. 36, 272-276. [2] Halley, E. (694). A new, exact, and easy method of nding roots of any equations generally, and that without any previous reduction (Latin). Philos. Trans. Roy. Soc. London 8 36-48. [English translation: Philos. Trans. Roy. Soc. London(abridged) 3, 640-649 (809).] [3] Hamilton, H.J.(950). A type of variation of Newton's method. Amer. Math. Monthly 57,57-522. [4] Hansen, E. and Patrick, M.(977). A family of root nding methods. Numer. Math. 27, 257-269. [5] Kalantari, B. and Kalantari, I.(993). High order iterative methods for approximating square and cube roots. Technical Report LCSR-TR-23, Department of Computer Science, Rutgers University, New Brunswick, NJ. [6] Kalantari, B. and Kalantari, I.(993). High order iterative methods for approximating roots of polynomials. Technical Report LCSR-TR-24, Department of Computer Science, Rutgers University, New Brunswick, NJ. [7] Popovski, D.B.(980). A family of one-point iteration formulae for nding roots. Internat.. J. Computer Math. 8, 85-88. 2

[8] Scavo, T.R. and Thoo, J.B.(995). On the geometry of Halley's method. Amer. Math. Monthly 02, 47-426. [9] Shub, M. and Smale, S.(985). On the geometry of polynomials and a theory of cost: Part I. Ann. Scient. Ec. Norm. Sup, 8, 07-42. [20] Smale, S.(994). Private communications with the rst author. [2] Snyder, R.W.(955). One more correction formula. Amer. Math. Monthly 62, 722-725. [22] Stewart, J.K.(95). Another variation of Newton's method. Amer. Math. Monthly 58, 33-334. [23] Traub, J.F.(964). Iterative Methods for the Solution of Equations. New Jersey, Prentice Hall. [24] van der Waerden, B.L.(970) Algebra, Volume. F. Ungar Publishing Co., New York. [25] Wall, H.S.(948). A modication of Newton's method. Amer. Math. Monthly 55, 90-94. 22