Institute for Advanced Computer Studies. Department of Computer Science. On the Convergence of. Multipoint Iterations. G. W. Stewart y.

Similar documents
Institute for Advanced Computer Studies. Department of Computer Science. On Markov Chains with Sluggish Transients. G. W. Stewart y.

Institute for Advanced Computer Studies. Department of Computer Science. On the Perturbation of. LU and Cholesky Factors. G. W.

UMIACS-TR July CS-TR 2721 Revised March Perturbation Theory for. Rectangular Matrix Pencils. G. W. Stewart.

1 Introduction It will be convenient to use the inx operators a b and a b to stand for maximum (least upper bound) and minimum (greatest lower bound)

Institute for Advanced Computer Studies. Department of Computer Science. Two Algorithms for the The Ecient Computation of

Automatica, 33(9): , September 1997.

Linear Algebra (part 1) : Vector Spaces (by Evan Dummit, 2017, v. 1.07) 1.1 The Formal Denition of a Vector Space

J. Symbolic Computation (1995) 11, 1{000 An algorithm for computing an integral basis in an algebraic function eld Mark van Hoeij Department of mathem

Spurious Chaotic Solutions of Dierential. Equations. Sigitas Keras. September Department of Applied Mathematics and Theoretical Physics

The iterative convex minorant algorithm for nonparametric estimation

A BASIC FAMILY OF ITERATION FUNCTIONS FOR POLYNOMIAL ROOT FINDING AND ITS CHARACTERIZATIONS. Bahman Kalantari 3. Department of Computer Science

Numerical Results on the Transcendence of Constants. David H. Bailey 275{281

Generalization of Hensel lemma: nding of roots of p-adic Lipschitz functions

Solutions Preliminary Examination in Numerical Analysis January, 2017

4.1 Eigenvalues, Eigenvectors, and The Characteristic Polynomial

Abstract Minimal degree interpolation spaces with respect to a nite set of

Novel Approach to Analysis of Nonlinear Recursions. 1 Department of Physics, Bar-Ilan University, Ramat-Gan, ISRAEL

STOCHASTIC DIFFERENTIAL EQUATIONS WITH EXTRA PROPERTIES H. JEROME KEISLER. Department of Mathematics. University of Wisconsin.

University of Maryland at College Park. limited amount of computer memory, thereby allowing problems with a very large number

Garrett: `Bernstein's analytic continuation of complex powers' 2 Let f be a polynomial in x 1 ; : : : ; x n with real coecients. For complex s, let f

Domain and range of exponential and logarithmic function *

Minimum and maximum values *

1 Solutions to selected problems

Lifting to non-integral idempotents

X. Hu, R. Shonkwiler, and M.C. Spruill. School of Mathematics. Georgia Institute of Technology. Atlanta, GA 30332

1. Introduction Let the least value of an objective function F (x), x2r n, be required, where F (x) can be calculated for any vector of variables x2r

Numerical Comparisons of. Path-Following Strategies for a. Basic Interior-Point Method for. Revised August Rice University

On reaching head-to-tail ratios for balanced and unbalanced coins

Linear Algebra, 4th day, Thursday 7/1/04 REU Info:

Test 2 Review Math 1111 College Algebra

Simple Lie subalgebras of locally nite associative algebras

NP-hardness of the stable matrix in unit interval family problem in discrete time

National Institute of Standards and Technology USA. Jon W. Tolle. Departments of Mathematics and Operations Research. University of North Carolina USA

Chapter 1: A Brief Review of Maximum Likelihood, GMM, and Numerical Tools. Joan Llull. Microeconometrics IDEA PhD Program

5 Eigenvalues and Diagonalization

Department of Applied Mathematics Preliminary Examination in Numerical Analysis August, 2013

Roots of Unity, Cyclotomic Polynomials and Applications

Analog Neural Nets with Gaussian or other Common. Noise Distributions cannot Recognize Arbitrary. Regular Languages.

An Attempt To Understand Tilling Approach In proving The Littlewood Conjecture

Chapter 3: Root Finding. September 26, 2005

BFGS WITH UPDATE SKIPPING AND VARYING MEMORY. July 9, 1996

G METHOD IN ACTION: FROM EXACT SAMPLING TO APPROXIMATE ONE

Contents. 2 Partial Derivatives. 2.1 Limits and Continuity. Calculus III (part 2): Partial Derivatives (by Evan Dummit, 2017, v. 2.

Decidability of Existence and Construction of a Complement of a given function

IE 5531: Engineering Optimization I

RATIONAL POINTS ON CURVES. Contents

290 J.M. Carnicer, J.M. Pe~na basis (u 1 ; : : : ; u n ) consisting of minimally supported elements, yet also has a basis (v 1 ; : : : ; v n ) which f

Nets Hawk Katz Theorem. There existsaconstant C>so that for any number >, whenever E [ ] [ ] is a set which does not contain the vertices of any axis

Adaptive linear quadratic control using policy. iteration. Steven J. Bradtke. University of Massachusetts.

Richard DiSalvo. Dr. Elmer. Mathematical Foundations of Economics. Fall/Spring,

Faculteit der Technische Wiskunde en Informatica Faculty of Technical Mathematics and Informatics

(10) What is the domain of log 123 (x)+ x

A MULTIGRID ALGORITHM FOR. Richard E. Ewing and Jian Shen. Institute for Scientic Computation. Texas A&M University. College Station, Texas SUMMARY

Quantum logics with given centres and variable state spaces Mirko Navara 1, Pavel Ptak 2 Abstract We ask which logics with a given centre allow for en

A version of for which ZFC can not predict a single bit Robert M. Solovay May 16, Introduction In [2], Chaitin introd

form, but that fails as soon as one has an object greater than every natural number. Induction in the < form frequently goes under the fancy name \tra

A Graph Based Parsing Algorithm for Context-free Languages

New concepts: Span of a vector set, matrix column space (range) Linearly dependent set of vectors Matrix null space

Increasing and decreasing intervals *

Order of convergence

Notes on Iterated Expectations Stephen Morris February 2002

THE REAL POSITIVE DEFINITE COMPLETION PROBLEM. WAYNE BARRETT**, CHARLES R. JOHNSONy and PABLO TARAZAGAz

Contents. 1 Introduction to Dynamics. 1.1 Examples of Dynamical Systems

1 Ordinary points and singular points

quantitative information on the error caused by using the solution of the linear problem to describe the response of the elastic material on a corner

Realization of set functions as cut functions of graphs and hypergraphs

Lecture Notes: Mathematical/Discrete Structures

Fraction-free Row Reduction of Matrices of Skew Polynomials

f(z)dz = 0. P dx + Qdy = D u dx v dy + i u dy + v dx. dxdy + i x = v

Our goal is to solve a general constant coecient linear second order. this way but that will not always happen). Once we have y 1, it will always

Congurations of periodic orbits for equations with delayed positive feedback

Our rst result is the direct analogue of the main theorem of [5], and can be roughly stated as follows: Theorem. Let R be a complex reection of order

Techinical Proofs for Nonlinear Learning using Local Coordinate Coding

Math 314 Lecture Notes Section 006 Fall 2006

Lie Groups for 2D and 3D Transformations

Contents. 2.1 Vectors in R n. Linear Algebra (part 2) : Vector Spaces (by Evan Dummit, 2017, v. 2.50) 2 Vector Spaces

Taylor series - Solutions

Pointwise convergence rate for nonlinear conservation. Eitan Tadmor and Tao Tang

Global Dynamics of Some Periodically Forced, Monotone Di erence Equations

ARCHIVUM MATHEMATICUM (BRNO) Tomus 32 (1996), 13 { 27. ON THE OSCILLATION OF AN mth ORDER PERTURBED NONLINEAR DIFFERENCE EQUATION

5 Introduction to Complex Dynamics

Reduction of two-loop Feynman integrals. Rob Verheyen

UMIACS-TR July CS-TR 2494 Revised January An Updating Algorithm for. Subspace Tracking. G. W. Stewart. abstract

2 JOHAN JONASSON we address in this paper is if this pattern holds for any nvertex graph G, i.e. if the product graph G k is covered faster the larger

Gerhard Rein. Mathematisches Institut der Universitat Munchen, Alan D. Rendall. Max-Planck-Institut fur Astrophysik, and

SOLVING LINEAR RECURRENCE RELATIONS

Positive definiteness of tridiagonal matrices via the numerical range

R. Schaback. numerical method is proposed which rst minimizes each f j separately. and then applies a penalty strategy to gradually force the

460 HOLGER DETTE AND WILLIAM J STUDDEN order to examine how a given design behaves in the model g` with respect to the D-optimality criterion one uses

A Stable Finite Dierence Ansatz for Higher Order Dierentiation of Non-Exact. Data. Bob Anderssen and Frank de Hoog,

A Proof of the EOQ Formula Using Quasi-Variational. Inequalities. March 19, Abstract

DK-2800 Lyngby, Denmark, Mercierlaan 94, B{3001 Heverlee, Belgium,

S. Lucidi F. Rochetich M. Roma. Curvilinear Stabilization Techniques for. Truncated Newton Methods in. the complete results 1

2 Dynamics of One-Parameter Families

N.G.Bean, D.A.Green and P.G.Taylor. University of Adelaide. Adelaide. Abstract. process of an MMPP/M/1 queue is not a MAP unless the queue is a

ARTIFICIAL INTELLIGENCE LABORATORY. and CENTER FOR BIOLOGICAL INFORMATION PROCESSING. A.I. Memo No August Federico Girosi.

The super line graph L 2

Note An example of a computable absolutely normal number

Linear Algebra (part 1) : Matrices and Systems of Linear Equations (by Evan Dummit, 2016, v. 2.02)

1 Introduction A well-known result concerning the limits of parallel computation, and called the speedup theorem, states that when two or more process

Transcription:

University of Maryland Institute for Advanced Computer Studies Department of Computer Science College Park TR{93{10 TR{3030 On the Convergence of Multipoint Iterations G. W. Stewart y February, 1993 Reviseed, Janurary 1994 ABSTRACT This note gives a new convergence proof for iterations based on multipoint formulas. It rests on the very general assumption that if the desired xed point appears as an argument in the formula then the the formula returns the xed point. This report is available by anonymous ftp from thales.cs.umd.edu in the directory pub/reports. y Department of Computer Science and Institute for Advanced Computer Studies, University of Maryland, College Park, MD 20742.

On the Convergence of Multipoint Iterations G. W. Stewart ABSTRACT This note gives a new convergence proof for iterations based on multipoint formulas. It rests on the very general assumption that if the desired xed point appears as an argument in the formula then the the formula returns the xed point. Many useful iterations for nding a zero x of a function f generate iterates according to a scheme of the form x k+1 = '(x k ; x k 1 ; : : : ; x k n+1 ): (1) For example, the secant method proceeds according to the formula x k+1 = x k 1f(x k ) x k f(x k 1 ) : (2) f(x k ) f(x k 1 ) Since the method uses two previous iterates to generate the next, it is called a two-point method. Three-point methods are also in common use; e.g., Muller's method [3] and the method of inverse quadratic interpolation [2]. More generally, the iteration (1) is called an n-point iteration or generically a multipoint iteration. Proofs of local convergence of multipoint methods usually involve using the special form of the iteration function ' to show that the errors e k = x k x satisfy an inequality of the form je k+1 j Cje k e k 1 e k n+1 j; (3) from which it can be shown that the convergence is at least superlinear of order p, where p is the positive root of the equation p n p n 1 p 1 = 0: (4) Only the constant in (3) depends on the particular method; in other words, the order of convergence depends only on the number of points. There is a simple reason why these dierent methods have the same rate of convergence: namely, each of the iterative methods named above return the 1

2 Multipoint Iterations solution whenever one of the arguments is the solution. More precisely, in the case of the secant method '(u; x ) x and '(x ; v) x ; as can easily be seen from (2). The purpose of this note is to show that a general convergence proof for multipoint iterations can be based on this property. Throughout we will assume that ' has as many derivatives as is needed for the analysis. The ideas are suciently well illustrated for the general two-point iteration, and to avoid clutter we will analyze that case. At the end of the note, we will indicate the modications necessary for the general case. The rst step in the analysis is to observe that since '(u; x ) and '(x ; v) are constant their derivatives with respect to u and v are zero: ' u (u; x ) 0 and ' v (x ; v) 0: The same is true of the unmixed second derivatives: ' uu (u; x ) 0 and ' vv (x ; v) 0: The next step is to derive an error recursion. We begin by expanding ' about (x ; x ) in a Taylor series. Specically, '(x + p; x + q) = x + ' u (x ; x )p + ' v (x ; x )q + 1 2 [' uu(x + p; x + q)p 2 + 2' uv (x + p; x + q)pq + ' vv (x + p; x + q)q 2 ]; where 2 [0; 1]. Since ' u (x ; x ) = ' v (x ; x ) = 0, '(x + p; x + q) = x + 1 2 [' uu(x + p; x + q)p 2 + 2' uv (x + p; x + q)pq + ' vv (x + p; x + q)q 2 ]; (5) We would like to factor the product pq out of this expression to obtain a product of errors like (3); however, we must rst massage the terms in p 2 and q 2. Since ' uu (x + p; x ) = 0, it follows from a Taylor expansion in the second argument that ' uu (x + p; x + q) = ' uuv (x + p; x + q q)q;

Multipoint Iterations 3 where q 2 [0; 1]. Similarly, ' vv (x + p; x + q) = ' uvv (x + p p; x + q)p; where p 2 [0; 1]. Substituting these values in (5) gives '(x + p; x + q) = x + pq 2 [' uuv(x + p; x + q q)p + 2' uv (x + p; x + q) + ' uvv (x + p p; x + q)q]: (6) Turning now to the iteration itself, let the starting values be x 0 and x 1, and let their errors be e 0 = x 0 x and e 1 = x 1 x. Taking p = e 1 and q = e 0 in (6), we get e 2 = '(x + e 1 ; x + e 0 ) = x + e 1e 0 2 [' uuv(x + e 1 ; x + e0 e 0 )e 1 + ' uv (x + e 1 ; x + e 0 ) + ' uvv (x + e1 e 1 ; x + e 0 )e 0 ] e 1e 0 2 r(e 1; e 0 ): (7) This is the error recurrence we need. We are now ready to establish the convergence of the method. First note that r(0; 0) = 2' uv (x ; x ): Hence there is a > 0 such that if juj; jvj then jvr(u; v)j C < 1: Now let je 0 j; je 1 j. From the error recurrence (7) it follows that je 2 j Cje 1 j < je 1 j. Hence and je 3 j Cje 2 j C 2 je 1 j. By induction je 1 r(e 2 ; e 1 )j C < 1: je k j C k 1 je 1 j; and since the right-hand side of this inequality converges to zero, we have e k! 0; i.e., the general two-point iteration converges from any two starting values whose errors are less than in absolute value.

4 Multipoint Iterations We now turn to the convergence rate of two point methods. The rst thing to note is that since e k+1 = e ke k 1 r(e k ; e k 1 ) (8) 2 and r(0; 0) = 2' uv (x ; x ), we have lim k!1 e k+1 e k e k 1 = ' uv (x ; x ): (9) If ' uv (x ; x ) 6= 0, we shall say that the sequence fx k g exhibits two-point convergence. We are going to show that two-point convergence is superlinear of order p = 1 + p 5 2 Here p is the largest root of the equation = 1:618 : : : : p 2 p 1 = 0: (10) There are two ways to establish this fact. The rst is to derive (10) directly from (9), which is the usual approach. However, since we already know the value of p, we can instead set s k = je k+1j je k j p (11) and use (9) to verify that the s k have a nonzero limit, which is usual denition of pth order convergence. From (11) we have je k j = s k 1 je k 1 j p and From (8), je k+1 j = s k je p k j = s ks p k 1 je k 1j p2 : jr k j jr(e k ; e k 1 )j = s ks p k 1je k 1 j p2 s k 1 je k 1 j p je k 1 j = s ks p 1 k 1 je k 1j p2 p 1 : Since p 2 p 1 = 0, we have je k 1 j p2 p 1 = 1 and jr k j = s k s p 1 k 1 :

Multipoint Iterations 5 Let k = log jr k j and k = log s k. Then our problem is to show that the sequence dened by k = k (p 1) k 1 has a limit. Let = lim k!1 k. Then the limit, if it exists, must satisfy = (p 1) : Thus we must show that the sequence of errors dened by ( k ) = ( k ) (p 1)( k ) converges to zero. To do this we use the following easily established result from the theory of dierence equations. If the roots of the equation x m a 1 x m 1 a m = 0 all lie in the unit circle and lim k!1 k generated by the recursion = 0, then the sequence f k g k = k + a 1 k 1 + + a m k m converges to zero, whatever the starting values 0 ; : : : ; n 1. In our application m = 1, k = ( k ), and k = ( k ). The equation whose roots are to lie in the unit circle is x + (p 1) = 0. Since p 1 = 0:618, the conditions of the above result are satised, and k!. It follows that the numbers s k have a nonzero limit. In other words, two-point convergence is superlinear convergence of order p=1.618: : :. The generalization of this result to multipoint methods is for the most part routine. The assumption that the iteration function is identically equal to x whenever one of its arguments is equal to x implies that the Taylor series begins with the term containing (u 1 x )(u 2 x ) (u n x ). The other terms of the same order that are introduced by the mean value theorem can be reduced to ones containing the product (u 1 x )(u 2 x ) (u n x ) by additional Taylor expansions. The result is a recursion of the form e k+1 = e k e k 1 e k n+1 r(e k ; e k 1 ; : : : ; e k n+1 ):

6 Multipoint Iterations where r(0; 0; : : : ; 0) is nite. We can use this recursion as above to establish the convergence of the method. The only nontrivial dierence lies in establishing the rate of convergence. Proceeding as above, we obtain the following dierence equation relating the k = k and k = k : k = k (p 1) k 1 (p 2 p 1) k 2 (p n 1 p n 2 1) k n : Here p is the positive root of the equation (4) and is easily seen to be bounded by two. If we set b i = p i p i 1 1; then by the result quoted above on dierence equations the pth order convergence of the iteration will be established if we can show that the roots of the equation h(x) = x n 1 + b 1 x n 2 + : : : + b n 1 = 0 (12) lie in the interior of the unit circle. Now the coecients b i are positive. Moreover, b i b i+1 = p i (2 p) > 0. Thus the b i satisfy 1 b 0 > b 1 > b 2 > > b n 1 > 0: (13) By the Enestrom{Kakeya theorem (see [1]), the absolute values of the zeros of h are bounded by the largest of the ratios b i+1 =b i, which are all less than one. The following is a table of the the values of p and the magnitude q of the largest root of (12) for n = 2; : : : ; 10. n p q 2 1:6180 0:6180 3 1:8393 0:7374 4 1:9276 0:8183 5 1:9659 0:8710 6 1:9836 0:9062 7 1:9920 0:9303 8 1:9960 0:9472 9 1:9980 0:9593 10 1:9990 0:9682

Multipoint Iterations 7 References [1] N. Anderson, E. B. Sa, and R. S. Varga. On the Enestrom{Kakeya theorem and its sharpness. Linear Algebra and Its Applications, 28:5{16, 1979. [2] R. P. Brent. Algorithms for Minimization without Derivatives. Prentice{Hall, Englewood Clis, New Jersey, 1973. [3] D. E. Muller. A method for solving algebraic equations using an automatic computer. Mathematical Tables and Other Aids to Computation, 10:208{215, 1956.