Multivariate Interpolation by Polynomials and Radial Basis Functions

Similar documents
Scattered Data Interpolation with Polynomial Precision and Conditionally Positive Definite Functions

Kernel B Splines and Interpolation

On interpolation by radial polynomials C. de Boor Happy 60th and beyond, Charlie!

APPENDIX A. Background Mathematics. A.1 Linear Algebra. Vector algebra. Let x denote the n-dimensional column vector with components x 1 x 2.

Elementary linear algebra

Stability constants for kernel-based interpolation processes

Solving the 3D Laplace Equation by Meshless Collocation via Harmonic Kernels

MATH 205C: STATIONARY PHASE LEMMA

Theoretical and computational aspects of multivariate interpolation with increasingly flat radial basis functions

DS-GA 1002 Lecture notes 0 Fall Linear Algebra. These notes provide a review of basic concepts in linear algebra.

Stability of Kernel Based Interpolation

08a. Operators on Hilbert spaces. 1. Boundedness, continuity, operator norms

University of Colorado at Denver Mathematics Department Applied Linear Algebra Preliminary Exam With Solutions 16 January 2009, 10:00 am 2:00 pm

Linear Algebra Massoud Malek

Algebra II. Paulius Drungilas and Jonas Jankauskas

Approximation by Conditionally Positive Definite Functions with Finitely Many Centers

Bare-bones outline of eigenvalue theory and the Jordan canonical form

Here each term has degree 2 (the sum of exponents is 2 for all summands). A quadratic form of three variables looks as

Tangent spaces, normals and extrema

x i e i ) + Q(x n e n ) + ( i<n c ij x i x j

Lecture Notes 1: Vector spaces

Problems in Linear Algebra and Representation Theory

Chapter 3 Transformations

Chapter 7. Extremal Problems. 7.1 Extrema and Local Extrema

MTH Linear Algebra. Study Guide. Dr. Tony Yee Department of Mathematics and Information Technology The Hong Kong Institute of Education

CHAPTER 3 Further properties of splines and B-splines

Linear Algebra. Min Yan

Lecture 6 Positive Definite Matrices

Chapter Two Elements of Linear Algebra

Math 102, Winter Final Exam Review. Chapter 1. Matrices and Gaussian Elimination

SPRING 2006 PRELIMINARY EXAMINATION SOLUTIONS

Insights into the Geometry of the Gaussian Kernel and an Application in Geometric Modeling

The eigenvalues are the roots of the characteristic polynomial, det(a λi). We can compute

October 25, 2013 INNER PRODUCT SPACES

8. Diagonalization.

Math 61CM - Solutions to homework 6

We simply compute: for v = x i e i, bilinearity of B implies that Q B (v) = B(v, v) is given by xi x j B(e i, e j ) =

Some notes on Coxeter groups

1 9/5 Matrices, vectors, and their applications

Mathematics Department Stanford University Math 61CM/DM Inner products

Math 113 Final Exam: Solutions

Kernel Methods. Machine Learning A W VO

Linear Algebra Primer

implies that if we fix a basis v of V and let M and M be the associated invertible symmetric matrices computing, and, then M = (L L)M and the

Lecture Notes in Mathematics. Arkansas Tech University Department of Mathematics. The Basics of Linear Algebra

MATHEMATICS 217 NOTES

Linear Algebra: Matrix Eigenvalue Problems

Vector Spaces. Vector space, ν, over the field of complex numbers, C, is a set of elements a, b,..., satisfying the following axioms.

RKHS, Mercer s theorem, Unbounded domains, Frames and Wavelets Class 22, 2004 Tomaso Poggio and Sayan Mukherjee

Preliminary/Qualifying Exam in Numerical Analysis (Math 502a) Spring 2012

Math Linear Algebra II. 1. Inner Products and Norms

Final Review Sheet. B = (1, 1 + 3x, 1 + x 2 ) then 2 + 3x + 6x 2

TOEPLITZ OPERATORS. Toeplitz studied infinite matrices with NW-SE diagonals constant. f e C :

A PRIMER ON SESQUILINEAR FORMS

4.2. ORTHOGONALITY 161

MTH 2032 Semester II

Slide05 Haykin Chapter 5: Radial-Basis Function Networks

Part IB - Easter Term 2003 Numerical Analysis I

REVIEW OF DIFFERENTIAL CALCULUS

Interpolation by Basis Functions of Different Scales and Shapes

Symmetric matrices and dot products

5.6. PSEUDOINVERSES 101. A H w.

Matrix Theory. A.Holst, V.Ufnarovski

Exercise Solutions to Functional Analysis

Foundations of Matrix Analysis

Exercise Sheet 1.

7. Symmetric Matrices and Quadratic Forms

Extreme Values and Positive/ Negative Definite Matrix Conditions

Topics in linear algebra

Canonical lossless state-space systems: staircase forms and the Schur algorithm

The value of a problem is not so much coming up with the answer as in the ideas and attempted ideas it forces on the would be solver I.N.

Kernel Method: Data Analysis with Positive Definite Kernels

Class notes: Approximation

RIEMANN SURFACES. max(0, deg x f)x.

Vectors in Function Spaces

Review problems for MA 54, Fall 2004.

0.1 Rational Canonical Forms

EIGENVALUES AND EIGENVECTORS 3

Fundamentals of Linear Algebra. Marcel B. Finan Arkansas Tech University c All Rights Reserved

Practical Linear Algebra: A Geometry Toolbox

Matrix Factorization and Analysis

18. Cyclotomic polynomials II

The Solution of Linear Systems AX = B

A linear algebra proof of the fundamental theorem of algebra

Lecture 7: Positive Semidefinite Matrices

1. General Vector Spaces

Extra Problems for Math 2050 Linear Algebra I

HOMEWORK PROBLEMS FROM STRANG S LINEAR ALGEBRA AND ITS APPLICATIONS (4TH EDITION)

A linear algebra proof of the fundamental theorem of algebra

Vector Spaces ปร ภ ม เวกเตอร

PUTNAM TRAINING POLYNOMIALS. Exercises 1. Find a polynomial with integral coefficients whose zeros include

Generalized eigenvector - Wikipedia, the free encyclopedia

Math 25a Practice Final #1 Solutions

Linear Algebra March 16, 2019

David Hilbert was old and partly deaf in the nineteen thirties. Yet being a diligent

MATH 590: Meshfree Methods

MATRIX ALGEBRA AND SYSTEMS OF EQUATIONS. + + x 1 x 2. x n 8 (4) 3 4 2

Algebra Workshops 10 and 11

A Brief Outline of Math 355

Stability of Feedback Solutions for Infinite Horizon Noncooperative Differential Games

Transcription:

Multivariate Interpolation by Polynomials and Radial Basis Functions Robert Schaback April 4, 00 Abstract In many cases, multivariate interpolation by smooth radial basis functions converges towards polynomial interpolants, when the basis functions are scaled to become wide. In particular, examples show that interpolation by scaled Gaussians seems to converge towards the de Boor/Ron least polynomial interpolant. The paper starts by providing sufficient criteria for the convergence of radial interpolants, and the structure of the polynomial limit interpolation is investigated to some extent. The results lead to general questions about radial polynomials x y l and the properties of spaces spanned by linear combinations of their shifts. For their investigation a number of helpful results are collected. In particular, the new notion of a discrete moment basis turns out to be rather useful. With these tools, a variety of well posed multivariate polynomial interpolation processes can be formulated, leading to interesting questions about their relationships. Part of them can be proven to be least in the sense of de Boor and Ron. Finally, the paper generalizes the de Boor/Ron interpolation process and shows that it occurs as the limit of interpolation by Gaussian radial basis functions. As a byproduct, we get a stable method for preconditioning the matrices arising with interpolation by smooth radial basis functions. Keywords: radial basis functions, Gaussian, moment conditions. Classification: 4A5, 4A5, 4A30, 4A63, 65D0 Introduction Let φ : [0, ) IR be a smooth radial basis function that can be written as φ(r) f(r ) with a smooth function f : IR IR, and in particular we have in mind the Gaussians and inverse multiquadrics, i.e. φ(r) exp( r ) and φ(r) ( + r ) β/, β < 0. We scale φ in such a way that the functions get wider, i.e. we define φ c(r) : φ(r c) f(cr ), c, r 0 () and since we want to consider small c, we assume that f is analytic around zero. We fix a set X {x,..., x M } IR d of scattered centers for interpolation, and consider the behaviour of the Lagrange interpolation basis for c 0. It is obtainable as the solution (u c (x),..., u c M (x)) IR M of the system φ c( x j x k )u c j(x) φ c( x x k ) for all k M. () By a surprising observation of Driscoll/Fornberg [6] and Danzeglocke [5] there are many cases where the limits of the Lagrange basis functions u c j(x) for c 0 exist and are multivariate polynomials in x. Our first goal is to prove this fact under certain assumptions on φ and X. From a recent paper by Fornberg, Wright and Larsson [7] it is known that convergence may depend critically on the geometry of X and certain properties of φ. We study these connections to some extent, and we want to characterize the final polynomial interpolant in a way that is independent of the limit process. This poses some interesting questions about multivariate polynomials and geometric properties of scattered data sets. The investigations about the limit

polynomial interpolants are based on radial polynomials of the form x y l because the matrix entries in () have series expansions and the right hand side contains φ c( x j x k ) f(c x j x k ) φ c( x x k ) f(c x x k ) l0 l0 f (l) (0) c l x j x k l l! f (l) (0) c l x x k l. l! Thus this paper contains a somewhat nonstandard approach to multivariate interpolation, namely via linear combinations of radial polynomials. In particular, we define two different classes of multivariate polynomial interpolation schemes that can be formulated without recurring to limits of radial basis functions. Examples show that the various methods are actually different. For their analysis, some useful theoretical notions are introduced, i.e. discrete moment conditions and discrete moment bases. To establish the link from interpolation by scaled Gaussians to the de Boor/Ron least polynomial interpolation [, 3, 4], we generalize the latter and, in particular, introduce a scaling and relate the theory to reproducing kernel Hilbert spaces. Using the new notion of a discrete moment basis we can prove that the algorithm of de Boor and Ron is the limit of radial basis function interpolation using the Gaussian kernel, if the kernel gets wide. Finally, we prove that properly scaled discrete moment bases can be used for preconditioning the systems arising in radial basis function interpolation. Limits of Radial Basis Functions Because we shall be working with determinants, we fix the numbering of the points in X now. For a second ordered set Y {y,..., y M } IR d with the same number M of data points we define the matrix A c,x,y : (φ c( x j y k )) j,k M ( f(c x j y k ) ) j,k M. Note that A c,x,x is symmetric and has a determinant that is independent of the order of the points in X. If φ is positive definite, the matrices A c,x,x are positive definite and have a positive determinant for all c > 0. Since f is analytic around the origin, the matrices A c,x,y have a determinant with a convergent series expansion det A c,x,y c k p k (X, Y ) (3) k0 for small c, where the functions p k (X, Y ) are polynomials in the points of X and Y. In particular they are sums of powers of terms of the form x j y k. They can be determined by symbolic computation, and we shall give an explicit formula in section 3 and prove the upper bound k for their total degree in Lemma 5. We define X j : X \ {x j}, where x j is deleted and the order of the remaining points is kept. Furthermore, in the sets X j(x) : (X \ {x j}) {x}, j M the point x j is replaced by x, keeping the order. The general structure of Lagrange basis functions is described by a standard technique: Lemma [7] The Lagrange basis functions u c j(x), j M for interpolation in X {x,..., x M } by a scaled positive definite radial function φ c have the form u c j(x) : det A c,x,x j (x) k0 ck p k (X, X j(x)) det A, j M. (4) c,x,x k0 ck p k (X, X) Proof: The quotient of determinants is in the span of the functions φ c( x x j ), j M, and it satisfies u j(x k ) δ jk, j, k M. Since interpolation is unique, we are done. From (4) it is clear that the convergence behaviour of the Lagrange basis function u c j(x) for c 0 crucially depends on the smallest values of k such that the real numbers p k (X, X) or p k (X, X j(x)) are nonzero. Examples show that this number in turn depends on the geometry of X, getting large when the set degenerates from general position.

Definition Let κ(d, M) be the minimal k 0 such that the multivariate polynomial X p k (X, X) is nonzero on the space R Md. A set X {x,..., x M } IR d is in general position with respect to φ if p κ(d,m) (X, X) 0. A set X {x,..., x M } IR d has a degeneration order j with respect to φ if p k (X, X) 0 for all 0 k < κ(d, M) + j. The maximal degeneration order of a set X {x,..., x M } IR d will be denoted by δ(x). We then have p k (X, X) 0 for all 0 k < κ(d, M) + δ(x) p k (X, X) 0 k κ(d, M) + δ(x). The degeneration order δ(x) is dependent on φ and the geometry of X. For convenience, we also use the notation k 0(X) : κ(d, M) + δ(x) (5) to describe the smallest k 0 such that p k (X, X) 0. If φ is positive definite, we can conclude that p k0 (X)(X, X) > 0 holds for all X. With this notion, the formula (4) immediately yields Theorem [7] If x IR d and j {,..., M} are such that then the limit of u c j(x) for c 0 is the value of the polynomial If (6) fails, then the limit is infinite. p k (X, X j(x)) 0 for all k < k 0(X), (6) p k0 (X)(X, X j(x)). (7) p k0 (X)(X, X) In the paper [7] of Fornberg et. al. there are cases where (6) fails for certain geometries, e.g. when φ is a multiquadric (inverse or not), when the set X consists of 5 points on a line in IR and when the evaluation point x does not lie on that line. Strangely enough, the observations in [7] lead to the conjecture that the Gaussian is the only radial basis function where (6) never fails when data are on a line and evaluation takes place off that line. However, at the end of the paper we shall finish the proof of part of a related statement: Theorem Interpolation with scaled Gaussians always converges to the de Boor/Ron polynomial interpolant when the Gaussian widths increase. The proof needs a rather special technique, and thus we postpone it to the penultimate section, proceeding now with our investigation of convergence in general. Unfortunately, condition (6) contains an unsymmetric term, and we want to replace it by k 0(X j(x)) k 0(X), i.e. δ(x j(x)) δ(x), i.e..p k (X j(x), X j(x)) 0 for all k < k 0(X). (8) Then we can extend results by Fornberg et al. in [6, 7]. Theorem 3 If the degeneration order δ(x) of X is not larger than the degeneration order δ(x j(x)) of X j(x), then the polynomial limit of the Lagrange basis function u c j(x) for c 0 exists. In particular, convergence takes place when X is in general position with respect to φ. Proof: We assert boundedness of u c j(x) for c 0 and then use Theorem. Let us denote the standard power function for interpolation on data X and evaluation at x by P X(x) and let us write. φc for the norm in the native space of φ c (see e.g. [9] for a short introduction). Then Lemma The standard error bound of radial basis function interpolation yields the bound for all x IR d, all c > 0 and all j, j M. Proof: Zero is the interpolant to u c j on X j X \ {x j}. Lemma 3 for j M, c > 0. u c j(x) P Xj (x) u c j φc (9) u c j φ c det Ac,X j,x j det A c,x,x (0) 3

Proof: If α j is the coefficient of φ( x x j ) in the representation of u c j, we have u c j φ c α j, because of the general fact that an interpolant s to data f(x k ), k M has the native space norm s φ c α jφ c( x x j ) α jf(x j). Then (0) follows from Cramer s rule applied to the interpolation problem with Kronecker data δ jk, k M solved by u c j. Lemma 4 The power function has the representation φ c P X(x) det A c,x {x},x {x} det A c,x,x. Proof: By expansion of the numerator, using (4) and the representation P X(x) φ c(0) u c j(x)φ c( x x j ). This form is somewhat nonstandard. It follows from the optimality property of the power function, and it can be retrieved from [0], p. 9, (4.3.4). To finish the proof of Theorem 3, the above results yield and P X j (x j) det Ac,X,X, u c j φ det A c c,xj,x j PX j (x j) u c j(x) P Xj (x) u c j φc PX j (x) P Xj (x j). With the representation of the power function via determinants we get (u c j(x)) P X j (x) P X j (x j) det A c,x j (x),x j (x) det A c,x,x () The numerator and denominator of the right hand side contain sets of M points each. If we assume (8), we arrive at (u c j(x)) kk 0 (X j (x)) ck p k (X j(x), X j(x)) < kk 0 (X) ck p k (X, X) which concludes the proof of Theorem 3. Remark. The first part of () is an interesting bound on Lagrange basis functions in radial basis function interpolation. If the set X is formed recursively by adding to X the point x M+ where P X(x) is maximal (this adds the data location where the worst case error occurs), one gets a sequence of Lagrange basis functions that is strictly bounded by in absolute value. The implications on Lebesgue constants and stability of the interpolation process should be clear, but cannot be pursued here. 3 Basic Polynomial Determinants To derive a formula for the polynomials p k in (3) we need the expansion f(z) f k z k () k0 of f around the origin. If φ is positive definite, we know by the standard Bernstein-Widder representation (see [] for a short summary) that all ( ) k f k are positive. Furthermore, we use the standard notation for determinants M det (b ij) i,j M π S M ( ) π b jπ(j) 4

where π varies over all permutations in the symmetric group S M and ( ) π is the number of inversions in π. Then M det A c,x,y ( ) π f(c x j y π(j) ) π S M M ( ) π f mc m x j y π(j) m π S M m0 M ( ) ( ) π f ρj c ρ j x j y π(j) ρ j π S M ρ 0 ρ 0 ρ M 0 M f ρc ρ x j y π(j) ρ j with multi index notation π S M ( ) π ρ IN M 0 k0 c k ρ IN M 0 M f ρc ρ ( ) π x j y π(j) ρ j π S M f ρd ρ(x, Y ) ρ IN M 0 ρ k f ρ : M fρ j d ρ(x, Y ) : det ( ) x i y j ρ i p k (X, Y ) : f ρd ρ(x, Y ). ρ IN M 0 ρ k To see a bound on the degree, consider ρ k and conclude that i,j M d ρ(x, Y ) : det ( x i y j ρ i ) i,j M ( ) π π S M has total degree at most ρ k. Altogether we have M x j y π(j) ρ j Lemma 5 The polynomials p k (X, Y ) have maximal degree k as polynomials in X and Y. In Lemma we shall provide a better result, but it requires more tools. We can also deduce that k 0(X) for X M increases with M. In particular, we get ( ) k0(x) + d M d from Lemma 6 If p k (X, Y ) is nonzero for some special sets X, Y with M X Y, then M ( ) k+d d. Conversely, if M > ( ) k+d d, then pk (X, Y ) 0 for all sets X, Y with M X Y. Proof: If some d ρ(x, Y ) is nonzero for M X Y, there are M linearly independent d variate polynomials of degree at most ρ k. This proves the first assertion, because these polynomials span a space of dimension ( ) k+d d. The second assertion is the contraposition of the first. Example Let us look at some special cases that we prepared with MAPLE. We reproduce the results in [7], but we have a somewhat different background and notation. The D case with M has in general p 0(X, X) 0, p (X, X) f(0)f (0)(x x ). Thus κ(, ) and there is no degeneration except coalescence. The bound in Lemma 5 turns out to be sharp here. The case M 3 leads to κ(, 3) 3 with p 3(X, X) f (0)(3f(0)f (0) f (0) )(x x ) (x x 3) (x x 3). Geometrically, there is no degeneration except coalescence. The factor 3f(0)f (0) f (0) could possibly lead to some discussion, but for positive definite φ it must be positive because we know that f(0), f (0), f (0) and p 3(X, X) are positive. We find further κ(, 4) 6 with p 6(X, X) 4 3 (3f (0) 5f (0)f (0))(3f(0)f (0) f (0) ) (x x ) (x x 3) (x x 4) (x x 3) (x x 4) (x 3 x 4). 5

The general situation seems to be κ(, M) M(M )/ with p κ(,m) being (up to a factor) the polynomial that consists of a product of all (x j x k ) for j < k M, which is of degree κ(, M) M(M ). Thus the maximal degree in Lemma 5 is actually attained again. Note that the D situation also carries over to the case when X and X j(x) lie on the same line in R d. Now let us look at D situations. The simplest nontrivial D case is for M when the evaluation is not on the line connecting the points of X. But from the D case we can infer κ(, ) k 0(X) k 0(X j(x)) and do not run into problems, because we have Theorem 3. In particular, we find p (X, X) f(0)f (0)((x x ) + (y y ) ). Now we look at M 3 in D. The general expansion yields κ(, 3) with p (X, X) 4f(0)f (0) (det B X) and B X being the standard 3 3 matrix for calculation of barycentric coordinates based on X. Its determinant vanishes iff the points in X are collinear. Thus nondegeneracy of 3 point sets with respect to positive definite radial basis functions is equivalent to the standard notion of general position of 3 points in IR. To look for higher order degeneration, we consider 3 collinear points now, and since everything is invariant under shifts and orthogonal transformations, we can assume that the data lie on the x axis. This boils down to the D case, and we get p 3(X, X) > 0 with no further possibility of degeneration. But now we have to look into the first critical case, i.e. when X is collinear but X j(x) is not. This means that we evaluate the interpolant off the line defined by X. Theorem 3 does not help here. If we explicitly go back to (6), we still get convergence if we prove that p (X, X j(x)) 0 for all collinear point sets X and all x IR. Fortunately, MAPLE calculates p (X, X j(x)) 4f(0)f (0) (det B X)(det B Xj (x)) and thus there are no convergence problems. However, the ratio of the terms p 3(X, X j(x)) and p 3(X, X) now depends on φ. Now we go for M 4 in D and first find κ(, 4) 4 from MAPLE, but it cannot factor the polynomial p 4(X, X) properly or write it as a sum of squares. Taking special cases of 3 points not on a line, the polynomial p 4(X, X) seems to be always positive except for coalescence. In particular, it does not vanish for 4 non collinear points on a circle or a conic, as one would suspect. Taking cases of 3 points on a line, the polynomial p 4(X, X) vanishes iff the fourth point also lies on that line. Thus there is some experience supporting the conjecture that nondegeneracy of 4 points in D with respect to positive definite functions just means that the points are not on a line. But if they are on a line, we find k 0(X) 6 due to the D case, and thus p 5(X, X) also vanishes. This is confirmed by MAPLE, and we now check the case where the points of X are on a line but those of X j(x) not. It turns out that then (6) holds for k 0(X) 6, and the case does not show divergence. The M 5 situation in IR has κ(, 5) 6. The geometric interpretation of points in general position wrt. φ is unknown, because the zero set of p 6(X, X) is hard to determine in general. If 4 points are fixed at the corners of the square [0, ], and if the polynomial p6(x, X) is evaluated for inverse multiquadrics with 3 β as a function of the remaining point x 5 (ξ, η) IR, we get the nonnegative polynomial 3ξ ( ξ) + 3η ( η) + (ξ( ξ) + η( η)) which vanishes only at the corners of the square. Thus it can be ruled out that degeneracy systematically occurs when 4 or 5 points are on a circle or three points are on a line. However, it turns out that p 6(X, X) always vanishes if 4 points are on a line. The next coefficient p 7(X, X), if calculated for 4 points on a line, vanishes either if the fifth point also lies on the line, or for β 0,, 3, 7, or for coalescence. The final degeneration case thus occurs when all 5 points are on a line, and from D we then expect k 0(X) 0. Let us examine the divergence case described by Fornberg et. al. in [7]. It occurs when X consists of 5 points on a line, while evaluation takes place off that line. The D case teaches us that we should get k 0(X) 0 for 5 collinear points, and MAPLE verifies this, at least for the fixed 5 collinear equidistant points on [0, ] {0}. However, we also find that p 9(X, X (x)) 9 ( 8388608 η 5f (0)f (0) 3f (0) ) ( f(0)f (0)f (0) + f (0)f (0) f (0) f (0) ) 6

for points x (ξ, η) IR. If we put in multiquadrics, i.e. f(t) ( + t) β/, we get the same result as in [7], which reads 9 p 9(X, X (x)) 68435456 η β 4 (β 7)(β ) in our notation, proving that divergence occurs for multiquadrics except for the strange case β 7. Another curiosity is that for multiquadrics the value p 0(X, X) vanishes for the conditionally positive definite cases β 7 and β.790. As expected, this polynomial is positive for the positive definite cases, e.g. for negative β. Checking the case where exactly 4 points of X are on a line, we find that (6) holds for k 0(X) 7, and thus there is no convergence problem. 4 A Related Class of Polynomial Interpolation Methods We can avoid all convergence problems if we boldly take (7) to define u j(x) : p k 0 (X)(X, X j(x)) p k0 (X)(X, X) (3) for all j M and all x IR d. The denominator will always be positive if we start with a positive definite function, and the discussion at the beginning of section 3 shows that the polynomials p k (X, Y ) will always vanish if either X or Y have two or more coalescing points. Thus we get Lagrange interpolation polynomials for any kind of geometry. The result will be dependent on the function f and its Taylor expansion, and thus there is a full scale of polynomial interpolation methods which is available without any limit process. However, it is clear from (7) that polynomial limits of radial basis function interpolants, if they exist, will usually have the above form. It will be interesting to study how the technique of de Boor and Ron [, 3, 4] relates to this. However, it uses a different truncation strategy. Example Let us check how the above technique overcomes the five point degeneration case in Example. If we take the 5 equidistant points on [0, ] {0} and classical multiquadrics, the Lagrange basis function u 0 corresponding to the origin becomes u 0(ξ, η) 3 (ξ )(4ξ 3)(ξ )(4ξ ) + 8 ξη (8ξ 5), and the second term is missing if we take the Gaussian. For f(t) log( + t) the additional term is 3339 η (595 + 540ξ 44ξ + 008η ). There is dependence on f, but no degeneration. We simply ignore p 9 and focus on the quotient of values of p 0. 5 Point Sets, Polynomials, and Moments Our results so far require knowledge and numerical availability of k 0(X) and p k0 (X)(X, X j(x)). Section 3 gives a first idea for the evaluation of these quantities, but it still uses the limit process. It suggests that one looks at polynomials of the form x y l, and we shall use this section to make a fresh start into multivariate polynomials and point sets. The relation to the earlier sections will turn up later. Let IP m d be the space of all d variate real valued polynomials of order (degree+) up to m, and let X {x,..., x M } be a fixed set of M points in IR d. With the dimension Q ( ) m+d d and a basis p,..., p Q of IP m d we can form the Q M matrices P m and the M M matrices A l with P m : (p i(x j)) i Q, j M, A l (( ) l x j x k l ) j,k M, l 0 (4) to provide a very useful notion that is closely related to multivariate divided differences (see C. de Boor []): Definition A vector α IR M satisfies discrete moment conditions of order m with respect to X if P mα 0 or α jp(x j) 0 for all p IP m d holds. These vectors form a linear subspace MC m(x) : ker P m of IR M for M X. 7

Note that the definition involves all polynomials of order up to m, while the following involves radial polynomials of the form x x j l for 0 l < m. Theorem 4 A vector α IR M satisfies discrete moment conditions of order m with respect to X iff holds for all 0 l < m. α T A l α 0 (5) Note that the condition A l α 0 would be more restrictive. It will come up later. The proof of Theorem 4 uses Micchelli s lemma from [8], which we restate here because we make frequent use of its proof technique later. Lemma 7 If α IR M satisfies discrete moment conditions of order m, then the numbers α T A l α vanish for all l < m and α T A mα is nonnegative. The latter quantity vanishes iff α satisfies discrete moment conditions of order m +. Proof: Let us take a vector α IR M satisfying discrete moment conditions of order m, and pick any l m to form ( ) l α T A l α α jα k ( x j x k ) l k k k α jα k α jα k l +l +l 3 l l + l + l 3 l l + l 3 m l + l m x j l ( (x j, x k )) l x k l 3 x j l ( (x j, x k )) l x k l 3. This value vanishes for l < m, and this also proves one direction of the second statement, if we formulate it for m. For l m the two inequalities can only hold if l l 3. Thus ( ) l ( ) l, and we can write in multi index notation α T A l α If this vanishes, all expressions k α jα k ( ) l x j x k l l l 0 l l ZZ l l 0 l l ZZ l l 0 l l ZZ l l 0 l l ZZ l l l l k k α jα k x j l l x k l l (x j, x k ) l α jα k x j l l x k l l i IN 0 d i l i IN d 0 i l k x i jx i k i IN 0 d i l α jα k x j l l x k l l x i jx i k ( M α j x j l l xj) i 0. α j x j l l x i j with 0 l l, l l ZZ, i IN 0 d, i l must vanish, and this implies that α satisfies discrete moment conditions of order l + m +. It is now easy to prove Theorem 4. If α satisfies discrete moment conditions up to order m, Micchelli s lemma proves that (5) holds. For the converse, assume that (5) is true for some α IR M and proceed by induction. There is nothing to prove for order zero, and if we assume that we have the assertion up to order 8

m 0, then we use it to conclude that α satisfies discrete moment conditions of order m because it satisfies (5) up to l m. Then we apply Micchelli s lemma again on the level m, and since we have α T A m α 0, we conclude that α satisfies discrete moment conditions of order m. There is another equivalent form of discrete moment conditions, taking the form of degree reduction of linear combinations of high degree radial polynomials: Lemma 8 A vector α IR M satisfies discrete moment conditions of order m, iff for all l m the polynomials α k x x k l have degree at most l m. k Proof: Let us first assume that α IR M satisfies discrete moment conditions of order m. We look at k k k α k x x k l α k α k l +l +l 3 l l + l + l 3 l l + l 3 m x l ( (x, x k )) l x k l 3 and this is of degree at most l + l l l 3 l l m. x l ( (x, x k )) l x k l 3 We now prove the converse and apply the same idea as in the proof of Micchelli s lemma to get k α k x x k l k l +l +l 3 l l l 0 l l l + l 3 ZZ α k x l ( (x, x k )) l x k l 3 ( ) l i IN d 0 i l x l x i k α k x k l 3 x i k for arbitrary vectors α IR M. If this is a polynomial of degree at most l m, then all sums k α k x k l 3 x i k with l + l > l m or, equivalently, l + l 3 < m must vanish. Thus α satisfies moment conditions of order m. Note that in the above argument it suffices to pick just one l with l m. Thus Lemma 9 A vector α IR M satisfies discrete moment conditions of order m, iff for some l with l m the polynomial α k x x k l has degree at most l m. k Now we use that the discrete moment spaces for a finite point set X {x,..., x M } IR d form a decreasing sequence MC m+(x) MC m(x) MC 0(X) R M. (6) This sequence must stop with some zero space at least at order M, because we can separate M points always by polynomials of degree M, using properly placed hyperplanes. 9

Definition 3 For any finite point set X {x,..., x M } IR d there is a unique largest natural number µ µ(x) such that MC µ(x) {0} MC µ+(x). We call µ(x) the maximal discrete moment order of X. With this notion, the sequence (6) can be written as MC µ+(x) {0} MC µ(x) MC µ (X) MC 0(X) R M. (7) There is a fundamental observation linked to the maximal discrete moment order. Theorem 5 If there is a polynomial interpolation process based on a set X, it cannot work exclusively with polynomials of degree less than µ(x). Proof: If we take a nonzero vector α from MC µ, we see that it is in the kernel of all matrices P m from (4) for all m µ. Thus these matrices can have full column rank M only if m > µ.. The following notion is borrowed from papers of de Boor and Ron [, 3, 4]. Definition 4 We call a polynomial interpolation process for a point set X least, if it works with polynomials of degree at most µ(x). Remark: Below we shall see couple of least polynomial interpolation processes on X, including the one by de Boor and Ron. We now go back to where we started from, and relate µ(x) with the quantity k 0(X) defined in (5). Lemma 0 For all sets X and Y of M points in IR d we have and in particular k 0(X) µ(x). Proof: Take a vector ρ ZZ M 0 p k (X, Y ) 0 unless k µ(y ) and the matrix ( ) xi y j ρ i i,j M. We multiply by a nonzero vector α MC µ for µ : µ(y ) and get ρ i α j x i y j ρ i α j x i l l 0 x i l ( (x T i y j)) l y j l3 l +l +l 3 ρ i α j( (x T i y j)) l y j l 3 l + l 3 ρ i l l + l 3 µ for all i, i M. Since l + l 3 µ means ρ i µ l + l, this vanishes for those i where ρ i < µ. Thus the matrix can be nonsingular only if ρ i µ for some i, and this implies ρ µ. Since the polynomials p k (X, Y ) are superpositions of determinants of such matrices with ρ k, the assertion is proven. Lemma The Lagrange basis polynomials of (7) are of degree at most k 0(X) µ(x). Proof: We look at the above argument, but swap the meaning of X and Y there, replacing X by X j(x) and Y by X. The determinants vanish unless ρ µ(x), and the remaining terms are of degree at most l + l ρ i µ(x) ρ µ(x) ρ µ(x) k µ(x). We note that there is a lot of leeway between the result of Lemma () and the actually observed degrees of the p k0 (X)(X, X j(x)). The latter seem to be bounded above by µ(x) instead of k 0(X) µ(x). Theorem 6 k 0(X) µ(x) j(dim MC j dim MC j+) µ(x). (8) 0

Proof: Let us take a nonzero vector α MC µ and evaluate the quadratic form α T A c,x,xα c s f sα T A sα c s f sα T A sα s0 sµ c µ f µα T A µα + sµ+ c s f sα T A sα. By Courant s minimum maximum principle, this implies that A c,x,x has at least dim MC µ eigenvalues that decay at least as fast as c µ to zero for c 0. But there are no eigenvalues that decay faster than that. To see this, take for each c > 0 a normalized nonzero eigenvector α c such that the unique smallest eigenvalue α T c A c,x,xα c c s f sαc T A sα c : λ c s0 decays like c µ or faster. The coefficients α T c A sα c can increase with s only like ( diam(x)) s, and thus we have a stable limit of the analytic function λ c of c with respect to c 0. If we pick sequences of c s that converge to zero such that α c converges to some nonzero normalized vector α, we see that necessary α MC µ. But then λ c cannot decay faster than c µ for c 0. Going back to Courant s minimum maximum principle, we now know that A c,x,x has precisely dim MC µ eigenvalues that decay exactly like c µ to zero for c 0. We can now repeat this argument on the subspace of MC µ which is orthogonal to MC µ. For each nonzero vector of this space, the quadratic form decays like c µ, and there are dim MC µ dim MC µ linear independent vectors with this property. Now we look for arbitrary vectors α c that are orthogonal to the already determined dim MC µ eigenvectors of A c,x,x with eigenvalues of decay c µ, and we assume that they provide eigenvalues with fastest possible decay. This decay cannot be of type c µ or faster due to the assumed orthogonality, which allows passing to the limit. It must thus be of exact decay c µ. Induction now establishes the fact that for each j, 0 j µ there are dim MC j dim MC j+ eigenvalues of A c,x,x with exact decay like c j for 0. Thus the determinant decays exactly like the product of these, and this proves our assertion. Note that the above discussion fails to prove that the limiting polynomial interpolation process coming from a smooth radial basis function is least in all cases. We have to leave this problem open. Though k 0(X) will exceed µ(x), for instance in D situations, there is plenty of cancellation in the polynomials p k0 (X)(X, X j(x)) that we have not accounted for, so far. On the other hand, we have not found any example where the polynomial limit of a radial basis function interpolation is not of least degree. There is another interesting relation of µ to the spaces spanned by radial polynomials: Lemma Define the M vectors F l (x) : ( ) ( ) l x x k l. k M Then the M M(s + ) matrix with columns F l (x j), j M, 0 l s has full rank M if s µ, and µ is the smallest possible number with this property. Proof: Assume that the matrix does not have full rank M for a fixed s. Then there is a nonzero vector α IR M such that A l α 0 for all 0 l s and this implies discrete moment conditions of order s +. Thus s + µ. This teaches us that when aiming at interpolation by radial polynomials of the form x x k l one has to go up to l µ to get anywhere. But in view of Theorem 5 this reservoir of radial polynomials is way too large if we focus on the degree. We have to find a useful basis of an M dimensional subspace of polynomials of degree at most µ, if we want a least interpolation method. The following notion will be very helpful for the rest of the paper.

Definition 5 A discrete moment basis of R M with respect to X is a basis α,..., α M such that α j MC tj \ MC tj + for the decomposition sequence (7) and t 0... t M µ. Remark. A discrete moment basis α,..., α M of IR M can be chosen to be orthonormal, when starting with α M, spanning the spaces MC µ MC µ... M 0 R M one after the other. But there are other normalizations that make sense, in particular the one that uses conjugation via A l on MC l \MC l+, because this matrix is positive definite there due to Micchelli s lemma. There is a hidden theoretical and numerical connection of discrete moment bases to properly pivoted LU factorizations of matrices as in (4) of values of polynomials (see also the papers [, 3, 4], of de Boor and Ron), but we shall neither go into details nor require the reader to figure this out before we proceed. We now consider the polynomials v j(x) : α j i x xi t j, j M (9) i that are of degree at most t j µ due to Lemma 9 and the definition of the discrete moment basis. They are low degree linear combinations of radial polynomials, and their definition depends crucially on the geometry of X. Lemma 3 The M M matrix with entries v j(x k ) is nonsingular. Proof: We multiply this matrix with the nonsingular M M matrix containing the discrete moment basis α,..., α M and get a matrix with entries γ jm i k α j i αm k x i x k t j for j, m M. Consider m > j and use t m t j to see that γ jm 0 as soon as t m > t j, because the entries can be written as values of a polynomial of degree t j t j t m < 0. Thus the matrix is block triangular, and the diagonal blocks consist of entries α j i αm k x i x k t with t t j t m. But these symmetric submatrices must be definite due to our construction of a discrete moment basis. We even could have chosen the basis such that the diagonal blocks are unit matrices, if we had used conjugation with respect to A t. Theorem 7 A least polynomial interpolation on X is possible using the functions v j of Lemma 3. These are of degree at most µ µ(x), and thus ( ) ( ) ( ) µ(x) + d µ(x) + d k0(x) + d < M. d d d Example 3 Let us look at the special case with M 4, d and points x : (0, 0) T, x : (, 0) T, x 3 : (0, ) T, x 4 : (/, ) T. The discrete moment conditions on vectors α IR 4 are 4 α j 0 for all α MC, 4 α j 0, α + α 4/ 0, α 3 + α 4 0 for all α MC. Furthermore we find MC 3 {0}, MC span {α 4 : (,,, ) T }, MC 0 IR 4 and MC \ MC span {α : (,, 0, 0) T, α 3 : (, 0,, 0) T }, such that a discrete moment basis of IR 4 can be formed by α : (, 0, 0, 0) T with α, α 3, and α 4 together with t 0 0 < t t < t 3 µ. From Theorem 6 we conclude that k 0(X) + 4. MAPLE confirms this, and the Lagrange basis of the form (3) comes out to be quadratic for all f that one could start with, but the result depends on f. For example, the Lagrange basis function for the origin is 9 x + 8 9 xy 7 9 x y for f(t) e t φ Gaussian ( 37 + 3xy 35y 0x y 7x ) for f(t) /( + t) φ inverse multiquadric. 37

The Gaussian case coincides with the de Boor/Ron solution from section 6 of [3]. The method based on (9) yields the basis function ( 9 3x 7y 6x y + 6xy ). 9 Thus we have different methods, but we note that the de Boor/Ron interpolation method coincides with the limit of interpolation with shifted and scaled Gaussians. We shall prove this in the penultimate section. 6 Polynomial Reproduction By (4), the Lagrange basis functions, if they exist, have analytic expansions u c j(x) b jm(x)c m m0 for small c with certain multivariate polynomials b jm depending on X and φ. We put this and the expansion () into the defining equations () to get b jm(x)c m f l c l x j x k l m0 s0 c s s l0 f l l0 b j,s l (x) x j x k l s0 s0 f sc s x x k s, k M f sc s x x k s, k M and by comparison of coefficients we arrive at s l0 f l b j,s l (x) x j x k l f s x x k s, k M, s 0. (0) In view of Lemma it seems to be the upshot of (0) that the reproduction of all vectors F s(x) of radial polynomials (where for this section we introduce the factors f l instead of ( ) l into the A l matrices) is possible from the data of all F l (x j), j M, 0 l s. This works if s µ by that Lemma, but the reconstruction is not unique, and equations (0) describe a special selection of a reconstruction. It is strange that the reconstruction via the above equations always works if convergence of the radial basis function interpolation takes place, also for small s < µ, and that it can be chosen to have a convolution type structure. We are interested in the special Lagrange basis polynomials u 0 j(x) b j0(x), j M with b j0(x k ) δ jk, j, k M that arise in (0). We want to derive properties and defining equations. To this end, define the set { } D X,φ : (α, s) IR M IN 0 : f l α k x j x k l 0 for all 0 l < s, j M and polynomials p α,s(x) : f s k α k x x k s for all (α, s) D X,φ. () Applying such a linear combination to (0) with respect to the points x k yields k b j,0(x)p α,s(x j) p α,s(x), (α, s) D X,φ. () Theorem 8 The limiting polynomial interpolation process reproduces all polynomials of the form (). Lemma 4 If f 0 0, the limiting polynomial interpolation process reproduces constants. Proof: Consider elements (α, 0) D X,φ in (). Lemma 5 If f 0, f 0, the limiting polynomial interpolation process reproduces all linear functions in the linear span of the data. 3

Proof: Consider elements (α, ) D X,φ in (). The restriction is k α k 0, and due to Lemma 4 we have reproduction of constants. Thus we can reproduce any polynomial of the form p(x) x x k x x l x k + x l x T (x l x k ) for k < l M. Let us generalize this. For all s 0, define the spaces and the maps T s D s : { α IR M : (α, s) D X,φ }, Rs : span {p α,s : α D s} (3) : D s R s with T s(α) : p α,s for all α D s. We have T s(d s) R s by definition, and we find that ker T s D s+. Note further that D 0 IR M. This yields a decomposition sequence of IR M into a sequence of spaces isomorphic to R 0, R,... and we have to see whether this sequence exhausts all of IR M. The decomposition can only fail if D s D t for some s and all t s, while D s still is nonzero. But this cannot happen, because then there is some nonzero α IR M which is in all D s for all s 0. But then 0 α jα k φ( x j x k ) j,k and this cannot hold for positive definite functions φ. Thus we know that there is a minimal finite decomposition D σ+ {0} D σ... D D 0 R M (4) with factor spaces R s T s(d s) D s/d s+ for 0 s σ. Unfortunately, the spaces R 0, R,... turn out to be overlapping, and we only know that σ R : is a polynomial space of dimension at most M that is reproduced by our interpolation. s0 The relation to discrete moments is that the D l spaces are the intersection of kernels of matrices A 0,..., A l. This is a different type of discrete moment condition, and in general σ µ. See Example 4 below. If () were sufficient to determine the functions b j0 completely, one could cancel the factors f s for positive definite functions and get a construction technique that is even independent of f and φ. But by looking at cases like in Example 3 for positive definite functions one can see that the results actually depend on f and φ. Thus one has to go back to (0). Here, one could write the equations for k M and 0 s S for some large positive S µ(x), and then by Lemma one has a solvable system of M(S + ) equations for M(S + ) unknowns. However, it turns out that there is some serious rank loss, while for large S the particular functions b j0 come out uniquely when running a symbolic MAPLE program, leaving many b jl for larger l undetermined. An example follows below. It is a challenging open problem to find a good technique to determine the b jl for all j M and small l 0 in a finite and stable way. We finally want to collect more information on the polynomials that are reproduced. Define, by splitting (0), the polynomials z k,s (x) : f s x x k s s f l l0 R s b j,s l (x) x j x k l f s for all k M, s 0. Then z k,s (x j) f s x j x k s, j, k M and (0) implies b j,0(x)z k,s (x j) b j,0(x)f s x j x k s f s x x k s z k,s (x), s f l l0 b j,s l (x) x j x k l b j,0(x) x j x k s (5) 4

proving that all the z k,s (x) are reproduced. There are infinitely many of them, but they span a space of dimension at most M, and thus their degree cannot increase indefinitely with s. So far, all examples indicate that the limiting polynomial interpolants have degrees not exceeding the maximal discrete moment order µ, but a proof is still missing. Example 4 Let us continue our previous case in Example 3 of 4 points in D in view of the above terminology. Looking at σ instead of µ, we find that D {0}, D MC, D 0 IR 4 MC 0, and thus µ > σ. We now look at (0) and () to determine the polynomials b j0(x), j 4. Here, we assume all f l to be nonzero. The system () leads to no more than 3 equations, and we can use (α, 0), (α, ), (α 3, ) D X,φ as pairs (α, s) there. The result consists of the three equations that describe reproduction of linear polynomials: b 0(x, y) + b 0(x, y) + b 30(x, y) + b 40(x, y) b 0(x, y) + b 40(x, y)/ x b 30(x, y) + b 40(x, y) y. (6) We thus turn to (0) in general, and discuss the cases s 0,,,... one after the other. To start with, the case s 0 yields the first equation of (6). Assuming that this condition is already satisfied, we consider s of (0) for α MC and find the two other equations of (6). Taking these for granted also, we consider a general α there and get a new equation f (x, y) f 4 b j0(x, y) x j + f 0 4 b j(x, y) that contains the new polynomial b (x, y) : 4 bj(x, y) and has the explicit form ( f (x + y ) f b 0(x, y) + b 30(x, y) + 5 ) 4 b40(x, y) + f 0b (x, y). We now have a new equation, but also a new variable. Thus the cases s 0 and s do not yet lead to a unique determination of the b j0. We have to turn to s in (0), but we do not want to introduce more variables than absolutely necessary. We find f 4 4 α k (x, y) x k 4 f b j0(x, y) k + f + f 0 4 b j(x, y) 4 b j(x, y) and we would like to find a nonzero vector α IR 4 such that 4 α k 0, k 4 α k x j x k 4 k 4 α k x j x k k 4 k α k 4 α k x j x k c for all j, j 4 k with a constant c, because then we would have no new variable. Fortunately, the vector α 4 (,,, ) T spanning MC does the job with c /, and with our new equation f 4 4 αk (x, 4 y) x k 4 f b j0(x, y) k f b(x, y) 4 αk x 4 j x k 4 we now have five equations for five variables. This is the point where the solution of (0) leads to unique determination of the b j0 and the sum over the b j. Note that we are still far from determining all the b jl, but we are interested in the b j0 only and succeeded to get away with small s in (0). This principle worked in all cases that we tested with MAPLE, and it directly yielded the Lagrange polynomial bases via symbolic computation. k 5

7 The method of de Boor and Ron revisited The goal of this section is to prove Theorem. For this we require at least a scaled version of the de Boor/Ron technique, and we take the opportunity to rephrase with slightly increased generality. For all α ZZ 0 d let w α be a positive real number, and consider the inner product (p, q) w : (D α p)(0)(d α q)(0) w α α Z d 0 on the space IP d of all d variate polynomials. The de Boor/Ron interpolant arises in the special case w α. Now we want to link the theory to radial basis function techniques. If we assume w α < α Z d 0 and define the kernel K w(x, y) : α Z d 0 x α y α w α, (7) all polynomials p IP d are reproduced on [, ] d via p(x) (p, K w(x, )) w (8) and this identity carries over to the Hilbert space completion H w : g C (IR d ) : g(x) (D α g)(0) xα, (D α g) (0) < w α α Z d 0 of the polynomials under the above inner product. The kernel K w is positive definite on [, ] d, and larger domains can be treated by scaling. Since polynomials separate points, it is clear that for all finite sets X {x,..., x M } [, ] d we have linear independence of the functions K w(, x j), and interpolation in X by the span of these functions is uniquely possible. So far, we used standard arguments of radial basis function theory. In the papers of de Boor and Ron, transition to a polynomial interpolation process is done via truncation, not via passing to the limit of a scaling. For all functions g from H w the notation g [k] (x) : (D α g)(0) xα α ZZ 0 d α k is introduced, while g stands for the nonzero function g [k] with minimal k. For a finite set X {x,..., x M } [, ] d the spaces α Z d 0 E w,x : span {K w(x, ) : x X}, P w,x : span {g : g E w,x} are introduced, and P w,x is a space of polynomials. Theorem 9 Interpolation on X by functions in P w,x is uniquely possible. Proof: Assume that there is some nonzero p P w,x such that p g p and p X 0. Then (8) yields orthogonality (p, g) w 0 for all g E w,x, and we have the contradiction 0 (p, g p) w (g p, g p) w (g p, g p ) w. So far, we have followed the proof of de Boor and Ron, but now we want to use a discrete moment basis α,..., α M to link the process with what we have done in previous sections. We define functions v r(y) : αjk r α y w(x j, y) w α α t r αj r x α j, gr : vr, r M. (9) 6

Due to the property of a discrete moment basis we see that not all of the quantities c α,r : αj r x α j for α t r can vanish, because otherwise α r MC tr+. Thus we have the homogeneous representation α y g r(y) v r w αc α,r (30) α t r and (g r, g s) w 0 for t r t s. The matrix formed by the (g r, g s) w is a positive semidefinite block diagonal Gramian. To prove its definiteness, we can focus on a single diagonal block with t t r t s. Collecting the indices r with t r t into a set I t, we assert linear independence of the functions g r for r I t. For a vanishing linear combination 0 γ rg r(y) r I t y α w αc α,r r I t γ r α t α t α t α t y α wα y α wα y α wα r I t γ rc α,r r I t γ r ( αj r x α j r I t γ rα r j we conclude that r I t γ rα r is a vector in MC t+, and this can hold only if the coefficients are zero. Thus the space P w,x contains the M linearly independent homogeneous polynomials g,..., g M of increasing degrees 0 t t M µ, and the theorem is proven. Due to Theorem 5, the degree is least, as known from the de Boor/Ron papers. We now proceed towards proving that the limit of interpolants by Gaussians is equal to the de Boor/Ron polynomial interpolant. We need something that links kernels of the form (7) to radial kernels. Lemma 6 If φ is a positive definite analytic radial basis function that can be written via an analytic function f satisfying () and (), then for all x, y IR d. f(x T y) α Z d 0 Proof: We use the Bernstein Widder representation to get f(r) f α (0) x α y α ( ) j f (j) (0) ( ) j j!f j 0 α Z d 0 e rt dµ(t), r 0 and similarly, factoring the exponential in the integral, and 0 ) x α j f α x α y α (3) t j dµ(t) (0, ) for all j 0 Dx α f(x T y) y α ( ) α t α e txt y dµ(t) 0 D α x f(x T y) x0 y α f α (0) y α f α where Dx α takes derivatives of order α with respect to x. The assertion follows from putting the result into the power series expansion at zero. At first sight, the above result is disappointing, because one cannot easily use (3) in (7), since the coefficients in (3) are alternating. However, a closer look reveals that the major part of the de Boor/Ron 7

theory does not rely on the signs of the coefficients. It is the link to radial basis functions as reproducing kernels that does not work without further arguments. This has a positive consequence: the generalized de Boor/Ron approach as given at the start of the section will yield many new cases of positive definite non radial interpolants with polynomial truncations that furnish least polynomial interpolants. On the downside, we cannot expect to find a direct link between interpolation by general positive definite radial basis functions and the generalized de Boor/Ron method. But for Gaussians, we can add some more work, factoring exp( c x j x k ) exp( c x j ) exp(cx T j x k ) exp( c x k ). We then rewrite the Lagrange system u c j(x) exp( c x j x k ) exp( c x x k ), k M in the form zj c (x) exp(cx T j x k ) u c j(x) exp( c( x j x )) exp(cx T j x k ) exp(cx T x k ), k M with z c j (x) : u c j(x) exp( c( x j x )), j M. Now we can use the technique of the previous lemma directly, putting in the expansions for the exponential and working with the kernel K c(x, y) : exp(cx t y) in our presentation of the de Boor/Ron technique, with a slight abuse of notation. This implies that the functions z c j (x) are Lagrangian interpolants in the span of the K c(x j, x) exp(cx t jx). Any interpolation by scaled Gaussians can be converted by the above transformation to and from an interpolation using the kernel K c. We now look at what happens if the de Boor/Ron truncation process is carried out on interpolants defined via K c. The functions in (30) come out as gr(y) c w c α y αc α,r α t r α (c) y α c α,r α t r y α (c) tr cα,r α t r (c) tr g dbr r i.e. they are just scalar multiples of the functions gr dbr of the de Boor/Ron process. Thus the polynomial space spanned by truncation of the K c is independent of c and coincides with the de Boor/Ron polynomial interpolation space. We have successfully moved from interpolation by Gaussian radial basis functions to interpolation by scaled exponentials, and we have seen that the truncation of the latter is the de Boor/Ron polynomial space. But we now have to investigate the limit of the interpolants spanned by the scaled exponentials K c(x j, ) for c 0 to see whether they converge towards the de Boor/Ron truncation. We go back to (9) to define functions v c r as such that v c r(y) : v lim r(y) c c 0 (c) tr (y) αjk r c(x j, y) t (c) r α t r (c) s s t r α s y α cα,r y α cα,r gdbr r (y), r M. This means that the space spanned by the K c(x j, ) contains a basis that converges towards a basis of the de Boor/Ron polynomial space for c. If the Lagrange basis for interpolation in the span of the K c(x j, ) is written in terms of this basis, it converges towards a polynomial limit. This ends the proof of Theorem. 8

8 Preconditioning The transition from the Gaussian system to the scaled basis vc r (y) (c) tr should be useful as a preconditioning technique. In general, we show in this section how to use a discrete moment basis for preconditioning badly conditioned matrices arising from interpolation by smooth radial basis functions. We go back to the beginning of the paper and precondition the matrix A c,x,x arising in () by a scaled discrete moment basis c t/ α,..., c t M / α M (3) in the following way. If we put the discrete moment basis into an M M matrix B c and form the positive definite symmetric matrix Z(c) : B ca c,x,xbc T, the matrix entries will be z rs(c) : αjc r tr/ αkc s ts/ φ c( x j x k ) r, s M with well defined limits z rs(0) j,k f nc n tr/ ts/ n0 n 0 n t r + t s j,k f nc n tr/ ts/ α r jα s k x j x k n j,k α r jα s k x j x k n { M f(tr+t s)/ j,k αr jαk x s j x k tr+ts t r + t s even 0 else for c 0. The matrix Z(0) is positive semidefinite by construction, and we assert Theorem 0 The matrix Z(0) is positive definite. Proof: We use the proof technique of Theorem 6. The product of all eigenvalues of A c,x,x decays with exponent k 0(X) as in (8), while the maximum eigenvalue stays bounded above independent of c. But our matrix transformation performs a multiplication of the spectral range by c k0(x), because k 0(X) M tj is just another way to write (8). Thus the smallest eigenvalue of the product must stay away from zero when c 0. But since the matrix Z(0) is well defined, the maximal eigenvalue of the product Z(c) must be bounded, and Z(0) altogether has a strictly positive spectrum. Example 5 If we go back to the four points in IR of Examples 3 and 4 and scale the discrete moment basis as in (3) via 0 0 0 B c : c c 0 0 c 0 c 0, MAPLE produces a limit matrix for Gaussians and c c c c 0 0 0 0 0 Z(0) : 0 0 0. 9 0 0 4 Z(0) : 0 0 4 0 0 0 0 0 0 0 0 9 4 6 for (negative definite) inverse multiquadrics with β. If we take four equidistant points on the line [0, ] {0}, we find 0 0 9 0 0 4 Z(0) : 9 7 4 0 0 9 7 0 4 7 0 40 43 for B c : 0 0 0 c c 0 0 c c c 3 c c c } 0 c 3 c c c c in case of Gaussians. The discrete moment basis now contains divided differences, and the zero structure is different from the previous case, because we have t j j, j 4 here. Acknowledgements: There were some very helpful e mail exchanges with Bengt Fornberg.. 9