Improved Fast Gauss Transform and Efficient Kernel Density Estimation

Size: px
Start display at page:

Download "Improved Fast Gauss Transform and Efficient Kernel Density Estimation"

Transcription

1 Improved Fast Gauss Transform and Efficient Kernel Density Estimation Cangjiang Yang, Ramani Duraiswami, Nail A. Gumerov and Larry Davis Perceptual Interfaces and Reality Laboratory University of Maryland, College Park, MD 07, USA Abstract Evaluating sums of multivariate Gaussians is a common computational task in computer vision and pattern recognition, including in te general and powerful kernel density estimation tecnique. Te quadratic computational complexity of te summation is a significant barrier to te scalability of tis algoritm to practical applications. Te fast Gauss transform (FGT) as successfully accelerated te kernel density estimation to linear running time for lowdimensional problems. Unfortunately, te cost of a direct extension of te FGT to iger-dimensional problems grows exponentially wit dimension, making it impractical for dimensions above 3. We develop an improved fast Gauss transform to efficiently estimate sums of Gaussians in iger dimensions, were a new multivariate expansion sceme and an adaptive space subdivision tecnique dramatically improve te performance. Te improved FGT as been applied to te mean sift algoritm acieving linear computational complexity. Experimental results demonstrate te efficiency and effectiveness of our algoritm. Introduction In most computer vision and pattern recognition applications, te feature space is complex, noisy and rarely can be described by te common parametric models [7], since te forms of te underlying density functions are in general unknown. In particular, data in ig-dimensional feature space becomes more sparse and scattered, making it muc more difficult to fit tem wit a single ig-dimensional density function. By contrast, witout te assumption tat te form of te underlying densities are known, nonparametric density estimation tecniques [, 0] ave been widely used to analyze arbitrarily structured feature spaces. Te most widely studied and used nonparametric tecnique is kernel density estimation (KDE), first introduced by Rosenblatt [], ten discussed in detail by Parzen [0] and Cacoullos [3]. In tis tecnique te density function is estimated by a sum of kernel functions (typically Gaussians) centered at te data points. A bandwidt associated wit te kernel function is cosen to control te smootness of te estimated densities. In general, more data points allow a narrower bandwidt and a better density estimate. Many approaces in computer vision and pattern recognition use kernel density estimation, including support vector macines [3], M-estimation [], normalized cut [] and mean sift analysis [5]. Wit enoug samples, te kernel density estimates provably converge to any arbitrary density function. On te oter and, te number of samples needed may be very large and muc greater tan would be required for parametric models. Moreover, te demand for large number of samples grows rapidly wit te dimension of te feature space. Given N source data points, te direct evaluation of densities at M target points takes O(MN) time. Te large dataset also leads to severe requirements for computational time and/or storage. Various metods ave been proposed to make te process of kernel density estimation more efficient. Te existing approaces can be rougly divided into two categories. One is based on te k-nearest neigbor searcing, were spatial data structures and/or branc and bound are employed to acieve te computational saving [,, 0, 9]. One is based on te fast Fourier transform (FFT) for evaluating density estimates on gridded data wic, owever, are unavailable for most applications [5]. Recently te fast multipole metod (FMM) and fast Gauss transform (FGT) ave been used to reduce te computational time of kernel density estimation to O(M + N) time, were te data are not necessarily on grids [5, ]. As faster computers and better video cameras become ceaper, te collection of sufficient data is becoming possible, wic results in a steady increase in te size of te dataset and te number of te features. Unfortunately te existing approaces including te fast Gauss transform suffer from te curse of dimensionality. Te complexity of computation and storage of te FGT grows exponentially wit dimension. In tis paper, we proposed an improved fast Gauss transform (IFGT) to efficiently evaluate te sum of Gaussians in iger dimensions. By iger dimensions,

2 we mean dimensions up to ten. Suc ig dimensional spaces are commonly used in many applications suc as in video sequence analysis and eigenspace based approaces. We also sow ow te IFGT can be applied to te kernel density estimation. Specifically te mean sift algoritm [,, 5] is cosen as a case study for te IFGT. Te mean sift algoritm is based on te KDE and recently rediscovered as a robust clustering metod. However, te mean sift algoritm suffers from te quadratic computational complexity, especially in iger dimensions. Te proposed IFGT successfully reduced te computational complexity into linear time. FMM and FGT Te fast Gauss transform introduced by Greengard and Strain [5, ] is an important variant of te more general fast multipole metod [3, ]. Originally te FMM was developed for te fast summation of potential fields generated by a large number of sources, suc as tose arising in gravitational or electrostatic potential problems in two or tree dimensions. Tereafter, tis metod was extended to oter potential problems, suc as tose arising in te solution of te Helmoltz and Maxwell equations, tose in cemistry and interpolation of scattered data [].. Fast Multipole Metod We briefly describe te FMM ere. Consider te sum v(y j ) = u i φ i (y j ), j =,...,M. () Direct evaluation requires O(M N) operations. In te FMM, we assume tat te functions φ i can be expanded in multipole (singular) series and local (regular) series tat are centered at locations x and y as follows: φ(y) = b n (x )S n (y x ) + ǫ(p), () φ(y) = a n (y )R n (y y ) + ǫ(p), (3) were S n and R n respectively are multipole (singular) and local (regular) basis functions, x and y are expansion centers, a n,b n are te expansion coefficients, and ǫ is te error introduced by truncating a possibly infinite series after p terms. Te operation reduction trick of te FMM relies on expressing te sum () using te series expansions () and (3). Ten te reexpansion for (3) is: v(y j ) = u i φ i (y j ) = c ni R n (y j x ), () u i p for j =,...,M. A similar expression can be obtained for (). Consolidating te N series into one p term series, by rearranging te order of summations, we get [ N v(y j )= u i c ni ]R n (y j x )= C n R n (y j x ). (5) Te single consolidated p term series (5) can be evaluated at all te M evaluation points. Te total number of operations required is ten O(Mp + Np) O(Np) for N M. Te truncation number p depends on te desired accuracy alone, and is independent of M, N. Te functions φ i in te FMM are not valid over te wole domain. So te singular expansions () are generated around clusters of sources. In a fine-to-coarse pass, te generated coefficients are translated into coarser level singular expansions troug a tree data structure by translation operators. In a coarse-to-fine pass, te coefficients of te singular expansions at coarser level are converted via a sequence of translations to coefficients of regular expansions at finer levels, ten evaluated at eac evaluation point.. Fast Gauss Transform Te fast Gauss transform was introduced in [5] for efficient computation of te weigted sum of Gaussians G(y j ) = q i e yj xi / () were q i are te weigt coefficients, {x i },...,N are te centers of te Gaussians (called sources ), is te bandwidt parameter of te Gaussians. Te sum of te Gaussians is evaluated at eac of te target points {y j } j=,...,m. Direct evaluation of te sum at M target points due to N sources requires O(MN) operations. Te original FGT directly applies te FMM idea by using te following expansions for te Gaussian: e y xi / = e y xi / = n! ( ) n ( xi x y x n n! n ) + ǫ(p), (7) ( )( ) n xi y y y + ǫ(p), were te Hermite functions n (x) are defined by n (x) = ( ) n dn dx n ( e x). Te two expansions (7) and () are identical, except tat te arguments of te Hermite functions and te monomials (Taylor series) are flipped. Te first is used as te counterpart of te multipole expansion, wile te second is used as te local expansion. Te FGT ten uses tese ()

3 expansions and applies te FMM mecanism to acieve its speedup. Conversion of a Hermite series into a Taylor series is acieved via a translation operation. Te error bound estimate given by Greengard and Strain [5] is incorrect, and a new and more complicated error bound estimate was presented in []. Te extension to iger dimensions was done by treating te multivariate Gaussian as a product of univariate Gaussians, applying te series factorizations (7) and () to eac dimension. For convenience s sake, we adopt te multiindex notation of te original FGT papers [5]. A multiindex α = (α,...,α d ) is a d-tuple of nonnegative integers. For any multi-index α N d and any x R d, we ave te monomial x α = x α xα xα d d. Te lengt and te factorial of α are defined as α = α + α α d, α! = α!α! α d!. Te multidimensional Hermite functions are defined by α (x) = α (x ) α (x ) αd (x d ). Te sum () is ten equal to te Hermite expansion about center x : G(y j ) = ( ) yj x C α α, (9) α 0 were te coefficients C α are given by C α = α! ( ) α xi x q i. (0) Te FGT in iger dimensions is ten just an accumulation of te product of te Hermite expansions along eac dimension. If we truncate eac of te Hermite series after p terms (or equivalently order p ), ten eac of te coefficients C α is a d-dimensional matrix wit p d terms. Te total computational complexity for a single Hermite expansion is O((M + N)p d ). Te factor O(p d ) grows exponentially as te dimensionality d increases. Despite tis defect in iger dimensions, te FGT is quite effective for two and tree-dimensional problems, and as already acieved success in some pysics, computer vision and pattern recognition problems [, ]. Anoter serious defect of te original FGT is te use of te box data structure. Te original FGT subdivides te space into boxes using a uniform mes. However, suc a simple space subdivision sceme is not appropriate in iger dimensions, especially in applications were te data migt be clustered on low dimensional manifolds. First of all, it may generate too many boxes (largely empty) in iger dimensions to store and manipulate. Suppose te unit box in 0 dimensional space is divided into tents along eac dimension, tere are 0 0 boxes wic may cause trouble in storage and waste time on processing empty boxes. Secondly, and more importantly, aving so many boxes makes it more difficult for searcing nonempty neigbor boxes. Finally, and most importantly te worst property of tis sceme is tat te ratio of volume of te ypercube to tat of te inscribed spere grows exponentially wit dimension. In oter words, te points ave a ig probability of falling into te area inside te box and outside te spere. Te truncation error of te above Hermite expansions (7) and () are muc larger near te boundary tan near te expansion center, wic will bring large truncation errors on most of te points. In brief, te original FGT suffers from te following two defects tat are te motivation beind tis paper:. Te exponential growt of complexity wit dimensionality.. Te use of te box data structure in te FGT is inefficient in iger dimensions. 3 Improved Fast Gauss Transform 3. A Different Factorization Te defects listed above can be tougt as a result of applying te FMM metodology to te FGT blindly. As sown in section, te FMM was developed for singular potential functions wose forces are long-ranged and nonsmoot (at least locally), ence it is necessary to make use of te tree data structures, multipole expansions, local expansions and translation operators. In contrast, te Gaussian is far from singular it is infinitely differentiable! Tere is no need to perform te multipole expansions wic account for te far-field contributions. Instead we present a simple new factorization and space subdivision sceme for te FGT. Te new approac is based on te fact tat te Gaussian, especially in iger dimensions, decays so rapidly tat te contributions outside of a certain radius can be safely ignored. Assuming we ave N sources {x i } centered at x and M target points {y j }, we can rewrite te exponential term as e yj xi / = e yj / e xi / e yj xi/, () were y j = y j x, x i = x i x. In expression () te first two exponential terms can be evaluated individually at eiter te source points or te target points. Te only problem left is to evaluate te last term were sources and targets are entangled. One way of breaking te entanglement is to expand it into te series e yj xi/ = Φ n ( y j )Ψ n ( x i ), ()

4 were Φ n and Ψ n are te expansion functions and will be defined in te next section. Denoting φ( y j ) = e yj /, ψ( x i ) = e xi /, we can rewrite te sum () as G(y j ) = q j φ( y j )ψ( x i ) Φ n ( y j )Ψ n ( x i ). (3) If te infinite series () absolutely converges, we can truncate it after p terms so as to obtain a desired precision. Excanging te summations in (3), we obtain G(y j ) = φ( y j ) C n Φ n ( y j ) + ǫ(p), () C n = q i ψ( x i )Ψ n ( x i ). (5) Te factorization () is te basis of our algoritm. In te following sections, we will discuss ow to implement it in an efficient way. 3. Multivariate Taylor Expansions Te key issue to speed up te FGT is to reduce te factor p d in te computational complexity. Te factor p d arises from te way tat te multivariate Gaussian is treated as te product of univariate Gaussian functions and expanded along eac dimension. To reduce tis factor, we treat te dot product in () as a scalar variable and expand it via te Taylor expansion. Te expansion functions Φ and Ψ are expressed as multivariate polynomials. We denote by Π d n te space of all real polynomials in d variables of total degree less tan or equal to n; its dimensionality is r nd = ( ) n+d d. To store, manipulate and evaluate te multivariate polynomials, we consider te monomial representation of polynomials. A polynomial p Π d n can be written as p(x) = C α x α, C α R. () α n It is computationally convenient and efficient to stack all te coefficients into a vector. To store all te r nd coefficients C α in a vector of lengt r nd, we sort te coefficient terms according to Graded lexicograpic order. Graded refers to te fact tat te total degree α is te main criterion. Graded lexicograpic ordering means tat te multi-indices are arranged as (0,0,...,0),(,0,...,0),(0,,...,0),...,(0,0,...,), (,0,...,0),(,,...,0),...,(0,0,...,),...,(0,0,...,n). Te power of te dot product of two vectors x and y can be expanded into multivariate polynomial: (x y) n = ( ) n x α y α, (7) α α =n a b c a b c a b c a ab ac b bc c a b c a 3 a b a c ab abc ac b 3 b c bc c 3 Figure : Efficient expansion of te multivariate polynomials. Te arrows point to te leading terms. were ( ) n α = n! α! α d! are te multinomial coefficients. So we ave te following multivariate Taylor expansion of te Gaussian functions: e x y = α 0 α α! xα y α. () From Eqs.(), () and (), te weigted sum of Gaussians () can be expressed as a multivariate Taylor expansions about center x : G(y j ) = α 0 C α e yj x / ( yj x were te coefficients C α are given by C α = α α! ) α, (9) ( ) α q i e xi x / xi x. (0) If we truncate te series after total degree p, te number of te terms r p,d = ( ) p+d d is muc less tan p d in iger dimensions (as sown in Table ). For instance, wen d = and p = 0, te original FGT needs 0 terms, te multivariate Taylor expansion needs only For d and moderate p, te number of terms becomes O(d p ), a substantial reduction. One of te benefits of te graded lexicograpic order is tat te expansion of multivariate polynomials can be computed efficiently. For a d-variate polynomial of order n, we can store all terms in a vector of lengt r nd. Starting from te order zero term (constant ), we take te following steps recursively. Assume we ave already evaluated terms of order k. Ten terms of order k can be obtained by multiplying eac of te d variables wit all te terms between te variable s leading term and te end, as sown in te Figure. Te required storage is r nd and te computations of te terms require r nd multiplications. 3.3 Spatial Data Structures As discussed above, we need to subdivide space into cells and collect te influence of te sources witin eac

5 Table : Number of terms in d-variate Taylor expansion truncated after order p. p\d cell. Te influence on eac target can be summarized from its neigboring cells tat lie witin a certain radius from te target. To efficiently subdivide te space, we need to devise a sceme tat adaptively subdivides te space according to te distribution of points. It is also desirable to generate cells as compact as possible. Based on te above considerations, we model te space subdivision task as a k-center problem, wic is defined as follows: given a set of n points and a predefined number of te clusters k, find a partition of te points into clusters S,..., Sk, and also te cluster centers c,..., ck, so as to minimize te cost function te maximum radius of clusters: max max kv ci k. i v Si Te k-center problem is known to be N P -ard []. Gonzalez [] proposed a very simple greedy algoritm, called fartest-point clustering. Initially pick an arbitrary point v0 as te center of te first cluster and add it to te center set C. Ten for i = to k do te follows: in iteration i, for every point, compute its distance to te set C: di (v, C) = minc C kv ck. Let vi be a point tat is fartest away from C, i.e., a point for wic di (vi, C) = maxv di (v, C). Add vi to set C. Report te points v0, v,..., vk as te cluster centers. Eac point is assigned to its nearest center. Gonzalez [] proved tat fartest-point clustering is a approximation algoritm wic computes a partition wit maximum radius at most twice te optimum. Te proof uses no geometry beyond te triangle inequity, so it old for any metric space. Hocbaum and Smoys [7] proved tat te factor cannot be improved unless P = N P. Te direct implementation of fartest-point clustering as running time O(nk). Feder and Greene [9] give a two-pase algoritm wit optimal running time O(n log k). Te predefined number of clusters k can be determined as follows: run te fartest-point algoritm until te maximum radius of clusters decreases to a given distance. In practice, te initial point as little influence on te final radius of te approximation, if number of te points n is sufficiently large. Figure displays te results of fartestpoint algoritm. In two dimensions, te algoritm leads to a Voronoi tessellation of te space. In tree dimensions, te partition boundary resembles te surface of a crystal Figure : Te fartest-point algoritm divides 0000 points into clusters (wit te centers indicated by te crosses) in 0. seconds on a 900MHZ PIII PC. Left: normal distributions; Rigt: Uniform distribution. 3. Te Algoritm Te improved fast Gauss transform consists of te following steps: Step Assign te N sources into K clusters using te fartest-point clustering algoritm suc tat te radius is less tan ρx. Step Coose p sufficiently large suc tat te error estimate () in appendix is less tan te desired precision ǫ. Step 3 For eac cluster Sk wit center ck, compute te coefficients given by te expression (0): Cαk α µ α X xi ck kxi ck k /. = qi e α! xi Sk Step Repeat for eac target yj, find its neigbor clusters wose centers lie witin te range ρy. Ten te sum of Gaussians () can be evaluated by te expression (9): G(yj ) = X X kyj ck k ρy α <p Cαk e kyj ck k / µ yj ck α Te amount of work required in step is O(N K) (for large K, we can use Feder and Greene s O(N log K) algoritm [9] instead). Te amount of work required in step 3 is of O(N rpd ). Te work required in step is O(M n rpd ), were n is te maximum number of te neigbor clusters for eac target. For most nonparametric statistics, computer.

6 vision and pattern recognition applications, te precision required is moderate, we can get small K and small r pd. Since n K, te improved fast Gauss transform acieves linear running time. Te algoritm needs to store te K coefficients of size r pd, so te storage complexity is reduced to O(Kr pd ). Mean Sift Analysis wit IFGT Segmentation using te mean sift analysis is cosen as a case study for te IFGT. Mean sift is a clustering tecnique based on kernel density estimation, wic is very effective and robust for te analysis of complex feature spaces. Te mean sift procedure employing a Gaussian kernel converges to te stationary point following a smoot trajectory, wic is teoretically important for convergence [5]. In practice, te quality of te results almost always improves wen te Gaussian kernel is employed. Despite its superior performance, te Gaussian kernel is not as widely used in mean sift as it sould be. In part tis may be due to te ig computational costs wic we try to alleviate in te paper. Given n data points x,...,x n in te d-dimensional space R d, te kernel density estimator wit kernel function K(x) and a window bandwidt, is given by [, 0, 7] ˆf n (x) = n d n ( ) x xi K, () were te d-variate kernel K(x) is nonnegative and integrates to one. Te Gaussian kernel is a common coice. Te mean sift algoritm is a steepest ascent procedure wic requires estimation of te density gradient: ˆf,K (x) = c n k,d x x i n d+ (x i x)g( ) ( n x x x ig i ) () = c k,g ˆf,G (x) n g ( x x i ) x, were g(x) = k N (x) = k N(x) wic can in turn be used as profile to define a Gaussian kernel G(x). Te kernel K(x) is called te sadow of G(x) []. Bot ave te same expression. ˆf,G (x) is te density estimation wit te kernel G. c k,g is te normalization coefficient. Te last term is te mean sift ( n x x x ig i ) m(x) = n g ( x x i ) x, (3) wic is proportional to te normalized density gradient and always points toward te steepest ascent direction of te function. Te mean sift algoritm iteratively performs te following two steps till it reaces te stationary point: Table : Running times in milliseconds for direct evaluation, fast Gauss transform and improved fast Gauss transform in tree dimensions. Case N = M Direct FGT IFGT Computation of te mean sift vector m(x k ). Updating te current position x k+ = x k + m(x k ). Te numerator in expression (3) is a weigted sum of Gaussians except tat te weigts are vectors. Te denominator is a uniform weigted sum of Gaussians. So bot can be evaluated by te improved fast Gauss transform as d + independent weigted sums of Gaussians. Te computation as been furter reduced because tey sare te same space subdivisions and series expansions. 5 Experimental Results Te first experiment compares te performance of our algoritm wit te original fast Gauss transform. Since tere is no practical fast Gauss transform in iger dimensions available, we only make comparisons in tree dimensions. Te sources and targets are uniformly distributed in a unit cube. Te weigts of te sources are uniformly distributed between 0 and. Te bandwidt of te Gaussian is = 0.. We set te relative error bound to % wic is reasonable for most kernel density estimation, because te estimated density function itself is an approximation. Table reports te CPU times using direct evaluation, te original fast Gauss transform (FGT) and te improved fast Gauss transform (IFGT). All te algoritms are programmed in C++ and were run on a 900MHz PIII PC. We can find te running time of te IFGT grows linearly as te number of sources and targets increases, wile te direct evaluation and te original FGT grows quadratically, toug it is lower tan te direct evaluation. Te poor performance of te FGT in 3D is also reported in []. Tis is probably due to te fact tat te number of boxes increases significantly by a uniform space subdivision in 3D, wic makes te cost to compute te interactions between te boxes grow quadratically. Te second experiment is to examine te performance of IFGT in iger dimensions. We randomly generate te source and target points in a unit ypercube based on a uniform distribution. Te weigts of te sources are uniformly distributed between 0 and. Te bandwidt is set to =. Te results are sown in Fig. 3. We compared te running

7 direct metod, D fast metod, D direct metod, D fast metod, D direct metod, D fast metod, D direct metod, 0D fast metod, 0D Table 3: Image sizes v.s. te running time of te mean sift. House Cooking Base Dive Zebra Size 55x9 0x53 3x9 x3 Time (s) CPU time N 0 3 D D D 0D Max abs error N Figure 3: Te running times in seconds (Top) and maximum absolute errors (Bottom) of te IFGT ( = ) v.s. direct evaluation in dimensions,,, 0. time of te direct evaluation to te IFGT wit = and N = M = 00,...,0000. Te comparisons are performed in dimensions from to 0 and results in dimensions,,, 0 are reported in Fig. 3. From te figure we notice tat te running time of te direct evaluation grows quadratically wit te size of points. Te running time of te IFGT grows linearly wit te size of te points. In,,, 0 dimensions, te IFGT takes 5ms, 0ms, 9 ms, 5ms to evaluate te sums on 0000 points. Te maximum relative absolute error as defined in [5] increases wit te dimensionality but not wit te number of points. Te worst error occurs in dimension 0, and is below 0 3. We can see tat for a 0-D problem involving more tan 700 Gaussians, te IFGT is faster tan direct evaluation, wile for a -D problem te IFGT is faster from almost te outset. Te tird experiment is to apply te improved fast Gauss transform to te mean sift algoritm. We first transform te images to L*u*v* color space and normalize to a unit cube. Ten we apply te mean sift algoritm wit = 0. to all te points in te 3D color space. After 5 iterations, te convergence points are grouped by a simple k-means algoritm [7]. We do not perform any postprocessing procedure as in [5]. Te code is written in C++ wit Matlab interfaces Figure : Segmentation results: (Rigt Column) Te original images. (Left Column) Segmented images labelled wit different colors. (Top Row) House image. (Second Row) Cooking image. (Tird Row) Base Dive image. (Bottom Row) Zebra image. and run on a 900MHz PIII PC. Te results are sown in Fig.. Te running time of te mean sift in seconds and te sizes of te images are sown in Table 3. Te speed of our implementation is at least as fast as any reported. We find tat te mean sift algoritm wit te improved fast Gauss transform already acieves clear boundaries witout any postprocessing. Tis is partly because tat we apply te mean sift algoritm to all feature points witout subsampling te feature space as in [5]. Tis leads to easily distinguisable valleys in our estimated densities. Anoter reason is tat in our metod te density evaluation at eac target point as contributions from a muc larger neigborood because of te Gaussian kernel, wic generates a smooter and better density estimate.

8 Conclusions In tis paper we present an improved fast Gauss transform to speed up te summations of Gaussians in iger dimensions. Te success in acceleration of te FGT comes from two innovations: te use of te fartest-point algoritm to adaptively subdivide te ig dimensional space, and te use of a new multivariate Taylor expansion we developed to dramatically reduce te computational and storage cost of te fast Gauss transform. Te recursive computation of te multivariate Taylor expansion furter reduces te computational cost and necessary storage. Te improved fast Gauss transform is applied to speed up te kernel density estimation wic is te keystone for many applications in computer vision and pattern recognition. Te general and powerful feature space analysis tool te mean sift algoritm is cosen as a case study for te IFGT. Using te IFGT, we acieved a linear running time mean sift algoritm. Witout using any euristic vicinity information between te points, te mean sift based image segmentation acieve satisfactory results. In future work, we will study te capability of te IFGT in applications suc as learning kernel classifiers (SVM), object tracking. We also plan to combine te IFGT wit oter tecniques [7] to furter improve te performance of mean sift algoritm. Appendix: Error Bound of Improved FGT Te error due to te truncation of series (9) after order p and te cutoff error te satisfies te bound ( ) p E(y) Q p! ρp xρ p y + e ρ y. () were Q = q j. Acknowledgments Te support by te NSF under te awards 9979, and 09 is gratefully acknowledged. References [] B. J. C. Baxter and G. Roussos. A new error estimate of te fast Gauss transform. SIAM Journal on Scientific Computing, ():57 59, 00. [] M. Bern and D. Eppstein. Approximation algoritms for geometric problems. In D. Hocbaum, editor, Approximation Algoritms for NP-Hard Problems, capter, pages PWS Publising Company, Boston, 997. [3] T. Cacoullos. Estimation of a multivariate density. Ann. Inst. Stat. Mat., ():79 9, 9. [] Y. Ceng. Mean sift, mode seeking, and clustering. IEEE Trans. Pattern Anal. Mac. Intell., 7(): , 995. [5] D. Comaniciu and P. Meer. Mean sift: A robust approac toward feature space analysis. IEEE Trans. Pattern Anal. Mac. Intell., (5):03 9, May 00. [] L. Devroye and F. Macell. Data structures in kernel density estimation. IEEE Trans. Pattern Anal. Mac. Intell., 7(3):30 3, 95. [7] R. O. Duda, P. E. Hart, and D. G. Stork. Pattern Classification. Jon Wiley & Sons, New York, nd edition, 00. [] A. Elgammal, R. Duraiswami, and L. Davis. Efficient nonparametric adaptive color modeling using fast Gauss transform. In Proc. IEEE Conf. Computer Vision and Pattern Recognition, Kauai, Hawaii, 00. [9] T. Feder and D. Greene. Optimal algoritms for approximate clustering. In Proc. 0t ACM Symp. Teory of computing, pages 3, Cicago, Illinois, 9. [0] K. Fukunaga and R. R. Hayes. Te reduced Parzen classifier. IEEE Trans. Pattern Anal. Mac. Intell., ():3 5, 99. [] K. Fukunaga and L. D. Hostetler. Te estimation of te gradient of a density function, wit applications in pattern recognition. IEEE Trans. Inform. Teory, :3 0, 975. [] T. Gonzalez. Clustering to minimize te maximum intercluster distance. Teoretical Computer Science, 3:93 30, 95. [3] L. Greengard and V. Roklin. A fast algoritm for particle simulations. J. Comput. Pys., 73():35 3, 97. [] L. Greengard and J. Strain. A fast algoritm for te evaluation of eat potentials. Comm. Pure Appl. Mat., 3():99 93, 990. [5] L. Greengard and J. Strain. Te fast Gauss transform. SIAM J. Sci. Statist. Comput., ():79 9, 99. [] N. A. Gumerov, R. Duraiswami, and E. A. Borovikov. Data structures, optimal coice of parameters, and complexity results for generalized multilevel fast multipole metods in d dimensions. Tecnical Report UMIACS-TR-003-, UMI- ACS, University of Maryland, College Park, 003. [7] D. S. Hocbaum and D. B. Smoys. A best possible euristic for te k-center problem. Matematics of Operations Researc, 0():0, 95. [] P. Huber. Robust Statistical Procedures. SIAM, edition, 99. [9] B. Jeon and D. A. Landgrebe. Fast Parzen density estimation using clustering-based branc and bound. IEEE Trans. Pattern Anal. Mac. Intell., (9):950 95, 99. [0] E. Parzen. On estimation of a probability density function and mode. Ann. Mat. Stat., 33(3):05 07, 9. [] J. G. Postaire and C. Vasseur. A fast algoritm for nonparametric probability density estimation. IEEE Trans. Pattern Anal. Mac. Intell., ():3, 9. [] M. Rosenblatt. Remarks on some nonparametric estimates of a density function. Ann. Mat. Stat., 7(3):3 37, 95. [3] B. Sclkopf and A. Smola. Learning wit Kernels: Support Vector Macines, Regularization, Optimization and Beyond. MIT Press, Cambridge, MA, 00. [] J. Si and J. Malik. Normalized cuts and image segmentation. IEEE Trans. Pattern Anal. Mac. Intell., (): 905, 000. [5] B. W. Silverman. Algoritm AS 7: Kernel density estimation using te fast Fourier transform. Appl. Stat., 3():93 99, 9. [] J. Strain. Te fast Gauss transform wit variable scales. SIAM J. Sci. Statist. Comput., (5), 99. [7] C. Yang, R. Duraiswami, D. DeMenton, and L. Davis. Mean-sift analysis using quasi-newton metods. In Proc. Int l Conf. Image Processing, Barcelona, Spain, 003.

Efficient Kernel Machines Using the Improved Fast Gauss Transform

Efficient Kernel Machines Using the Improved Fast Gauss Transform Efficient Kernel Machines Using the Improved Fast Gauss Transform Changjiang Yang, Ramani Duraiswami and Larry Davis Department of Computer Science University of Maryland College Park, MD 20742 {yangcj,ramani,lsd}@umiacs.umd.edu

More information

Fast optimal bandwidth selection for kernel density estimation

Fast optimal bandwidth selection for kernel density estimation Fast optimal bandwidt selection for kernel density estimation Vikas Candrakant Raykar and Ramani Duraiswami Dept of computer science and UMIACS, University of Maryland, CollegePark {vikas,ramani}@csumdedu

More information

Improved Fast Gauss Transform. (based on work by Changjiang Yang and Vikas Raykar)

Improved Fast Gauss Transform. (based on work by Changjiang Yang and Vikas Raykar) Improved Fast Gauss Transform (based on work by Changjiang Yang and Vikas Raykar) Fast Gauss Transform (FGT) Originally proposed by Greengard and Strain (1991) to efficiently evaluate the weighted sum

More information

Improved Fast Gauss Transform

Improved Fast Gauss Transform Improved Fast Gauss Transform Changjiang Yang Department of Computer Science University of Maryland Fast Gauss Transform (FGT) Originally proposed by Greengard and Strain (1991) to efficiently evaluate

More information

Improved Fast Gauss Transform. Fast Gauss Transform (FGT)

Improved Fast Gauss Transform. Fast Gauss Transform (FGT) 10/11/011 Improved Fast Gauss Transform Based on work by Changjiang Yang (00), Vikas Raykar (005), and Vlad Morariu (007) Fast Gauss Transform (FGT) Originally proposed by Greengard and Strain (1991) to

More information

lecture 26: Richardson extrapolation

lecture 26: Richardson extrapolation 43 lecture 26: Ricardson extrapolation 35 Ricardson extrapolation, Romberg integration Trougout numerical analysis, one encounters procedures tat apply some simple approximation (eg, linear interpolation)

More information

Numerical Differentiation

Numerical Differentiation Numerical Differentiation Finite Difference Formulas for te first derivative (Using Taylor Expansion tecnique) (section 8.3.) Suppose tat f() = g() is a function of te variable, and tat as 0 te function

More information

Lecture 15. Interpolation II. 2 Piecewise polynomial interpolation Hermite splines

Lecture 15. Interpolation II. 2 Piecewise polynomial interpolation Hermite splines Lecture 5 Interpolation II Introduction In te previous lecture we focused primarily on polynomial interpolation of a set of n points. A difficulty we observed is tat wen n is large, our polynomial as to

More information

Copyright c 2008 Kevin Long

Copyright c 2008 Kevin Long Lecture 4 Numerical solution of initial value problems Te metods you ve learned so far ave obtained closed-form solutions to initial value problems. A closedform solution is an explicit algebriac formula

More information

Consider a function f we ll specify which assumptions we need to make about it in a minute. Let us reformulate the integral. 1 f(x) dx.

Consider a function f we ll specify which assumptions we need to make about it in a minute. Let us reformulate the integral. 1 f(x) dx. Capter 2 Integrals as sums and derivatives as differences We now switc to te simplest metods for integrating or differentiating a function from its function samples. A careful study of Taylor expansions

More information

Fast Exact Univariate Kernel Density Estimation

Fast Exact Univariate Kernel Density Estimation Fast Exact Univariate Kernel Density Estimation David P. Hofmeyr Department of Statistics and Actuarial Science, Stellenbosc University arxiv:1806.00690v2 [stat.co] 12 Jul 2018 July 13, 2018 Abstract Tis

More information

Polynomial Interpolation

Polynomial Interpolation Capter 4 Polynomial Interpolation In tis capter, we consider te important problem of approximatinga function fx, wose values at a set of distinct points x, x, x,, x n are known, by a polynomial P x suc

More information

Chapter 5 FINITE DIFFERENCE METHOD (FDM)

Chapter 5 FINITE DIFFERENCE METHOD (FDM) MEE7 Computer Modeling Tecniques in Engineering Capter 5 FINITE DIFFERENCE METHOD (FDM) 5. Introduction to FDM Te finite difference tecniques are based upon approximations wic permit replacing differential

More information

Polynomial Interpolation

Polynomial Interpolation Capter 4 Polynomial Interpolation In tis capter, we consider te important problem of approximating a function f(x, wose values at a set of distinct points x, x, x 2,,x n are known, by a polynomial P (x

More information

7 Semiparametric Methods and Partially Linear Regression

7 Semiparametric Methods and Partially Linear Regression 7 Semiparametric Metods and Partially Linear Regression 7. Overview A model is called semiparametric if it is described by and were is nite-dimensional (e.g. parametric) and is in nite-dimensional (nonparametric).

More information

Homework 1 Due: Wednesday, September 28, 2016

Homework 1 Due: Wednesday, September 28, 2016 0-704 Information Processing and Learning Fall 06 Homework Due: Wednesday, September 8, 06 Notes: For positive integers k, [k] := {,..., k} denotes te set of te first k positive integers. Wen p and Y q

More information

A = h w (1) Error Analysis Physics 141

A = h w (1) Error Analysis Physics 141 Introduction In all brances of pysical science and engineering one deals constantly wit numbers wic results more or less directly from experimental observations. Experimental observations always ave inaccuracies.

More information

The Verlet Algorithm for Molecular Dynamics Simulations

The Verlet Algorithm for Molecular Dynamics Simulations Cemistry 380.37 Fall 2015 Dr. Jean M. Standard November 9, 2015 Te Verlet Algoritm for Molecular Dynamics Simulations Equations of motion For a many-body system consisting of N particles, Newton's classical

More information

Lecture 21. Numerical differentiation. f ( x+h) f ( x) h h

Lecture 21. Numerical differentiation. f ( x+h) f ( x) h h Lecture Numerical differentiation Introduction We can analytically calculate te derivative of any elementary function, so tere migt seem to be no motivation for calculating derivatives numerically. However

More information

5 Ordinary Differential Equations: Finite Difference Methods for Boundary Problems

5 Ordinary Differential Equations: Finite Difference Methods for Boundary Problems 5 Ordinary Differential Equations: Finite Difference Metods for Boundary Problems Read sections 10.1, 10.2, 10.4 Review questions 10.1 10.4, 10.8 10.9, 10.13 5.1 Introduction In te previous capters we

More information

HOMEWORK HELP 2 FOR MATH 151

HOMEWORK HELP 2 FOR MATH 151 HOMEWORK HELP 2 FOR MATH 151 Here we go; te second round of omework elp. If tere are oters you would like to see, let me know! 2.4, 43 and 44 At wat points are te functions f(x) and g(x) = xf(x)continuous,

More information

The derivative function

The derivative function Roberto s Notes on Differential Calculus Capter : Definition of derivative Section Te derivative function Wat you need to know already: f is at a point on its grap and ow to compute it. Wat te derivative

More information

SECTION 3.2: DERIVATIVE FUNCTIONS and DIFFERENTIABILITY

SECTION 3.2: DERIVATIVE FUNCTIONS and DIFFERENTIABILITY (Section 3.2: Derivative Functions and Differentiability) 3.2.1 SECTION 3.2: DERIVATIVE FUNCTIONS and DIFFERENTIABILITY LEARNING OBJECTIVES Know, understand, and apply te Limit Definition of te Derivative

More information

The Priestley-Chao Estimator

The Priestley-Chao Estimator Te Priestley-Cao Estimator In tis section we will consider te Pristley-Cao estimator of te unknown regression function. It is assumed tat we ave a sample of observations (Y i, x i ), i = 1,..., n wic are

More information

The Laplace equation, cylindrically or spherically symmetric case

The Laplace equation, cylindrically or spherically symmetric case Numerisce Metoden II, 7 4, und Übungen, 7 5 Course Notes, Summer Term 7 Some material and exercises Te Laplace equation, cylindrically or sperically symmetric case Electric and gravitational potential,

More information

Basic Nonparametric Estimation Spring 2002

Basic Nonparametric Estimation Spring 2002 Basic Nonparametric Estimation Spring 2002 Te following topics are covered today: Basic Nonparametric Regression. Tere are four books tat you can find reference: Silverman986, Wand and Jones995, Hardle990,

More information

Polynomials 3: Powers of x 0 + h

Polynomials 3: Powers of x 0 + h near small binomial Capter 17 Polynomials 3: Powers of + Wile it is easy to compute wit powers of a counting-numerator, it is a lot more difficult to compute wit powers of a decimal-numerator. EXAMPLE

More information

How to Find the Derivative of a Function: Calculus 1

How to Find the Derivative of a Function: Calculus 1 Introduction How to Find te Derivative of a Function: Calculus 1 Calculus is not an easy matematics course Te fact tat you ave enrolled in suc a difficult subject indicates tat you are interested in te

More information

Optimal parameters for a hierarchical grid data structure for contact detection in arbitrarily polydisperse particle systems

Optimal parameters for a hierarchical grid data structure for contact detection in arbitrarily polydisperse particle systems Comp. Part. Mec. 04) :357 37 DOI 0.007/s4057-04-000-9 Optimal parameters for a ierarcical grid data structure for contact detection in arbitrarily polydisperse particle systems Dinant Krijgsman Vitaliy

More information

MATH745 Fall MATH745 Fall

MATH745 Fall MATH745 Fall MATH745 Fall 5 MATH745 Fall 5 INTRODUCTION WELCOME TO MATH 745 TOPICS IN NUMERICAL ANALYSIS Instructor: Dr Bartosz Protas Department of Matematics & Statistics Email: bprotas@mcmasterca Office HH 36, Ext

More information

ERROR BOUNDS FOR THE METHODS OF GLIMM, GODUNOV AND LEVEQUE BRADLEY J. LUCIER*

ERROR BOUNDS FOR THE METHODS OF GLIMM, GODUNOV AND LEVEQUE BRADLEY J. LUCIER* EO BOUNDS FO THE METHODS OF GLIMM, GODUNOV AND LEVEQUE BADLEY J. LUCIE* Abstract. Te expected error in L ) attimet for Glimm s sceme wen applied to a scalar conservation law is bounded by + 2 ) ) /2 T

More information

Computational tractability of machine learning algorithms for tall fat data

Computational tractability of machine learning algorithms for tall fat data Computational tractability of machine learning algorithms for tall fat data Getting good enough solutions as fast as possible Vikas Chandrakant Raykar vikas@cs.umd.edu University of Maryland, CollegePark

More information

Boosting Kernel Density Estimates: a Bias Reduction. Technique?

Boosting Kernel Density Estimates: a Bias Reduction. Technique? Boosting Kernel Density Estimates: a Bias Reduction Tecnique? Marco Di Marzio Dipartimento di Metodi Quantitativi e Teoria Economica, Università di Cieti-Pescara, Viale Pindaro 42, 65127 Pescara, Italy

More information

Parameter Fitted Scheme for Singularly Perturbed Delay Differential Equations

Parameter Fitted Scheme for Singularly Perturbed Delay Differential Equations International Journal of Applied Science and Engineering 2013. 11, 4: 361-373 Parameter Fitted Sceme for Singularly Perturbed Delay Differential Equations Awoke Andargiea* and Y. N. Reddyb a b Department

More information

LECTURE 14 NUMERICAL INTEGRATION. Find

LECTURE 14 NUMERICAL INTEGRATION. Find LECTURE 14 NUMERCAL NTEGRATON Find b a fxdx or b a vx ux fx ydy dx Often integration is required. However te form of fx may be suc tat analytical integration would be very difficult or impossible. Use

More information

Efficient algorithms for for clone items detection

Efficient algorithms for for clone items detection Efficient algoritms for for clone items detection Raoul Medina, Caroline Noyer, and Olivier Raynaud Raoul Medina, Caroline Noyer and Olivier Raynaud LIMOS - Université Blaise Pascal, Campus universitaire

More information

Exercises for numerical differentiation. Øyvind Ryan

Exercises for numerical differentiation. Øyvind Ryan Exercises for numerical differentiation Øyvind Ryan February 25, 2013 1. Mark eac of te following statements as true or false. a. Wen we use te approximation f (a) (f (a +) f (a))/ on a computer, we can

More information

Average Rate of Change

Average Rate of Change Te Derivative Tis can be tougt of as an attempt to draw a parallel (pysically and metaporically) between a line and a curve, applying te concept of slope to someting tat isn't actually straigt. Te slope

More information

New Streamfunction Approach for Magnetohydrodynamics

New Streamfunction Approach for Magnetohydrodynamics New Streamfunction Approac for Magnetoydrodynamics Kab Seo Kang Brooaven National Laboratory, Computational Science Center, Building 63, Room, Upton NY 973, USA. sang@bnl.gov Summary. We apply te finite

More information

A h u h = f h. 4.1 The CoarseGrid SystemandtheResidual Equation

A h u h = f h. 4.1 The CoarseGrid SystemandtheResidual Equation Capter Grid Transfer Remark. Contents of tis capter. Consider a grid wit grid size and te corresponding linear system of equations A u = f. Te summary given in Section 3. leads to te idea tat tere migt

More information

Solving Continuous Linear Least-Squares Problems by Iterated Projection

Solving Continuous Linear Least-Squares Problems by Iterated Projection Solving Continuous Linear Least-Squares Problems by Iterated Projection by Ral Juengling Department o Computer Science, Portland State University PO Box 75 Portland, OR 977 USA Email: juenglin@cs.pdx.edu

More information

THE STURM-LIOUVILLE-TRANSFORMATION FOR THE SOLUTION OF VECTOR PARTIAL DIFFERENTIAL EQUATIONS. L. Trautmann, R. Rabenstein

THE STURM-LIOUVILLE-TRANSFORMATION FOR THE SOLUTION OF VECTOR PARTIAL DIFFERENTIAL EQUATIONS. L. Trautmann, R. Rabenstein Worksop on Transforms and Filter Banks (WTFB),Brandenburg, Germany, Marc 999 THE STURM-LIOUVILLE-TRANSFORMATION FOR THE SOLUTION OF VECTOR PARTIAL DIFFERENTIAL EQUATIONS L. Trautmann, R. Rabenstein Lerstul

More information

Order of Accuracy. ũ h u Ch p, (1)

Order of Accuracy. ũ h u Ch p, (1) Order of Accuracy 1 Terminology We consider a numerical approximation of an exact value u. Te approximation depends on a small parameter, wic can be for instance te grid size or time step in a numerical

More information

Preconditioning in H(div) and Applications

Preconditioning in H(div) and Applications 1 Preconditioning in H(div) and Applications Douglas N. Arnold 1, Ricard S. Falk 2 and Ragnar Winter 3 4 Abstract. Summarizing te work of [AFW97], we sow ow to construct preconditioners using domain decomposition

More information

Math 102 TEST CHAPTERS 3 & 4 Solutions & Comments Fall 2006

Math 102 TEST CHAPTERS 3 & 4 Solutions & Comments Fall 2006 Mat 102 TEST CHAPTERS 3 & 4 Solutions & Comments Fall 2006 f(x+) f(x) 10 1. For f(x) = x 2 + 2x 5, find ))))))))) and simplify completely. NOTE: **f(x+) is NOT f(x)+! f(x+) f(x) (x+) 2 + 2(x+) 5 ( x 2

More information

Exam 1 Review Solutions

Exam 1 Review Solutions Exam Review Solutions Please also review te old quizzes, and be sure tat you understand te omework problems. General notes: () Always give an algebraic reason for your answer (graps are not sufficient),

More information

Learning based super-resolution land cover mapping

Learning based super-resolution land cover mapping earning based super-resolution land cover mapping Feng ing, Yiang Zang, Giles M. Foody IEEE Fellow, Xiaodong Xiuua Zang, Siming Fang, Wenbo Yun Du is work was supported in part by te National Basic Researc

More information

Data-Based Optimal Bandwidth for Kernel Density Estimation of Statistical Samples

Data-Based Optimal Bandwidth for Kernel Density Estimation of Statistical Samples Commun. Teor. Pys. 70 (208) 728 734 Vol. 70 No. 6 December 208 Data-Based Optimal Bandwidt for Kernel Density Estimation of Statistical Samples Zen-Wei Li ( 李振伟 ) 2 and Ping He ( 何平 ) 3 Center for Teoretical

More information

Chapter 4: Numerical Methods for Common Mathematical Problems

Chapter 4: Numerical Methods for Common Mathematical Problems 1 Capter 4: Numerical Metods for Common Matematical Problems Interpolation Problem: Suppose we ave data defined at a discrete set of points (x i, y i ), i = 0, 1,..., N. Often it is useful to ave a smoot

More information

INTRODUCTION AND MATHEMATICAL CONCEPTS

INTRODUCTION AND MATHEMATICAL CONCEPTS Capter 1 INTRODUCTION ND MTHEMTICL CONCEPTS PREVIEW Tis capter introduces you to te basic matematical tools for doing pysics. You will study units and converting between units, te trigonometric relationsips

More information

Digital Filter Structures

Digital Filter Structures Digital Filter Structures Te convolution sum description of an LTI discrete-time system can, in principle, be used to implement te system For an IIR finite-dimensional system tis approac is not practical

More information

HOW TO DEAL WITH FFT SAMPLING INFLUENCES ON ADEV CALCULATIONS

HOW TO DEAL WITH FFT SAMPLING INFLUENCES ON ADEV CALCULATIONS HOW TO DEAL WITH FFT SAMPLING INFLUENCES ON ADEV CALCULATIONS Po-Ceng Cang National Standard Time & Frequency Lab., TL, Taiwan 1, Lane 551, Min-Tsu Road, Sec. 5, Yang-Mei, Taoyuan, Taiwan 36 Tel: 886 3

More information

WYSE Academic Challenge 2004 Sectional Mathematics Solution Set

WYSE Academic Challenge 2004 Sectional Mathematics Solution Set WYSE Academic Callenge 00 Sectional Matematics Solution Set. Answer: B. Since te equation can be written in te form x + y, we ave a major 5 semi-axis of lengt 5 and minor semi-axis of lengt. Tis means

More information

IEOR 165 Lecture 10 Distribution Estimation

IEOR 165 Lecture 10 Distribution Estimation IEOR 165 Lecture 10 Distribution Estimation 1 Motivating Problem Consider a situation were we ave iid data x i from some unknown distribution. One problem of interest is estimating te distribution tat

More information

Volume 29, Issue 3. Existence of competitive equilibrium in economies with multi-member households

Volume 29, Issue 3. Existence of competitive equilibrium in economies with multi-member households Volume 29, Issue 3 Existence of competitive equilibrium in economies wit multi-member ouseolds Noriisa Sato Graduate Scool of Economics, Waseda University Abstract Tis paper focuses on te existence of

More information

Dedicated to the 70th birthday of Professor Lin Qun

Dedicated to the 70th birthday of Professor Lin Qun Journal of Computational Matematics, Vol.4, No.3, 6, 4 44. ACCELERATION METHODS OF NONLINEAR ITERATION FOR NONLINEAR PARABOLIC EQUATIONS Guang-wei Yuan Xu-deng Hang Laboratory of Computational Pysics,

More information

An Empirical Bayesian interpretation and generalization of NL-means

An Empirical Bayesian interpretation and generalization of NL-means Computer Science Tecnical Report TR2010-934, October 2010 Courant Institute of Matematical Sciences, New York University ttp://cs.nyu.edu/web/researc/tecreports/reports.tml An Empirical Bayesian interpretation

More information

1 Calculus. 1.1 Gradients and the Derivative. Q f(x+h) f(x)

1 Calculus. 1.1 Gradients and the Derivative. Q f(x+h) f(x) Calculus. Gradients and te Derivative Q f(x+) δy P T δx R f(x) 0 x x+ Let P (x, f(x)) and Q(x+, f(x+)) denote two points on te curve of te function y = f(x) and let R denote te point of intersection of

More information

Quantum Mechanics Chapter 1.5: An illustration using measurements of particle spin.

Quantum Mechanics Chapter 1.5: An illustration using measurements of particle spin. I Introduction. Quantum Mecanics Capter.5: An illustration using measurements of particle spin. Quantum mecanics is a teory of pysics tat as been very successful in explaining and predicting many pysical

More information

Regularized Regression

Regularized Regression Regularized Regression David M. Blei Columbia University December 5, 205 Modern regression problems are ig dimensional, wic means tat te number of covariates p is large. In practice statisticians regularize

More information

Taylor Series and the Mean Value Theorem of Derivatives

Taylor Series and the Mean Value Theorem of Derivatives 1 - Taylor Series and te Mean Value Teorem o Derivatives Te numerical solution o engineering and scientiic problems described by matematical models oten requires solving dierential equations. Dierential

More information

Cubic Functions: Local Analysis

Cubic Functions: Local Analysis Cubic function cubing coefficient Capter 13 Cubic Functions: Local Analysis Input-Output Pairs, 378 Normalized Input-Output Rule, 380 Local I-O Rule Near, 382 Local Grap Near, 384 Types of Local Graps

More information

Differentiation in higher dimensions

Differentiation in higher dimensions Capter 2 Differentiation in iger dimensions 2.1 Te Total Derivative Recall tat if f : R R is a 1-variable function, and a R, we say tat f is differentiable at x = a if and only if te ratio f(a+) f(a) tends

More information

Handling Missing Data on Asymmetric Distribution

Handling Missing Data on Asymmetric Distribution International Matematical Forum, Vol. 8, 03, no. 4, 53-65 Handling Missing Data on Asymmetric Distribution Amad M. H. Al-Kazale Department of Matematics, Faculty of Science Al-albayt University, Al-Mafraq-Jordan

More information

Finding and Using Derivative The shortcuts

Finding and Using Derivative The shortcuts Calculus 1 Lia Vas Finding and Using Derivative Te sortcuts We ave seen tat te formula f f(x+) f(x) (x) = lim 0 is manageable for relatively simple functions like a linear or quadratic. For more complex

More information

Combining functions: algebraic methods

Combining functions: algebraic methods Combining functions: algebraic metods Functions can be added, subtracted, multiplied, divided, and raised to a power, just like numbers or algebra expressions. If f(x) = x 2 and g(x) = x + 2, clearly f(x)

More information

Notes on wavefunctions II: momentum wavefunctions

Notes on wavefunctions II: momentum wavefunctions Notes on wavefunctions II: momentum wavefunctions and uncertainty Te state of a particle at any time is described by a wavefunction ψ(x). Tese wavefunction must cange wit time, since we know tat particles

More information

Physically Based Modeling: Principles and Practice Implicit Methods for Differential Equations

Physically Based Modeling: Principles and Practice Implicit Methods for Differential Equations Pysically Based Modeling: Principles and Practice Implicit Metods for Differential Equations David Baraff Robotics Institute Carnegie Mellon University Please note: Tis document is 997 by David Baraff

More information

Material for Difference Quotient

Material for Difference Quotient Material for Difference Quotient Prepared by Stepanie Quintal, graduate student and Marvin Stick, professor Dept. of Matematical Sciences, UMass Lowell Summer 05 Preface Te following difference quotient

More information

Discontinuous Galerkin Methods for Relativistic Vlasov-Maxwell System

Discontinuous Galerkin Methods for Relativistic Vlasov-Maxwell System Discontinuous Galerkin Metods for Relativistic Vlasov-Maxwell System He Yang and Fengyan Li December 1, 16 Abstract e relativistic Vlasov-Maxwell (RVM) system is a kinetic model tat describes te dynamics

More information

5.1 We will begin this section with the definition of a rational expression. We

5.1 We will begin this section with the definition of a rational expression. We Basic Properties and Reducing to Lowest Terms 5.1 We will begin tis section wit te definition of a rational epression. We will ten state te two basic properties associated wit rational epressions and go

More information

Introduction to Derivatives

Introduction to Derivatives Introduction to Derivatives 5-Minute Review: Instantaneous Rates and Tangent Slope Recall te analogy tat we developed earlier First we saw tat te secant slope of te line troug te two points (a, f (a))

More information

. If lim. x 2 x 1. f(x+h) f(x)

. If lim. x 2 x 1. f(x+h) f(x) Review of Differential Calculus Wen te value of one variable y is uniquely determined by te value of anoter variable x, ten te relationsip between x and y is described by a function f tat assigns a value

More information

The Complexity of Computing the MCD-Estimator

The Complexity of Computing the MCD-Estimator Te Complexity of Computing te MCD-Estimator Torsten Bernolt Lerstul Informatik 2 Universität Dortmund, Germany torstenbernolt@uni-dortmundde Paul Fiscer IMM, Danisc Tecnical University Kongens Lyngby,

More information

INTRODUCTION AND MATHEMATICAL CONCEPTS

INTRODUCTION AND MATHEMATICAL CONCEPTS INTODUCTION ND MTHEMTICL CONCEPTS PEVIEW Tis capter introduces you to te basic matematical tools for doing pysics. You will study units and converting between units, te trigonometric relationsips of sine,

More information

Robotic manipulation project

Robotic manipulation project Robotic manipulation project Bin Nguyen December 5, 2006 Abstract Tis is te draft report for Robotic Manipulation s class project. Te cosen project aims to understand and implement Kevin Egan s non-convex

More information

Chapter 1. Density Estimation

Chapter 1. Density Estimation Capter 1 Density Estimation Let X 1, X,..., X n be observations from a density f X x. Te aim is to use only tis data to obtain an estimate ˆf X x of f X x. Properties of f f X x x, Parametric metods f

More information

EXTENSION OF A POSTPROCESSING TECHNIQUE FOR THE DISCONTINUOUS GALERKIN METHOD FOR HYPERBOLIC EQUATIONS WITH APPLICATION TO AN AEROACOUSTIC PROBLEM

EXTENSION OF A POSTPROCESSING TECHNIQUE FOR THE DISCONTINUOUS GALERKIN METHOD FOR HYPERBOLIC EQUATIONS WITH APPLICATION TO AN AEROACOUSTIC PROBLEM SIAM J. SCI. COMPUT. Vol. 26, No. 3, pp. 821 843 c 2005 Society for Industrial and Applied Matematics ETENSION OF A POSTPROCESSING TECHNIQUE FOR THE DISCONTINUOUS GALERKIN METHOD FOR HYPERBOLIC EQUATIONS

More information

A MONTE CARLO ANALYSIS OF THE EFFECTS OF COVARIANCE ON PROPAGATED UNCERTAINTIES

A MONTE CARLO ANALYSIS OF THE EFFECTS OF COVARIANCE ON PROPAGATED UNCERTAINTIES A MONTE CARLO ANALYSIS OF THE EFFECTS OF COVARIANCE ON PROPAGATED UNCERTAINTIES Ronald Ainswort Hart Scientific, American Fork UT, USA ABSTRACT Reports of calibration typically provide total combined uncertainties

More information

Investigating Euler s Method and Differential Equations to Approximate π. Lindsay Crowl August 2, 2001

Investigating Euler s Method and Differential Equations to Approximate π. Lindsay Crowl August 2, 2001 Investigating Euler s Metod and Differential Equations to Approximate π Lindsa Crowl August 2, 2001 Tis researc paper focuses on finding a more efficient and accurate wa to approximate π. Suppose tat x

More information

GRID CONVERGENCE ERROR ANALYSIS FOR MIXED-ORDER NUMERICAL SCHEMES

GRID CONVERGENCE ERROR ANALYSIS FOR MIXED-ORDER NUMERICAL SCHEMES GRID CONVERGENCE ERROR ANALYSIS FOR MIXED-ORDER NUMERICAL SCHEMES Cristoper J. Roy Sandia National Laboratories* P. O. Box 5800, MS 085 Albuquerque, NM 8785-085 AIAA Paper 00-606 Abstract New developments

More information

Convergence and Descent Properties for a Class of Multilevel Optimization Algorithms

Convergence and Descent Properties for a Class of Multilevel Optimization Algorithms Convergence and Descent Properties for a Class of Multilevel Optimization Algoritms Stepen G. Nas April 28, 2010 Abstract I present a multilevel optimization approac (termed MG/Opt) for te solution of

More information

A First-Order System Approach for Diffusion Equation. I. Second-Order Residual-Distribution Schemes

A First-Order System Approach for Diffusion Equation. I. Second-Order Residual-Distribution Schemes A First-Order System Approac for Diffusion Equation. I. Second-Order Residual-Distribution Scemes Hiroaki Nisikawa W. M. Keck Foundation Laboratory for Computational Fluid Dynamics, Department of Aerospace

More information

Chapter 2 Ising Model for Ferromagnetism

Chapter 2 Ising Model for Ferromagnetism Capter Ising Model for Ferromagnetism Abstract Tis capter presents te Ising model for ferromagnetism, wic is a standard simple model of a pase transition. Using te approximation of mean-field teory, te

More information

Derivation Of The Schwarzschild Radius Without General Relativity

Derivation Of The Schwarzschild Radius Without General Relativity Derivation Of Te Scwarzscild Radius Witout General Relativity In tis paper I present an alternative metod of deriving te Scwarzscild radius of a black ole. Te metod uses tree of te Planck units formulas:

More information

= 0 and states ''hence there is a stationary point'' All aspects of the proof dx must be correct (c)

= 0 and states ''hence there is a stationary point'' All aspects of the proof dx must be correct (c) Paper 1: Pure Matematics 1 Mark Sceme 1(a) (i) (ii) d d y 3 1x 4x x M1 A1 d y dx 1.1b 1.1b 36x 48x A1ft 1.1b Substitutes x = into teir dx (3) 3 1 4 Sows d y 0 and states ''ence tere is a stationary point''

More information

HARMONIC ALLOCATION TO MV CUSTOMERS IN RURAL DISTRIBUTION SYSTEMS

HARMONIC ALLOCATION TO MV CUSTOMERS IN RURAL DISTRIBUTION SYSTEMS HARMONIC ALLOCATION TO MV CUSTOMERS IN RURAL DISTRIBUTION SYSTEMS V Gosbell University of Wollongong Department of Electrical, Computer & Telecommunications Engineering, Wollongong, NSW 2522, Australia

More information

1watt=1W=1kg m 2 /s 3

1watt=1W=1kg m 2 /s 3 Appendix A Matematics Appendix A.1 Units To measure a pysical quantity, you need a standard. Eac pysical quantity as certain units. A unit is just a standard we use to compare, e.g. a ruler. In tis laboratory

More information

ch (for some fixed positive number c) reaching c

ch (for some fixed positive number c) reaching c GSTF Journal of Matematics Statistics and Operations Researc (JMSOR) Vol. No. September 05 DOI 0.60/s4086-05-000-z Nonlinear Piecewise-defined Difference Equations wit Reciprocal and Cubic Terms Ramadan

More information

Very fast optimal bandwidth selection for univariate kernel density estimation

Very fast optimal bandwidth selection for univariate kernel density estimation Very fast optimal bandwidt selection for univariate kernel density estimation VIKAS CHANDAKANT AYKA and AMANI DUAISWAMI Perceptual Interfaces and eality Laboratory Department of Computer Science and Institute

More information

Finite Difference Methods Assignments

Finite Difference Methods Assignments Finite Difference Metods Assignments Anders Söberg and Aay Saxena, Micael Tuné, and Maria Westermarck Revised: Jarmo Rantakokko June 6, 1999 Teknisk databeandling Assignment 1: A one-dimensional eat equation

More information

Simulation and verification of a plate heat exchanger with a built-in tap water accumulator

Simulation and verification of a plate heat exchanger with a built-in tap water accumulator Simulation and verification of a plate eat excanger wit a built-in tap water accumulator Anders Eriksson Abstract In order to test and verify a compact brazed eat excanger (CBE wit a built-in accumulation

More information

Inf sup testing of upwind methods

Inf sup testing of upwind methods INTERNATIONAL JOURNAL FOR NUMERICAL METHODS IN ENGINEERING Int. J. Numer. Met. Engng 000; 48:745 760 Inf sup testing of upwind metods Klaus-Jurgen Bate 1; ;, Dena Hendriana 1, Franco Brezzi and Giancarlo

More information

Bootstrap prediction intervals for Markov processes

Bootstrap prediction intervals for Markov processes arxiv: arxiv:0000.0000 Bootstrap prediction intervals for Markov processes Li Pan and Dimitris N. Politis Li Pan Department of Matematics University of California San Diego La Jolla, CA 92093-0112, USA

More information

Long Term Time Series Prediction with Multi-Input Multi-Output Local Learning

Long Term Time Series Prediction with Multi-Input Multi-Output Local Learning Long Term Time Series Prediction wit Multi-Input Multi-Output Local Learning Gianluca Bontempi Macine Learning Group, Département d Informatique Faculté des Sciences, ULB, Université Libre de Bruxelles

More information

Click here to see an animation of the derivative

Click here to see an animation of the derivative Differentiation Massoud Malek Derivative Te concept of derivative is at te core of Calculus; It is a very powerful tool for understanding te beavior of matematical functions. It allows us to optimize functions,

More information

Faster Gaussian Summation: Theory and Experiment

Faster Gaussian Summation: Theory and Experiment Faster Gaussian Summation: Teory and Experiment Dongryeol Lee College of Computing Georgia Institute of Tecnology Atlanta, GA 3033 Alexander Gray College of Computing Georgia Institute of Tecnology Atlanta,

More information

NONLINEAR SYSTEMS IDENTIFICATION USING THE VOLTERRA MODEL. Georgeta Budura

NONLINEAR SYSTEMS IDENTIFICATION USING THE VOLTERRA MODEL. Georgeta Budura NONLINEAR SYSTEMS IDENTIFICATION USING THE VOLTERRA MODEL Georgeta Budura Politenica University of Timisoara, Faculty of Electronics and Telecommunications, Comm. Dep., georgeta.budura@etc.utt.ro Abstract:

More information

Numerical Experiments Using MATLAB: Superconvergence of Nonconforming Finite Element Approximation for Second-Order Elliptic Problems

Numerical Experiments Using MATLAB: Superconvergence of Nonconforming Finite Element Approximation for Second-Order Elliptic Problems Applied Matematics, 06, 7, 74-8 ttp://wwwscirporg/journal/am ISSN Online: 5-7393 ISSN Print: 5-7385 Numerical Experiments Using MATLAB: Superconvergence of Nonconforming Finite Element Approximation for

More information

Numerical Solution to Parabolic PDE Using Implicit Finite Difference Approach

Numerical Solution to Parabolic PDE Using Implicit Finite Difference Approach Numerical Solution to arabolic DE Using Implicit Finite Difference Approac Jon Amoa-Mensa, Francis Oene Boateng, Kwame Bonsu Department of Matematics and Statistics, Sunyani Tecnical University, Sunyani,

More information