Dual-Tree Fast Gauss Transforms

Size: px
Start display at page:

Download "Dual-Tree Fast Gauss Transforms"

Transcription

1 Dual-Tree Fast Gauss Transforms Dongryeol Lee Computer Science Carnegie Mellon Univ. Alexander Gray Computer Science Carnegie Mellon Univ. Andrew Moore Computer Science Carnegie Mellon Univ. Abstract In previous work we presented an efficient approach to computing kernel summations which arise in many machine learning methods such as kernel density estimation. This approach, dual-tree recursion with finitedifference approximation, generalized existing methods for similar problems arising in computational physics in two ways appropriate for statistical problems: toward distribution sensitivity and general dimension, partly by avoiding series expansions. While this proved to be the fastest practical method for multivariate kernel density estimation at the optimal bandwidth, it is much less efficient at larger-than-optimal bandwidths. In this work, we explore the extent to which the dual-tree approach can be integrated with multipole-like Hermite expansions in order to achieve reasonable efficiency across all bandwidth scales, though only for low dimensionalities. In the process, we derive and demonstrate the first truly hierarchical fast Gauss transforms, effectively combining the best tools from discrete algorithms and continuous approximation theory. Fast Gaussian Summation Kernel summations are fundamental in both statistics/learning and computational physics. This paper will focus on the common form G(x q = N xq xr e h i.e. where the kernel is the Gaussian kernel with scaling parameter, or bandwidth h, there are N reference points x r, and we desire the sum for N Q different query points x q. Such kernel summations appear in a wide array of statistical/learning methods [5], perhaps most obviously in kernel density estimation [], the most widely used distribution-free method for the fundamental task of density estimation, which will be our main example. Understanding kernel summation algorithms from a recently developed unified perspective [5] begins with the picture of Figure, then separately considers the discrete and continuous aspects. Discrete/geometric aspect. In terms of discrete algorithmic structure, the dual-tree framework of [5], in the context of kernel summation, generalizes all of the well-known algorithms. It was applied to the problem of kernel density estimation in [7] using a simple These include the Barnes-Hut algorithm [], the Fast Multipole Method [8], Appel s algorithm [], and the WSPD [4]: the dual-tree method is a node-node algorithm (considers query regions rather than points, is fully recursive, can use distribution-sensitive data structures such as kd-trees, and is bichromatic (can specialize for differing query and reference sets. r=

2 Figure : The basic idea is to approximate the kernel sum contribution of some subset of the reference points, lying in some compact region of space with centroid x, to a query point. In more efficient schemes a query region is considered, i.e. the approximate contribution is made to an entire subset of the query points Q lying in some region of space Q, with centroid x Q. finite-difference approximation, which is tantamount to a centroid approximation. Partially by avoiding series expansions, which depend explicitly on the dimension, the result was the fastest such algorithm for general dimension, when operating at the optimal bandwidth. Unfortunately, when performing cross-validation to determine the (initially unknown optimal bandwidth, both suboptimally small and large bandwidths must be evaluated. The finite-difference-based dual-tree method tends to be efficient at or below the optimal bandwidth, and at very large bandwidths, but for intermediately-large bandwidths it suffers. Continuous/approximation aspect. This motivates investigating a multipole-like series approximation which is appropriate for the Gaussian kernel, as introduced by [9], which can be shown the generalize the centroid approximation. We define the Hermite functions h n (t by h n (t = e t H n (t, where the Hermite polynomials H n (t are defined by the odrigues formula: H n (t = ( n e t D n e t, t. After scaling and shifting the argument t appropriately, then taking the product of univariate functions for each dimension, we obtain the multivariate Hermite expansion N xq xr N G(x q = e h = r= r= r= r= α! ( xr x h α ( xq x h α h where we ve adopted the usual multi-index notation as in [9]. This can be re-written as N xq xr N ( ( α G(x q = e h = α! h xr x Q xq x Q α ( h h to express the sum as a Taylor (local expansion about a nearby representative centroid x Q in the query region. We will be using both types of expansions simultaneously. Since series approximations only hold locally, Greengard and okhlin [8] showed that it is useful to think in terms of a set of three translation operators for converting between expansions centered at different points, in order to create their celebrated hierarchical algorithm. This was done in the context of the Coulombic kernel, but the Gaussian kernel has importantly different mathematical properties. The original Fast Gauss Transform (FGT [9] was based on a flat grid, and thus provided only one operator ( HL of the next section, with an associated error bound (which was unfortunately incorrect. The Improved Fast Gauss Transform (IFGT [4] was based on a flat set of clusters and provided no operators with a rearranged series approximation, which intended to be more favorable in higher dimensions but had an incorrect error bound. We will show the derivations of all the translation operators and associated error bounds needed to obtain, for the first time, a hierarchical algorithm for the Gaussian kernel. (

3 Translation Operators and Error Bounds The first operator converts a multipole expansion of a reference node to form a local expansion centered at the centroid of the query node, and is our main approximation workhorse. Lemma.. Hermite-to-local (HL translation operator for Gaussian kernel (as presented in Lemma. in [9, 0]: Given a reference node, a query node Q, and the Hermite expansion centered at a centroid x of : G(x q = ( α h, the Taylor expansion of the Hermite expansion at the centroid x Q of the query node Q is given by G(x q = ( β B Q ( β h where Bβ = ( β xq x α+β h. Proof. (sketch The proof consists of replacing the Hermite function portion of the expansion with its Taylor series. [ N ( α! ] α x h r x ( h α h by interchanging Note that we can rewrite G(x q = r= the summation order, such that the term in the brackets depends only on the reference points, and can thus be computed indepedent of any query location we will call such terms Hermite moments. The next operator allows the efficient pre-computation of the Hermite moments in the reference tree in a bottom-up fashion from its children. Lemma.. Hermite-to-Hermite (HH translation operator for Gaussian kernel: Given the Hermite expansion centered at a centroid x in a reference node : G(x q = A ( α h α h, this same Hermite expansion shifted to a new location x of the parent node of is given by G(x q = ( A γ h γ h where γ 0 A γ = ( x x γ α. h 0 α γ (γ α! A α Proof. We simply replace the Hermite function part of the expansion by a new Taylor series, as follows: G(x q = «A xq x αh α h = «β «A x x α ( β xq x h α+β h h where γ = α + β. = A α = A α = 4 γ 0 0 α γ «β x x ( β h α+β h «β x x h α+β h (γ α! A α «xq x h 3 «γ α x x h «xq x h 5 h γ xq x h «

4 The next operator acts as a clean-up routine in a hierarchical algorithm. Since we can approximate at different scales in the query tree, we must somehow combine all the approximations at the end of the computation. By performing a breadth-first traversal of the query tree, the LL operator shifts a node s local expansion to the centroid of each child. β α α!(β α! B β Lemma.3. Local-to-local (LL translation operator for Gaussian kernel: Given a Taylor expansion centered at a centroid x Q of a query node Q : G(x q = ( B Q β, β h the Taylor expansion obtained by shifting [ this expansion to the new centroid ] x Q of the child node Q is G(x q = ( xq x Q β α ( Q α. h h Proof. Applying the multinomial theorem to to expand about the new center x Q yields: G(x q = «xq x β Q B β h = «xq x β α «α Q xq x Q B β. α!(β α! α β h h whose summation order can be interchanged to achieve the result. Because the Hermite and the Taylor expansion are truncated after taking p D terms, we incur an error in approximation. The original error bounds for the Gaussian kernel in [9, 0] were wrong and corrections were shown in [3]. Here, we will present all necessary three error bounds incurred in performing translation operators. We note that these error bounds place limits on the size of the query node and the reference node. Lemma.4. Error Bound for Truncating an Hermite Expansion (as presented in [3]: Suppose we are given an Hermite expansion of a reference node about its centroid x : G(x q = ( α h where A α = N ( xr x α. h α! For any query point xq, the r= error due to truncating the series after the first p D D ( term is ǫ M (p N D ( r D k ( ( D k r p k r p p! where xr satisfies x r x < rh for r <. Proof. (sketch We expand the Hermite expansion as a product of one-dimensional Hermite functions, and utilize a bound on one-dimensional Hermite functions due to [3]: n! h n(x n n! e x, n 0, x. Lemma.5. Error Bound for Truncating a Taylor Expansion Converted from an Hermite Expansion of Infinite Order: Suppose we are given the following Taylor expansion about the centroid x Q of a query node G(x q = ( β B Q β h where Strain [] proposed the interesting idea of using Stirling s formula (for any non-negative integer n: ` n+ n n! to lift the node size constraint; one might imagine that this could allow approximation of larger regions in a tree-based algorithm. Unfortunately, the error bounds developed in [] e were also incorrect. We have derived the three necessary corrected error bounds based on the techniques in [3]. However, due to space, and because using these bounds actually degraded performance slightly, we do not include those lemmas here.

5 B β = ( β ( xq x α+β h and A α s are the coefficients of the Hermite expansion centered at the reference node centroid x. Then, truncating the series after p D ( ( D k terms satisfies the error bound ǫ L (p D k ( r p k r p p! where x q x Q < rh for r <, x q Q. = D N ( r D Proof. Taylor expansion of the Hermite function yields xq xr e h = ( β «α xr x xq x h α+β α! h h = ( β «α x x r ( α h α+β α! h ««( β β xq x r xq x Q h β h h Use e r h ««β xq x Q h xq x h ««β xq x Q h = D (u p (x qi, x ri, x Qi + v p (x qi, x ri, x Qi for i D, where i= u p(x qi, x ri, x Qi = v p(x qi, x ri, x Qi = p n i =0 n i =p ( n i xqi x ri h ni h ( n i xqi x ri h ni h ««ni h ««ni h. These univariate functions respectively satisfy u p (x qi, x ri, x Qi rp r v p (x qi, x ri, x Qi r p p! r, for i D, achieving the multivariate bound. Lemma.6. Error Bound for Truncating a Taylor Expansion Converted from an Already Truncated Hermite Expansion: A truncated Hermite expansion centered about the centroid x of a reference node G(x q = ( α h has the following α<p Taylor expansion about the centroid x Q of a query node: G(x q = ( β C Q β h where the coefficients C β are given by C β = ( β α<p α+β ( xq x h and. Truncating the series after p D D N terms satisfies the error bound ǫ L (p ( D ( r D k (( ( (r p k ((r p ( (r p D k p! for a query node Q for which x q x Q < rh, and a reference node for which x r x < rh for r <, x q Q, x r. Proof. We define u pi = u p (x qi, x ri, x Qi, x i, v pi = v p (x qi, x ri, x Qi, x i, w pi = w p (x qi, x ri, x Qi, x i for i D: u pi = v pi = p n i =0 p n i =0 ( n i ( n i p n j =0 n j =p n j! n j! «nj xi x ri h ( n j h ni+nj «nj xi x ri h ( n j h ni+nj xqi x i h xqi x i h ««ni h ««ni h

6 w pi = n i =p Note that e ( n i n j =0 r h n j! «nj xi x ri h ( n j h ni+nj xqi x i h ««ni h = D (u pi + v pi + w pi for i D. Using the bound for i= Hermite functions and the property of geometric series, we obtain the following upper bounds: p p «u pi (r n i (r n j (r p = r Therefore, h e xq xr v pi p p! G(xq β<p n i =0 n j =0 n i =0 n j =p w pi p! DY i= (r n i (r n j = ««(r p (r p p! r r n i =p n j =0 D u ( r pi D «xq x Q C β h β (r n i (r n j = p! N ( r D r ««(r p r! «D ((r (( (r p k p ( (r p D k k p! D D ((r (( (r p k p ( (r p «D k k p! 3 Algorithm and esults Algorithm. The algorithm mainly consists of making the function call DFGT(Q.root,.root, i.e. calling the recursive function DFGT( with the root nodes of the query tree and reference tree. After the DFGT( routine is completed, the pre-order traversal of the query tree implied by the LL operator is performed. Before the DFGT( routine is called, the reference tree could be initialized with Hermite coefficients stored in each node using the HH translation operator, but instead we will compute them as needed on the fly. It adaptively chooses among three possible methods for approximating the summation contribution of the points in node to the queries in node Q, which are self-explanatory, based on crude operation count estimates. G min Q, a running lower bound on the kernel sum G(x q for any x q Q, is used to ensure locally that the global relative error is ǫ or less. This automatic mechanism allows the user to specify only an error tolerance ǫ rather than other tweak parameters. Upon approximation, the upper and lower bounds on G for Q and all its children are updated; the latter can be done in an O( delayed fashion as in [7]. The remainder of the routine implements the characteristic four-way dual-tree recursion. We also tested a hybrid method (DFGTH which approximates if either of the DFD or DFGT approximation criteria are met. Experimental results. We empirically studied the runtime 3 performance of five algorithms on five real-world datasets for kernel density estimation at every query point with a range of bandwidths, from 3 orders of magnitude smaller than optimal to three orders larger than optimal, according to the standard least-squares cross-validation score []. The naive 3 All times include all preprocessing costs including any data structure construction. Times are measured in CPU seconds on a dual-processor AMD Opteron 4 machine with 8 Gb of main memory and Mb of CPU cache. All the codes that we have written and obtained are written in C and C++, and was compiled under -O6 -funroll-loops flags on Linux kernel.4.6.

7 algorithm computes the sum explicitly and ( thus exactly. We have limited all datasets to 50K points so that true relative error, i.e. Ĝ(x q G true (x q /G true (x q, can be evaluated, and set the tolerance at % relative error for all query points. When any method fails to achieve the error tolerance in less time than twice that of the naive method, we give up. Codes for the FGT [9] and for the IFGT [4] were obtained from the authors websites. Note that both of these methods require the user to tweak parameters, while the others are automatic. 4 DFD refers to the depth-first dual-tree finite-difference method [7]. DFGT(Q, p DH = p DL = p HL = if.maxside < h, p DH = the smallest p such that N ( ( D k D k ( r p k r p p! < ǫg min ( r D D Q. if Q.maxside < h, p DL = the smallest p such that N ( ( D k D k ( r p k r p p! < ǫg min ( r D D D N ( r D Q. if max(q.maxside,.maxside < h, p HL = the smallest p such that ( ( D k (( (r p k ((r p ( (r p D k p! < ǫg min Q. c DH = p D DH N Q. c DL = p D DL N. c HL = Dp D+ HL. c Direct = DN Q N. if no Hermite coefficient of order p DH exists for, Compute it. c DH = c DH + p D DH N. if no Hermite coefficient of order p HL exists for, Compute it. c HL = c HL + p D HL N. c = min(c DH, c DL, c HL, c Direct. if c = c DH <, (Direct Hermite Evaluate each x q at the Hermite series of order p DH centered about x of using Equation. if c = c DL <, (Direct Local Accumulate each x r as the Taylor series of order p DL about the center x Q of Q using Equation. if c = c HL <, (Hermite-to-Local Convert the Hermite series of order p HL centered about x of to the Taylor series of the same order centered about x Q of Q using Lemma.. if c c Direct, Update G min and G max in Q and all its children. return. if leaf(q and leaf(, Perform the naive algorithm on every pair of points in Q and. else DFGT(Q.left,.left. DFGT(Q.left,.right. DFGT(Q.right,.left. DFGT(Q.right,.right. 4 For the FGT, note that the algorithm only ensures: bg(x q G true(x q τ. Therefore, we first set τ = ǫ, halving τ until the error tolerance ǫ was met. For the IFGT, which has multiple parameters that must be tweaked simultaneously, an automatic scheme was created, based on the recommendations given in the paper and software documentation: For D =, use p = 8; for D = 3, use p = 6; set ρ x =.5; start with K = N and double K until the error tolerance is met. When this failed to meet the tolerance, we resorted to additional trial and error by hand. The costs of parameter selection for these methods in both computer and human time is not included in the table.

8 Algorithm \ scale sj (astronomy: positions, D =, N = 50000, h = Naive FGT out of AM out of AM out of AM IFGT > Naive > Naive > Naive > Naive > Naive > Naive DFD DFGT DFGTH colors50k (astronomy: colors, D =, N = 50000, h = Naive FGT out of AM out of AM out of AM > Naive > Naive IFGT > Naive > Naive > Naive > Naive > Naive > Naive DFD DFGT DFGTH edsgc-radec-rnd (astronomy: angles, D =, N = 50000, h = Naive FGT out of AM out of AM out of AM IFGT > Naive > Naive > Naive > Naive > Naive > Naive DFD DFGT DFGTH mockgalaxy-d-m-rnd (cosmology: positions, D = 3, N = 50000, h = Naive FGT out of AM out of AM out of AM out of AM > Naive > Naive > Naive IFGT > Naive > Naive > Naive > Naive > Naive > Naive > Naive DFD DFGT DFGTH bio5-rnd (biology: drug activity, D = 5, N = 50000, h = Naive FGT out of AM out of AM out of AM out of AM out of AM out of AM out of AM IFGT > Naive > Naive > Naive > Naive > Naive > Naive > Naive DFD DFGT > Naive > Naive > Naive > Naive > Naive > Naive > Naive DFGTH > Naive > Naive > Naive > Naive > Naive > Naive > Naive Discussion. The experiments indicate that the DFGTH method is able to achieve reasonable performance across all bandwidth scales. Unfortunately none of the series approximation-based methods do well on the 5-dimensional data, as expected, highlighting the main weakness of the approach presented. Pursuing corrections to the error bounds necessary to use the intriguing series form of [4] may allow an increase in dimensionality. eferences [] A. W. Appel. An Efficient Program for Many-Body Simulations. SIAM Journal on Scientific and Statistical Computing, 6(:85 03, 985. [] J. Barnes and P. Hut. A Hierarchical O(NlogN Force-Calculation Algorithm. Nature, 34, 986. [3] B. Baxter and G. oussos. A new error estimate of the fast gauss transform. SIAM Journal on Scientific Computing, 4(:57 59, 00. [4] P. Callahan and S. Kosaraju. A decomposition of multidimensional point sets with applications to k-nearest-neighbors and n-body potential fields. Journal of the ACM, 6(:67 90, January 995. [5] A. Gray and A. W. Moore. N-Body Problems in Statistical Learning. In T. K. Leen, T. G. Dietterich, and V. Tresp, editors, Advances in Neural Information Processing Systems 3 (December 000. MIT Press, 00. [6] A. G. Gray. Bringing Tractability to Generalized N-Body Problems in Statistical and Scientific Computation. PhD thesis, Carnegie Mellon University, 003. [7] A. G. Gray and A. W. Moore. apid Evaluation of Multiple Density Models. In Artificial Intelligence and Statistics 003, 003. [8] L. Greengard and V. okhlin. A Fast Algorithm for Particle Simulations. Journal of Computational Physics, 73, 987. [9] L. Greengard and J. Strain. The fast gauss transform. SIAM Journal on Scientific and Statistical Computing, (:79 94, 99. [0] L. Greengard and. Sun. A new version of the fast gauss transform. Documenta Mathematica, Extra Volume ICM(III: , 998. [] B. W. Silverman. Density Estimation for Statistics and Data Analysis. Chapman and Hall, 986. [] J. Strain. The fast gauss transform with variable scales. SIAM Journal on Scientific and Statistical Computing, :3 39, 99. [3] O. Szász. On the relative extrema of the hermite orthogonal functions. J. Indian Math. Soc., 5:9 34, 95. [4] C. Yang,. Duraiswami, N. A. Gumerov, and L. Davis. Improved fast gauss transform and efficient kernel density estimation. International Conference on Computer Vision, 003.

Fast High-dimensional Kernel Summations Using the Monte Carlo Multipole Method

Fast High-dimensional Kernel Summations Using the Monte Carlo Multipole Method Fast High-dimensional Kernel Summations Using the Monte Carlo Multipole Method Dongryeol Lee Computational Science and Engineering Georgia Institute of Technology Atlanta, GA 30332 dongryel@cc.gatech.edu

More information

Faster Gaussian Summation: Theory and Experiment

Faster Gaussian Summation: Theory and Experiment Faster Gaussian Summation: Teory and Experiment Dongryeol Lee College of Computing Georgia Institute of Tecnology Atlanta, GA 3033 Alexander Gray College of Computing Georgia Institute of Tecnology Atlanta,

More information

Computational tractability of machine learning algorithms for tall fat data

Computational tractability of machine learning algorithms for tall fat data Computational tractability of machine learning algorithms for tall fat data Getting good enough solutions as fast as possible Vikas Chandrakant Raykar vikas@cs.umd.edu University of Maryland, CollegePark

More information

Single-tree GMM training

Single-tree GMM training Single-tree GMM training Ryan R. Curtin May 27, 2015 1 Introduction In this short document, we derive a tree-independent single-tree algorithm for Gaussian mixture model training, based on a technique

More information

Improved Fast Gauss Transform. Fast Gauss Transform (FGT)

Improved Fast Gauss Transform. Fast Gauss Transform (FGT) 10/11/011 Improved Fast Gauss Transform Based on work by Changjiang Yang (00), Vikas Raykar (005), and Vlad Morariu (007) Fast Gauss Transform (FGT) Originally proposed by Greengard and Strain (1991) to

More information

Improved Fast Gauss Transform. (based on work by Changjiang Yang and Vikas Raykar)

Improved Fast Gauss Transform. (based on work by Changjiang Yang and Vikas Raykar) Improved Fast Gauss Transform (based on work by Changjiang Yang and Vikas Raykar) Fast Gauss Transform (FGT) Originally proposed by Greengard and Strain (1991) to efficiently evaluate the weighted sum

More information

Efficient Kernel Machines Using the Improved Fast Gauss Transform

Efficient Kernel Machines Using the Improved Fast Gauss Transform Efficient Kernel Machines Using the Improved Fast Gauss Transform Changjiang Yang, Ramani Duraiswami and Larry Davis Department of Computer Science University of Maryland College Park, MD 20742 {yangcj,ramani,lsd}@umiacs.umd.edu

More information

APPROXIMATING GAUSSIAN PROCESSES

APPROXIMATING GAUSSIAN PROCESSES 1 / 23 APPROXIMATING GAUSSIAN PROCESSES WITH H 2 -MATRICES Steffen Börm 1 Jochen Garcke 2 1 Christian-Albrechts-Universität zu Kiel 2 Universität Bonn and Fraunhofer SCAI 2 / 23 OUTLINE 1 GAUSSIAN PROCESSES

More information

Scalable machine learning for massive datasets: Fast summation algorithms

Scalable machine learning for massive datasets: Fast summation algorithms Scalable machine learning for massive datasets: Fast summation algorithms Getting good enough solutions as fast as possible Vikas Chandrakant Raykar vikas@cs.umd.edu University of Maryland, CollegePark

More information

Fast multipole method

Fast multipole method 203/0/ page Chapter 4 Fast multipole method Originally designed for particle simulations One of the most well nown way to tae advantage of the low-ran properties Applications include Simulation of physical

More information

The Fast Multipole Method and other Fast Summation Techniques

The Fast Multipole Method and other Fast Summation Techniques The Fast Multipole Method and other Fast Summation Techniques Gunnar Martinsson The University of Colorado at Boulder (The factor of 1/2π is suppressed.) Problem definition: Consider the task of evaluation

More information

Fast Krylov Methods for N-Body Learning

Fast Krylov Methods for N-Body Learning Fast Krylov Methods for N-Body Learning Nando de Freitas Department of Computer Science University of British Columbia nando@cs.ubc.ca Maryam Mahdaviani Department of Computer Science University of British

More information

A Multipole Based Treecode Using Spherical Harmonics for Potentials of the Form r λ

A Multipole Based Treecode Using Spherical Harmonics for Potentials of the Form r λ A Multipole Based Treecode Using Spherical Harmonics for Potentials of the Form r λ Kasthuri Srinivasan, Hemant Mahawar, and Vivek Sarin Department of Computer Science, Texas A&M University, College Station,

More information

INTRODUCTION TO FAST MULTIPOLE METHODS LONG CHEN

INTRODUCTION TO FAST MULTIPOLE METHODS LONG CHEN INTRODUCTION TO FAST MULTIPOLE METHODS LONG CHEN ABSTRACT. This is a simple introduction to fast multipole methods for the N-body summation problems. Low rank approximation plus hierarchical decomposition

More information

CS 450 Numerical Analysis. Chapter 8: Numerical Integration and Differentiation

CS 450 Numerical Analysis. Chapter 8: Numerical Integration and Differentiation Lecture slides based on the textbook Scientific Computing: An Introductory Survey by Michael T. Heath, copyright c 2018 by the Society for Industrial and Applied Mathematics. http://www.siam.org/books/cl80

More information

Introduction to Machine Learning

Introduction to Machine Learning Introduction to Machine Learning 3. Instance Based Learning Alex Smola Carnegie Mellon University http://alex.smola.org/teaching/cmu2013-10-701 10-701 Outline Parzen Windows Kernels, algorithm Model selection

More information

5746 IEEE TRANSACTIONS ON SIGNAL PROCESSING, VOL. 56, NO. 12, DECEMBER X/$ IEEE

5746 IEEE TRANSACTIONS ON SIGNAL PROCESSING, VOL. 56, NO. 12, DECEMBER X/$ IEEE 5746 IEEE TRANSACTIONS ON SIGNAL PROCESSING, VOL. 56, NO. 12, DECEMBER 2008 Nonlinear Filtering Using a New Proposal Distribution and the Improved Fast Gauss Transform With Tighter Performance Bounds Roni

More information

Fast Multipole Methods: Fundamentals & Applications. Ramani Duraiswami Nail A. Gumerov

Fast Multipole Methods: Fundamentals & Applications. Ramani Duraiswami Nail A. Gumerov Fast Multipole Methods: Fundamentals & Applications Ramani Duraiswami Nail A. Gumerov Week 1. Introduction. What are multipole methods and what is this course about. Problems from physics, mathematics,

More information

Improved Fast Gauss Transform

Improved Fast Gauss Transform Improved Fast Gauss Transform Changjiang Yang Department of Computer Science University of Maryland Fast Gauss Transform (FGT) Originally proposed by Greengard and Strain (1991) to efficiently evaluate

More information

Lecture 2: The Fast Multipole Method

Lecture 2: The Fast Multipole Method CBMS Conference on Fast Direct Solvers Dartmouth College June 23 June 27, 2014 Lecture 2: The Fast Multipole Method Gunnar Martinsson The University of Colorado at Boulder Research support by: Recall:

More information

Lecture 6: Backpropagation

Lecture 6: Backpropagation Lecture 6: Backpropagation Roger Grosse 1 Introduction So far, we ve seen how to train shallow models, where the predictions are computed as a linear function of the inputs. We ve also observed that deeper

More information

Generative v. Discriminative classifiers Intuition

Generative v. Discriminative classifiers Intuition Logistic Regression (Continued) Generative v. Discriminative Decision rees Machine Learning 10701/15781 Carlos Guestrin Carnegie Mellon University January 31 st, 2007 2005-2007 Carlos Guestrin 1 Generative

More information

A Remark on the Fast Gauss Transform

A Remark on the Fast Gauss Transform Publ. RIMS, Kyoto Univ. 39 (2003), 785 796 A Remark on the Fast Gauss Transform By Kenta Kobayashi Abstract We propose an improvement on the Fast Gauss Transform which was presented by Greengard and Sun

More information

A Crash Course on Fast Multipole Method (FMM)

A Crash Course on Fast Multipole Method (FMM) A Crash Course on Fast Multipole Method (FMM) Hanliang Guo Kanso Biodynamical System Lab, 2018 April What is FMM? A fast algorithm to study N body interactions (Gravitational, Coulombic interactions )

More information

Fast Multipole Methods

Fast Multipole Methods An Introduction to Fast Multipole Methods Ramani Duraiswami Institute for Advanced Computer Studies University of Maryland, College Park http://www.umiacs.umd.edu/~ramani Joint work with Nail A. Gumerov

More information

Midterm: CS 6375 Spring 2015 Solutions

Midterm: CS 6375 Spring 2015 Solutions Midterm: CS 6375 Spring 2015 Solutions The exam is closed book. You are allowed a one-page cheat sheet. Answer the questions in the spaces provided on the question sheets. If you run out of room for an

More information

Compute the Fourier transform on the first register to get x {0,1} n x 0.

Compute the Fourier transform on the first register to get x {0,1} n x 0. CS 94 Recursive Fourier Sampling, Simon s Algorithm /5/009 Spring 009 Lecture 3 1 Review Recall that we can write any classical circuit x f(x) as a reversible circuit R f. We can view R f as a unitary

More information

Hierarchical Matrices. Jon Cockayne April 18, 2017

Hierarchical Matrices. Jon Cockayne April 18, 2017 Hierarchical Matrices Jon Cockayne April 18, 2017 1 Sources Introduction to Hierarchical Matrices with Applications [Börm et al., 2003] 2 Sources Introduction to Hierarchical Matrices with Applications

More information

Lecture 2: Linear regression

Lecture 2: Linear regression Lecture 2: Linear regression Roger Grosse 1 Introduction Let s ump right in and look at our first machine learning algorithm, linear regression. In regression, we are interested in predicting a scalar-valued

More information

Verified Global Optimization with Taylor Model based Range Bounders

Verified Global Optimization with Taylor Model based Range Bounders Verified Global Optimization with Taylor Model based Range Bounders KYOKO MAKINO and MARTIN BERZ Department of Physics and Astronomy Michigan State University East Lansing, MI 88 USA makino@msu.edu, berz@msu.edu

More information

Scalability Metrics for Parallel Molecular Dynamics

Scalability Metrics for Parallel Molecular Dynamics Scalability Metrics for arallel Molecular Dynamics arallel Efficiency We define the efficiency of a parallel program running on processors to solve a problem of size W Let T(W, ) be the execution time

More information

Lars Schmidt-Thieme, Information Systems and Machine Learning Lab (ISMLL), University of Hildesheim, Germany

Lars Schmidt-Thieme, Information Systems and Machine Learning Lab (ISMLL), University of Hildesheim, Germany Syllabus Fri. 21.10. (1) 0. Introduction A. Supervised Learning: Linear Models & Fundamentals Fri. 27.10. (2) A.1 Linear Regression Fri. 3.11. (3) A.2 Linear Classification Fri. 10.11. (4) A.3 Regularization

More information

ALGEBRAIC STRUCTURE AND DEGREE REDUCTION

ALGEBRAIC STRUCTURE AND DEGREE REDUCTION ALGEBRAIC STRUCTURE AND DEGREE REDUCTION Let S F n. We define deg(s) to be the minimal degree of a non-zero polynomial that vanishes on S. We have seen that for a finite set S, deg(s) n S 1/n. In fact,

More information

are the q-versions of n, n! and . The falling factorial is (x) k = x(x 1)(x 2)... (x k + 1).

are the q-versions of n, n! and . The falling factorial is (x) k = x(x 1)(x 2)... (x k + 1). Lecture A jacques@ucsd.edu Notation: N, R, Z, F, C naturals, reals, integers, a field, complex numbers. p(n), S n,, b(n), s n, partition numbers, Stirling of the second ind, Bell numbers, Stirling of the

More information

Belief Update in CLG Bayesian Networks With Lazy Propagation

Belief Update in CLG Bayesian Networks With Lazy Propagation Belief Update in CLG Bayesian Networks With Lazy Propagation Anders L Madsen HUGIN Expert A/S Gasværksvej 5 9000 Aalborg, Denmark Anders.L.Madsen@hugin.com Abstract In recent years Bayesian networks (BNs)

More information

Notes on Tabular Methods

Notes on Tabular Methods Notes on Tabular ethods Nan Jiang September 28, 208 Overview of the methods. Tabular certainty-equivalence Certainty-equivalence is a model-based RL algorithm, that is, it first estimates an DP model from

More information

Particle Methods. Lecture Reduce and Broadcast: A function viewpoint

Particle Methods. Lecture Reduce and Broadcast: A function viewpoint Lecture 9 Particle Methods 9.1 Reduce and Broadcast: A function viewpoint [This section is being rewritten with what we hope will be the world s clearest explanation of the fast multipole algorithm. Readers

More information

A fast method for solving the Heat equation by Layer Potentials

A fast method for solving the Heat equation by Layer Potentials A fast method for solving the Heat equation by Layer Potentials Johannes Tausch Abstract Boundary integral formulations of the heat equation involve time convolutions in addition to surface potentials.

More information

Wolf-Tilo Balke Silviu Homoceanu Institut für Informationssysteme Technische Universität Braunschweig

Wolf-Tilo Balke Silviu Homoceanu Institut für Informationssysteme Technische Universität Braunschweig Multimedia Databases Wolf-Tilo Balke Silviu Homoceanu Institut für Informationssysteme Technische Universität Braunschweig http://www.ifis.cs.tu-bs.de 13 Indexes for Multimedia Data 13 Indexes for Multimedia

More information

A Polynomial Chaos Approach to Robust Multiobjective Optimization

A Polynomial Chaos Approach to Robust Multiobjective Optimization A Polynomial Chaos Approach to Robust Multiobjective Optimization Silvia Poles 1, Alberto Lovison 2 1 EnginSoft S.p.A., Optimization Consulting Via Giambellino, 7 35129 Padova, Italy s.poles@enginsoft.it

More information

Midterm, Fall 2003

Midterm, Fall 2003 5-78 Midterm, Fall 2003 YOUR ANDREW USERID IN CAPITAL LETTERS: YOUR NAME: There are 9 questions. The ninth may be more time-consuming and is worth only three points, so do not attempt 9 unless you are

More information

Text Mining. Dr. Yanjun Li. Associate Professor. Department of Computer and Information Sciences Fordham University

Text Mining. Dr. Yanjun Li. Associate Professor. Department of Computer and Information Sciences Fordham University Text Mining Dr. Yanjun Li Associate Professor Department of Computer and Information Sciences Fordham University Outline Introduction: Data Mining Part One: Text Mining Part Two: Preprocessing Text Data

More information

Distance between multinomial and multivariate normal models

Distance between multinomial and multivariate normal models Chapter 9 Distance between multinomial and multivariate normal models SECTION 1 introduces Andrew Carter s recursive procedure for bounding the Le Cam distance between a multinomialmodeland its approximating

More information

Notes on Paths, Trees and Lagrange Inversion

Notes on Paths, Trees and Lagrange Inversion Notes on Paths, Trees and Lagrange Inversion Today we are going to start with a problem that may seem somewhat unmotivated, and solve it in two ways. From there, we will proceed to discuss applications

More information

Notes on Logarithmic Lower Bounds in the Cell Probe Model

Notes on Logarithmic Lower Bounds in the Cell Probe Model Notes on Logarithmic Lower Bounds in the Cell Probe Model Kevin Zatloukal November 10, 2010 1 Overview Paper is by Mihai Pâtraşcu and Erik Demaine. Both were at MIT at the time. (Mihai is now at AT&T Labs.)

More information

Decision Trees. Machine Learning 10701/15781 Carlos Guestrin Carnegie Mellon University. February 5 th, Carlos Guestrin 1

Decision Trees. Machine Learning 10701/15781 Carlos Guestrin Carnegie Mellon University. February 5 th, Carlos Guestrin 1 Decision Trees Machine Learning 10701/15781 Carlos Guestrin Carnegie Mellon University February 5 th, 2007 2005-2007 Carlos Guestrin 1 Linear separability A dataset is linearly separable iff 9 a separating

More information

N-body simulations. Phys 750 Lecture 10

N-body simulations. Phys 750 Lecture 10 N-body simulations Phys 750 Lecture 10 N-body simulations Forward integration in time of strongly coupled classical particles or (rigid-body composite objects) dissipation Brownian motion Very general

More information

Expectation Propagation in Factor Graphs: A Tutorial

Expectation Propagation in Factor Graphs: A Tutorial DRAFT: Version 0.1, 28 October 2005. Do not distribute. Expectation Propagation in Factor Graphs: A Tutorial Charles Sutton October 28, 2005 Abstract Expectation propagation is an important variational

More information

CS4495/6495 Introduction to Computer Vision. 8C-L3 Support Vector Machines

CS4495/6495 Introduction to Computer Vision. 8C-L3 Support Vector Machines CS4495/6495 Introduction to Computer Vision 8C-L3 Support Vector Machines Discriminative classifiers Discriminative classifiers find a division (surface) in feature space that separates the classes Several

More information

Multimedia Databases 1/29/ Indexes for Multimedia Data Indexes for Multimedia Data Indexes for Multimedia Data

Multimedia Databases 1/29/ Indexes for Multimedia Data Indexes for Multimedia Data Indexes for Multimedia Data 1/29/2010 13 Indexes for Multimedia Data 13 Indexes for Multimedia Data 13.1 R-Trees Multimedia Databases Wolf-Tilo Balke Silviu Homoceanu Institut für Informationssysteme Technische Universität Braunschweig

More information

Alpha-Beta Pruning: Algorithm and Analysis

Alpha-Beta Pruning: Algorithm and Analysis Alpha-Beta Pruning: Algorithm and Analysis Tsan-sheng Hsu tshsu@iis.sinica.edu.tw http://www.iis.sinica.edu.tw/~tshsu 1 Introduction Alpha-beta pruning is the standard searching procedure used for 2-person

More information

Transforming Hierarchical Trees on Metric Spaces

Transforming Hierarchical Trees on Metric Spaces CCCG 016, Vancouver, British Columbia, August 3 5, 016 Transforming Hierarchical Trees on Metric Spaces Mahmoodreza Jahanseir Donald R. Sheehy Abstract We show how a simple hierarchical tree called a cover

More information

More Dynamic Programming

More Dynamic Programming CS 374: Algorithms & Models of Computation, Spring 2017 More Dynamic Programming Lecture 14 March 9, 2017 Chandra Chekuri (UIUC) CS374 1 Spring 2017 1 / 42 What is the running time of the following? Consider

More information

Math Models of OR: Branch-and-Bound

Math Models of OR: Branch-and-Bound Math Models of OR: Branch-and-Bound John E. Mitchell Department of Mathematical Sciences RPI, Troy, NY 12180 USA November 2018 Mitchell Branch-and-Bound 1 / 15 Branch-and-Bound Outline 1 Branch-and-Bound

More information

More Dynamic Programming

More Dynamic Programming Algorithms & Models of Computation CS/ECE 374, Fall 2017 More Dynamic Programming Lecture 14 Tuesday, October 17, 2017 Sariel Har-Peled (UIUC) CS374 1 Fall 2017 1 / 48 What is the running time of the following?

More information

Technical Vignette 5: Understanding intrinsic Gaussian Markov random field spatial models, including intrinsic conditional autoregressive models

Technical Vignette 5: Understanding intrinsic Gaussian Markov random field spatial models, including intrinsic conditional autoregressive models Technical Vignette 5: Understanding intrinsic Gaussian Markov random field spatial models, including intrinsic conditional autoregressive models Christopher Paciorek, Department of Statistics, University

More information

Fast Multipole Methods for The Laplace Equation. Outline

Fast Multipole Methods for The Laplace Equation. Outline Fast Multipole Methods for The Laplace Equation Ramani Duraiswami Nail Gumerov Outline 3D Laplace equation and Coulomb potentials Multipole and local expansions Special functions Legendre polynomials Associated

More information

Alpha-Beta Pruning: Algorithm and Analysis

Alpha-Beta Pruning: Algorithm and Analysis Alpha-Beta Pruning: Algorithm and Analysis Tsan-sheng Hsu tshsu@iis.sinica.edu.tw http://www.iis.sinica.edu.tw/~tshsu 1 Introduction Alpha-beta pruning is the standard searching procedure used for 2-person

More information

STA 4273H: Statistical Machine Learning

STA 4273H: Statistical Machine Learning STA 4273H: Statistical Machine Learning Russ Salakhutdinov Department of Statistics! rsalakhu@utstat.toronto.edu! http://www.utstat.utoronto.ca/~rsalakhu/ Sidney Smith Hall, Room 6002 Lecture 3 Linear

More information

Bayesian Networks: Independencies and Inference

Bayesian Networks: Independencies and Inference Bayesian Networks: Independencies and Inference Scott Davies and Andrew Moore Note to other teachers and users of these slides. Andrew and Scott would be delighted if you found this source material useful

More information

Analyzing the Error Bounds of Multipole-Based

Analyzing the Error Bounds of Multipole-Based Page 1 of 12 Analyzing the Error Bounds of Multipole-Based Treecodes Vivek Sarin, Ananth Grama, and Ahmed Sameh Department of Computer Sciences Purdue University W. Lafayette, IN 47906 Email: Sarin, Grama,

More information

Extreme Point Solutions for Infinite Network Flow Problems

Extreme Point Solutions for Infinite Network Flow Problems Extreme Point Solutions for Infinite Network Flow Problems H. Edwin Romeijn Dushyant Sharma Robert L. Smith January 3, 004 Abstract We study capacitated network flow problems with supplies and demands

More information

Maintaining Nets and Net Trees under Incremental Motion

Maintaining Nets and Net Trees under Incremental Motion Maintaining Nets and Net Trees under Incremental Motion Minkyoung Cho, David Mount, and Eunhui Park Department of Computer Science University of Maryland, College Park August 25, 2009 Latent Space Embedding

More information

27 : Distributed Monte Carlo Markov Chain. 1 Recap of MCMC and Naive Parallel Gibbs Sampling

27 : Distributed Monte Carlo Markov Chain. 1 Recap of MCMC and Naive Parallel Gibbs Sampling 10-708: Probabilistic Graphical Models 10-708, Spring 2014 27 : Distributed Monte Carlo Markov Chain Lecturer: Eric P. Xing Scribes: Pengtao Xie, Khoa Luu In this scribe, we are going to review the Parallel

More information

Alpha-Beta Pruning: Algorithm and Analysis

Alpha-Beta Pruning: Algorithm and Analysis Alpha-Beta Pruning: Algorithm and Analysis Tsan-sheng Hsu tshsu@iis.sinica.edu.tw http://www.iis.sinica.edu.tw/~tshsu 1 Introduction Alpha-beta pruning is the standard searching procedure used for solving

More information

Vikas Chandrakant Raykar Doctor of Philosophy, 2007

Vikas Chandrakant Raykar Doctor of Philosophy, 2007 ABSTRACT Title of dissertation: SCALABLE MACHINE LEARNING FOR MASSIVE DATASETS : FAST SUMMATION ALGORITHMS Vikas Chandrakant Raykar Doctor of Philosophy, 2007 Dissertation directed by: Professor Dr. Ramani

More information

Clustering. CSL465/603 - Fall 2016 Narayanan C Krishnan

Clustering. CSL465/603 - Fall 2016 Narayanan C Krishnan Clustering CSL465/603 - Fall 2016 Narayanan C Krishnan ckn@iitrpr.ac.in Supervised vs Unsupervised Learning Supervised learning Given x ", y " "%& ', learn a function f: X Y Categorical output classification

More information

Introduction to Machine Learning Lecture 13. Mehryar Mohri Courant Institute and Google Research

Introduction to Machine Learning Lecture 13. Mehryar Mohri Courant Institute and Google Research Introduction to Machine Learning Lecture 13 Mehryar Mohri Courant Institute and Google Research mohri@cims.nyu.edu Multi-Class Classification Mehryar Mohri - Introduction to Machine Learning page 2 Motivation

More information

On the Average Path Length of Complete m-ary Trees

On the Average Path Length of Complete m-ary Trees 1 2 3 47 6 23 11 Journal of Integer Sequences, Vol. 17 2014, Article 14.6.3 On the Average Path Length of Complete m-ary Trees M. A. Nyblom School of Mathematics and Geospatial Science RMIT University

More information

The Fast Optimal Voltage Partitioning Algorithm For Peak Power Density Minimization

The Fast Optimal Voltage Partitioning Algorithm For Peak Power Density Minimization The Fast Optimal Voltage Partitioning Algorithm For Peak Power Density Minimization Jia Wang, Shiyan Hu Department of Electrical and Computer Engineering Michigan Technological University Houghton, Michigan

More information

Chenhan D. Yu The 3rd BLIS Retreat Sep 28, 2015

Chenhan D. Yu The 3rd BLIS Retreat Sep 28, 2015 GSKS GSKNN BLIS-Based High Performance Computing Kernels in N-body Problems Chenhan D. Yu The 3rd BLIS Retreat Sep 28, 2015 N-body Problems Hellstorm Astronomy and 3D https://youtu.be/bllwkx_mrfk 2 N-body

More information

arxiv: v1 [cs.lg] 26 Jul 2017

arxiv: v1 [cs.lg] 26 Jul 2017 Updating Singular Value Decomposition for Rank One Matrix Perturbation Ratnik Gandhi, Amoli Rajgor School of Engineering & Applied Science, Ahmedabad University, Ahmedabad-380009, India arxiv:70708369v

More information

Prediction of double gene knockout measurements

Prediction of double gene knockout measurements Prediction of double gene knockout measurements Sofia Kyriazopoulou-Panagiotopoulou sofiakp@stanford.edu December 12, 2008 Abstract One way to get an insight into the potential interaction between a pair

More information

CSC321 Lecture 2: Linear Regression

CSC321 Lecture 2: Linear Regression CSC32 Lecture 2: Linear Regression Roger Grosse Roger Grosse CSC32 Lecture 2: Linear Regression / 26 Overview First learning algorithm of the course: linear regression Task: predict scalar-valued targets,

More information

FINAL: CS 6375 (Machine Learning) Fall 2014

FINAL: CS 6375 (Machine Learning) Fall 2014 FINAL: CS 6375 (Machine Learning) Fall 2014 The exam is closed book. You are allowed a one-page cheat sheet. Answer the questions in the spaces provided on the question sheets. If you run out of room for

More information

A fast algorithm to generate necklaces with xed content

A fast algorithm to generate necklaces with xed content Theoretical Computer Science 301 (003) 477 489 www.elsevier.com/locate/tcs Note A fast algorithm to generate necklaces with xed content Joe Sawada 1 Department of Computer Science, University of Toronto,

More information

Math 671: Tensor Train decomposition methods

Math 671: Tensor Train decomposition methods Math 671: Eduardo Corona 1 1 University of Michigan at Ann Arbor December 8, 2016 Table of Contents 1 Preliminaries and goal 2 Unfolding matrices for tensorized arrays The Tensor Train decomposition 3

More information

Intensity Analysis of Spatial Point Patterns Geog 210C Introduction to Spatial Data Analysis

Intensity Analysis of Spatial Point Patterns Geog 210C Introduction to Spatial Data Analysis Intensity Analysis of Spatial Point Patterns Geog 210C Introduction to Spatial Data Analysis Chris Funk Lecture 4 Spatial Point Patterns Definition Set of point locations with recorded events" within study

More information

k-protected VERTICES IN BINARY SEARCH TREES

k-protected VERTICES IN BINARY SEARCH TREES k-protected VERTICES IN BINARY SEARCH TREES MIKLÓS BÓNA Abstract. We show that for every k, the probability that a randomly selected vertex of a random binary search tree on n nodes is at distance k from

More information

High-Dimensional Indexing by Distributed Aggregation

High-Dimensional Indexing by Distributed Aggregation High-Dimensional Indexing by Distributed Aggregation Yufei Tao ITEE University of Queensland In this lecture, we will learn a new approach for indexing high-dimensional points. The approach borrows ideas

More information

ERLANGEN REGIONAL COMPUTING CENTER

ERLANGEN REGIONAL COMPUTING CENTER ERLANGEN REGIONAL COMPUTING CENTER Making Sense of Performance Numbers Georg Hager Erlangen Regional Computing Center (RRZE) Friedrich-Alexander-Universität Erlangen-Nürnberg OpenMPCon 2018 Barcelona,

More information

Final Exam, Fall 2002

Final Exam, Fall 2002 15-781 Final Exam, Fall 22 1. Write your name and your andrew email address below. Name: Andrew ID: 2. There should be 17 pages in this exam (excluding this cover sheet). 3. If you need more room to work

More information

Notes on Computer Systems for Solving Symbolic Equations

Notes on Computer Systems for Solving Symbolic Equations Notes on Computer Systems for Solving Symbolic Equations Richard J. Fateman March, 1991, revisited 2005 Abstract Math students learn to solve single equations symbolically beginning in elementary or junior

More information

Entropic Affinities: Properties and Efficient Numerical Computation

Entropic Affinities: Properties and Efficient Numerical Computation Entropic Affinities: Properties and Efficient Numerical Computation Max Vladymyrov and Miguel Carreira-Perpiñán Electrical Engineering and Computer Science University of California, Merced http://eecs.ucmerced.edu

More information

An Overview of Fast Multipole Methods

An Overview of Fast Multipole Methods An Overview of Fast Multipole Methods Alexander T. Ihler April 23, 2004 Sing, Muse, of the low-rank approximations Of spherical harmonics and O(N) computation... Abstract We present some background and

More information

Power series and Taylor series

Power series and Taylor series Power series and Taylor series D. DeTurck University of Pennsylvania March 29, 2018 D. DeTurck Math 104 002 2018A: Series 1 / 42 Series First... a review of what we have done so far: 1 We examined series

More information

SMT 2013 Power Round Solutions February 2, 2013

SMT 2013 Power Round Solutions February 2, 2013 Introduction This Power Round is an exploration of numerical semigroups, mathematical structures which appear very naturally out of answers to simple questions. For example, suppose McDonald s sells Chicken

More information

Direct and Incomplete Cholesky Factorizations with Static Supernodes

Direct and Incomplete Cholesky Factorizations with Static Supernodes Direct and Incomplete Cholesky Factorizations with Static Supernodes AMSC 661 Term Project Report Yuancheng Luo 2010-05-14 Introduction Incomplete factorizations of sparse symmetric positive definite (SSPD)

More information

Nearest Neighbors Methods for Support Vector Machines

Nearest Neighbors Methods for Support Vector Machines Nearest Neighbors Methods for Support Vector Machines A. J. Quiroz, Dpto. de Matemáticas. Universidad de Los Andes joint work with María González-Lima, Universidad Simón Boĺıvar and Sergio A. Camelo, Universidad

More information

Appendix C: Recapitulation of Numerical schemes

Appendix C: Recapitulation of Numerical schemes Appendix C: Recapitulation of Numerical schemes August 31, 2009) SUMMARY: Certain numerical schemes of general use are regrouped here in order to facilitate implementations of simple models C1 The tridiagonal

More information

CS145: INTRODUCTION TO DATA MINING

CS145: INTRODUCTION TO DATA MINING CS145: INTRODUCTION TO DATA MINING 5: Vector Data: Support Vector Machine Instructor: Yizhou Sun yzsun@cs.ucla.edu October 18, 2017 Homework 1 Announcements Due end of the day of this Thursday (11:59pm)

More information

Recent Results for Moving Least Squares Approximation

Recent Results for Moving Least Squares Approximation Recent Results for Moving Least Squares Approximation Gregory E. Fasshauer and Jack G. Zhang Abstract. We describe two experiments recently conducted with the approximate moving least squares (MLS) approximation

More information

Tight Bounds on the Diameter of Gaussian Cubes

Tight Bounds on the Diameter of Gaussian Cubes Tight Bounds on the Diameter of Gaussian Cubes DING-MING KWAI AND BEHROOZ PARHAMI Department of Electrical and Computer Engineering, University of California, Santa Barbara, CA 93106 9560, USA Email: parhami@ece.ucsb.edu

More information

Learning from Examples

Learning from Examples Learning from Examples Data fitting Decision trees Cross validation Computational learning theory Linear classifiers Neural networks Nonparametric methods: nearest neighbor Support vector machines Ensemble

More information

On the Relationship between Sum-Product Networks and Bayesian Networks

On the Relationship between Sum-Product Networks and Bayesian Networks On the Relationship between Sum-Product Networks and Bayesian Networks International Conference on Machine Learning, 2015 Han Zhao Mazen Melibari Pascal Poupart University of Waterloo, Waterloo, ON, Canada

More information

Testing Problems with Sub-Learning Sample Complexity

Testing Problems with Sub-Learning Sample Complexity Testing Problems with Sub-Learning Sample Complexity Michael Kearns AT&T Labs Research 180 Park Avenue Florham Park, NJ, 07932 mkearns@researchattcom Dana Ron Laboratory for Computer Science, MIT 545 Technology

More information

Communication avoiding parallel algorithms for dense matrix factorizations

Communication avoiding parallel algorithms for dense matrix factorizations Communication avoiding parallel dense matrix factorizations 1/ 44 Communication avoiding parallel algorithms for dense matrix factorizations Edgar Solomonik Department of EECS, UC Berkeley October 2013

More information

Support Vector Machines.

Support Vector Machines. Support Vector Machines www.cs.wisc.edu/~dpage 1 Goals for the lecture you should understand the following concepts the margin slack variables the linear support vector machine nonlinear SVMs the kernel

More information

Statistical Machine Learning

Statistical Machine Learning Statistical Machine Learning Christoph Lampert Spring Semester 2015/2016 // Lecture 12 1 / 36 Unsupervised Learning Dimensionality Reduction 2 / 36 Dimensionality Reduction Given: data X = {x 1,..., x

More information

1 Differential Privacy and Statistical Query Learning

1 Differential Privacy and Statistical Query Learning 10-806 Foundations of Machine Learning and Data Science Lecturer: Maria-Florina Balcan Lecture 5: December 07, 015 1 Differential Privacy and Statistical Query Learning 1.1 Differential Privacy Suppose

More information