Distributions of Persistence Diagrams and Approximations
|
|
- Theodore Walters
- 5 years ago
- Views:
Transcription
1 Distributions of Persistence Diagrams and Approximations Vasileios Maroulas Department of Mathematics Department of Business Analytics and Statistics University of Tennessee August 31, 2018
2 Thanks V. Maroulas (UTK) KDE of random persistence diagrams August 31, / 45
3 Joint with: Funded by: Josh Mike (now at Michigan State) Andrew Marchese (now at Plated) John Sgouralis (now at Arizona State) Chris Oballe V. Maroulas (UTK) KDE of random persistence diagrams August 31, / 45
4 Acoustic Signals at the ARL Two classes representing two different types of weapons. Goal is to help military officers make tactical decisions based on the type of weapon system. V. Maroulas (UTK) KDE of random persistence diagrams August 31, / 45
5 Signals from ARL Dataset V. Maroulas (UTK) KDE of random persistence diagrams August 31, / 45
6 Merge statistics and topology to understand the geometry of signals and classify them. TDA has recently been introduced to the field of signal and time-series classification. Biological Signals (Zhang et al. (2015)) Action Recognition (Venkataraman et al. (2016)) Wheeze Detection (Emrani et al. (2014)) V. Maroulas (UTK) KDE of random persistence diagrams August 31, / 45
7 Motivation Data has shape and shape matters. Latent topological features in scientific data. VM and A. Nebenführ, Tracking rapid intracellular movements: a Bayesian random set approach, Annals of Applied Statistics, I. Sgouralis, A. Nebenführ and VM, A Bayesian Topological Framework for the Identification and Reconstruction of Subcellular Motion, SIAM on Imaging Sciences, J. Mike, C. Sumrall, VM and F. Schwartz. Non-Landmark Classification in Paleobiology, Paleobiology, V. Maroulas (UTK) KDE of random persistence diagrams August 31, / 45
8 Key Picture Data lies in a topological space. Take measurements, sampling that space. Reconstruct it by using an approximation. Compute the invariants to understand it. V. Maroulas (UTK) KDE of random persistence diagrams August 31, / 45
9 From Signals to Point Clouds (Taken s Thm) Suppose that w : [0, T] R are the realizations of a discrete time series in [0, T]. Consider a set of delay indices τ 1, τ 2,, τ n 1 The n dimensional delay embedding of W is the concatenation of time-delayed samples: W(t) = (w(t), w(t + τ 1 ), w(t + τ 2 ),, w(t + τ n 1 )) (1) 4.4 Point Cloud w(t) w(t+τ 2 ) Time (t) w(t+τ 1 ) w(t) Figure: Signal evolution in time domain Figure: 3D delay embedding V. Maroulas (UTK) KDE of random persistence diagrams August 31, / 45
10 From Point Clouds to Inference and Classification Now we have turned our time-series into a point cloud living in N-dimensional space. Point Cloud 4.5 w(t+τ 2 ) w(t+τ 1 ) w(t) How can we extract information from this data and use it for classification and statistical inference? V. Maroulas (UTK) KDE of random persistence diagrams August 31, / 45
11 Outline Data Analysis using Persistence Homology Distribution of random persistence diagrams Kernel Density Estimation Conclusion V. Maroulas (UTK) KDE of random persistence diagrams August 31, / 45
12 Simplicial Complex Simplicial complexes are discretizations of real-life shapes Generalization of graphs with higher order relationships among the nodes. A simplicial complex is the union of simple pieces (simplices) i.e. vertices, edges, triangles etc. A face of k simplex are all the (k 1) simplex. Two simplices must intersect at a common face or not at all. V. Maroulas (UTK) KDE of random persistence diagrams August 31, / 45
13 Construction of Simplicial complexes for data Start with a point-cloud Π and create an abstract representation of vertices one for each point in your Π. Figure: Left: Point Cloud; Right: Simplicial Complex V. Maroulas (UTK) KDE of random persistence diagrams August 31, / 45
14 Construction of Simplicial complexes for data Create spheres" of radius r centered at each point. Figure: Left: Point Cloud; Right: Simplicial Complex V. Maroulas (UTK) KDE of random persistence diagrams August 31, / 45
15 Construction of Simplicial complexes for data Increase radius r. Figure: Left: Point Cloud; Right: Simplicial Complex V. Maroulas (UTK) KDE of random persistence diagrams August 31, / 45
16 Construction of Simplicial complexes for data Add edges between vertices v i and v j if the corresponding circles intersect. Figure: Left: Point Cloud; Right: Simplicial Complex V. Maroulas (UTK) KDE of random persistence diagrams August 31, / 45
17 Construction of Simplicial complexes for data Add triangles between vertices v i, v j and v k if all three circles intersect, etc. Figure: Left: Point Cloud; Right: Simplicial Complex V. Maroulas (UTK) KDE of random persistence diagrams August 31, / 45
18 Based on the simplicial complex we retrieve the Betti numbers (the dimensions of some vector spaces associated to our topological space) Betti 0: number of clusters" Betti 1: number of holes Figure: β 0 = 2, β 1 = 0 V. Maroulas (UTK) KDE of random persistence diagrams August 31, / 45
19 Based on the simplicial complex we retrieve the Betti numbers (the dimensions of some vector spaces associated to our topological space) Betti 0: number of clusters" Betti 1: number of holes Figure: β 0 = 2, β 1 = 0 Figure: β 0 = 2, β 1 = 1 V. Maroulas (UTK) KDE of random persistence diagrams August 31, / 45
20 Persistence Diagrams Interested in is the persistence of the Betti numbers. When do different connected components/holes form and how long do they last (with respect to r)? The Betti numbers compactly encoded in a 2-dim plot which provides the birth time vs death time of these features V. Maroulas (UTK) KDE of random persistence diagrams August 31, / 45
21 Results on Signals from ARL Dataset (a) (b) (c) (d) (e) (f) A. Marchese and VM. Signal classification with a point process distance on the space of persistence diagrams. Advances in Data Analysis and Classification, pp. 1-26, 2017 V. Maroulas (UTK) KDE of random persistence diagrams August 31, / 45
22 Classifier V. Maroulas (UTK) KDE of random persistence diagrams August 31, / 45
23 Statistics and Persistence Diagrams Summary statistics such as center and variance (Bobrowski et al., 2014; Mileyko et al., 2011; Turner et al., 2014; Marchese and VM, 2017) Birth and death estimates (Emmett et al., 2014) Confidence sets (Fasy et al., 2014) Need a framework to understand the above summary statistics through a single viewpoint V. Maroulas (UTK) KDE of random persistence diagrams August 31, / 45
24 Novel framework A complete and consistent framework of how to construct distributions of persistence diagrams. Capture the important information of these diagrams in terms of their inherent set properties Set membership Cardinality V. Maroulas (UTK) KDE of random persistence diagrams August 31, / 45
25 Setup Take data X = x j generated by some random process Associated (random) persistence diagram D with features ξ i = (b i, d i ), such that the a hole" appears at scale b i and is filled at scale d i V. Maroulas (UTK) KDE of random persistence diagrams August 31, / 45
26 Lemma 2.1 (J. Mike & VM, 2018) Consider a multiset of independent singleton random persistence diagrams { D j} M j=1. If each singleton Dj is described by the value q (j) = P[D j ] and the subsequent conditional pdf, p (j) (ξ), given D j = 1, then the global pdf for D = M j=1 D j is given by f D (ξ 1,..., ξ N ) = γ I(N,M) Q(γ) N p (γ(k)) (ξ k ), (2) for each N {0,..., M} where I(N, M) consists of all increasing injections γ : {1,..., N} {1,..., M}, and The sum over γ I(N, M) in Eq. (2) accounts for each possible combination of singleton presence. The weights Q(γ) proportional to the probability for each singleton to be either present, q(j), or absent, 1 q(j), for each j. k=1 J. Mike and VM. Nonparametric Estimation of Probability Density Functions of Random Persistence Diagrams. arxiv: V. Maroulas (UTK) KDE of random persistence diagrams August 31, / 45
27 Example 2.2 Consider two 1-dimensional singleton diagrams, D 1 and D 2, with probabilities of being nonempty q (1) = 0.6 and q (2) = 0.8. Local densities when nonempty: p (1) (x) = 1 2π e (x+1)2 /2 and p (2) (x) = 1 2π e (x 1)2 /2. Global pdf for D = D 1 D 2 through a set of local densities { f0, f 1 (x), f 2 (x, y) } such that f 0 = P[ D = 0] = (1 q (1) )(1 q (2) ) = 0.08 f 1 (x) = (1 q (2) )q (1) p (1) (x) + (1 q (1) )q (2) p (2) (x) = 0.12 e (x+1)2 / e (x 1)2 /2, 2π 2π f 2 (x, y) = q(1) q (2) 2 = π [ ] p (1) (x)p (2) (y) + p (1) (y)p (2) (x) ( e ((x 1)2 +(y+1) 2 )/2 + e ((x+1)2 +(y 1) 2 )/2 ). (3a) (3b) (3c) V. Maroulas (UTK) KDE of random persistence diagrams August 31, / 45
28 Figure: Left: Plot of the local density f 1(x) in Eq. (3b); Right: Contour plot of the local density f 2(x, y) in Eq. (3c). These pdfs cover the different possible input dimensions and are symmetric under permutations of the input. V. Maroulas (UTK) KDE of random persistence diagrams August 31, / 45
29 Past Studies-KDE Extensive work to devise various maps from persistence diagrams into Hilbert spaces Mapping into a Hilbert space, these studies allow the application of statistical learning methods such as principal component analysis, random forest, support vector machine, etc. Chepushtanova et al. (2015) discretizes persistence diagrams via bins, yielding vectors in a high dimensional Euclidean space. Reininghaus et al., (2014) and Kusano et al., (2016) define kernels between persistence diagrams in a Reproducing Kernel Hilbert Spaces Adler et al. (2017) utilizes Gibbs distributions in order to replicate similar persistence diagrams, e.g. for use in MCMC type sampling. Kernel density estimation on the underlying data to estimate a target diagram Bobrowski et al. (2014) constructs an estimator for the target diagram Fasy et al. (2014) defines a confidence set. V. Maroulas (UTK) KDE of random persistence diagrams August 31, / 45
30 Building a Kernel Density Goal: kernel density K σ (Z, D). Center Diagram D Bandwidth σ Input Z. Split D into upper and lower halves: D u = { (b, d) D : d b > σ } D l = { (b, d) D : d b σ } Define random diagrams: D u centered at D u. D l centered at D l. V. Maroulas (UTK) KDE of random persistence diagrams August 31, / 45
31 Building the Upper Density. Split into singletons: D u = j D j,u Each D j,u is described by: q j = P[D j,u ]. Local pdf, p j (b, d), Restricted Gaussian V. Maroulas (UTK) KDE of random persistence diagrams August 31, / 45
32 Building the Lower Density. p l (b, d) = 1 N l (b i,d i) D l 1 πσ 2 e Lower cardinality N l = D l. Cardinality probability mass: ν(j). Chosen with mean N l. Single kernel density p l : Project D l onto the diagonal. Kernel estimate for these points. D l has number according to ν, with draws i.i.d. according to p l. ( ( ) b b i +d i 2+ ( 2 Global pdf D l : f D l(ξ 1,..., ξ N ) = ν(n) d b i +d i 2 ) 2 ) /2σ 2. (4) N p l (ξ j ). (5) j=1 V. Maroulas (UTK) KDE of random persistence diagrams August 31, / 45
33 The Kernel Density Theorem 3.1 (J. Mike & VM, 2018) The random diagram D = D u D l with D u and D l defined according to the previous construction with center D and bandwidth σ has the following global pdf, or kernel density, evaluated at Z = {ξ 1,..., ξ N }: N u K σ (Z, D) = ν(n j) j=0 γ I(j,N) Q(γ) j p γ(k) (ξ k ) k=1 N k=j+1 where I(j, N) = { γ : {1,..., j} {1,..., N} : γ is increasing }, and Q(γ) = p l (ξ k ) (6) N ( ) j=1 1 q j j ( ) q γ(k). (7) 1 q γ(k) j k=1 k=1 V. Maroulas (UTK) KDE of random persistence diagrams August 31, / 45
34 Example 3.2 Consider D = ((1, 3), (2, 4), (1, 1.3), (3, 3.2)) and σ = 1/2. V. Maroulas (UTK) KDE of random persistence diagrams August 31, / 45
35 Kernel with input cardinality 1 Kernel K σ ((b 1, d 1 ), D) equals to: [ ν(0) (1 q (2) )q (1) p (1) (b 1, d 1 ) + (1 q (1) )q (2) p (2) ] (b 1, d 1 ) [ + ν(1) (1 q (1) )(1 q (2) )p l ] (b 1, d 1 ) (8) Figure: Contour map for the kernel density restricted to a single input feature (Eq. (8)). The center diagram is indicated by red (upper) and green (lower) points. Scale bars at the right of each plot indicate the range of probability density in each shaded region p l (b, d) = 2 [e ((b 1.15)2 +(d 1.15) 2 ) + e ((b 3.1) 2 +(d 3.1) 2 ) ] π p (1) (b 1, d 1 ) e 2((b1 2)2 +(d 1 4) 2 ) p (2) (b 1, d 1 ) e 2((b1 1)2 +(d 1 3) 2 ) V. Maroulas (UTK) KDE of random persistence diagrams August 31, / 45
36 Kernel with input cardinality 2 Consider Z = (ξ 1, ξ 2 ) = ((b 1, d 1 ), (b 2, d 2 )) K σ ((ξ 1, ξ 2 ), D) = ν(0)q (1) q (2) p (1) (b 1, d 1 )p (2) (b 2, d 2 ) [ + ν(1) (1 q (2) )q (1) p (1) (b 1, d 1 ) ] + (1 q (1) )q (2) p (2) (b 1, d 1 ) p l (b 2, d 2 ) (9) + ν(2)(1 q (1) )(1 q (2) )p l (b 1, d 1 )p l (b 2, d 2 ) V. Maroulas (UTK) KDE of random persistence diagrams August 31, / 45
37 Kernel with input cardinality 2 (a) (b) (c) Figure: Contour maps for slices of the kernel density K σ((ξ, ξ 2), D) with input cardinality 2. A single feature ξ 2, indicated by white crosshairs, is fixed to restrict to a 2D subspace as follows: (a) ξ 2 = (1, 3) (b) ξ 2 = (2, 4) and (c) ξ 2 = (2.5, 2.7). The center diagram is indicated by red (upper) and green (lower) points. Scale bars at the right of each plot indicate the range of probability density in each shaded region. V. Maroulas (UTK) KDE of random persistence diagrams August 31, / 45
38 Kernel with input cardinality 3 [ ] K σ ((ξ 1, ξ 2, ξ 3 ), D) = ν(1) q (1) q (2) p (1) (b 1, d 1 )p (2) (b 2, d 2 ) p l (b 3, d 3 ) [ + ν(2) (1 q (2) )q (1) p (1) (b 1, d 1 ) ] + ν(2)(1 q (1) )q (2) p (2) (b 1, d 1 ) p l (b 2, d 2 )p l (b 3, d 3 ) + ν(3)(1 q (1) )(1 q (2) )p l (b 1, d 1 )p l (b 2, d 2 )p l (b 3, d 3 ). (10) V. Maroulas (UTK) KDE of random persistence diagrams August 31, / 45
39 Kernel with input cardinality 3 (a) Figure: Contour maps for slices of the kernel density K σ((ξ, ξ 2, ξ 3), D) with input cardinality 3. A pair of features ξ 2 and ξ 3, indicated by white crosshairs, are fixed to restrict to a 2D subspace as follows: (a) (ξ 2, ξ 3) = ((1, 3), (2, 4)) and (b) (ξ 2, ξ 3) = ((1, 3), (2.5, 3.5)). The center diagram is indicated by red (upper) and green (lower) points. Scale bars at the right of each plot indicate the range of probability density in each shaded region. (b) V. Maroulas (UTK) KDE of random persistence diagrams August 31, / 45
40 Kernel Density Estimation Theorem 3.3 (J. Mike & VM, 2018) Random persistence diagram with global pdf f : f must satisfy decay and boundedness conditions. Diagrams {Di} n i=1 sampled i.i.d. according to f. Yield a KDE: ˆf (Z) = 1 n n i=1 K σ(z, D i ) σ = O(n α ) chosen with 0 < α α 2M, ˆf f uniformly on compact subsets of the space of PDs. J. Mike and VM. Nonparametric Estimation of Probability Density Functions of Random Persistence Diagrams. arxiv: V. Maroulas (UTK) KDE of random persistence diagrams August 31, / 45
41 Example 3.4 Generate samples which each consist of 10 points uniformly from the unit circle with additive Gaussian noise, N((0, 0), ( 1 50 )2 I 2 ). Toy dataset for signal analysis. (a) (b) V. Maroulas (UTK) KDE of random persistence diagrams August 31, / 45
42 Plots of persistence diagram KDEs. Color indicates the probability density. White regions above the diagonal indicate portions of very low probability density. Each column is a particular slice, while each row is a particular KDE with n and σ. Left: Local KDEs ˆf n,σ((b, d)) evaluated at a diagram with only one feature. The mode of the converged density is approximately (b 2, d 2) = (0.77, 0.98). Right: Local KDEs ˆf n,σ((b, d), (0.77, 0.98)) evaluated at a diagram with two features and one feature fixed.these slices have two modes which are very close to the diagonal at (0, 0) and (1, 1). V. Maroulas (UTK) KDE of random persistence diagrams August 31, / 45
43 Summary Considered the problem of estimating the distribution of persistence diagrams Established a novel kernel density We focused on set properties-membership and cardinality We established convergence and verified several synthetic examples. With a pdf at hand, we can start implementing Monte Carlo sampling and move on to further probabilistic settings Bayesian formulation Applications in biology, defense, material science and chemistry. V. Maroulas (UTK) KDE of random persistence diagrams August 31, / 45
44 Bayesian Framework In principle, we can compute posterior distributions using Bayes theorem for random sets π(d X D Y ) l(d Y D X )π(d X ) V. Maroulas (UTK) KDE of random persistence diagrams August 31, / 45
45 Posterior Approximation Mahler (2003) Singh, Vo, Baddeley and Zuyev (2007) Caron, Del Moral, Doucet, Pace (2011) Prior Prior Posterior Persistence Intensity Persistence Intensity Persistence Intensity Birth (a) Birth (b) Birth (c) V. Maroulas (UTK) KDE of random persistence diagrams August 31, / 45
46 Thank you-questions? V. Maroulas (UTK) KDE of random persistence diagrams August 31, / 45
Neuronal structure detection using Persistent Homology
Neuronal structure detection using Persistent Homology J. Heras, G. Mata and J. Rubio Department of Mathematics and Computer Science, University of La Rioja Seminario de Informática Mirian Andrés March
More informationProbabilistic Machine Learning. Industrial AI Lab.
Probabilistic Machine Learning Industrial AI Lab. Probabilistic Linear Regression Outline Probabilistic Classification Probabilistic Clustering Probabilistic Dimension Reduction 2 Probabilistic Linear
More informationGaussian Processes. Le Song. Machine Learning II: Advanced Topics CSE 8803ML, Spring 2012
Gaussian Processes Le Song Machine Learning II: Advanced Topics CSE 8803ML, Spring 01 Pictorial view of embedding distribution Transform the entire distribution to expected features Feature space Feature
More informationLearning from persistence diagrams
Learning from persistence diagrams Ulrich Bauer TUM July 7, 2015 ACAT final meeting IST Austria Joint work with: Jan Reininghaus, Stefan Huber, Roland Kwitt 1 / 28 ? 2 / 28 ? 2 / 28 3 / 28 3 / 28 3 / 28
More informationECE521 week 3: 23/26 January 2017
ECE521 week 3: 23/26 January 2017 Outline Probabilistic interpretation of linear regression - Maximum likelihood estimation (MLE) - Maximum a posteriori (MAP) estimation Bias-variance trade-off Linear
More informationKernel methods, kernel SVM and ridge regression
Kernel methods, kernel SVM and ridge regression Le Song Machine Learning II: Advanced Topics CSE 8803ML, Spring 2012 Collaborative Filtering 2 Collaborative Filtering R: rating matrix; U: user factor;
More informationMatthew L. Wright. St. Olaf College. June 15, 2017
Matthew L. Wright St. Olaf College June 15, 217 Persistent homology detects topological features of data. For this data, the persistence barcode reveals one significant hole in the point cloud. Problem:
More informationTopological Data Analysis - II. Afra Zomorodian Department of Computer Science Dartmouth College
Topological Data Analysis - II Afra Zomorodian Department of Computer Science Dartmouth College September 4, 2007 1 Plan Yesterday: Motivation Topology Simplicial Complexes Invariants Homology Algebraic
More informationCOURSE INTRODUCTION. J. Elder CSE 6390/PSYC 6225 Computational Modeling of Visual Perception
COURSE INTRODUCTION COMPUTATIONAL MODELING OF VISUAL PERCEPTION 2 The goal of this course is to provide a framework and computational tools for modeling visual inference, motivated by interesting examples
More informationPattern Recognition and Machine Learning
Christopher M. Bishop Pattern Recognition and Machine Learning ÖSpri inger Contents Preface Mathematical notation Contents vii xi xiii 1 Introduction 1 1.1 Example: Polynomial Curve Fitting 4 1.2 Probability
More informationStatistics: Learning models from data
DS-GA 1002 Lecture notes 5 October 19, 2015 Statistics: Learning models from data Learning models from data that are assumed to be generated probabilistically from a certain unknown distribution is a crucial
More informationNonparametric Bayesian Methods
Nonparametric Bayesian Methods Debdeep Pati Florida State University October 2, 2014 Large spatial datasets (Problem of big n) Large observational and computer-generated datasets: Often have spatial and
More informationApplications of Persistent Homology to Time-Varying Systems
Applications of Persistent Homology to Time-Varying Systems Elizabeth Munch Duke University :: Dept of Mathematics March 28, 2013 Time Death Radius Birth Radius Liz Munch (Duke) PhD Defense March 28, 2013
More informationSTAT 518 Intro Student Presentation
STAT 518 Intro Student Presentation Wen Wei Loh April 11, 2013 Title of paper Radford M. Neal [1999] Bayesian Statistics, 6: 475-501, 1999 What the paper is about Regression and Classification Flexible
More informationSome Topics in Computational Topology. Yusu Wang. AMS Short Course 2014
Some Topics in Computational Topology Yusu Wang AMS Short Course 2014 Introduction Much recent developments in computational topology Both in theory and in their applications E.g, the theory of persistence
More informationBayesian Nonparametrics
Bayesian Nonparametrics Peter Orbanz Columbia University PARAMETERS AND PATTERNS Parameters P(X θ) = Probability[data pattern] 3 2 1 0 1 2 3 5 0 5 Inference idea data = underlying pattern + independent
More informationA Kernel on Persistence Diagrams for Machine Learning
A Kernel on Persistence Diagrams for Machine Learning Jan Reininghaus 1 Stefan Huber 1 Roland Kwitt 2 Ulrich Bauer 1 1 Institute of Science and Technology Austria 2 FB Computer Science Universität Salzburg,
More informationChapter 3: Homology Groups Topics in Computational Topology: An Algorithmic View
Chapter 3: Homology Groups Topics in Computational Topology: An Algorithmic View As discussed in Chapter 2, we have complete topological information about 2-manifolds. How about higher dimensional manifolds?
More informationNonparametric Bayesian Methods (Gaussian Processes)
[70240413 Statistical Machine Learning, Spring, 2015] Nonparametric Bayesian Methods (Gaussian Processes) Jun Zhu dcszj@mail.tsinghua.edu.cn http://bigml.cs.tsinghua.edu.cn/~jun State Key Lab of Intelligent
More informationData Mining Techniques
Data Mining Techniques CS 6220 - Section 3 - Fall 2016 Lecture 12 Jan-Willem van de Meent (credit: Yijun Zhao, Percy Liang) DIMENSIONALITY REDUCTION Borrowing from: Percy Liang (Stanford) Linear Dimensionality
More informationSome Topics in Computational Topology
Some Topics in Computational Topology Yusu Wang Ohio State University AMS Short Course 2014 Introduction Much recent developments in computational topology Both in theory and in their applications E.g,
More informationIntroduction to Machine Learning CMU-10701
Introduction to Machine Learning CMU-10701 Markov Chain Monte Carlo Methods Barnabás Póczos & Aarti Singh Contents Markov Chain Monte Carlo Methods Goal & Motivation Sampling Rejection Importance Markov
More informationSTA 4273H: Statistical Machine Learning
STA 4273H: Statistical Machine Learning Russ Salakhutdinov Department of Computer Science! Department of Statistical Sciences! rsalakhu@cs.toronto.edu! h0p://www.cs.utoronto.ca/~rsalakhu/ Lecture 7 Approximate
More informationSTA414/2104 Statistical Methods for Machine Learning II
STA414/2104 Statistical Methods for Machine Learning II Murat A. Erdogdu & David Duvenaud Department of Computer Science Department of Statistical Sciences Lecture 3 Slide credits: Russ Salakhutdinov Announcements
More informationThe Intersection of Statistics and Topology:
The Intersection of Statistics and Topology: Confidence Sets for Persistence Diagrams Brittany Terese Fasy joint work with S. Balakrishnan, F. Lecci, A. Rinaldo, A. Singh, L. Wasserman 3 June 2014 [FLRWBS]
More informationIntroduction to Bayesian Learning
Course Information Introduction Introduction to Bayesian Learning Davide Bacciu Dipartimento di Informatica Università di Pisa bacciu@di.unipi.it Apprendimento Automatico: Fondamenti - A.A. 2016/2017 Outline
More information7.3 Singular Homology Groups
184 CHAPTER 7. HOMOLOGY THEORY 7.3 Singular Homology Groups 7.3.1 Cycles, Boundaries and Homology Groups We can define the singular p-chains with coefficients in a field K. Furthermore, we can define the
More informationRecent Advances in Bayesian Inference Techniques
Recent Advances in Bayesian Inference Techniques Christopher M. Bishop Microsoft Research, Cambridge, U.K. research.microsoft.com/~cmbishop SIAM Conference on Data Mining, April 2004 Abstract Bayesian
More informationPATTERN RECOGNITION AND MACHINE LEARNING CHAPTER 13: SEQUENTIAL DATA
PATTERN RECOGNITION AND MACHINE LEARNING CHAPTER 13: SEQUENTIAL DATA Contents in latter part Linear Dynamical Systems What is different from HMM? Kalman filter Its strength and limitation Particle Filter
More informationSTA 4273H: Statistical Machine Learning
STA 4273H: Statistical Machine Learning Russ Salakhutdinov Department of Statistics! rsalakhu@utstat.toronto.edu! http://www.utstat.utoronto.ca/~rsalakhu/ Sidney Smith Hall, Room 6002 Lecture 3 Linear
More informationIntroduction to Machine Learning Midterm Exam
10-701 Introduction to Machine Learning Midterm Exam Instructors: Eric Xing, Ziv Bar-Joseph 17 November, 2015 There are 11 questions, for a total of 100 points. This exam is open book, open notes, but
More informationBeyond Scalar Affinities for Network Analysis or Vector Diffusion Maps and the Connection Laplacian
Beyond Scalar Affinities for Network Analysis or Vector Diffusion Maps and the Connection Laplacian Amit Singer Princeton University Department of Mathematics and Program in Applied and Computational Mathematics
More information27 : Distributed Monte Carlo Markov Chain. 1 Recap of MCMC and Naive Parallel Gibbs Sampling
10-708: Probabilistic Graphical Models 10-708, Spring 2014 27 : Distributed Monte Carlo Markov Chain Lecturer: Eric P. Xing Scribes: Pengtao Xie, Khoa Luu In this scribe, we are going to review the Parallel
More informationDensity Propagation for Continuous Temporal Chains Generative and Discriminative Models
$ Technical Report, University of Toronto, CSRG-501, October 2004 Density Propagation for Continuous Temporal Chains Generative and Discriminative Models Cristian Sminchisescu and Allan Jepson Department
More informationUnsupervised Learning Methods
Structural Health Monitoring Using Statistical Pattern Recognition Unsupervised Learning Methods Keith Worden and Graeme Manson Presented by Keith Worden The Structural Health Monitoring Process 1. Operational
More informationBayesian Networks BY: MOHAMAD ALSABBAGH
Bayesian Networks BY: MOHAMAD ALSABBAGH Outlines Introduction Bayes Rule Bayesian Networks (BN) Representation Size of a Bayesian Network Inference via BN BN Learning Dynamic BN Introduction Conditional
More informationMachine Learning. Lecture 4: Regularization and Bayesian Statistics. Feng Li. https://funglee.github.io
Machine Learning Lecture 4: Regularization and Bayesian Statistics Feng Li fli@sdu.edu.cn https://funglee.github.io School of Computer Science and Technology Shandong University Fall 207 Overfitting Problem
More informationReview. DS GA 1002 Statistical and Mathematical Models. Carlos Fernandez-Granda
Review DS GA 1002 Statistical and Mathematical Models http://www.cims.nyu.edu/~cfgranda/pages/dsga1002_fall16 Carlos Fernandez-Granda Probability and statistics Probability: Framework for dealing with
More informationMetric Thickenings of Euclidean Submanifolds
Metric Thickenings of Euclidean Submanifolds Advisor: Dr. Henry Adams Committe: Dr. Chris Peterson, Dr. Daniel Cooley Joshua Mirth Masters Thesis Defense October 3, 2017 Introduction Motivation Can we
More informationDynamic Data Modeling, Recognition, and Synthesis. Rui Zhao Thesis Defense Advisor: Professor Qiang Ji
Dynamic Data Modeling, Recognition, and Synthesis Rui Zhao Thesis Defense Advisor: Professor Qiang Ji Contents Introduction Related Work Dynamic Data Modeling & Analysis Temporal localization Insufficient
More informationMachine learning for pervasive systems Classification in high-dimensional spaces
Machine learning for pervasive systems Classification in high-dimensional spaces Department of Communications and Networking Aalto University, School of Electrical Engineering stephan.sigg@aalto.fi Version
More informationAdvanced Introduction to Machine Learning
10-715 Advanced Introduction to Machine Learning Homework 3 Due Nov 12, 10.30 am Rules 1. Homework is due on the due date at 10.30 am. Please hand over your homework at the beginning of class. Please see
More informationCSC 2541: Bayesian Methods for Machine Learning
CSC 2541: Bayesian Methods for Machine Learning Radford M. Neal, University of Toronto, 2011 Lecture 3 More Markov Chain Monte Carlo Methods The Metropolis algorithm isn t the only way to do MCMC. We ll
More informationAn Introduction to Topological Data Analysis
An Introduction to Topological Data Analysis Elizabeth Munch Duke University :: Dept of Mathematics Tuesday, June 11, 2013 Elizabeth Munch (Duke) CompTop-SUNYIT Tuesday, June 11, 2013 1 / 43 There is no
More informationTutorial on Approximate Bayesian Computation
Tutorial on Approximate Bayesian Computation Michael Gutmann https://sites.google.com/site/michaelgutmann University of Helsinki Aalto University Helsinki Institute for Information Technology 16 May 2016
More informationKernel adaptive Sequential Monte Carlo
Kernel adaptive Sequential Monte Carlo Ingmar Schuster (Paris Dauphine) Heiko Strathmann (University College London) Brooks Paige (Oxford) Dino Sejdinovic (Oxford) December 7, 2015 1 / 36 Section 1 Outline
More informationTopological Data Analysis for Brain Networks
Topological Data Analysis for Brain Networks Relating Functional Brain Network Topology to Clinical Measures of Behavior Bei Wang Phillips 1,2 University of Utah Joint work with Eleanor Wong 1,2, Sourabh
More informationUnsupervised Machine Learning and Data Mining. DS 5230 / DS Fall Lecture 7. Jan-Willem van de Meent
Unsupervised Machine Learning and Data Mining DS 5230 / DS 4420 - Fall 2018 Lecture 7 Jan-Willem van de Meent DIMENSIONALITY REDUCTION Borrowing from: Percy Liang (Stanford) Dimensionality Reduction Goal:
More informationIntroduction to Probability and Statistics (Continued)
Introduction to Probability and Statistics (Continued) Prof. icholas Zabaras Center for Informatics and Computational Science https://cics.nd.edu/ University of otre Dame otre Dame, Indiana, USA Email:
More informationBayesian Linear Regression [DRAFT - In Progress]
Bayesian Linear Regression [DRAFT - In Progress] David S. Rosenberg Abstract Here we develop some basics of Bayesian linear regression. Most of the calculations for this document come from the basic theory
More informationBased on slides by Richard Zemel
CSC 412/2506 Winter 2018 Probabilistic Learning and Reasoning Lecture 3: Directed Graphical Models and Latent Variables Based on slides by Richard Zemel Learning outcomes What aspects of a model can we
More informationVector spaces. DS-GA 1013 / MATH-GA 2824 Optimization-based Data Analysis.
Vector spaces DS-GA 1013 / MATH-GA 2824 Optimization-based Data Analysis http://www.cims.nyu.edu/~cfgranda/pages/obda_fall17/index.html Carlos Fernandez-Granda Vector space Consists of: A set V A scalar
More informationConsistent Manifold Representation for Topological Data Analysis
Consistent Manifold Representation for Topological Data Analysis Tyrus Berry (tberry@gmu.edu) and Timothy Sauer Dept. of Mathematical Sciences, George Mason University, Fairfax, VA 3 Abstract For data
More informationL26: Advanced dimensionality reduction
L26: Advanced dimensionality reduction The snapshot CA approach Oriented rincipal Components Analysis Non-linear dimensionality reduction (manifold learning) ISOMA Locally Linear Embedding CSCE 666 attern
More informationRandom projection trees and low dimensional manifolds. Sanjoy Dasgupta and Yoav Freund University of California, San Diego
Random projection trees and low dimensional manifolds Sanjoy Dasgupta and Yoav Freund University of California, San Diego I. The new nonparametrics The new nonparametrics The traditional bane of nonparametric
More informationNonparametric Density Estimation. October 1, 2018
Nonparametric Density Estimation October 1, 2018 Introduction If we can t fit a distribution to our data, then we use nonparametric density estimation. Start with a histogram. But there are problems with
More informationChapter 4 Dynamic Bayesian Networks Fall Jin Gu, Michael Zhang
Chapter 4 Dynamic Bayesian Networks 2016 Fall Jin Gu, Michael Zhang Reviews: BN Representation Basic steps for BN representations Define variables Define the preliminary relations between variables Check
More informationBeyond the Point Cloud: From Transductive to Semi-Supervised Learning
Beyond the Point Cloud: From Transductive to Semi-Supervised Learning Vikas Sindhwani, Partha Niyogi, Mikhail Belkin Andrew B. Goldberg goldberg@cs.wisc.edu Department of Computer Sciences University of
More informationNotes on Machine Learning for and
Notes on Machine Learning for 16.410 and 16.413 (Notes adapted from Tom Mitchell and Andrew Moore.) Choosing Hypotheses Generally want the most probable hypothesis given the training data Maximum a posteriori
More informationSolving Classification Problems By Knowledge Sets
Solving Classification Problems By Knowledge Sets Marcin Orchel a, a Department of Computer Science, AGH University of Science and Technology, Al. A. Mickiewicza 30, 30-059 Kraków, Poland Abstract We propose
More informationStatistical Rock Physics
Statistical - Introduction Book review 3.1-3.3 Min Sun March. 13, 2009 Outline. What is Statistical. Why we need Statistical. How Statistical works Statistical Rock physics Information theory Statistics
More informationarxiv: v5 [math.st] 20 Dec 2018
Tropical Sufficient Statistics for Persistent Homology Anthea Monod 1,, Sara Kališnik 2, Juan Ángel Patiño-Galindo1, and Lorin Crawford 3 5 1 Department of Systems Biology, Columbia University, New York,
More informationPersistent homology and nonparametric regression
Cleveland State University March 10, 2009, BIRS: Data Analysis using Computational Topology and Geometric Statistics joint work with Gunnar Carlsson (Stanford), Moo Chung (Wisconsin Madison), Peter Kim
More informationPersistent Homology. 128 VI Persistence
8 VI Persistence VI. Persistent Homology A main purpose of persistent homology is the measurement of the scale or resolution of a topological feature. There are two ingredients, one geometric, assigning
More informationarxiv: v2 [math.pr] 15 May 2016
MAXIMALLY PERSISTENT CYCLES IN RANDOM GEOMETRIC COMPLEXES OMER BOBROWSKI, MATTHEW KAHLE, AND PRIMOZ SKRABA arxiv:1509.04347v2 [math.pr] 15 May 2016 Abstract. We initiate the study of persistent homology
More informationSTA 4273H: Sta-s-cal Machine Learning
STA 4273H: Sta-s-cal Machine Learning Russ Salakhutdinov Department of Computer Science! Department of Statistical Sciences! rsalakhu@cs.toronto.edu! h0p://www.cs.utoronto.ca/~rsalakhu/ Lecture 2 In our
More informationKernel-Based Contrast Functions for Sufficient Dimension Reduction
Kernel-Based Contrast Functions for Sufficient Dimension Reduction Michael I. Jordan Departments of Statistics and EECS University of California, Berkeley Joint work with Kenji Fukumizu and Francis Bach
More informationSTA 4273H: Statistical Machine Learning
STA 4273H: Statistical Machine Learning Russ Salakhutdinov Department of Statistics! rsalakhu@utstat.toronto.edu! http://www.utstat.utoronto.ca/~rsalakhu/ Sidney Smith Hall, Room 6002 Lecture 11 Project
More informationAdvanced Introduction to Machine Learning CMU-10715
Advanced Introduction to Machine Learning CMU-10715 Gaussian Processes Barnabás Póczos http://www.gaussianprocess.org/ 2 Some of these slides in the intro are taken from D. Lizotte, R. Parr, C. Guesterin
More informationStabilizing the unstable output of persistent homology computations
Stabilizing the unstable output of persistent homology computations Peter Bubenik with Paul Bendich (Duke) and Alex Wagner (Florida) May 5, 2017 Conference on Applied and Computational Algebraic Topology
More informationStatistical Learning. Philipp Koehn. 10 November 2015
Statistical Learning Philipp Koehn 10 November 2015 Outline 1 Learning agents Inductive learning Decision tree learning Measuring learning performance Bayesian learning Maximum a posteriori and maximum
More informationOn Inverse Problems in TDA
Abel Symposium Geiranger, June 2018 On Inverse Problems in TDA Steve Oudot joint work with E. Solomon (Brown Math.) arxiv: 1712.03630 Features Data Rn The preimage problem in the data Sciences feature
More information20: Gaussian Processes
10-708: Probabilistic Graphical Models 10-708, Spring 2016 20: Gaussian Processes Lecturer: Andrew Gordon Wilson Scribes: Sai Ganesh Bandiatmakuri 1 Discussion about ML Here we discuss an introduction
More informationApproximate Inference Part 1 of 2
Approximate Inference Part 1 of 2 Tom Minka Microsoft Research, Cambridge, UK Machine Learning Summer School 2009 http://mlg.eng.cam.ac.uk/mlss09/ Bayesian paradigm Consistent use of probability theory
More informationNonparametric Bayes Inference on Manifolds with Applications
Nonparametric Bayes Inference on Manifolds with Applications Abhishek Bhattacharya Indian Statistical Institute Based on the book Nonparametric Statistics On Manifolds With Applications To Shape Spaces
More informationComputer Vision Group Prof. Daniel Cremers. 10a. Markov Chain Monte Carlo
Group Prof. Daniel Cremers 10a. Markov Chain Monte Carlo Markov Chain Monte Carlo In high-dimensional spaces, rejection sampling and importance sampling are very inefficient An alternative is Markov Chain
More informationAdaptive Crowdsourcing via EM with Prior
Adaptive Crowdsourcing via EM with Prior Peter Maginnis and Tanmay Gupta May, 205 In this work, we make two primary contributions: derivation of the EM update for the shifted and rescaled beta prior and
More informationApproximate Inference Part 1 of 2
Approximate Inference Part 1 of 2 Tom Minka Microsoft Research, Cambridge, UK Machine Learning Summer School 2009 http://mlg.eng.cam.ac.uk/mlss09/ 1 Bayesian paradigm Consistent use of probability theory
More informationStatistical techniques for data analysis in Cosmology
Statistical techniques for data analysis in Cosmology arxiv:0712.3028; arxiv:0911.3105 Numerical recipes (the bible ) Licia Verde ICREA & ICC UB-IEEC http://icc.ub.edu/~liciaverde outline Lecture 1: Introduction
More informationThe Jackknife-Like Method for Assessing Uncertainty of Point Estimates for Bayesian Estimation in a Finite Gaussian Mixture Model
Thai Journal of Mathematics : 45 58 Special Issue: Annual Meeting in Mathematics 207 http://thaijmath.in.cmu.ac.th ISSN 686-0209 The Jackknife-Like Method for Assessing Uncertainty of Point Estimates for
More informationPractical Bayesian Optimization of Machine Learning. Learning Algorithms
Practical Bayesian Optimization of Machine Learning Algorithms CS 294 University of California, Berkeley Tuesday, April 20, 2016 Motivation Machine Learning Algorithms (MLA s) have hyperparameters that
More informationMachine Learning - MT & 14. PCA and MDS
Machine Learning - MT 2016 13 & 14. PCA and MDS Varun Kanade University of Oxford November 21 & 23, 2016 Announcements Sheet 4 due this Friday by noon Practical 3 this week (continue next week if necessary)
More informationA Gaussian Type Kernel for Persistence Diagrams
TRIPODS Summer Bootcamp: Topology and Machine Learning Brown University, August 2018 A Gaussian Type Kernel for Persistence Diagrams Mathieu Carrière joint work with S. Oudot and M. Cuturi Persistence
More informationIntroduction to Machine Learning
Outline Introduction to Machine Learning Bayesian Classification Varun Chandola March 8, 017 1. {circular,large,light,smooth,thick}, malignant. {circular,large,light,irregular,thick}, malignant 3. {oval,large,dark,smooth,thin},
More informationTutorial on Gaussian Processes and the Gaussian Process Latent Variable Model
Tutorial on Gaussian Processes and the Gaussian Process Latent Variable Model (& discussion on the GPLVM tech. report by Prof. N. Lawrence, 06) Andreas Damianou Department of Neuro- and Computer Science,
More informationClustering by Mixture Models. General background on clustering Example method: k-means Mixture model based clustering Model estimation
Clustering by Mixture Models General bacground on clustering Example method: -means Mixture model based clustering Model estimation 1 Clustering A basic tool in data mining/pattern recognition: Divide
More informationChris Bishop s PRML Ch. 8: Graphical Models
Chris Bishop s PRML Ch. 8: Graphical Models January 24, 2008 Introduction Visualize the structure of a probabilistic model Design and motivate new models Insights into the model s properties, in particular
More informationInstance-based Learning CE-717: Machine Learning Sharif University of Technology. M. Soleymani Fall 2016
Instance-based Learning CE-717: Machine Learning Sharif University of Technology M. Soleymani Fall 2016 Outline Non-parametric approach Unsupervised: Non-parametric density estimation Parzen Windows Kn-Nearest
More informationA Review of Pseudo-Marginal Markov Chain Monte Carlo
A Review of Pseudo-Marginal Markov Chain Monte Carlo Discussed by: Yizhe Zhang October 21, 2016 Outline 1 Overview 2 Paper review 3 experiment 4 conclusion Motivation & overview Notation: θ denotes the
More informationSYDE 372 Introduction to Pattern Recognition. Probability Measures for Classification: Part I
SYDE 372 Introduction to Pattern Recognition Probability Measures for Classification: Part I Alexander Wong Department of Systems Design Engineering University of Waterloo Outline 1 2 3 4 Why use probability
More informationLinear vs Non-linear classifier. CS789: Machine Learning and Neural Network. Introduction
Linear vs Non-linear classifier CS789: Machine Learning and Neural Network Support Vector Machine Jakramate Bootkrajang Department of Computer Science Chiang Mai University Linear classifier is in the
More informationClustering VS Classification
MCQ Clustering VS Classification 1. What is the relation between the distance between clusters and the corresponding class discriminability? a. proportional b. inversely-proportional c. no-relation Ans:
More informationHomework in Topology, Spring 2009.
Homework in Topology, Spring 2009. Björn Gustafsson April 29, 2009 1 Generalities To pass the course you should hand in correct and well-written solutions of approximately 10-15 of the problems. For higher
More informationKernel Density Topic Models: Visual Topics Without Visual Words
Kernel Density Topic Models: Visual Topics Without Visual Words Konstantinos Rematas K.U. Leuven ESAT-iMinds krematas@esat.kuleuven.be Mario Fritz Max Planck Institute for Informatics mfrtiz@mpi-inf.mpg.de
More information6.867 Machine Learning
6.867 Machine Learning Problem Set 2 Due date: Wednesday October 6 Please address all questions and comments about this problem set to 6867-staff@csail.mit.edu. You will need to use MATLAB for some of
More informationComputational Genomics. Systems biology. Putting it together: Data integration using graphical models
02-710 Computational Genomics Systems biology Putting it together: Data integration using graphical models High throughput data So far in this class we discussed several different types of high throughput
More informationProbabilistic Graphical Networks: Definitions and Basic Results
This document gives a cursory overview of Probabilistic Graphical Networks. The material has been gleaned from different sources. I make no claim to original authorship of this material. Bayesian Graphical
More informationLecture Notes 1: Vector spaces
Optimization-based data analysis Fall 2017 Lecture Notes 1: Vector spaces In this chapter we review certain basic concepts of linear algebra, highlighting their application to signal processing. 1 Vector
More informationProbabilistic Graphical Models Lecture 20: Gaussian Processes
Probabilistic Graphical Models Lecture 20: Gaussian Processes Andrew Gordon Wilson www.cs.cmu.edu/~andrewgw Carnegie Mellon University March 30, 2015 1 / 53 What is Machine Learning? Machine learning algorithms
More informationW vs. QCD Jet Tagging at the Large Hadron Collider
W vs. QCD Jet Tagging at the Large Hadron Collider Bryan Anenberg: anenberg@stanford.edu; CS229 December 13, 2013 Problem Statement High energy collisions of protons at the Large Hadron Collider (LHC)
More information