Distributions of Persistence Diagrams and Approximations

Size: px
Start display at page:

Download "Distributions of Persistence Diagrams and Approximations"

Transcription

1 Distributions of Persistence Diagrams and Approximations Vasileios Maroulas Department of Mathematics Department of Business Analytics and Statistics University of Tennessee August 31, 2018

2 Thanks V. Maroulas (UTK) KDE of random persistence diagrams August 31, / 45

3 Joint with: Funded by: Josh Mike (now at Michigan State) Andrew Marchese (now at Plated) John Sgouralis (now at Arizona State) Chris Oballe V. Maroulas (UTK) KDE of random persistence diagrams August 31, / 45

4 Acoustic Signals at the ARL Two classes representing two different types of weapons. Goal is to help military officers make tactical decisions based on the type of weapon system. V. Maroulas (UTK) KDE of random persistence diagrams August 31, / 45

5 Signals from ARL Dataset V. Maroulas (UTK) KDE of random persistence diagrams August 31, / 45

6 Merge statistics and topology to understand the geometry of signals and classify them. TDA has recently been introduced to the field of signal and time-series classification. Biological Signals (Zhang et al. (2015)) Action Recognition (Venkataraman et al. (2016)) Wheeze Detection (Emrani et al. (2014)) V. Maroulas (UTK) KDE of random persistence diagrams August 31, / 45

7 Motivation Data has shape and shape matters. Latent topological features in scientific data. VM and A. Nebenführ, Tracking rapid intracellular movements: a Bayesian random set approach, Annals of Applied Statistics, I. Sgouralis, A. Nebenführ and VM, A Bayesian Topological Framework for the Identification and Reconstruction of Subcellular Motion, SIAM on Imaging Sciences, J. Mike, C. Sumrall, VM and F. Schwartz. Non-Landmark Classification in Paleobiology, Paleobiology, V. Maroulas (UTK) KDE of random persistence diagrams August 31, / 45

8 Key Picture Data lies in a topological space. Take measurements, sampling that space. Reconstruct it by using an approximation. Compute the invariants to understand it. V. Maroulas (UTK) KDE of random persistence diagrams August 31, / 45

9 From Signals to Point Clouds (Taken s Thm) Suppose that w : [0, T] R are the realizations of a discrete time series in [0, T]. Consider a set of delay indices τ 1, τ 2,, τ n 1 The n dimensional delay embedding of W is the concatenation of time-delayed samples: W(t) = (w(t), w(t + τ 1 ), w(t + τ 2 ),, w(t + τ n 1 )) (1) 4.4 Point Cloud w(t) w(t+τ 2 ) Time (t) w(t+τ 1 ) w(t) Figure: Signal evolution in time domain Figure: 3D delay embedding V. Maroulas (UTK) KDE of random persistence diagrams August 31, / 45

10 From Point Clouds to Inference and Classification Now we have turned our time-series into a point cloud living in N-dimensional space. Point Cloud 4.5 w(t+τ 2 ) w(t+τ 1 ) w(t) How can we extract information from this data and use it for classification and statistical inference? V. Maroulas (UTK) KDE of random persistence diagrams August 31, / 45

11 Outline Data Analysis using Persistence Homology Distribution of random persistence diagrams Kernel Density Estimation Conclusion V. Maroulas (UTK) KDE of random persistence diagrams August 31, / 45

12 Simplicial Complex Simplicial complexes are discretizations of real-life shapes Generalization of graphs with higher order relationships among the nodes. A simplicial complex is the union of simple pieces (simplices) i.e. vertices, edges, triangles etc. A face of k simplex are all the (k 1) simplex. Two simplices must intersect at a common face or not at all. V. Maroulas (UTK) KDE of random persistence diagrams August 31, / 45

13 Construction of Simplicial complexes for data Start with a point-cloud Π and create an abstract representation of vertices one for each point in your Π. Figure: Left: Point Cloud; Right: Simplicial Complex V. Maroulas (UTK) KDE of random persistence diagrams August 31, / 45

14 Construction of Simplicial complexes for data Create spheres" of radius r centered at each point. Figure: Left: Point Cloud; Right: Simplicial Complex V. Maroulas (UTK) KDE of random persistence diagrams August 31, / 45

15 Construction of Simplicial complexes for data Increase radius r. Figure: Left: Point Cloud; Right: Simplicial Complex V. Maroulas (UTK) KDE of random persistence diagrams August 31, / 45

16 Construction of Simplicial complexes for data Add edges between vertices v i and v j if the corresponding circles intersect. Figure: Left: Point Cloud; Right: Simplicial Complex V. Maroulas (UTK) KDE of random persistence diagrams August 31, / 45

17 Construction of Simplicial complexes for data Add triangles between vertices v i, v j and v k if all three circles intersect, etc. Figure: Left: Point Cloud; Right: Simplicial Complex V. Maroulas (UTK) KDE of random persistence diagrams August 31, / 45

18 Based on the simplicial complex we retrieve the Betti numbers (the dimensions of some vector spaces associated to our topological space) Betti 0: number of clusters" Betti 1: number of holes Figure: β 0 = 2, β 1 = 0 V. Maroulas (UTK) KDE of random persistence diagrams August 31, / 45

19 Based on the simplicial complex we retrieve the Betti numbers (the dimensions of some vector spaces associated to our topological space) Betti 0: number of clusters" Betti 1: number of holes Figure: β 0 = 2, β 1 = 0 Figure: β 0 = 2, β 1 = 1 V. Maroulas (UTK) KDE of random persistence diagrams August 31, / 45

20 Persistence Diagrams Interested in is the persistence of the Betti numbers. When do different connected components/holes form and how long do they last (with respect to r)? The Betti numbers compactly encoded in a 2-dim plot which provides the birth time vs death time of these features V. Maroulas (UTK) KDE of random persistence diagrams August 31, / 45

21 Results on Signals from ARL Dataset (a) (b) (c) (d) (e) (f) A. Marchese and VM. Signal classification with a point process distance on the space of persistence diagrams. Advances in Data Analysis and Classification, pp. 1-26, 2017 V. Maroulas (UTK) KDE of random persistence diagrams August 31, / 45

22 Classifier V. Maroulas (UTK) KDE of random persistence diagrams August 31, / 45

23 Statistics and Persistence Diagrams Summary statistics such as center and variance (Bobrowski et al., 2014; Mileyko et al., 2011; Turner et al., 2014; Marchese and VM, 2017) Birth and death estimates (Emmett et al., 2014) Confidence sets (Fasy et al., 2014) Need a framework to understand the above summary statistics through a single viewpoint V. Maroulas (UTK) KDE of random persistence diagrams August 31, / 45

24 Novel framework A complete and consistent framework of how to construct distributions of persistence diagrams. Capture the important information of these diagrams in terms of their inherent set properties Set membership Cardinality V. Maroulas (UTK) KDE of random persistence diagrams August 31, / 45

25 Setup Take data X = x j generated by some random process Associated (random) persistence diagram D with features ξ i = (b i, d i ), such that the a hole" appears at scale b i and is filled at scale d i V. Maroulas (UTK) KDE of random persistence diagrams August 31, / 45

26 Lemma 2.1 (J. Mike & VM, 2018) Consider a multiset of independent singleton random persistence diagrams { D j} M j=1. If each singleton Dj is described by the value q (j) = P[D j ] and the subsequent conditional pdf, p (j) (ξ), given D j = 1, then the global pdf for D = M j=1 D j is given by f D (ξ 1,..., ξ N ) = γ I(N,M) Q(γ) N p (γ(k)) (ξ k ), (2) for each N {0,..., M} where I(N, M) consists of all increasing injections γ : {1,..., N} {1,..., M}, and The sum over γ I(N, M) in Eq. (2) accounts for each possible combination of singleton presence. The weights Q(γ) proportional to the probability for each singleton to be either present, q(j), or absent, 1 q(j), for each j. k=1 J. Mike and VM. Nonparametric Estimation of Probability Density Functions of Random Persistence Diagrams. arxiv: V. Maroulas (UTK) KDE of random persistence diagrams August 31, / 45

27 Example 2.2 Consider two 1-dimensional singleton diagrams, D 1 and D 2, with probabilities of being nonempty q (1) = 0.6 and q (2) = 0.8. Local densities when nonempty: p (1) (x) = 1 2π e (x+1)2 /2 and p (2) (x) = 1 2π e (x 1)2 /2. Global pdf for D = D 1 D 2 through a set of local densities { f0, f 1 (x), f 2 (x, y) } such that f 0 = P[ D = 0] = (1 q (1) )(1 q (2) ) = 0.08 f 1 (x) = (1 q (2) )q (1) p (1) (x) + (1 q (1) )q (2) p (2) (x) = 0.12 e (x+1)2 / e (x 1)2 /2, 2π 2π f 2 (x, y) = q(1) q (2) 2 = π [ ] p (1) (x)p (2) (y) + p (1) (y)p (2) (x) ( e ((x 1)2 +(y+1) 2 )/2 + e ((x+1)2 +(y 1) 2 )/2 ). (3a) (3b) (3c) V. Maroulas (UTK) KDE of random persistence diagrams August 31, / 45

28 Figure: Left: Plot of the local density f 1(x) in Eq. (3b); Right: Contour plot of the local density f 2(x, y) in Eq. (3c). These pdfs cover the different possible input dimensions and are symmetric under permutations of the input. V. Maroulas (UTK) KDE of random persistence diagrams August 31, / 45

29 Past Studies-KDE Extensive work to devise various maps from persistence diagrams into Hilbert spaces Mapping into a Hilbert space, these studies allow the application of statistical learning methods such as principal component analysis, random forest, support vector machine, etc. Chepushtanova et al. (2015) discretizes persistence diagrams via bins, yielding vectors in a high dimensional Euclidean space. Reininghaus et al., (2014) and Kusano et al., (2016) define kernels between persistence diagrams in a Reproducing Kernel Hilbert Spaces Adler et al. (2017) utilizes Gibbs distributions in order to replicate similar persistence diagrams, e.g. for use in MCMC type sampling. Kernel density estimation on the underlying data to estimate a target diagram Bobrowski et al. (2014) constructs an estimator for the target diagram Fasy et al. (2014) defines a confidence set. V. Maroulas (UTK) KDE of random persistence diagrams August 31, / 45

30 Building a Kernel Density Goal: kernel density K σ (Z, D). Center Diagram D Bandwidth σ Input Z. Split D into upper and lower halves: D u = { (b, d) D : d b > σ } D l = { (b, d) D : d b σ } Define random diagrams: D u centered at D u. D l centered at D l. V. Maroulas (UTK) KDE of random persistence diagrams August 31, / 45

31 Building the Upper Density. Split into singletons: D u = j D j,u Each D j,u is described by: q j = P[D j,u ]. Local pdf, p j (b, d), Restricted Gaussian V. Maroulas (UTK) KDE of random persistence diagrams August 31, / 45

32 Building the Lower Density. p l (b, d) = 1 N l (b i,d i) D l 1 πσ 2 e Lower cardinality N l = D l. Cardinality probability mass: ν(j). Chosen with mean N l. Single kernel density p l : Project D l onto the diagonal. Kernel estimate for these points. D l has number according to ν, with draws i.i.d. according to p l. ( ( ) b b i +d i 2+ ( 2 Global pdf D l : f D l(ξ 1,..., ξ N ) = ν(n) d b i +d i 2 ) 2 ) /2σ 2. (4) N p l (ξ j ). (5) j=1 V. Maroulas (UTK) KDE of random persistence diagrams August 31, / 45

33 The Kernel Density Theorem 3.1 (J. Mike & VM, 2018) The random diagram D = D u D l with D u and D l defined according to the previous construction with center D and bandwidth σ has the following global pdf, or kernel density, evaluated at Z = {ξ 1,..., ξ N }: N u K σ (Z, D) = ν(n j) j=0 γ I(j,N) Q(γ) j p γ(k) (ξ k ) k=1 N k=j+1 where I(j, N) = { γ : {1,..., j} {1,..., N} : γ is increasing }, and Q(γ) = p l (ξ k ) (6) N ( ) j=1 1 q j j ( ) q γ(k). (7) 1 q γ(k) j k=1 k=1 V. Maroulas (UTK) KDE of random persistence diagrams August 31, / 45

34 Example 3.2 Consider D = ((1, 3), (2, 4), (1, 1.3), (3, 3.2)) and σ = 1/2. V. Maroulas (UTK) KDE of random persistence diagrams August 31, / 45

35 Kernel with input cardinality 1 Kernel K σ ((b 1, d 1 ), D) equals to: [ ν(0) (1 q (2) )q (1) p (1) (b 1, d 1 ) + (1 q (1) )q (2) p (2) ] (b 1, d 1 ) [ + ν(1) (1 q (1) )(1 q (2) )p l ] (b 1, d 1 ) (8) Figure: Contour map for the kernel density restricted to a single input feature (Eq. (8)). The center diagram is indicated by red (upper) and green (lower) points. Scale bars at the right of each plot indicate the range of probability density in each shaded region p l (b, d) = 2 [e ((b 1.15)2 +(d 1.15) 2 ) + e ((b 3.1) 2 +(d 3.1) 2 ) ] π p (1) (b 1, d 1 ) e 2((b1 2)2 +(d 1 4) 2 ) p (2) (b 1, d 1 ) e 2((b1 1)2 +(d 1 3) 2 ) V. Maroulas (UTK) KDE of random persistence diagrams August 31, / 45

36 Kernel with input cardinality 2 Consider Z = (ξ 1, ξ 2 ) = ((b 1, d 1 ), (b 2, d 2 )) K σ ((ξ 1, ξ 2 ), D) = ν(0)q (1) q (2) p (1) (b 1, d 1 )p (2) (b 2, d 2 ) [ + ν(1) (1 q (2) )q (1) p (1) (b 1, d 1 ) ] + (1 q (1) )q (2) p (2) (b 1, d 1 ) p l (b 2, d 2 ) (9) + ν(2)(1 q (1) )(1 q (2) )p l (b 1, d 1 )p l (b 2, d 2 ) V. Maroulas (UTK) KDE of random persistence diagrams August 31, / 45

37 Kernel with input cardinality 2 (a) (b) (c) Figure: Contour maps for slices of the kernel density K σ((ξ, ξ 2), D) with input cardinality 2. A single feature ξ 2, indicated by white crosshairs, is fixed to restrict to a 2D subspace as follows: (a) ξ 2 = (1, 3) (b) ξ 2 = (2, 4) and (c) ξ 2 = (2.5, 2.7). The center diagram is indicated by red (upper) and green (lower) points. Scale bars at the right of each plot indicate the range of probability density in each shaded region. V. Maroulas (UTK) KDE of random persistence diagrams August 31, / 45

38 Kernel with input cardinality 3 [ ] K σ ((ξ 1, ξ 2, ξ 3 ), D) = ν(1) q (1) q (2) p (1) (b 1, d 1 )p (2) (b 2, d 2 ) p l (b 3, d 3 ) [ + ν(2) (1 q (2) )q (1) p (1) (b 1, d 1 ) ] + ν(2)(1 q (1) )q (2) p (2) (b 1, d 1 ) p l (b 2, d 2 )p l (b 3, d 3 ) + ν(3)(1 q (1) )(1 q (2) )p l (b 1, d 1 )p l (b 2, d 2 )p l (b 3, d 3 ). (10) V. Maroulas (UTK) KDE of random persistence diagrams August 31, / 45

39 Kernel with input cardinality 3 (a) Figure: Contour maps for slices of the kernel density K σ((ξ, ξ 2, ξ 3), D) with input cardinality 3. A pair of features ξ 2 and ξ 3, indicated by white crosshairs, are fixed to restrict to a 2D subspace as follows: (a) (ξ 2, ξ 3) = ((1, 3), (2, 4)) and (b) (ξ 2, ξ 3) = ((1, 3), (2.5, 3.5)). The center diagram is indicated by red (upper) and green (lower) points. Scale bars at the right of each plot indicate the range of probability density in each shaded region. (b) V. Maroulas (UTK) KDE of random persistence diagrams August 31, / 45

40 Kernel Density Estimation Theorem 3.3 (J. Mike & VM, 2018) Random persistence diagram with global pdf f : f must satisfy decay and boundedness conditions. Diagrams {Di} n i=1 sampled i.i.d. according to f. Yield a KDE: ˆf (Z) = 1 n n i=1 K σ(z, D i ) σ = O(n α ) chosen with 0 < α α 2M, ˆf f uniformly on compact subsets of the space of PDs. J. Mike and VM. Nonparametric Estimation of Probability Density Functions of Random Persistence Diagrams. arxiv: V. Maroulas (UTK) KDE of random persistence diagrams August 31, / 45

41 Example 3.4 Generate samples which each consist of 10 points uniformly from the unit circle with additive Gaussian noise, N((0, 0), ( 1 50 )2 I 2 ). Toy dataset for signal analysis. (a) (b) V. Maroulas (UTK) KDE of random persistence diagrams August 31, / 45

42 Plots of persistence diagram KDEs. Color indicates the probability density. White regions above the diagonal indicate portions of very low probability density. Each column is a particular slice, while each row is a particular KDE with n and σ. Left: Local KDEs ˆf n,σ((b, d)) evaluated at a diagram with only one feature. The mode of the converged density is approximately (b 2, d 2) = (0.77, 0.98). Right: Local KDEs ˆf n,σ((b, d), (0.77, 0.98)) evaluated at a diagram with two features and one feature fixed.these slices have two modes which are very close to the diagonal at (0, 0) and (1, 1). V. Maroulas (UTK) KDE of random persistence diagrams August 31, / 45

43 Summary Considered the problem of estimating the distribution of persistence diagrams Established a novel kernel density We focused on set properties-membership and cardinality We established convergence and verified several synthetic examples. With a pdf at hand, we can start implementing Monte Carlo sampling and move on to further probabilistic settings Bayesian formulation Applications in biology, defense, material science and chemistry. V. Maroulas (UTK) KDE of random persistence diagrams August 31, / 45

44 Bayesian Framework In principle, we can compute posterior distributions using Bayes theorem for random sets π(d X D Y ) l(d Y D X )π(d X ) V. Maroulas (UTK) KDE of random persistence diagrams August 31, / 45

45 Posterior Approximation Mahler (2003) Singh, Vo, Baddeley and Zuyev (2007) Caron, Del Moral, Doucet, Pace (2011) Prior Prior Posterior Persistence Intensity Persistence Intensity Persistence Intensity Birth (a) Birth (b) Birth (c) V. Maroulas (UTK) KDE of random persistence diagrams August 31, / 45

46 Thank you-questions? V. Maroulas (UTK) KDE of random persistence diagrams August 31, / 45

Neuronal structure detection using Persistent Homology

Neuronal structure detection using Persistent Homology Neuronal structure detection using Persistent Homology J. Heras, G. Mata and J. Rubio Department of Mathematics and Computer Science, University of La Rioja Seminario de Informática Mirian Andrés March

More information

Probabilistic Machine Learning. Industrial AI Lab.

Probabilistic Machine Learning. Industrial AI Lab. Probabilistic Machine Learning Industrial AI Lab. Probabilistic Linear Regression Outline Probabilistic Classification Probabilistic Clustering Probabilistic Dimension Reduction 2 Probabilistic Linear

More information

Gaussian Processes. Le Song. Machine Learning II: Advanced Topics CSE 8803ML, Spring 2012

Gaussian Processes. Le Song. Machine Learning II: Advanced Topics CSE 8803ML, Spring 2012 Gaussian Processes Le Song Machine Learning II: Advanced Topics CSE 8803ML, Spring 01 Pictorial view of embedding distribution Transform the entire distribution to expected features Feature space Feature

More information

Learning from persistence diagrams

Learning from persistence diagrams Learning from persistence diagrams Ulrich Bauer TUM July 7, 2015 ACAT final meeting IST Austria Joint work with: Jan Reininghaus, Stefan Huber, Roland Kwitt 1 / 28 ? 2 / 28 ? 2 / 28 3 / 28 3 / 28 3 / 28

More information

ECE521 week 3: 23/26 January 2017

ECE521 week 3: 23/26 January 2017 ECE521 week 3: 23/26 January 2017 Outline Probabilistic interpretation of linear regression - Maximum likelihood estimation (MLE) - Maximum a posteriori (MAP) estimation Bias-variance trade-off Linear

More information

Kernel methods, kernel SVM and ridge regression

Kernel methods, kernel SVM and ridge regression Kernel methods, kernel SVM and ridge regression Le Song Machine Learning II: Advanced Topics CSE 8803ML, Spring 2012 Collaborative Filtering 2 Collaborative Filtering R: rating matrix; U: user factor;

More information

Matthew L. Wright. St. Olaf College. June 15, 2017

Matthew L. Wright. St. Olaf College. June 15, 2017 Matthew L. Wright St. Olaf College June 15, 217 Persistent homology detects topological features of data. For this data, the persistence barcode reveals one significant hole in the point cloud. Problem:

More information

Topological Data Analysis - II. Afra Zomorodian Department of Computer Science Dartmouth College

Topological Data Analysis - II. Afra Zomorodian Department of Computer Science Dartmouth College Topological Data Analysis - II Afra Zomorodian Department of Computer Science Dartmouth College September 4, 2007 1 Plan Yesterday: Motivation Topology Simplicial Complexes Invariants Homology Algebraic

More information

COURSE INTRODUCTION. J. Elder CSE 6390/PSYC 6225 Computational Modeling of Visual Perception

COURSE INTRODUCTION. J. Elder CSE 6390/PSYC 6225 Computational Modeling of Visual Perception COURSE INTRODUCTION COMPUTATIONAL MODELING OF VISUAL PERCEPTION 2 The goal of this course is to provide a framework and computational tools for modeling visual inference, motivated by interesting examples

More information

Pattern Recognition and Machine Learning

Pattern Recognition and Machine Learning Christopher M. Bishop Pattern Recognition and Machine Learning ÖSpri inger Contents Preface Mathematical notation Contents vii xi xiii 1 Introduction 1 1.1 Example: Polynomial Curve Fitting 4 1.2 Probability

More information

Statistics: Learning models from data

Statistics: Learning models from data DS-GA 1002 Lecture notes 5 October 19, 2015 Statistics: Learning models from data Learning models from data that are assumed to be generated probabilistically from a certain unknown distribution is a crucial

More information

Nonparametric Bayesian Methods

Nonparametric Bayesian Methods Nonparametric Bayesian Methods Debdeep Pati Florida State University October 2, 2014 Large spatial datasets (Problem of big n) Large observational and computer-generated datasets: Often have spatial and

More information

Applications of Persistent Homology to Time-Varying Systems

Applications of Persistent Homology to Time-Varying Systems Applications of Persistent Homology to Time-Varying Systems Elizabeth Munch Duke University :: Dept of Mathematics March 28, 2013 Time Death Radius Birth Radius Liz Munch (Duke) PhD Defense March 28, 2013

More information

STAT 518 Intro Student Presentation

STAT 518 Intro Student Presentation STAT 518 Intro Student Presentation Wen Wei Loh April 11, 2013 Title of paper Radford M. Neal [1999] Bayesian Statistics, 6: 475-501, 1999 What the paper is about Regression and Classification Flexible

More information

Some Topics in Computational Topology. Yusu Wang. AMS Short Course 2014

Some Topics in Computational Topology. Yusu Wang. AMS Short Course 2014 Some Topics in Computational Topology Yusu Wang AMS Short Course 2014 Introduction Much recent developments in computational topology Both in theory and in their applications E.g, the theory of persistence

More information

Bayesian Nonparametrics

Bayesian Nonparametrics Bayesian Nonparametrics Peter Orbanz Columbia University PARAMETERS AND PATTERNS Parameters P(X θ) = Probability[data pattern] 3 2 1 0 1 2 3 5 0 5 Inference idea data = underlying pattern + independent

More information

A Kernel on Persistence Diagrams for Machine Learning

A Kernel on Persistence Diagrams for Machine Learning A Kernel on Persistence Diagrams for Machine Learning Jan Reininghaus 1 Stefan Huber 1 Roland Kwitt 2 Ulrich Bauer 1 1 Institute of Science and Technology Austria 2 FB Computer Science Universität Salzburg,

More information

Chapter 3: Homology Groups Topics in Computational Topology: An Algorithmic View

Chapter 3: Homology Groups Topics in Computational Topology: An Algorithmic View Chapter 3: Homology Groups Topics in Computational Topology: An Algorithmic View As discussed in Chapter 2, we have complete topological information about 2-manifolds. How about higher dimensional manifolds?

More information

Nonparametric Bayesian Methods (Gaussian Processes)

Nonparametric Bayesian Methods (Gaussian Processes) [70240413 Statistical Machine Learning, Spring, 2015] Nonparametric Bayesian Methods (Gaussian Processes) Jun Zhu dcszj@mail.tsinghua.edu.cn http://bigml.cs.tsinghua.edu.cn/~jun State Key Lab of Intelligent

More information

Data Mining Techniques

Data Mining Techniques Data Mining Techniques CS 6220 - Section 3 - Fall 2016 Lecture 12 Jan-Willem van de Meent (credit: Yijun Zhao, Percy Liang) DIMENSIONALITY REDUCTION Borrowing from: Percy Liang (Stanford) Linear Dimensionality

More information

Some Topics in Computational Topology

Some Topics in Computational Topology Some Topics in Computational Topology Yusu Wang Ohio State University AMS Short Course 2014 Introduction Much recent developments in computational topology Both in theory and in their applications E.g,

More information

Introduction to Machine Learning CMU-10701

Introduction to Machine Learning CMU-10701 Introduction to Machine Learning CMU-10701 Markov Chain Monte Carlo Methods Barnabás Póczos & Aarti Singh Contents Markov Chain Monte Carlo Methods Goal & Motivation Sampling Rejection Importance Markov

More information

STA 4273H: Statistical Machine Learning

STA 4273H: Statistical Machine Learning STA 4273H: Statistical Machine Learning Russ Salakhutdinov Department of Computer Science! Department of Statistical Sciences! rsalakhu@cs.toronto.edu! h0p://www.cs.utoronto.ca/~rsalakhu/ Lecture 7 Approximate

More information

STA414/2104 Statistical Methods for Machine Learning II

STA414/2104 Statistical Methods for Machine Learning II STA414/2104 Statistical Methods for Machine Learning II Murat A. Erdogdu & David Duvenaud Department of Computer Science Department of Statistical Sciences Lecture 3 Slide credits: Russ Salakhutdinov Announcements

More information

The Intersection of Statistics and Topology:

The Intersection of Statistics and Topology: The Intersection of Statistics and Topology: Confidence Sets for Persistence Diagrams Brittany Terese Fasy joint work with S. Balakrishnan, F. Lecci, A. Rinaldo, A. Singh, L. Wasserman 3 June 2014 [FLRWBS]

More information

Introduction to Bayesian Learning

Introduction to Bayesian Learning Course Information Introduction Introduction to Bayesian Learning Davide Bacciu Dipartimento di Informatica Università di Pisa bacciu@di.unipi.it Apprendimento Automatico: Fondamenti - A.A. 2016/2017 Outline

More information

7.3 Singular Homology Groups

7.3 Singular Homology Groups 184 CHAPTER 7. HOMOLOGY THEORY 7.3 Singular Homology Groups 7.3.1 Cycles, Boundaries and Homology Groups We can define the singular p-chains with coefficients in a field K. Furthermore, we can define the

More information

Recent Advances in Bayesian Inference Techniques

Recent Advances in Bayesian Inference Techniques Recent Advances in Bayesian Inference Techniques Christopher M. Bishop Microsoft Research, Cambridge, U.K. research.microsoft.com/~cmbishop SIAM Conference on Data Mining, April 2004 Abstract Bayesian

More information

PATTERN RECOGNITION AND MACHINE LEARNING CHAPTER 13: SEQUENTIAL DATA

PATTERN RECOGNITION AND MACHINE LEARNING CHAPTER 13: SEQUENTIAL DATA PATTERN RECOGNITION AND MACHINE LEARNING CHAPTER 13: SEQUENTIAL DATA Contents in latter part Linear Dynamical Systems What is different from HMM? Kalman filter Its strength and limitation Particle Filter

More information

STA 4273H: Statistical Machine Learning

STA 4273H: Statistical Machine Learning STA 4273H: Statistical Machine Learning Russ Salakhutdinov Department of Statistics! rsalakhu@utstat.toronto.edu! http://www.utstat.utoronto.ca/~rsalakhu/ Sidney Smith Hall, Room 6002 Lecture 3 Linear

More information

Introduction to Machine Learning Midterm Exam

Introduction to Machine Learning Midterm Exam 10-701 Introduction to Machine Learning Midterm Exam Instructors: Eric Xing, Ziv Bar-Joseph 17 November, 2015 There are 11 questions, for a total of 100 points. This exam is open book, open notes, but

More information

Beyond Scalar Affinities for Network Analysis or Vector Diffusion Maps and the Connection Laplacian

Beyond Scalar Affinities for Network Analysis or Vector Diffusion Maps and the Connection Laplacian Beyond Scalar Affinities for Network Analysis or Vector Diffusion Maps and the Connection Laplacian Amit Singer Princeton University Department of Mathematics and Program in Applied and Computational Mathematics

More information

27 : Distributed Monte Carlo Markov Chain. 1 Recap of MCMC and Naive Parallel Gibbs Sampling

27 : Distributed Monte Carlo Markov Chain. 1 Recap of MCMC and Naive Parallel Gibbs Sampling 10-708: Probabilistic Graphical Models 10-708, Spring 2014 27 : Distributed Monte Carlo Markov Chain Lecturer: Eric P. Xing Scribes: Pengtao Xie, Khoa Luu In this scribe, we are going to review the Parallel

More information

Density Propagation for Continuous Temporal Chains Generative and Discriminative Models

Density Propagation for Continuous Temporal Chains Generative and Discriminative Models $ Technical Report, University of Toronto, CSRG-501, October 2004 Density Propagation for Continuous Temporal Chains Generative and Discriminative Models Cristian Sminchisescu and Allan Jepson Department

More information

Unsupervised Learning Methods

Unsupervised Learning Methods Structural Health Monitoring Using Statistical Pattern Recognition Unsupervised Learning Methods Keith Worden and Graeme Manson Presented by Keith Worden The Structural Health Monitoring Process 1. Operational

More information

Bayesian Networks BY: MOHAMAD ALSABBAGH

Bayesian Networks BY: MOHAMAD ALSABBAGH Bayesian Networks BY: MOHAMAD ALSABBAGH Outlines Introduction Bayes Rule Bayesian Networks (BN) Representation Size of a Bayesian Network Inference via BN BN Learning Dynamic BN Introduction Conditional

More information

Machine Learning. Lecture 4: Regularization and Bayesian Statistics. Feng Li. https://funglee.github.io

Machine Learning. Lecture 4: Regularization and Bayesian Statistics. Feng Li. https://funglee.github.io Machine Learning Lecture 4: Regularization and Bayesian Statistics Feng Li fli@sdu.edu.cn https://funglee.github.io School of Computer Science and Technology Shandong University Fall 207 Overfitting Problem

More information

Review. DS GA 1002 Statistical and Mathematical Models. Carlos Fernandez-Granda

Review. DS GA 1002 Statistical and Mathematical Models.   Carlos Fernandez-Granda Review DS GA 1002 Statistical and Mathematical Models http://www.cims.nyu.edu/~cfgranda/pages/dsga1002_fall16 Carlos Fernandez-Granda Probability and statistics Probability: Framework for dealing with

More information

Metric Thickenings of Euclidean Submanifolds

Metric Thickenings of Euclidean Submanifolds Metric Thickenings of Euclidean Submanifolds Advisor: Dr. Henry Adams Committe: Dr. Chris Peterson, Dr. Daniel Cooley Joshua Mirth Masters Thesis Defense October 3, 2017 Introduction Motivation Can we

More information

Dynamic Data Modeling, Recognition, and Synthesis. Rui Zhao Thesis Defense Advisor: Professor Qiang Ji

Dynamic Data Modeling, Recognition, and Synthesis. Rui Zhao Thesis Defense Advisor: Professor Qiang Ji Dynamic Data Modeling, Recognition, and Synthesis Rui Zhao Thesis Defense Advisor: Professor Qiang Ji Contents Introduction Related Work Dynamic Data Modeling & Analysis Temporal localization Insufficient

More information

Machine learning for pervasive systems Classification in high-dimensional spaces

Machine learning for pervasive systems Classification in high-dimensional spaces Machine learning for pervasive systems Classification in high-dimensional spaces Department of Communications and Networking Aalto University, School of Electrical Engineering stephan.sigg@aalto.fi Version

More information

Advanced Introduction to Machine Learning

Advanced Introduction to Machine Learning 10-715 Advanced Introduction to Machine Learning Homework 3 Due Nov 12, 10.30 am Rules 1. Homework is due on the due date at 10.30 am. Please hand over your homework at the beginning of class. Please see

More information

CSC 2541: Bayesian Methods for Machine Learning

CSC 2541: Bayesian Methods for Machine Learning CSC 2541: Bayesian Methods for Machine Learning Radford M. Neal, University of Toronto, 2011 Lecture 3 More Markov Chain Monte Carlo Methods The Metropolis algorithm isn t the only way to do MCMC. We ll

More information

An Introduction to Topological Data Analysis

An Introduction to Topological Data Analysis An Introduction to Topological Data Analysis Elizabeth Munch Duke University :: Dept of Mathematics Tuesday, June 11, 2013 Elizabeth Munch (Duke) CompTop-SUNYIT Tuesday, June 11, 2013 1 / 43 There is no

More information

Tutorial on Approximate Bayesian Computation

Tutorial on Approximate Bayesian Computation Tutorial on Approximate Bayesian Computation Michael Gutmann https://sites.google.com/site/michaelgutmann University of Helsinki Aalto University Helsinki Institute for Information Technology 16 May 2016

More information

Kernel adaptive Sequential Monte Carlo

Kernel adaptive Sequential Monte Carlo Kernel adaptive Sequential Monte Carlo Ingmar Schuster (Paris Dauphine) Heiko Strathmann (University College London) Brooks Paige (Oxford) Dino Sejdinovic (Oxford) December 7, 2015 1 / 36 Section 1 Outline

More information

Topological Data Analysis for Brain Networks

Topological Data Analysis for Brain Networks Topological Data Analysis for Brain Networks Relating Functional Brain Network Topology to Clinical Measures of Behavior Bei Wang Phillips 1,2 University of Utah Joint work with Eleanor Wong 1,2, Sourabh

More information

Unsupervised Machine Learning and Data Mining. DS 5230 / DS Fall Lecture 7. Jan-Willem van de Meent

Unsupervised Machine Learning and Data Mining. DS 5230 / DS Fall Lecture 7. Jan-Willem van de Meent Unsupervised Machine Learning and Data Mining DS 5230 / DS 4420 - Fall 2018 Lecture 7 Jan-Willem van de Meent DIMENSIONALITY REDUCTION Borrowing from: Percy Liang (Stanford) Dimensionality Reduction Goal:

More information

Introduction to Probability and Statistics (Continued)

Introduction to Probability and Statistics (Continued) Introduction to Probability and Statistics (Continued) Prof. icholas Zabaras Center for Informatics and Computational Science https://cics.nd.edu/ University of otre Dame otre Dame, Indiana, USA Email:

More information

Bayesian Linear Regression [DRAFT - In Progress]

Bayesian Linear Regression [DRAFT - In Progress] Bayesian Linear Regression [DRAFT - In Progress] David S. Rosenberg Abstract Here we develop some basics of Bayesian linear regression. Most of the calculations for this document come from the basic theory

More information

Based on slides by Richard Zemel

Based on slides by Richard Zemel CSC 412/2506 Winter 2018 Probabilistic Learning and Reasoning Lecture 3: Directed Graphical Models and Latent Variables Based on slides by Richard Zemel Learning outcomes What aspects of a model can we

More information

Vector spaces. DS-GA 1013 / MATH-GA 2824 Optimization-based Data Analysis.

Vector spaces. DS-GA 1013 / MATH-GA 2824 Optimization-based Data Analysis. Vector spaces DS-GA 1013 / MATH-GA 2824 Optimization-based Data Analysis http://www.cims.nyu.edu/~cfgranda/pages/obda_fall17/index.html Carlos Fernandez-Granda Vector space Consists of: A set V A scalar

More information

Consistent Manifold Representation for Topological Data Analysis

Consistent Manifold Representation for Topological Data Analysis Consistent Manifold Representation for Topological Data Analysis Tyrus Berry (tberry@gmu.edu) and Timothy Sauer Dept. of Mathematical Sciences, George Mason University, Fairfax, VA 3 Abstract For data

More information

L26: Advanced dimensionality reduction

L26: Advanced dimensionality reduction L26: Advanced dimensionality reduction The snapshot CA approach Oriented rincipal Components Analysis Non-linear dimensionality reduction (manifold learning) ISOMA Locally Linear Embedding CSCE 666 attern

More information

Random projection trees and low dimensional manifolds. Sanjoy Dasgupta and Yoav Freund University of California, San Diego

Random projection trees and low dimensional manifolds. Sanjoy Dasgupta and Yoav Freund University of California, San Diego Random projection trees and low dimensional manifolds Sanjoy Dasgupta and Yoav Freund University of California, San Diego I. The new nonparametrics The new nonparametrics The traditional bane of nonparametric

More information

Nonparametric Density Estimation. October 1, 2018

Nonparametric Density Estimation. October 1, 2018 Nonparametric Density Estimation October 1, 2018 Introduction If we can t fit a distribution to our data, then we use nonparametric density estimation. Start with a histogram. But there are problems with

More information

Chapter 4 Dynamic Bayesian Networks Fall Jin Gu, Michael Zhang

Chapter 4 Dynamic Bayesian Networks Fall Jin Gu, Michael Zhang Chapter 4 Dynamic Bayesian Networks 2016 Fall Jin Gu, Michael Zhang Reviews: BN Representation Basic steps for BN representations Define variables Define the preliminary relations between variables Check

More information

Beyond the Point Cloud: From Transductive to Semi-Supervised Learning

Beyond the Point Cloud: From Transductive to Semi-Supervised Learning Beyond the Point Cloud: From Transductive to Semi-Supervised Learning Vikas Sindhwani, Partha Niyogi, Mikhail Belkin Andrew B. Goldberg goldberg@cs.wisc.edu Department of Computer Sciences University of

More information

Notes on Machine Learning for and

Notes on Machine Learning for and Notes on Machine Learning for 16.410 and 16.413 (Notes adapted from Tom Mitchell and Andrew Moore.) Choosing Hypotheses Generally want the most probable hypothesis given the training data Maximum a posteriori

More information

Solving Classification Problems By Knowledge Sets

Solving Classification Problems By Knowledge Sets Solving Classification Problems By Knowledge Sets Marcin Orchel a, a Department of Computer Science, AGH University of Science and Technology, Al. A. Mickiewicza 30, 30-059 Kraków, Poland Abstract We propose

More information

Statistical Rock Physics

Statistical Rock Physics Statistical - Introduction Book review 3.1-3.3 Min Sun March. 13, 2009 Outline. What is Statistical. Why we need Statistical. How Statistical works Statistical Rock physics Information theory Statistics

More information

arxiv: v5 [math.st] 20 Dec 2018

arxiv: v5 [math.st] 20 Dec 2018 Tropical Sufficient Statistics for Persistent Homology Anthea Monod 1,, Sara Kališnik 2, Juan Ángel Patiño-Galindo1, and Lorin Crawford 3 5 1 Department of Systems Biology, Columbia University, New York,

More information

Persistent homology and nonparametric regression

Persistent homology and nonparametric regression Cleveland State University March 10, 2009, BIRS: Data Analysis using Computational Topology and Geometric Statistics joint work with Gunnar Carlsson (Stanford), Moo Chung (Wisconsin Madison), Peter Kim

More information

Persistent Homology. 128 VI Persistence

Persistent Homology. 128 VI Persistence 8 VI Persistence VI. Persistent Homology A main purpose of persistent homology is the measurement of the scale or resolution of a topological feature. There are two ingredients, one geometric, assigning

More information

arxiv: v2 [math.pr] 15 May 2016

arxiv: v2 [math.pr] 15 May 2016 MAXIMALLY PERSISTENT CYCLES IN RANDOM GEOMETRIC COMPLEXES OMER BOBROWSKI, MATTHEW KAHLE, AND PRIMOZ SKRABA arxiv:1509.04347v2 [math.pr] 15 May 2016 Abstract. We initiate the study of persistent homology

More information

STA 4273H: Sta-s-cal Machine Learning

STA 4273H: Sta-s-cal Machine Learning STA 4273H: Sta-s-cal Machine Learning Russ Salakhutdinov Department of Computer Science! Department of Statistical Sciences! rsalakhu@cs.toronto.edu! h0p://www.cs.utoronto.ca/~rsalakhu/ Lecture 2 In our

More information

Kernel-Based Contrast Functions for Sufficient Dimension Reduction

Kernel-Based Contrast Functions for Sufficient Dimension Reduction Kernel-Based Contrast Functions for Sufficient Dimension Reduction Michael I. Jordan Departments of Statistics and EECS University of California, Berkeley Joint work with Kenji Fukumizu and Francis Bach

More information

STA 4273H: Statistical Machine Learning

STA 4273H: Statistical Machine Learning STA 4273H: Statistical Machine Learning Russ Salakhutdinov Department of Statistics! rsalakhu@utstat.toronto.edu! http://www.utstat.utoronto.ca/~rsalakhu/ Sidney Smith Hall, Room 6002 Lecture 11 Project

More information

Advanced Introduction to Machine Learning CMU-10715

Advanced Introduction to Machine Learning CMU-10715 Advanced Introduction to Machine Learning CMU-10715 Gaussian Processes Barnabás Póczos http://www.gaussianprocess.org/ 2 Some of these slides in the intro are taken from D. Lizotte, R. Parr, C. Guesterin

More information

Stabilizing the unstable output of persistent homology computations

Stabilizing the unstable output of persistent homology computations Stabilizing the unstable output of persistent homology computations Peter Bubenik with Paul Bendich (Duke) and Alex Wagner (Florida) May 5, 2017 Conference on Applied and Computational Algebraic Topology

More information

Statistical Learning. Philipp Koehn. 10 November 2015

Statistical Learning. Philipp Koehn. 10 November 2015 Statistical Learning Philipp Koehn 10 November 2015 Outline 1 Learning agents Inductive learning Decision tree learning Measuring learning performance Bayesian learning Maximum a posteriori and maximum

More information

On Inverse Problems in TDA

On Inverse Problems in TDA Abel Symposium Geiranger, June 2018 On Inverse Problems in TDA Steve Oudot joint work with E. Solomon (Brown Math.) arxiv: 1712.03630 Features Data Rn The preimage problem in the data Sciences feature

More information

20: Gaussian Processes

20: Gaussian Processes 10-708: Probabilistic Graphical Models 10-708, Spring 2016 20: Gaussian Processes Lecturer: Andrew Gordon Wilson Scribes: Sai Ganesh Bandiatmakuri 1 Discussion about ML Here we discuss an introduction

More information

Approximate Inference Part 1 of 2

Approximate Inference Part 1 of 2 Approximate Inference Part 1 of 2 Tom Minka Microsoft Research, Cambridge, UK Machine Learning Summer School 2009 http://mlg.eng.cam.ac.uk/mlss09/ Bayesian paradigm Consistent use of probability theory

More information

Nonparametric Bayes Inference on Manifolds with Applications

Nonparametric Bayes Inference on Manifolds with Applications Nonparametric Bayes Inference on Manifolds with Applications Abhishek Bhattacharya Indian Statistical Institute Based on the book Nonparametric Statistics On Manifolds With Applications To Shape Spaces

More information

Computer Vision Group Prof. Daniel Cremers. 10a. Markov Chain Monte Carlo

Computer Vision Group Prof. Daniel Cremers. 10a. Markov Chain Monte Carlo Group Prof. Daniel Cremers 10a. Markov Chain Monte Carlo Markov Chain Monte Carlo In high-dimensional spaces, rejection sampling and importance sampling are very inefficient An alternative is Markov Chain

More information

Adaptive Crowdsourcing via EM with Prior

Adaptive Crowdsourcing via EM with Prior Adaptive Crowdsourcing via EM with Prior Peter Maginnis and Tanmay Gupta May, 205 In this work, we make two primary contributions: derivation of the EM update for the shifted and rescaled beta prior and

More information

Approximate Inference Part 1 of 2

Approximate Inference Part 1 of 2 Approximate Inference Part 1 of 2 Tom Minka Microsoft Research, Cambridge, UK Machine Learning Summer School 2009 http://mlg.eng.cam.ac.uk/mlss09/ 1 Bayesian paradigm Consistent use of probability theory

More information

Statistical techniques for data analysis in Cosmology

Statistical techniques for data analysis in Cosmology Statistical techniques for data analysis in Cosmology arxiv:0712.3028; arxiv:0911.3105 Numerical recipes (the bible ) Licia Verde ICREA & ICC UB-IEEC http://icc.ub.edu/~liciaverde outline Lecture 1: Introduction

More information

The Jackknife-Like Method for Assessing Uncertainty of Point Estimates for Bayesian Estimation in a Finite Gaussian Mixture Model

The Jackknife-Like Method for Assessing Uncertainty of Point Estimates for Bayesian Estimation in a Finite Gaussian Mixture Model Thai Journal of Mathematics : 45 58 Special Issue: Annual Meeting in Mathematics 207 http://thaijmath.in.cmu.ac.th ISSN 686-0209 The Jackknife-Like Method for Assessing Uncertainty of Point Estimates for

More information

Practical Bayesian Optimization of Machine Learning. Learning Algorithms

Practical Bayesian Optimization of Machine Learning. Learning Algorithms Practical Bayesian Optimization of Machine Learning Algorithms CS 294 University of California, Berkeley Tuesday, April 20, 2016 Motivation Machine Learning Algorithms (MLA s) have hyperparameters that

More information

Machine Learning - MT & 14. PCA and MDS

Machine Learning - MT & 14. PCA and MDS Machine Learning - MT 2016 13 & 14. PCA and MDS Varun Kanade University of Oxford November 21 & 23, 2016 Announcements Sheet 4 due this Friday by noon Practical 3 this week (continue next week if necessary)

More information

A Gaussian Type Kernel for Persistence Diagrams

A Gaussian Type Kernel for Persistence Diagrams TRIPODS Summer Bootcamp: Topology and Machine Learning Brown University, August 2018 A Gaussian Type Kernel for Persistence Diagrams Mathieu Carrière joint work with S. Oudot and M. Cuturi Persistence

More information

Introduction to Machine Learning

Introduction to Machine Learning Outline Introduction to Machine Learning Bayesian Classification Varun Chandola March 8, 017 1. {circular,large,light,smooth,thick}, malignant. {circular,large,light,irregular,thick}, malignant 3. {oval,large,dark,smooth,thin},

More information

Tutorial on Gaussian Processes and the Gaussian Process Latent Variable Model

Tutorial on Gaussian Processes and the Gaussian Process Latent Variable Model Tutorial on Gaussian Processes and the Gaussian Process Latent Variable Model (& discussion on the GPLVM tech. report by Prof. N. Lawrence, 06) Andreas Damianou Department of Neuro- and Computer Science,

More information

Clustering by Mixture Models. General background on clustering Example method: k-means Mixture model based clustering Model estimation

Clustering by Mixture Models. General background on clustering Example method: k-means Mixture model based clustering Model estimation Clustering by Mixture Models General bacground on clustering Example method: -means Mixture model based clustering Model estimation 1 Clustering A basic tool in data mining/pattern recognition: Divide

More information

Chris Bishop s PRML Ch. 8: Graphical Models

Chris Bishop s PRML Ch. 8: Graphical Models Chris Bishop s PRML Ch. 8: Graphical Models January 24, 2008 Introduction Visualize the structure of a probabilistic model Design and motivate new models Insights into the model s properties, in particular

More information

Instance-based Learning CE-717: Machine Learning Sharif University of Technology. M. Soleymani Fall 2016

Instance-based Learning CE-717: Machine Learning Sharif University of Technology. M. Soleymani Fall 2016 Instance-based Learning CE-717: Machine Learning Sharif University of Technology M. Soleymani Fall 2016 Outline Non-parametric approach Unsupervised: Non-parametric density estimation Parzen Windows Kn-Nearest

More information

A Review of Pseudo-Marginal Markov Chain Monte Carlo

A Review of Pseudo-Marginal Markov Chain Monte Carlo A Review of Pseudo-Marginal Markov Chain Monte Carlo Discussed by: Yizhe Zhang October 21, 2016 Outline 1 Overview 2 Paper review 3 experiment 4 conclusion Motivation & overview Notation: θ denotes the

More information

SYDE 372 Introduction to Pattern Recognition. Probability Measures for Classification: Part I

SYDE 372 Introduction to Pattern Recognition. Probability Measures for Classification: Part I SYDE 372 Introduction to Pattern Recognition Probability Measures for Classification: Part I Alexander Wong Department of Systems Design Engineering University of Waterloo Outline 1 2 3 4 Why use probability

More information

Linear vs Non-linear classifier. CS789: Machine Learning and Neural Network. Introduction

Linear vs Non-linear classifier. CS789: Machine Learning and Neural Network. Introduction Linear vs Non-linear classifier CS789: Machine Learning and Neural Network Support Vector Machine Jakramate Bootkrajang Department of Computer Science Chiang Mai University Linear classifier is in the

More information

Clustering VS Classification

Clustering VS Classification MCQ Clustering VS Classification 1. What is the relation between the distance between clusters and the corresponding class discriminability? a. proportional b. inversely-proportional c. no-relation Ans:

More information

Homework in Topology, Spring 2009.

Homework in Topology, Spring 2009. Homework in Topology, Spring 2009. Björn Gustafsson April 29, 2009 1 Generalities To pass the course you should hand in correct and well-written solutions of approximately 10-15 of the problems. For higher

More information

Kernel Density Topic Models: Visual Topics Without Visual Words

Kernel Density Topic Models: Visual Topics Without Visual Words Kernel Density Topic Models: Visual Topics Without Visual Words Konstantinos Rematas K.U. Leuven ESAT-iMinds krematas@esat.kuleuven.be Mario Fritz Max Planck Institute for Informatics mfrtiz@mpi-inf.mpg.de

More information

6.867 Machine Learning

6.867 Machine Learning 6.867 Machine Learning Problem Set 2 Due date: Wednesday October 6 Please address all questions and comments about this problem set to 6867-staff@csail.mit.edu. You will need to use MATLAB for some of

More information

Computational Genomics. Systems biology. Putting it together: Data integration using graphical models

Computational Genomics. Systems biology. Putting it together: Data integration using graphical models 02-710 Computational Genomics Systems biology Putting it together: Data integration using graphical models High throughput data So far in this class we discussed several different types of high throughput

More information

Probabilistic Graphical Networks: Definitions and Basic Results

Probabilistic Graphical Networks: Definitions and Basic Results This document gives a cursory overview of Probabilistic Graphical Networks. The material has been gleaned from different sources. I make no claim to original authorship of this material. Bayesian Graphical

More information

Lecture Notes 1: Vector spaces

Lecture Notes 1: Vector spaces Optimization-based data analysis Fall 2017 Lecture Notes 1: Vector spaces In this chapter we review certain basic concepts of linear algebra, highlighting their application to signal processing. 1 Vector

More information

Probabilistic Graphical Models Lecture 20: Gaussian Processes

Probabilistic Graphical Models Lecture 20: Gaussian Processes Probabilistic Graphical Models Lecture 20: Gaussian Processes Andrew Gordon Wilson www.cs.cmu.edu/~andrewgw Carnegie Mellon University March 30, 2015 1 / 53 What is Machine Learning? Machine learning algorithms

More information

W vs. QCD Jet Tagging at the Large Hadron Collider

W vs. QCD Jet Tagging at the Large Hadron Collider W vs. QCD Jet Tagging at the Large Hadron Collider Bryan Anenberg: anenberg@stanford.edu; CS229 December 13, 2013 Problem Statement High energy collisions of protons at the Large Hadron Collider (LHC)

More information