Model-based cluster analysis: a Defence. Gilles Celeux Inria Futurs
|
|
- Clifton Murphy
- 5 years ago
- Views:
Transcription
1 Model-based cluster analysis: a Defence Gilles Celeux Inria Futurs
2 Model-based cluster analysis Model-based clustering (MBC) consists of assuming that the data come from a source with several subpopulations. Each subpopulation is modeled separetaly. The overall population is a mixture of these subpopulations. The resulting model is a finite mixture models.
3 Finite Mixture Models The general form of a mixture model with K groups is f(x) = K p k f k (x) k=1 the p k s are the mixing proportions and the f k (.) s denote the densities of the groups. The parameterisation of the group densities depends of the nature (continuous or discrete) of the observed data.
4 An hidden structure model The mixture model is an incomplete data structure model: The complete data are y = (y 1,..., y n ) = ((x 1, z 1 ),..., (x n, z n )) where the missing data are z = (z 1,..., z n ), with z i = (z i1,..., z ik ) are binary vectors such that z ik = 1 iff x i arises from group k. The z s define a partition P = (P 1,..., P K ) of the observed data x with P k = {x i z ik = 1}.
5 Multivariate Gaussian Mixture (MGM) Multidimensional observations x = (x 1,..., x n ) in R d are assumed to be a sample from a probability distribution with density f(x i K, θ) = K p k φ(x i m k, Σ k ) k=1 where the p k s are the mixing proportions and φ(. m k, Σ k ) denotes a Gaussian density with mean m k and variance matrix Σ k. This is the most popular model for clustering of quantitative data.
6 Discrete Data Observations to be classified are described with d discrete variables. Each variable j has m j response levels. Data are represented in the following way: (x 1,..., x n ) where x i = (x jh i ; j = 1,..., d; h = 1,..., m j ) with { x jh i = 1 if i has response level h for variable j x jh i = 0 otherwise.
7 The standard latent class model (LCM) Data are supposed to arise from a mixture of g multivariate multinomial distributions with pdf g f(x i ; θ) = p k (α jh i where k=1 p k m k (x i ; α k ) = k j,h k )xjh α jh k is denoting the probability that variable xj has level h if object i in cluster k, and α k = (α jh k ; j = 1,..., p; h = 1,..., m j), p = (p 1,..., p g ) is denoting the vector of mixing proportions of the g latent clusters, θ = (p k, α k, k = 1,..., g) denoting the vector parameter of the latent class model to be estimated. Latent class model is assuming that the variables are conditionnally independent knowing the latent clusters.
8 First Interest of MBC Many versatile or parsimonious models available
9 MGM: The variance matrix eigenvalue decomposition where Σ k = V k D t k A kd k V k = Σ k ) 1/d defines the component volume (d is the dimension of the observation space), D k the matrix of eigenvectors of Σ defines the component orientation A k the diagonal matrix of normalised eigenvalues defines the component shape. By allowing some of these quantities to vary between components, we get different and easily interpreted models.
10 28 different models Following Banfield & Raftery (1993) or Celeux & Govaert (1995), a large range of 28 versatile (from the most complex to the simplest one) models derived from this eigenvalue decomposition can be considered. The general family: Assuming equal or free proportions, volumes orientations and shapes leads to 16 different models. The diagonal family: Assuming in addition that the component variances matrices are diagonal leads to 8 models. The spherical family: Assuming in addition that the component variance matrices are proportional to the identity matrix leads to 4 models.
11 LMC: a Reparameterization α (a, ε) where binary vector a k = (a 1 k,..., ad k ) provides the mode levels in cluster k for variable j { (a jh 1 if h = arg maxh α ) = jh 0 otherwise. and the ε j k can be regarded as scattering values. (ε jh ) = { 1 α jh if a jh = 1 α jh if a jh = 0. For instance, if α j = (0.7, 0.2, 0.1), the new parameters are a j = (1, 0, 0) and ε j = (0.3, 0.2, 0.1).
12 Five latent class models Denoting h(ij) the level of object i for the variable j, the model can be written f(x i ; θ) = ( ) p k (1 ε jh(jk) k ) xjh(jk) i (ε jh(ij) k ) xjh(ij) i x jh(jk) k. k j Using this form, it is possible to impose various constraints to the scattering parameters ε jh k. The models we consider are the following the standard latent class model [ε jh k ]: The scattering is depending upon clusters, variables and levels. [ε j k ]: The scattering is depending upon clusters and variables but not upon levels. [ε k ]: The scattering is depending upon clusters, but not upon variables. [ε j ]: The scattering is depending upon variables, but not upon clusters. [ε]: The scattering is constant over variables and clusters.
13 Second interest of MBC Many algorithms to estimate the mixture model from different points of view
14 Maximum likelihood estimation The EM algorithm is the reference tool to derive the ML estimates in a mixture model. E step Compute the conditional probabilities t ik, i = 1,..., n, k = 1,..., K that x i arises from the kth component for the current value of the mixture parameters. M step Update the mixture parameter estimates maximising the expected value of the completed likelihood. It leads to weight the observation i for group k with the conditional probability t ik.
15 Classification EM E step Compute the conditional probabilities t ik, i = 1,..., n, k = 1,..., K that x i arises from the kth component for the current value of the mixture parameters. C step Assign each point x i to the component maximising the conditional probability t ik (MAP principle). M step Update the mixture parameter estimates maximising the completed likelihood.
16 Features of CEM CEM aims maximising the complete likelihood where the component label of each sample point is included in the data set. Contrary to EM, CEM converges in a finite number of iterations CEM is a K-means-like algorithm. CEM provides biased estimates of the mixture parameters.
17 Stochastic EM E step Compute the conditional probabilities t ik, i = 1,..., n, k = 1,..., K that x i arises from the kth component for the current value of the mixture parameters. S step Assign each point x i at random to one of the component according to the distribution defined by the (t ik, k = 1,..., K). M step Update the mixture parameter estimates maximising the completed likelihood.
18 Features of SEM SEM generates a Markov chain whose stationary distribution is (more or less) concentrated around the ML paramater estimator. Thus a natural parameter estimate from a SEM sequence is the mean of the iterates values obtain after a burn-in period (SEMmean). An alternative estimate is to consider the parameter value leading to the largest likelihood in a SEM sequence (SEMmax). Different variants (Monte Carlo EM, Simulated Annealing EM) are possible.
19 Strategies for getting sensible ml estimate Aim: Find the maximum likelihood value with specified model and component number. Biernacki, Celeux & Govaert (2003) proposed simple strategies acting in two ways to deal with this problem. Starting values: use of CEM or SEM. Repeating: Run several times EM or algorithms combining CEM cdcccand EM.
20 Third interest of MBC Finite mixture models can be compared and assessed in an objective way
21 Model selection Choosing a parsimonious model in a collection of models. The problem is to solve the bias-variance dilemma. A too simple model leads to a large approximation error. A too complex model leads to a large estimation error. Standard criteria of model selection are AIC and BIC criteria. Both criteria are penalized likelihood criteria.
22 The AIC criterion AIC is approximating the deviance of a model m with ν m free parameters d(x) = 2[log p(x) log p(x m, ˆθ m )] on a single test observation X. of nd(x) E(d(x)) where The penalization is an estimation is the expected deviance on X. D(X) = 2E[log p(x) log p(x m, ˆθ m )] Assuming that the data arose from a distribution belonging to the collection of models in competition, AIC is AIC(m) = 2 log p(x m, ˆθ m ) 2ν m.
23 The BIC criterion BIC is a Bayesian criterion. It is approximating the integrated likelihood of the data p(x m) = p(x m, θ m )π(θ m )dθ m, π(θ m ) being a prior distribution for parameter θ m. BIC is BIC(m) = log p(x m, ˆθ m ) ν m 2 log(n). This approximation is appropriate when one and only one of the competing models is true.
24 Choosing a mixture model in a density estimation context Despite theoretical difficulties in the mixture context Simulation experiments (see Roeder & Wasserman 1997) show that BIC works well at a practical level to choose a sensible Gaussian mixture model, See also the good performances of a MML criterion proposed by Figuereido & Jain (2002). Simulations experiments (Nadolski & Viele 2004) show that a variant of AIC, AIC3 which replaces the penalty 2ν m with 3ν m works well at a practical level to choose a sensible latent class mixture model.
25 Assessing K in a clustering context Assessing the number of groups K is an important but difficult problem. Mixture modelling can be regarded as a semi parametric tool for density estimation purpose or as a model for cluster analysis. In the density estimation context, BIC is doing the job quite well. In the cluster analysis context, since BIC does not take into account the clustering purpose for assessing K, BIC has a tendency to overestimate K regardless of the separation of the clusters. To overcome this limitation, it can be advantageous to choose K in order to get the mixture giving rise to partitioning data with the greatest evidence (Biernacki, Celeux & Govaert 2000).
26 The ICL criterion: definition It leads to consider the integrated likelihood of the complete data (x, z) (or integrated completed likelihood), p(x, z K) = p(x, z K, θ)π(θ K)dθ, Θ K where with p(x, z K, θ) = p(x i, z i K, θ) = n p(x i, z i K, θ) i=1 K k=1 p z ik k [φ(x i a k )] z ik. To approximate this integrated complete likelihood, a BIC-like approximation is possible. It leads to the criterion ICL(K) = log p(x, ẑ K, ˆθ) ν K log n, 2 where the missing data have been replaced by their most probable value for parameter estimate ˆθ.
27 Behavior of the ICL criterion Roughly speaking criterion ICL is the criterion BIC penalized by the estimated mean entropy E(K) = K k=1 n t ik log t ik 0, i=1 t ik denoting the conditional probability that x i arises from the kth mixture component (1 i n and 1 k K). Because of this additional entropy term, ICL favors K values giving rise to partitioning the data with the greatest evidence. ICL appears to provide a stable and reliable estimate of K for real data sets and also for simulated data sets from mixtures when the components are not too much overlapping. But ICL, which is not aiming to discover the true number of mixture components, can underestimate the number of components for simulated data arising from mixture with poorly separated components.
28 Fourth interest of MBC Special questions can be tackled in a proper way in the MBC context
29 Robust Cluster Analysis Using multivariate Student distributions instead of Multivariate Gaussian distributions lead to attenuate the influence of outliers (McLachlan & Peel 2000). Including in the mixture a group from a uniform distribution allows to take into account noisy data (DasGupta & Raftery 1998). Shrinking the group variance matrix in a proper way (see for instance Ciuperca, Idier & Ridolfi 2002)
30 Some other special questions solved with MBC Imposing specific constraints as groups with known distributions. Dealing with missing data at random in a proper way (Hunt & Basford 1999, 2001). Variable Selection for Model-Based Clustering with a stepwise procedure using BIC criterion (Raftery & Dean 2006) (see also Law, Figueiredo & Jain 2004 for an alternative approach). Simple and efficient models in semi-supevised Classification (see for instance Ganesalingam & McLachlan 1978,... ). Assuming the variables are conditionally independent knowing the groups makes valid, in a simple way, the treatment of continuous and discrete data with the same MBC.
31 Concluding Remarks Finite Mixture analysis is definitively a powerful framework for model Cluster analysis. Many free and valuable softwares are available: C.A.MAN, EM- MIX, FLEXMIX, MClust, MIXMOD a, MULTIMIX, SNOB... Since the seminal paper of Diebolt & Robert (1993), Bayesian Inference for mixtures have received a lot of attention. Bayesian analysis of univariate mixtures has became the standard Bayesian tool for density estimation. It cannot be regarded as a standard tool for Bayesian clustering of multivariate data since a lot of problems are not completely solved: slow convergence, subjective weakly informative problems, identifiability (label switching),... a
Model selection criteria in Classification contexts. Gilles Celeux INRIA Futurs (orsay)
Model selection criteria in Classification contexts Gilles Celeux INRIA Futurs (orsay) Cluster analysis Exploratory data analysis tools which aim is to find clusters in a large set of data (many observations
More informationDifferent points of view for selecting a latent structure model
Different points of view for selecting a latent structure model Gilles Celeux Inria Saclay-Île-de-France, Université Paris-Sud Latent structure models: two different point of views Density estimation LSM
More informationChoosing a model in a Classification purpose. Guillaume Bouchard, Gilles Celeux
Choosing a model in a Classification purpose Guillaume Bouchard, Gilles Celeux Abstract: We advocate the usefulness of taking into account the modelling purpose when selecting a model. Two situations are
More informationFinite Mixture Models and Clustering
Finite Mixture Models and Clustering Mohamed Nadif LIPADE, Université Paris Descartes, France Nadif (LIPADE) EPAT, May, 2010 Course 3 1 / 40 Introduction Outline 1 Introduction Mixture Approach 2 Finite
More informationVariable selection for model-based clustering
Variable selection for model-based clustering Matthieu Marbac (Ensai - Crest) Joint works with: M. Sedki (Univ. Paris-sud) and V. Vandewalle (Univ. Lille 2) The problem Objective: Estimation of a partition
More informationmixmod Statistical Documentation
mixmod Statistical Documentation February 10, 2016 Contents 1 Introduction 3 2 Mixture model 3 2.1 Density estimation from a mixture model............ 4 2.2 Clustering with mixture model..................
More informationExact and Monte Carlo Calculations of Integrated Likelihoods for the Latent Class Model
Exact and Monte Carlo Calculations of Integrated Likelihoods for the Latent Class Model C. Biernacki a,, G. Celeux b, G. Govaert c a CNRS & Université de Lille 1, Villeneuve d Ascq, France b INRIA, Orsay,
More informationMixture models for analysing transcriptome and ChIP-chip data
Mixture models for analysing transcriptome and ChIP-chip data Marie-Laure Martin-Magniette French National Institute for agricultural research (INRA) Unit of Applied Mathematics and Informatics at AgroParisTech,
More informationMixtures and Hidden Markov Models for genomics data
Mixtures and Hidden Markov Models for genomics data M.-L. Martin-Magniette marie laure.martin@agroparistech.fr Researcher at the French National Institut for Agricultural Research Plant Science Institut
More informationINRIA Rh^one-Alpes. Abstract. Friedman (1989) has proposed a regularization technique (RDA) of discriminant analysis
Regularized Gaussian Discriminant Analysis through Eigenvalue Decomposition Halima Bensmail Universite Paris 6 Gilles Celeux INRIA Rh^one-Alpes Abstract Friedman (1989) has proposed a regularization technique
More informationAn Introduction to mixture models
An Introduction to mixture models by Franck Picard Research Report No. 7 March 2007 Statistics for Systems Biology Group Jouy-en-Josas/Paris/Evry, France http://genome.jouy.inra.fr/ssb/ An introduction
More informationMixtures and Hidden Markov Models for analyzing genomic data
Mixtures and Hidden Markov Models for analyzing genomic data Marie-Laure Martin-Magniette UMR AgroParisTech/INRA Mathématique et Informatique Appliquées, Paris UMR INRA/UEVE ERL CNRS Unité de Recherche
More informationThe Jackknife-Like Method for Assessing Uncertainty of Point Estimates for Bayesian Estimation in a Finite Gaussian Mixture Model
Thai Journal of Mathematics : 45 58 Special Issue: Annual Meeting in Mathematics 207 http://thaijmath.in.cmu.ac.th ISSN 686-0209 The Jackknife-Like Method for Assessing Uncertainty of Point Estimates for
More informationSTA 4273H: Statistical Machine Learning
STA 4273H: Statistical Machine Learning Russ Salakhutdinov Department of Statistics! rsalakhu@utstat.toronto.edu! http://www.utstat.utoronto.ca/~rsalakhu/ Sidney Smith Hall, Room 6002 Lecture 3 Linear
More informationUnsupervised machine learning
Chapter 9 Unsupervised machine learning Unsupervised machine learning (a.k.a. cluster analysis) is a set of methods to assign objects into clusters under a predefined distance measure when class labels
More informationAdaptive Mixture Discriminant Analysis for. Supervised Learning with Unobserved Classes
Adaptive Mixture Discriminant Analysis for Supervised Learning with Unobserved Classes Charles Bouveyron SAMOS-MATISSE, CES, UMR CNRS 8174 Université Paris 1 (Panthéon-Sorbonne), Paris, France Abstract
More informationMore on Unsupervised Learning
More on Unsupervised Learning Two types of problems are to find association rules for occurrences in common in observations (market basket analysis), and finding the groups of values of observational data
More informationMODEL BASED CLUSTERING FOR COUNT DATA
MODEL BASED CLUSTERING FOR COUNT DATA Dimitris Karlis Department of Statistics Athens University of Economics and Business, Athens April OUTLINE Clustering methods Model based clustering!"the general model!"algorithmic
More informationComputer Vision Group Prof. Daniel Cremers. 10a. Markov Chain Monte Carlo
Group Prof. Daniel Cremers 10a. Markov Chain Monte Carlo Markov Chain Monte Carlo In high-dimensional spaces, rejection sampling and importance sampling are very inefficient An alternative is Markov Chain
More informationUNIVERSITY of PENNSYLVANIA CIS 520: Machine Learning Final, Fall 2013
UNIVERSITY of PENNSYLVANIA CIS 520: Machine Learning Final, Fall 2013 Exam policy: This exam allows two one-page, two-sided cheat sheets; No other materials. Time: 2 hours. Be sure to write your name and
More informationLabel Switching and Its Simple Solutions for Frequentist Mixture Models
Label Switching and Its Simple Solutions for Frequentist Mixture Models Weixin Yao Department of Statistics, Kansas State University, Manhattan, Kansas 66506, U.S.A. wxyao@ksu.edu Abstract The label switching
More informationHmms with variable dimension structures and extensions
Hmm days/enst/january 21, 2002 1 Hmms with variable dimension structures and extensions Christian P. Robert Université Paris Dauphine www.ceremade.dauphine.fr/ xian Hmm days/enst/january 21, 2002 2 1 Estimating
More informationSTA 4273H: Statistical Machine Learning
STA 4273H: Statistical Machine Learning Russ Salakhutdinov Department of Statistics! rsalakhu@utstat.toronto.edu! http://www.utstat.utoronto.ca/~rsalakhu/ Sidney Smith Hall, Room 6002 Lecture 7 Approximate
More informationStatistics 202: Data Mining. c Jonathan Taylor. Model-based clustering Based in part on slides from textbook, slides of Susan Holmes.
Model-based clustering Based in part on slides from textbook, slides of Susan Holmes December 2, 2012 1 / 1 Model-based clustering General approach Choose a type of mixture model (e.g. multivariate Normal)
More informationAugmented Likelihood Estimators for Mixture Models
Augmented Likelihood Estimators for Mixture Models Markus Haas Jochen Krause Marc S. Paolella Swiss Banking Institute, University of Zurich What is mixture degeneracy? mixtures under study are finite convex
More informationJournal of Statistical Software
JSS Journal of Statistical Software January 2012, Volume 46, Issue 6. http://www.jstatsoft.org/ HDclassif: An R Package for Model-Based Clustering and Discriminant Analysis of High-Dimensional Data Laurent
More informationModel-based clustering for conditionally correlated categorical data
Model-based clustering for conditionally correlated categorical data arxiv:1401.5684v3 [stat.co] 10 Jul 2014 Matthieu Marbac Inria & DGA Christophe Biernacki University Lille 1 & CNRS & Inria Vincent Vandewalle
More informationMinimum message length estimation of mixtures of multivariate Gaussian and von Mises-Fisher distributions
Minimum message length estimation of mixtures of multivariate Gaussian and von Mises-Fisher distributions Parthan Kasarapu & Lloyd Allison Monash University, Australia September 8, 25 Parthan Kasarapu
More informationModel-Based Co-Clustering of Multivariate Functional Data
Model-Based Co-Clustering of Multivariate Functional Data Faicel Chamroukhi, Christophe Biernacki To cite this version: Faicel Chamroukhi, Christophe Biernacki. Model-Based Co-Clustering of Multivariate
More informationRegression Clustering
Regression Clustering In regression clustering, we assume a model of the form y = f g (x, θ g ) + ɛ g for observations y and x in the g th group. Usually, of course, we assume linear models of the form
More informationModel-based clustering for conditionally correlated categorical data
Model-based clustering for conditionally correlated categorical data Matthieu Marbac Inria & DGA Christophe Biernacki University Lille 1 & CNRS & Inria Vincent Vandewalle EA 2694 University Lille 2 & Inria
More informationTHE EMMIX SOFTWARE FOR THE FITTING OF MIXTURES OF NORMAL AND t-components
THE EMMIX SOFTWARE FOR THE FITTING OF MIXTURES OF NORMAL AND t-components G.J. McLachlan, D. Peel, K.E. Basford*, and P. Adams Department of Mathematics, University of Queensland, St. Lucia, Queensland
More informationarxiv: v1 [stat.ml] 22 Jun 2012
Hidden Markov Models with mixtures as emission distributions Stevenn Volant 1,2, Caroline Bérard 1,2, Marie-Laure Martin Magniette 1,2,3,4,5 and Stéphane Robin 1,2 arxiv:1206.5102v1 [stat.ml] 22 Jun 2012
More informationPattern Recognition and Machine Learning
Christopher M. Bishop Pattern Recognition and Machine Learning ÖSpri inger Contents Preface Mathematical notation Contents vii xi xiii 1 Introduction 1 1.1 Example: Polynomial Curve Fitting 4 1.2 Probability
More informationEnhancing the selection of a model-based clustering with external qualitative variables
Enhancing the selection of a model-based clustering with external qualitative variables Jean-Patrick Baudry, Margarida Cardoso, Gilles Celeux, Maria José Amorim, Ana Sousa Ferreira RESEARCH REPORT N 814
More informationEM for Spherical Gaussians
EM for Spherical Gaussians Karthekeyan Chandrasekaran Hassan Kingravi December 4, 2007 1 Introduction In this project, we examine two aspects of the behavior of the EM algorithm for mixtures of spherical
More informationSubmitted to the Brazilian Journal of Probability and Statistics
Submitted to the Brazilian Journal of Probability and Statistics A Bayesian sparse finite mixture model for clustering data from a heterogeneous population Erlandson F. Saraiva a, Adriano K. Suzuki b and
More informationRecursive Deviance Information Criterion for the Hidden Markov Model
International Journal of Statistics and Probability; Vol. 5, No. 1; 2016 ISSN 1927-7032 E-ISSN 1927-7040 Published by Canadian Center of Science and Education Recursive Deviance Information Criterion for
More informationA clustering view on ESS measures of political interest:
A clustering view on ESS measures of political interest: An EM-MML approach Cláudia Silvestre Margarida Cardoso Mário Figueiredo Escola Superior de Comunicação Social - IPL BRU_UNIDE, ISCTE-IUL Instituto
More informationReview of Maximum Likelihood Estimators
Libby MacKinnon CSE 527 notes Lecture 7, October 7, 2007 MLE and EM Review of Maximum Likelihood Estimators MLE is one of many approaches to parameter estimation. The likelihood of independent observations
More informationIntroduction to Machine Learning. Maximum Likelihood and Bayesian Inference. Lecturers: Eran Halperin, Lior Wolf
1 Introduction to Machine Learning Maximum Likelihood and Bayesian Inference Lecturers: Eran Halperin, Lior Wolf 2014-15 We know that X ~ B(n,p), but we do not know p. We get a random sample from X, a
More informationParametric Unsupervised Learning Expectation Maximization (EM) Lecture 20.a
Parametric Unsupervised Learning Expectation Maximization (EM) Lecture 20.a Some slides are due to Christopher Bishop Limitations of K-means Hard assignments of data points to clusters small shift of a
More informationSelecting Hidden Markov Model State Number with Cross-Validated Likelihood
Selecting Hidden Markov Model State Number with Cross-Validated Likelihood Gilles Celeux, Jean-Baptiste Durand To cite this version: Gilles Celeux, Jean-Baptiste Durand. Selecting Hidden Markov Model State
More informationOverview. and data transformations of gene expression data. Toy 2-d Clustering Example. K-Means. Motivation. Model-based clustering
Model-based clustering and data transformations of gene expression data Walter L. Ruzzo University of Washington UW CSE Computational Biology Group 2 Toy 2-d Clustering Example K-Means? 3 4 Hierarchical
More informationSeminar über Statistik FS2008: Model Selection
Seminar über Statistik FS2008: Model Selection Alessia Fenaroli, Ghazale Jazayeri Monday, April 2, 2008 Introduction Model Choice deals with the comparison of models and the selection of a model. It can
More informationLecture 6: Model Checking and Selection
Lecture 6: Model Checking and Selection Melih Kandemir melih.kandemir@iwr.uni-heidelberg.de May 27, 2014 Model selection We often have multiple modeling choices that are equally sensible: M 1,, M T. Which
More informationMACHINE LEARNING ADVANCED MACHINE LEARNING
MACHINE LEARNING ADVANCED MACHINE LEARNING Recap of Important Notions on Estimation of Probability Density Functions 2 2 MACHINE LEARNING Overview Definition pdf Definition joint, condition, marginal,
More informationStandard Errors & Confidence Intervals. N(0, I( β) 1 ), I( β) = [ 2 l(β, φ; y) β i β β= β j
Standard Errors & Confidence Intervals β β asy N(0, I( β) 1 ), where I( β) = [ 2 l(β, φ; y) ] β i β β= β j We can obtain asymptotic 100(1 α)% confidence intervals for β j using: β j ± Z 1 α/2 se( β j )
More informationPublication IV Springer-Verlag Berlin Heidelberg. Reprinted with permission.
Publication IV Emil Eirola and Amaury Lendasse. Gaussian Mixture Models for Time Series Modelling, Forecasting, and Interpolation. In Advances in Intelligent Data Analysis XII 12th International Symposium
More informationMaximum Likelihood Estimation. only training data is available to design a classifier
Introduction to Pattern Recognition [ Part 5 ] Mahdi Vasighi Introduction Bayesian Decision Theory shows that we could design an optimal classifier if we knew: P( i ) : priors p(x i ) : class-conditional
More informationNonstationary spatial process modeling Part II Paul D. Sampson --- Catherine Calder Univ of Washington --- Ohio State University
Nonstationary spatial process modeling Part II Paul D. Sampson --- Catherine Calder Univ of Washington --- Ohio State University this presentation derived from that presented at the Pan-American Advanced
More informationAdvanced Introduction to Machine Learning
10-715 Advanced Introduction to Machine Learning Homework 3 Due Nov 12, 10.30 am Rules 1. Homework is due on the due date at 10.30 am. Please hand over your homework at the beginning of class. Please see
More informationClustering by Mixture Models. General background on clustering Example method: k-means Mixture model based clustering Model estimation
Clustering by Mixture Models General bacground on clustering Example method: -means Mixture model based clustering Model estimation 1 Clustering A basic tool in data mining/pattern recognition: Divide
More informationMinimum Hellinger Distance Estimation in a. Semiparametric Mixture Model
Minimum Hellinger Distance Estimation in a Semiparametric Mixture Model Sijia Xiang 1, Weixin Yao 1, and Jingjing Wu 2 1 Department of Statistics, Kansas State University, Manhattan, Kansas, USA 66506-0802.
More informationPerformance Comparison of K-Means and Expectation Maximization with Gaussian Mixture Models for Clustering EE6540 Final Project
Performance Comparison of K-Means and Expectation Maximization with Gaussian Mixture Models for Clustering EE6540 Final Project Devin Cornell & Sushruth Sastry May 2015 1 Abstract In this article, we explore
More informationIntroduction to Machine Learning
Introduction to Machine Learning Brown University CSCI 1950-F, Spring 2012 Prof. Erik Sudderth Lecture 20: Expectation Maximization Algorithm EM for Mixture Models Many figures courtesy Kevin Murphy s
More informationL11: Pattern recognition principles
L11: Pattern recognition principles Bayesian decision theory Statistical classifiers Dimensionality reduction Clustering This lecture is partly based on [Huang, Acero and Hon, 2001, ch. 4] Introduction
More informationMinimum Error Rate Classification
Minimum Error Rate Classification Dr. K.Vijayarekha Associate Dean School of Electrical and Electronics Engineering SASTRA University, Thanjavur-613 401 Table of Contents 1.Minimum Error Rate Classification...
More informationBAYESIAN METHODS FOR VARIABLE SELECTION WITH APPLICATIONS TO HIGH-DIMENSIONAL DATA
BAYESIAN METHODS FOR VARIABLE SELECTION WITH APPLICATIONS TO HIGH-DIMENSIONAL DATA Intro: Course Outline and Brief Intro to Marina Vannucci Rice University, USA PASI-CIMAT 04/28-30/2010 Marina Vannucci
More informationPattern Recognition and Machine Learning. Bishop Chapter 2: Probability Distributions
Pattern Recognition and Machine Learning Chapter 2: Probability Distributions Cécile Amblard Alex Kläser Jakob Verbeek October 11, 27 Probability Distributions: General Density Estimation: given a finite
More informationCOS513 LECTURE 8 STATISTICAL CONCEPTS
COS513 LECTURE 8 STATISTICAL CONCEPTS NIKOLAI SLAVOV AND ANKUR PARIKH 1. MAKING MEANINGFUL STATEMENTS FROM JOINT PROBABILITY DISTRIBUTIONS. A graphical model (GM) represents a family of probability distributions
More informationAdaptive Mixture Discriminant Analysis for Supervised Learning with Unobserved Classes
Adaptive Mixture Discriminant Analysis for Supervised Learning with Unobserved Classes Charles Bouveyron To cite this version: Charles Bouveyron. Adaptive Mixture Discriminant Analysis for Supervised Learning
More information36-720: Latent Class Models
36-720: Latent Class Models Brian Junker October 17, 2007 Latent Class Models and Simpson s Paradox Latent Class Model for a Table of Counts An Intuitive E-M Algorithm for Fitting Latent Class Models Deriving
More informationDensity Estimation. Seungjin Choi
Density Estimation Seungjin Choi Department of Computer Science and Engineering Pohang University of Science and Technology 77 Cheongam-ro, Nam-gu, Pohang 37673, Korea seungjin@postech.ac.kr http://mlg.postech.ac.kr/
More informationDetermining the number of components in mixture models for hierarchical data
Determining the number of components in mixture models for hierarchical data Olga Lukočienė 1 and Jeroen K. Vermunt 2 1 Department of Methodology and Statistics, Tilburg University, P.O. Box 90153, 5000
More informationAlgorithmisches Lernen/Machine Learning
Algorithmisches Lernen/Machine Learning Part 1: Stefan Wermter Introduction Connectionist Learning (e.g. Neural Networks) Decision-Trees, Genetic Algorithms Part 2: Norman Hendrich Support-Vector Machines
More informationAn Extended BIC for Model Selection
An Extended BIC for Model Selection at the JSM meeting 2007 - Salt Lake City Surajit Ray Boston University (Dept of Mathematics and Statistics) Joint work with James Berger, Duke University; Susie Bayarri,
More informationMachine Learning. Gaussian Mixture Models. Zhiyao Duan & Bryan Pardo, Machine Learning: EECS 349 Fall
Machine Learning Gaussian Mixture Models Zhiyao Duan & Bryan Pardo, Machine Learning: EECS 349 Fall 2012 1 The Generative Model POV We think of the data as being generated from some process. We assume
More informationCommunications in Statistics - Simulation and Computation. Comparison of EM and SEM Algorithms in Poisson Regression Models: a simulation study
Comparison of EM and SEM Algorithms in Poisson Regression Models: a simulation study Journal: Manuscript ID: LSSP-00-0.R Manuscript Type: Original Paper Date Submitted by the Author: -May-0 Complete List
More informationNonparametric Bayesian Methods (Gaussian Processes)
[70240413 Statistical Machine Learning, Spring, 2015] Nonparametric Bayesian Methods (Gaussian Processes) Jun Zhu dcszj@mail.tsinghua.edu.cn http://bigml.cs.tsinghua.edu.cn/~jun State Key Lab of Intelligent
More informationPenalized Loss functions for Bayesian Model Choice
Penalized Loss functions for Bayesian Model Choice Martyn International Agency for Research on Cancer Lyon, France 13 November 2009 The pure approach For a Bayesian purist, all uncertainty is represented
More informationUnsupervised Learning
Unsupervised Learning Bayesian Model Comparison Zoubin Ghahramani zoubin@gatsby.ucl.ac.uk Gatsby Computational Neuroscience Unit, and MSc in Intelligent Systems, Dept Computer Science University College
More informationMonte Carlo Methods. Leon Gu CSD, CMU
Monte Carlo Methods Leon Gu CSD, CMU Approximate Inference EM: y-observed variables; x-hidden variables; θ-parameters; E-step: q(x) = p(x y, θ t 1 ) M-step: θ t = arg max E q(x) [log p(y, x θ)] θ Monte
More informationCovariance Matrix Enhancement Approach to Train Robust Gaussian Mixture Models of Speech Data
Covariance Matrix Enhancement Approach to Train Robust Gaussian Mixture Models of Speech Data Jan Vaněk, Lukáš Machlica, Josef V. Psutka, Josef Psutka University of West Bohemia in Pilsen, Univerzitní
More information9/12/17. Types of learning. Modeling data. Supervised learning: Classification. Supervised learning: Regression. Unsupervised learning: Clustering
Types of learning Modeling data Supervised: we know input and targets Goal is to learn a model that, given input data, accurately predicts target data Unsupervised: we know the input only and want to make
More informationLatent Variable Models and EM algorithm
Latent Variable Models and EM algorithm SC4/SM4 Data Mining and Machine Learning, Hilary Term 2017 Dino Sejdinovic 3.1 Clustering and Mixture Modelling K-means and hierarchical clustering are non-probabilistic
More informationDIC, AIC, BIC, PPL, MSPE Residuals Predictive residuals
DIC, AIC, BIC, PPL, MSPE Residuals Predictive residuals Overall Measures of GOF Deviance: this measures the overall likelihood of the model given a parameter vector D( θ) = 2 log L( θ) This can be averaged
More informationStatistics 203: Introduction to Regression and Analysis of Variance Course review
Statistics 203: Introduction to Regression and Analysis of Variance Course review Jonathan Taylor - p. 1/?? Today Review / overview of what we learned. - p. 2/?? General themes in regression models Specifying
More informationNew Global Optimization Algorithms for Model-Based Clustering
New Global Optimization Algorithms for Model-Based Clustering Jeffrey W. Heath Department of Mathematics University of Maryland, College Park, MD 7, jheath@math.umd.edu Michael C. Fu Robert H. Smith School
More informationLecture 16: Small Sample Size Problems (Covariance Estimation) Many thanks to Carlos Thomaz who authored the original version of these slides
Lecture 16: Small Sample Size Problems (Covariance Estimation) Many thanks to Carlos Thomaz who authored the original version of these slides Intelligent Data Analysis and Probabilistic Inference Lecture
More informationLecture 4: Probabilistic Learning
DD2431 Autumn, 2015 1 Maximum Likelihood Methods Maximum A Posteriori Methods Bayesian methods 2 Classification vs Clustering Heuristic Example: K-means Expectation Maximization 3 Maximum Likelihood Methods
More informationGaussian Mixture Models
Gaussian Mixture Models Pradeep Ravikumar Co-instructor: Manuela Veloso Machine Learning 10-701 Some slides courtesy of Eric Xing, Carlos Guestrin (One) bad case for K- means Clusters may overlap Some
More informationMultilevel Statistical Models: 3 rd edition, 2003 Contents
Multilevel Statistical Models: 3 rd edition, 2003 Contents Preface Acknowledgements Notation Two and three level models. A general classification notation and diagram Glossary Chapter 1 An introduction
More informationISSN Article. Selection Criteria in Regime Switching Conditional Volatility Models
Econometrics 2015, 3, 289-316; doi:10.3390/econometrics3020289 OPEN ACCESS econometrics ISSN 2225-1146 www.mdpi.com/journal/econometrics Article Selection Criteria in Regime Switching Conditional Volatility
More informationNon-Parametric Bayes
Non-Parametric Bayes Mark Schmidt UBC Machine Learning Reading Group January 2016 Current Hot Topics in Machine Learning Bayesian learning includes: Gaussian processes. Approximate inference. Bayesian
More informationMachine Learning Techniques for Computer Vision
Machine Learning Techniques for Computer Vision Part 2: Unsupervised Learning Microsoft Research Cambridge x 3 1 0.5 0.2 0 0.5 0.3 0 0.5 1 ECCV 2004, Prague x 2 x 1 Overview of Part 2 Mixture models EM
More informationEstimation and model selection for model-based clustering with the conditional classification likelihood
Electronic Journal of Statistics Vol. 9 (215) 141 177 ISSN: 1935-7524 DOI: 1.1214/15-EJS126 Estimation and model selection for model-based clustering with the conditional classification likelihood Jean-Patrick
More informationPATTERN CLASSIFICATION
PATTERN CLASSIFICATION Second Edition Richard O. Duda Peter E. Hart David G. Stork A Wiley-lnterscience Publication JOHN WILEY & SONS, INC. New York Chichester Weinheim Brisbane Singapore Toronto CONTENTS
More informationMathematical Formulation of Our Example
Mathematical Formulation of Our Example We define two binary random variables: open and, where is light on or light off. Our question is: What is? Computer Vision 1 Combining Evidence Suppose our robot
More informationLecture 13: Data Modelling and Distributions. Intelligent Data Analysis and Probabilistic Inference Lecture 13 Slide No 1
Lecture 13: Data Modelling and Distributions Intelligent Data Analysis and Probabilistic Inference Lecture 13 Slide No 1 Why data distributions? It is a well established fact that many naturally occurring
More informationExpectation Maximization Algorithm
Expectation Maximization Algorithm Vibhav Gogate The University of Texas at Dallas Slides adapted from Carlos Guestrin, Dan Klein, Luke Zettlemoyer and Dan Weld The Evils of Hard Assignments? Clusters
More informationDegenerate Expectation-Maximization Algorithm for Local Dimension Reduction
Degenerate Expectation-Maximization Algorithm for Local Dimension Reduction Xiaodong Lin 1 and Yu Zhu 2 1 Statistical and Applied Mathematical Science Institute, RTP, NC, 27709 USA University of Cincinnati,
More informationVariational Scoring of Graphical Model Structures
Variational Scoring of Graphical Model Structures Matthew J. Beal Work with Zoubin Ghahramani & Carl Rasmussen, Toronto. 15th September 2003 Overview Bayesian model selection Approximations using Variational
More informationModel Selection for Mixture Models-Perspectives and Strategies
Model Selection for Mixture Models-Perspectives and Strategies Gilles Celeux, Sylvia Frühwirth-Schnatter, Christian Robert To cite this version: Gilles Celeux, Sylvia Frühwirth-Schnatter, Christian Robert.
More informationParsimonious Gaussian Mixture Models
Parsimonious Gaussian Mixture Models Brendan Murphy Department of Statistics, Trinity College Dublin, Ireland. East Liguria West Liguria Umbria North Apulia Coast Sardina Inland Sardinia South Apulia Calabria
More informationLecture 4: Probabilistic Learning. Estimation Theory. Classification with Probability Distributions
DD2431 Autumn, 2014 1 2 3 Classification with Probability Distributions Estimation Theory Classification in the last lecture we assumed we new: P(y) Prior P(x y) Lielihood x2 x features y {ω 1,..., ω K
More informationMultimodal Nested Sampling
Multimodal Nested Sampling Farhan Feroz Astrophysics Group, Cavendish Lab, Cambridge Inverse Problems & Cosmology Most obvious example: standard CMB data analysis pipeline But many others: object detection,
More informationCurves clustering with approximation of the density of functional random variables
Curves clustering with approximation of the density of functional random variables Julien Jacques and Cristian Preda Laboratoire Paul Painlevé, UMR CNRS 8524, University Lille I, Lille, France INRIA Lille-Nord
More information7. Estimation and hypothesis testing. Objective. Recommended reading
7. Estimation and hypothesis testing Objective In this chapter, we show how the election of estimators can be represented as a decision problem. Secondly, we consider the problem of hypothesis testing
More information