Missing values imputation for mixed data based on principal component methods

Size: px
Start display at page:

Download "Missing values imputation for mixed data based on principal component methods"

Transcription

1 Missing values imputation for mixed data based on principal component methods Vincent Audigier, François Husson & Julie Josse Agrocampus Rennes Compstat' 2012, Limassol (Cyprus), / 21

2 A real dataset age weight size alcohol sex snore tobacco or 2 glasses/day M yes no or 2 glasses/day M no <= No W no <= or 2 glasses/day M no <= No W yes > >2 glasses/day M yes no or 2 glasses/day M yes <= No W no <= >2 glasses/day M no > or 2 glasses/day W no <= or 2 glasses/day M no no or 2 glasses/day W no <= or 2 glasses/day M yes <= >2 glasses/day M yes <= or 2 glasses/day M no <= No W no <= No M no <= or 2 glasses/day M yes <= or 2 glasses/day M yes no >2 glasses/day M no <= or 2 glasses/day W no <=1 2 / 21

3 A real dataset age weight size alcohol sex snore tobacco 51 NA 172 NA M yes no or 2 glasses/day M NA <=1 48 NA 164 No W no NA or 2 glasses/day M no <= No W yes > NA >2 glasses/day M yes no or 2 glasses/day M NA NA 49 NA 179 No W no <= >2 glasses/day M NA >1 NA or 2 glasses/day W no <= or 2 glasses/day M no no 34 NA 181 NA W no <= NA 1 or 2 glasses/day M yes <= NA >2 glasses/day M NA <= or 2 glasses/day M no NA NA No W NA <= No M no <= NA 1 or 2 glasses/day M NA NA 65 NA or 2 glasses/day M yes no >2 glasses/day M no <= NA W no <=1 2 / 21

4 A real dataset age weight size alcohol sex snore tobacco 51 NA 172 NA M yes no or 2 glasses/day M NA <=1 48 NA 164 No W no NA or 2 glasses/day M no <= No W yes > NA >2 glasses/day M yes no or 2 glasses/day M NA NA 49 NA 179 No W no <= >2 glasses/day M NA >1 NA or 2 glasses/day W no <= or 2 glasses/day M no no 34 NA 181 NA W no <= NA 1 or 2 glasses/day M yes <= NA >2 glasses/day M NA <= or 2 glasses/day M no NA NA No W NA <= No M no <= NA 1 or 2 glasses/day M NA NA 65 NA or 2 glasses/day M yes no >2 glasses/day M no <= NA W no <=1 Popular approach to deal with missing values: Single imputation 2 / 21

5 Single imputation methods Continuous variables: k-nearest neighbours, normal distribution (joint modelling), iterative regression (fully conditional specication), etc. Categorical variables: hot-deck imputation, multinomial model, latent class model (Vermunt et al., 2008), etc. Mixed data: transform the categorical variables into dummy variables and deal as continuous variables (package Amelia) MICE (multivariate imputation by chained equations, van Bureen 1999): a model must be specied for each variable random forest (Stekhoven & Bühlmann, 2011) New imputation method based on principal component methods 3 / 21

6 Imputation with PCA for continuous variables PCA minimizes: With missing values: C = X I J F I S U t S J 2 C = W (X FU t ) 2, with w ij = 0 if x ij is missing, w ij = 1 otherwise. Iterative PCA (Kiers, 1997) 4 / 21

7 Iterative PCA algorithm 1.5 NA x2 The data set x1 5 / 21

8 Iterative PCA algorithm 1.5 NA x2 Initialization step: mean imputation x1 5 / 21

9 1.5 NA Iterative PCA algorithm PCA performed on the completed data set; 1 dimension is kept x2 x1 5 / 21

10 Iterative PCA algorithm 1.5 NA x2 Calculation of the model prediction x1 5 / 21

11 Iterative PCA algorithm 1.5 NA x2 Imputation step: X l = W X + (1 W) ˆX l x1 5 / 21

12 Iterative PCA algorithm 1.5 NA x2 PCA is performed; 1 dimension is kept x1 5 / 21

13 Iterative PCA algorithm 1.5 NA x2 Imputation step: X l = W X + (1 W) ˆX l x1 5 / 21

14 Iterative PCA algorithm 1.5 NA x2 Iterate until convergence x1 5 / 21

15 x2 Iterative PCA - convergence Imputed values are obtained at convergence 1.5 NA x1 6 / 21

16 Iterative PCA 1 initialization l = 0: X 0 (mean imputation) 2 step l: (a) PCA on the completed matrix X l 1 ˆF l I S, Ûl K S S dimensions are kept; ˆX l = ˆF l Û l (estimation) (b) X l = W X + (1 W) ˆF l Û l (imputation) 3 Estimation and imputation are repeated until convergence The number of dimensions S has to be chosen a priori An imputation is performed during the algorithm PCA can be seen as an imputation method Overtting problems are managed with a regularized algorithm 7 / 21

17 Imputation with MCA for categorical variables MCA can be seen as the PCA on an indicator matrix X with specic weights depending on D Σ NA NA NA NA NA J I X= xik J D Σ = 0 I k NA NA I 1 I k I K J IJ I K 8 / 21

18 Iterative MCA Iterative MCA algorithm: 1 initialization: imputation of the indicator matrix (proportion) 2 iterate until convergence (a) MCA on the completed indicator matrix ˆF, Û (Estimation) (b) imputation of the missing values with the model matrix (c) column margins are updated V1 V2 V3 V14 V1_a V1_b V1_c V2_e V2_f V3_g V3_h ind 1 a NA g u ind ind 2 NA f g u ind ind 3 a e h v ind ind 4 a e h v ind ind 5 b f h u ind ind 6 c f h u ind ind 7 c f NA v ind ind 1232 c f h v ind imputed values can be seen as degree of membership 9 / 21

19 Principal component method for mixed data The core of principal component methods is PCA on particular matrices "Doing a data analysis, in good mathematics, is simply searching eigenvectors, all the science of it (the art) is just to nd the right matrix to diagonalize" (Benzécri) 10 / 21

20 Principal component method for mixed data (complete case) Factorial Analysis on Mixed Data (Escoer, 1979), PCAMIX (Kiers, 1991) Continuous variables Categorical variables Indicator matrix centring & scaling I 1 I 2 I k division by I/Ik and centring Matrix which balances the influence of each variable A PCA is performed on the weighted matrix 11 / 21

21 Properties of the method The distance between individuals is: K cont d 2 (i, l) = (x ik x lk ) 2 + k=1 K Q q q=1 k=1 The principal component F s maximises: K cont k=1 Q cat r 2 (F s, v k ) + η 2 (F s, v q ) q=1 1 I kq (x iq x lq ) 2 12 / 21

22 Imputation of mixed data The same kind of iterative algorithm as before age weight size alcohol sex snore tobacco NA NA M yes no gl/d M NA <=1 NA No W no NA gl/d M no <=1 NA NA NA NA NA NA NA NA NA NA imputeafdm age weight size alcohol sex snore tobacco gl/d M yes no gl/d M no <= No W no <= gl/d M no <= / 21

23 Properties on the imputations Imputation based on scores and loadings similarities between individuals and relationships between variables Relationships between continuous and categorical variables are taken into account Categorical variables evolve within many dimensions (k q 1 if they have k q categories) so they need many dimensions to be well predicted Compared to a PCA on the (unweighted) indicator matrix, small categories are better imputed 14 / 21

24 Simulation pattern Simulations 2 independent variables are drawn from a normal distribution 1 variable is replicated 4 times, the other 8 2 dimensions Random noise is added Half of the variables in each dimension are split in 3 clusters 10%, 20% or 30% of missing values are chosen at random Data are constructed (expected) to be in 4 dimensions Criterion for continuous data: N2RMSE = i missing mean ( ( X true i var (X true i ) ) ) 2 X imp i for categorical data: proportion of falsely classied entries 15 / 21

25 Simulations Imputation using continuous data only Imputation using both continuous and categorical data Error on continous data N2RMSE quanti 10% mixte 10% quanti 20% mixte 20% quanti 30% mixte 30% 16 / 21

26 Simulations Imputation using continuous data only Imputation using both continuous and categorical data Error on continous data N2RMSE quanti 10% mixte 10% quanti 20% mixte 20% quanti 30% mixte 30% Categorical data improved the imputation on continuous data / 21

27 Simulations Imputation using continuous data only Imputation using categorical data only Imputation using both continuous and categorical data Error on continous data Error on categorical data N2RMSE PFC quanti 10% mixte 10% quanti 20% mixte 20% quanti 30% mixte 30% quali 10% mixte 10% quali 20% mixte 20% quali 30% mixte 30% Categorical data improved the imputation on continuous data / 21

28 Simulations Imputation using continuous data only Imputation using categorical data only Imputation using both continuous and categorical data Error on continous data Error on categorical data N2RMSE PFC quanti 10% mixte 10% quanti 20% mixte 20% quanti 30% mixte 30% quali 10% mixte 10% quali 20% mixte 20% quali 30% mixte 30% Categorical data improved the imputation on continuous data and continuous data improved the imputation on categorical data 16 / 21

29 Simulations Error on continuous variables Error on the qualitative variables N2RMSE % 20% 30% PFC % 20% 30% Nb of dimensions Nb of dimensions The error on the estimation of the number of dimensions has not an important impact on the imputation error... if the estimation is not too bad 17 / 21

30 Comparison with random forest on real data sets Imputations obtained with random forest & iterative algorithm GBSG2 N2RMSE RF 10% AFDM 10% RF 20% AFDM 20% RF 30% AFDM 30% PFC RF 10% AFDM 10% RF 20% AFDM 20% RF 30% AFDM 30% 18 / 21

31 Comparison with random forest on real data sets Imputations obtained with random forest & iterative algorithm GBSG2 Ozone N2RMSE N2RMSE RF 10% AFDM 10% RF 20% AFDM 20% RF 30% AFDM 30% RF 10% AFDM 10% RF 20% AFDM 20% RF 30% AFDM 30% PFC PFC RF 10% AFDM 10% RF 20% AFDM 20% RF 30% AFDM 30% RF 10% AFDM 10% RF 20% AFDM 20% RF 30% AFDM 30% 18 / 21

32 Comparison with random forest Compared to random forest, imputations are quite similar Imputations are slightly better: for categorical variables especially for rare categories and imputations are worse: when there are non-linear relationships between continuous variables when there are interactions 19 / 21

33 Conclusion A new way to impute missing values is ecient when strong linear relationships between variables (you learn from the other variables) but needs tuning parameters, cv? approximation? is available in the missmda package. This package: handles missing values in principal component methods (PCA, MCA, MIXPCA, MFA) impute missing values for continuous, categorical and mixed variables performs multiple imputation for continuous variables 20 / 21

34 Perspective How to perform a statistical analysis from an incomplete dataset? we can modify the estimation process to apply it on an incomplete dataset (not always easy!) we can predict the missing entries with a single imputation method, but BE CAREFUL using the usual methods leads to underestimate the standard errors An alternative is to use multiple imputation... and single imputation is a rst step towards multiple imputation 21 / 21

missmda: a package to handle missing values in Multivariate exploratory Data Analysis methods

missmda: a package to handle missing values in Multivariate exploratory Data Analysis methods missmda: a package to handle missing values in Multivariate exploratory Data Analysis methods Julie Josse & Francois Husson Applied Mathematics Department Agrocampus Rennes - France user! 2011, Warwick,

More information

Handling missing values in Multiple Factor Analysis

Handling missing values in Multiple Factor Analysis Handling missing values in Multiple Factor Analysis François Husson, Julie Josse Applied mathematics department, Agrocampus Ouest, Rennes, France Rabat, 26 March 2014 1 / 10 Multi-blocks data set Groups

More information

Distributed multilevel matrix completion for medical databases

Distributed multilevel matrix completion for medical databases Distributed multilevel matrix completion for medical databases Julie Josse Ecole Polytechnique, INRIA Séminaire Parisien de Statistique, IHP January 15, 2018 1 / 38 Overview 1 Introduction 2 Single imputation

More information

CS281 Section 4: Factor Analysis and PCA

CS281 Section 4: Factor Analysis and PCA CS81 Section 4: Factor Analysis and PCA Scott Linderman At this point we have seen a variety of machine learning models, with a particular emphasis on models for supervised learning. In particular, we

More information

Comparison of multiple imputation methods for systematically and sporadically missing multilevel data

Comparison of multiple imputation methods for systematically and sporadically missing multilevel data Comparison of multiple imputation methods for systematically and sporadically missing multilevel data V. Audigier, I. White, S. Jolani, T. Debray, M. Quartagno, J. Carpenter, S. van Buuren, M. Resche-Rigon

More information

Statistical Methods. Missing Data snijders/sm.htm. Tom A.B. Snijders. November, University of Oxford 1 / 23

Statistical Methods. Missing Data  snijders/sm.htm. Tom A.B. Snijders. November, University of Oxford 1 / 23 1 / 23 Statistical Methods Missing Data http://www.stats.ox.ac.uk/ snijders/sm.htm Tom A.B. Snijders University of Oxford November, 2011 2 / 23 Literature: Joseph L. Schafer and John W. Graham, Missing

More information

Data set Science&Environment (ISSP,1993) Multiple categorical variables. Burt matrix. Indicator matrix V4a(A) Multiple Correspondence Analysis

Data set Science&Environment (ISSP,1993) Multiple categorical variables. Burt matrix. Indicator matrix V4a(A) Multiple Correspondence Analysis Data set Science&Environment (ISSP,1993) Q. SCIENCE AND ENVIRONMENT Multiple categorical variables Multiple Correspondence Analysis How much do you agree or disagree with each of these statements? Q.a

More information

Probabilistic Low-Rank Matrix Completion with Adaptive Spectral Regularization Algorithms

Probabilistic Low-Rank Matrix Completion with Adaptive Spectral Regularization Algorithms Probabilistic Low-Rank Matrix Completion with Adaptive Spectral Regularization Algorithms François Caron Department of Statistics, Oxford STATLEARN 2014, Paris April 7, 2014 Joint work with Adrien Todeschini,

More information

Structural Equation Modeling and Confirmatory Factor Analysis. Types of Variables

Structural Equation Modeling and Confirmatory Factor Analysis. Types of Variables /4/04 Structural Equation Modeling and Confirmatory Factor Analysis Advanced Statistics for Researchers Session 3 Dr. Chris Rakes Website: http://csrakes.yolasite.com Email: Rakes@umbc.edu Twitter: @RakesChris

More information

Threshold models with fixed and random effects for ordered categorical data

Threshold models with fixed and random effects for ordered categorical data Threshold models with fixed and random effects for ordered categorical data Hans-Peter Piepho Universität Hohenheim, Germany Edith Kalka Universität Kassel, Germany Contents 1. Introduction. Case studies

More information

Regularized PCA to denoise and visualise data

Regularized PCA to denoise and visualise data Regularized PCA to denoise and visualise data Marie Verbanck Julie Josse François Husson Laboratoire de statistique, Agrocampus Ouest, Rennes, France CNAM, Paris, 16 janvier 2013 1 / 30 Outline 1 PCA 2

More information

Principal Component Analysis (PCA) for Sparse High-Dimensional Data

Principal Component Analysis (PCA) for Sparse High-Dimensional Data AB Principal Component Analysis (PCA) for Sparse High-Dimensional Data Tapani Raiko, Alexander Ilin, and Juha Karhunen Helsinki University of Technology, Finland Adaptive Informatics Research Center Principal

More information

Sparse vectors recap. ANLP Lecture 22 Lexical Semantics with Dense Vectors. Before density, another approach to normalisation.

Sparse vectors recap. ANLP Lecture 22 Lexical Semantics with Dense Vectors. Before density, another approach to normalisation. ANLP Lecture 22 Lexical Semantics with Dense Vectors Henry S. Thompson Based on slides by Jurafsky & Martin, some via Dorota Glowacka 5 November 2018 Previous lectures: Sparse vectors recap How to represent

More information

ANLP Lecture 22 Lexical Semantics with Dense Vectors

ANLP Lecture 22 Lexical Semantics with Dense Vectors ANLP Lecture 22 Lexical Semantics with Dense Vectors Henry S. Thompson Based on slides by Jurafsky & Martin, some via Dorota Glowacka 5 November 2018 Henry S. Thompson ANLP Lecture 22 5 November 2018 Previous

More information

MULTILEVEL IMPUTATION 1

MULTILEVEL IMPUTATION 1 MULTILEVEL IMPUTATION 1 Supplement B: MCMC Sampling Steps and Distributions for Two-Level Imputation This document gives technical details of the full conditional distributions used to draw regression

More information

Advanced Introduction to Machine Learning

Advanced Introduction to Machine Learning 10-715 Advanced Introduction to Machine Learning Homework 3 Due Nov 12, 10.30 am Rules 1. Homework is due on the due date at 10.30 am. Please hand over your homework at the beginning of class. Please see

More information

Experimental Design and Data Analysis for Biologists

Experimental Design and Data Analysis for Biologists Experimental Design and Data Analysis for Biologists Gerry P. Quinn Monash University Michael J. Keough University of Melbourne CAMBRIDGE UNIVERSITY PRESS Contents Preface page xv I I Introduction 1 1.1

More information

Fractional Hot Deck Imputation for Robust Inference Under Item Nonresponse in Survey Sampling

Fractional Hot Deck Imputation for Robust Inference Under Item Nonresponse in Survey Sampling Fractional Hot Deck Imputation for Robust Inference Under Item Nonresponse in Survey Sampling Jae-Kwang Kim 1 Iowa State University June 26, 2013 1 Joint work with Shu Yang Introduction 1 Introduction

More information

Linear & Non-Linear Discriminant Analysis! Hugh R. Wilson

Linear & Non-Linear Discriminant Analysis! Hugh R. Wilson Linear & Non-Linear Discriminant Analysis! Hugh R. Wilson PCA Review! Supervised learning! Fisher linear discriminant analysis! Nonlinear discriminant analysis! Research example! Multiple Classes! Unsupervised

More information

Final Overview. Introduction to ML. Marek Petrik 4/25/2017

Final Overview. Introduction to ML. Marek Petrik 4/25/2017 Final Overview Introduction to ML Marek Petrik 4/25/2017 This Course: Introduction to Machine Learning Build a foundation for practice and research in ML Basic machine learning concepts: max likelihood,

More information

An ensemble learning method for variable selection

An ensemble learning method for variable selection An ensemble learning method for variable selection Vincent Audigier, Avner Bar-Hen CNAM, MSDMA team, Paris Journées de la statistique 2018 1 / 17 Context Y n 1 = X n p β p 1 + ε n 1 ε N (0, σ 2 I) β sparse

More information

Statistical aspects of prediction models with high-dimensional data

Statistical aspects of prediction models with high-dimensional data Statistical aspects of prediction models with high-dimensional data Anne Laure Boulesteix Institut für Medizinische Informationsverarbeitung, Biometrie und Epidemiologie February 15th, 2017 Typeset by

More information

Latent Variable Models and EM algorithm

Latent Variable Models and EM algorithm Latent Variable Models and EM algorithm SC4/SM4 Data Mining and Machine Learning, Hilary Term 2017 Dino Sejdinovic 3.1 Clustering and Mixture Modelling K-means and hierarchical clustering are non-probabilistic

More information

Generative Models for Discrete Data

Generative Models for Discrete Data Generative Models for Discrete Data ddebarr@uw.edu 2016-04-21 Agenda Bayesian Concept Learning Beta-Binomial Model Dirichlet-Multinomial Model Naïve Bayes Classifiers Bayesian Concept Learning Numbers

More information

Cheng Soon Ong & Christian Walder. Canberra February June 2018

Cheng Soon Ong & Christian Walder. Canberra February June 2018 Cheng Soon Ong & Christian Walder Research Group and College of Engineering and Computer Science Canberra February June 218 Outlines Overview Introduction Linear Algebra Probability Linear Regression 1

More information

Data Preprocessing Tasks

Data Preprocessing Tasks Data Tasks 1 2 3 Data Reduction 4 We re here. 1 Dimensionality Reduction Dimensionality reduction is a commonly used approach for generating fewer features. Typically used because too many features can

More information

Factor Analysis (10/2/13)

Factor Analysis (10/2/13) STA561: Probabilistic machine learning Factor Analysis (10/2/13) Lecturer: Barbara Engelhardt Scribes: Li Zhu, Fan Li, Ni Guan Factor Analysis Factor analysis is related to the mixture models we have studied.

More information

Pooling multiple imputations when the sample happens to be the population.

Pooling multiple imputations when the sample happens to be the population. Pooling multiple imputations when the sample happens to be the population. Gerko Vink 1,2, and Stef van Buuren 1,3 arxiv:1409.8542v1 [math.st] 30 Sep 2014 1 Department of Methodology and Statistics, Utrecht

More information

0.1 factor.ord: Ordinal Data Factor Analysis

0.1 factor.ord: Ordinal Data Factor Analysis 0.1 factor.ord: Ordinal Data Factor Analysis Given some unobserved explanatory variables and observed ordinal dependent variables, this model estimates latent factors using a Gibbs sampler with data augmentation.

More information

PRINCIPAL COMPONENTS ANALYSIS

PRINCIPAL COMPONENTS ANALYSIS 121 CHAPTER 11 PRINCIPAL COMPONENTS ANALYSIS We now have the tools necessary to discuss one of the most important concepts in mathematical statistics: Principal Components Analysis (PCA). PCA involves

More information

The mice package. Stef van Buuren 1,2. amst-r-dam, Oct. 29, TNO, Leiden. 2 Methodology and Statistics, FSBS, Utrecht University

The mice package. Stef van Buuren 1,2. amst-r-dam, Oct. 29, TNO, Leiden. 2 Methodology and Statistics, FSBS, Utrecht University The mice package 1,2 1 TNO, Leiden 2 Methodology and Statistics, FSBS, Utrecht University amst-r-dam, Oct. 29, 2012 > The problem of missing data Consequences of missing data Less information than planned

More information

Nonnegative Matrix Factorization

Nonnegative Matrix Factorization Nonnegative Matrix Factorization Seungjin Choi Department of Computer Science and Engineering Pohang University of Science and Technology 77 Cheongam-ro, Nam-gu, Pohang 37673, Korea seungjin@postech.ac.kr

More information

Final Exam, Machine Learning, Spring 2009

Final Exam, Machine Learning, Spring 2009 Name: Andrew ID: Final Exam, 10701 Machine Learning, Spring 2009 - The exam is open-book, open-notes, no electronics other than calculators. - The maximum possible score on this exam is 100. You have 3

More information

Design of Text Mining Experiments. Matt Taddy, University of Chicago Booth School of Business faculty.chicagobooth.edu/matt.

Design of Text Mining Experiments. Matt Taddy, University of Chicago Booth School of Business faculty.chicagobooth.edu/matt. Design of Text Mining Experiments Matt Taddy, University of Chicago Booth School of Business faculty.chicagobooth.edu/matt.taddy/research Active Learning: a flavor of design of experiments Optimal : consider

More information

Machine Learning - MT & 14. PCA and MDS

Machine Learning - MT & 14. PCA and MDS Machine Learning - MT 2016 13 & 14. PCA and MDS Varun Kanade University of Oxford November 21 & 23, 2016 Announcements Sheet 4 due this Friday by noon Practical 3 this week (continue next week if necessary)

More information

MMWS Software Program Manual

MMWS Software Program Manual MMWS Software Program Manual 1 Software Development The MMWS program is regularly updated. The latest beta version can be downloaded from http://hlmsoft.net/ghong/ MMWS Click here to get MMWS. For a sample

More information

A Modified Incremental Principal Component Analysis for On-line Learning of Feature Space and Classifier

A Modified Incremental Principal Component Analysis for On-line Learning of Feature Space and Classifier A Modified Incremental Principal Component Analysis for On-line Learning of Feature Space and Classifier Seiichi Ozawa, Shaoning Pang, and Nikola Kasabov Graduate School of Science and Technology, Kobe

More information

Real Estate Price Prediction with Regression and Classification CS 229 Autumn 2016 Project Final Report

Real Estate Price Prediction with Regression and Classification CS 229 Autumn 2016 Project Final Report Real Estate Price Prediction with Regression and Classification CS 229 Autumn 2016 Project Final Report Hujia Yu, Jiafu Wu [hujiay, jiafuwu]@stanford.edu 1. Introduction Housing prices are an important

More information

COMS 4721: Machine Learning for Data Science Lecture 19, 4/6/2017

COMS 4721: Machine Learning for Data Science Lecture 19, 4/6/2017 COMS 4721: Machine Learning for Data Science Lecture 19, 4/6/2017 Prof. John Paisley Department of Electrical Engineering & Data Science Institute Columbia University PRINCIPAL COMPONENT ANALYSIS DIMENSIONALITY

More information

Correspondence Analysis

Correspondence Analysis Correspondence Analysis Julie Josse, François Husson, Sébastien Lê Applied Mathematics Department, Agrocampus Ouest user-2008 Dortmund, August 11th 2008 1 / 22 History Theoretical principles: Fisher (1940)

More information

CS181 Midterm 2 Practice Solutions

CS181 Midterm 2 Practice Solutions CS181 Midterm 2 Practice Solutions 1. Convergence of -Means Consider Lloyd s algorithm for finding a -Means clustering of N data, i.e., minimizing the distortion measure objective function J({r n } N n=1,

More information

MACHINE LEARNING. Methods for feature extraction and reduction of dimensionality: Probabilistic PCA and kernel PCA

MACHINE LEARNING. Methods for feature extraction and reduction of dimensionality: Probabilistic PCA and kernel PCA 1 MACHINE LEARNING Methods for feature extraction and reduction of dimensionality: Probabilistic PCA and kernel PCA 2 Practicals Next Week Next Week, Practical Session on Computer Takes Place in Room GR

More information

Data Mining Techniques

Data Mining Techniques Data Mining Techniques CS 622 - Section 2 - Spring 27 Pre-final Review Jan-Willem van de Meent Feedback Feedback https://goo.gl/er7eo8 (also posted on Piazza) Also, please fill out your TRACE evaluations!

More information

A Partitioning Method for the Clustering of Categorical Variables

A Partitioning Method for the Clustering of Categorical Variables A Partitioning Method for the Clustering of Categorical Variables Marie Chavent 1,2, Vanessa Kuentz 1,2, and Jérôme Saracco 1,2,3 1 Université de Bordeaux, IMB, CNRS, UMR 5251, France chavent@math.u-bordeaux1.fr;kuentz@math.u-bordeaux1.fr

More information

Latent Variable Models and EM Algorithm

Latent Variable Models and EM Algorithm SC4/SM8 Advanced Topics in Statistical Machine Learning Latent Variable Models and EM Algorithm Dino Sejdinovic Department of Statistics Oxford Slides and other materials available at: http://www.stats.ox.ac.uk/~sejdinov/atsml/

More information

Analyzing Pilot Studies with Missing Observations

Analyzing Pilot Studies with Missing Observations Analyzing Pilot Studies with Missing Observations Monnie McGee mmcgee@smu.edu. Department of Statistical Science Southern Methodist University, Dallas, Texas Co-authored with N. Bergasa (SUNY Downstate

More information

PROBABILISTIC LATENT SEMANTIC ANALYSIS

PROBABILISTIC LATENT SEMANTIC ANALYSIS PROBABILISTIC LATENT SEMANTIC ANALYSIS Lingjia Deng Revised from slides of Shuguang Wang Outline Review of previous notes PCA/SVD HITS Latent Semantic Analysis Probabilistic Latent Semantic Analysis Applications

More information

An introduction to clustering techniques

An introduction to clustering techniques - ABSTRACT Cluster analysis has been used in a wide variety of fields, such as marketing, social science, biology, pattern recognition etc. It is used to identify homogenous groups of cases to better understand

More information

Journal of Statistical Software

Journal of Statistical Software JSS Journal of Statistical Software January 2012, Volume 46, Issue 6. http://www.jstatsoft.org/ HDclassif: An R Package for Model-Based Clustering and Discriminant Analysis of High-Dimensional Data Laurent

More information

Unconstrained Ordination

Unconstrained Ordination Unconstrained Ordination Sites Species A Species B Species C Species D Species E 1 0 (1) 5 (1) 1 (1) 10 (4) 10 (4) 2 2 (3) 8 (3) 4 (3) 12 (6) 20 (6) 3 8 (6) 20 (6) 10 (6) 1 (2) 3 (2) 4 4 (5) 11 (5) 8 (5)

More information

Machine Learning - MT Clustering

Machine Learning - MT Clustering Machine Learning - MT 2016 15. Clustering Varun Kanade University of Oxford November 28, 2016 Announcements No new practical this week All practicals must be signed off in sessions this week Firm Deadline:

More information

PACKAGE LMest FOR LATENT MARKOV ANALYSIS

PACKAGE LMest FOR LATENT MARKOV ANALYSIS PACKAGE LMest FOR LATENT MARKOV ANALYSIS OF LONGITUDINAL CATEGORICAL DATA Francesco Bartolucci 1, Silvia Pandofi 1, and Fulvia Pennoni 2 1 Department of Economics, University of Perugia (e-mail: francesco.bartolucci@unipg.it,

More information

2.3. Clustering or vector quantization 57

2.3. Clustering or vector quantization 57 Multivariate Statistics non-negative matrix factorisation and sparse dictionary learning The PCA decomposition is by construction optimal solution to argmin A R n q,h R q p X AH 2 2 under constraint :

More information

6. Fractional Imputation in Survey Sampling

6. Fractional Imputation in Survey Sampling 6. Fractional Imputation in Survey Sampling 1 Introduction Consider a finite population of N units identified by a set of indices U = {1, 2,, N} with N known. Associated with each unit i in the population

More information

Expectation Maximization

Expectation Maximization Expectation Maximization Aaron C. Courville Université de Montréal Note: Material for the slides is taken directly from a presentation prepared by Christopher M. Bishop Learning in DAGs Two things could

More information

Multilevel Component Analysis applied to the measurement of a complex product experience

Multilevel Component Analysis applied to the measurement of a complex product experience Multilevel Component Analysis applied to the measurement of a complex product experience Boucon, C.A., Petit-Jublot, C.E.F., Groeneschild C., Dijksterhuis, G.B. Outline Background Introduction to Simultaneous

More information

Noise & Data Reduction

Noise & Data Reduction Noise & Data Reduction Paired Sample t Test Data Transformation - Overview From Covariance Matrix to PCA and Dimension Reduction Fourier Analysis - Spectrum Dimension Reduction 1 Remember: Central Limit

More information

MULTIVARIATE STATISTICAL ANALYSIS OF SPECTROSCOPIC DATA. Haisheng Lin, Ognjen Marjanovic, Barry Lennox

MULTIVARIATE STATISTICAL ANALYSIS OF SPECTROSCOPIC DATA. Haisheng Lin, Ognjen Marjanovic, Barry Lennox MULTIVARIATE STATISTICAL ANALYSIS OF SPECTROSCOPIC DATA Haisheng Lin, Ognjen Marjanovic, Barry Lennox Control Systems Centre, School of Electrical and Electronic Engineering, University of Manchester Abstract:

More information

Nearest Neighbor Gaussian Processes for Large Spatial Data

Nearest Neighbor Gaussian Processes for Large Spatial Data Nearest Neighbor Gaussian Processes for Large Spatial Data Abhi Datta 1, Sudipto Banerjee 2 and Andrew O. Finley 3 July 31, 2017 1 Department of Biostatistics, Bloomberg School of Public Health, Johns

More information

Data Mining. Dimensionality reduction. Hamid Beigy. Sharif University of Technology. Fall 1395

Data Mining. Dimensionality reduction. Hamid Beigy. Sharif University of Technology. Fall 1395 Data Mining Dimensionality reduction Hamid Beigy Sharif University of Technology Fall 1395 Hamid Beigy (Sharif University of Technology) Data Mining Fall 1395 1 / 42 Outline 1 Introduction 2 Feature selection

More information

EXTENDING PARTIAL LEAST SQUARES REGRESSION

EXTENDING PARTIAL LEAST SQUARES REGRESSION EXTENDING PARTIAL LEAST SQUARES REGRESSION ATHANASSIOS KONDYLIS UNIVERSITY OF NEUCHÂTEL 1 Outline Multivariate Calibration in Chemometrics PLS regression (PLSR) and the PLS1 algorithm PLS1 from a statistical

More information

Summary and discussion of: Dropout Training as Adaptive Regularization

Summary and discussion of: Dropout Training as Adaptive Regularization Summary and discussion of: Dropout Training as Adaptive Regularization Statistics Journal Club, 36-825 Kirstin Early and Calvin Murdock November 21, 2014 1 Introduction Multi-layered (i.e. deep) artificial

More information

Two-stage acceleration for non-linear PCA

Two-stage acceleration for non-linear PCA Two-stage acceleration for non-linear PCA Masahiro Kuroda, Okayama University of Science, kuroda@soci.ous.ac.jp Michio Sakakihara, Okayama University of Science, sakaki@mis.ous.ac.jp Yuichi Mori, Okayama

More information

Appendix A : rational of the spatial Principal Component Analysis

Appendix A : rational of the spatial Principal Component Analysis Appendix A : rational of the spatial Principal Component Analysis In this appendix, the following notations are used : X is the n-by-p table of centred allelic frequencies, where rows are observations

More information

Chapter 2 Nonlinear Principal Component Analysis

Chapter 2 Nonlinear Principal Component Analysis Chapter 2 Nonlinear Principal Component Analysis Abstract Principal components analysis (PCA) is a commonly used descriptive multivariate method for handling quantitative data and can be extended to deal

More information

Basics of Multivariate Modelling and Data Analysis

Basics of Multivariate Modelling and Data Analysis Basics of Multivariate Modelling and Data Analysis Kurt-Erik Häggblom 6. Principal component analysis (PCA) 6.1 Overview 6.2 Essentials of PCA 6.3 Numerical calculation of PCs 6.4 Effects of data preprocessing

More information

A Program for Data Transformations and Kernel Density Estimation

A Program for Data Transformations and Kernel Density Estimation A Program for Data Transformations and Kernel Density Estimation John G. Manchuk and Clayton V. Deutsch Modeling applications in geostatistics often involve multiple variables that are not multivariate

More information

Dynamic Clustering-Based Estimation of Missing Values in Mixed Type Data

Dynamic Clustering-Based Estimation of Missing Values in Mixed Type Data Dynamic Clustering-Based Estimation of Missing Values in Mixed Type Data Vadim Ayuyev, Joseph Jupin, Philip Harris and Zoran Obradovic Temple University, Philadelphia, USA 2009 Real Life Data is Often

More information

Decision Support. Dr. Johan Hagelbäck.

Decision Support. Dr. Johan Hagelbäck. Decision Support Dr. Johan Hagelbäck johan.hagelback@lnu.se http://aiguy.org Decision Support One of the earliest AI problems was decision support The first solution to this problem was expert systems

More information

analysis of incomplete data in statistical surveys

analysis of incomplete data in statistical surveys analysis of incomplete data in statistical surveys Ugo Guarnera 1 1 Italian National Institute of Statistics, Italy guarnera@istat.it Jordan Twinning: Imputation - Amman, 6-13 Dec 2014 outline 1 origin

More information

Chemometrics. Matti Hotokka Physical chemistry Åbo Akademi University

Chemometrics. Matti Hotokka Physical chemistry Åbo Akademi University Chemometrics Matti Hotokka Physical chemistry Åbo Akademi University Linear regression Experiment Consider spectrophotometry as an example Beer-Lamberts law: A = cå Experiment Make three known references

More information

Correspondence Analysis of Longitudinal Data

Correspondence Analysis of Longitudinal Data Correspondence Analysis of Longitudinal Data Mark de Rooij* LEIDEN UNIVERSITY, LEIDEN, NETHERLANDS Peter van der G. M. Heijden UTRECHT UNIVERSITY, UTRECHT, NETHERLANDS *Corresponding author (rooijm@fsw.leidenuniv.nl)

More information

Widths. Center Fluctuations. Centers. Centers. Widths

Widths. Center Fluctuations. Centers. Centers. Widths Radial Basis Functions: a Bayesian treatment David Barber Bernhard Schottky Neural Computing Research Group Department of Applied Mathematics and Computer Science Aston University, Birmingham B4 7ET, U.K.

More information

MACHINE LEARNING ADVANCED MACHINE LEARNING

MACHINE LEARNING ADVANCED MACHINE LEARNING MACHINE LEARNING ADVANCED MACHINE LEARNING Recap of Important Notions on Estimation of Probability Density Functions 2 2 MACHINE LEARNING Overview Definition pdf Definition joint, condition, marginal,

More information

Overlapping Community Detection at Scale: A Nonnegative Matrix Factorization Approach

Overlapping Community Detection at Scale: A Nonnegative Matrix Factorization Approach Overlapping Community Detection at Scale: A Nonnegative Matrix Factorization Approach Author: Jaewon Yang, Jure Leskovec 1 1 Venue: WSDM 2013 Presenter: Yupeng Gu 1 Stanford University 1 Background Community

More information

Multivariate Statistics Fundamentals Part 1: Rotation-based Techniques

Multivariate Statistics Fundamentals Part 1: Rotation-based Techniques Multivariate Statistics Fundamentals Part 1: Rotation-based Techniques A reminded from a univariate statistics courses Population Class of things (What you want to learn about) Sample group representing

More information

MACHINE LEARNING ADVANCED MACHINE LEARNING

MACHINE LEARNING ADVANCED MACHINE LEARNING MACHINE LEARNING ADVANCED MACHINE LEARNING Recap of Important Notions on Estimation of Probability Density Functions 22 MACHINE LEARNING Discrete Probabilities Consider two variables and y taking discrete

More information

Principal Component Analysis

Principal Component Analysis Machine Learning Michaelmas 2017 James Worrell Principal Component Analysis 1 Introduction 1.1 Goals of PCA Principal components analysis (PCA) is a dimensionality reduction technique that can be used

More information

Correspondence Analysis & Related Methods

Correspondence Analysis & Related Methods Correspondence Analysis & Related Methods Michael Greenacre SESSION 9: CA applied to rankings, preferences & paired comparisons Correspondence analysis (CA) can also be applied to other types of data:

More information

Predictive analysis on Multivariate, Time Series datasets using Shapelets

Predictive analysis on Multivariate, Time Series datasets using Shapelets 1 Predictive analysis on Multivariate, Time Series datasets using Shapelets Hemal Thakkar Department of Computer Science, Stanford University hemal@stanford.edu hemal.tt@gmail.com Abstract Multivariate,

More information

A Modified Incremental Principal Component Analysis for On-Line Learning of Feature Space and Classifier

A Modified Incremental Principal Component Analysis for On-Line Learning of Feature Space and Classifier A Modified Incremental Principal Component Analysis for On-Line Learning of Feature Space and Classifier Seiichi Ozawa 1, Shaoning Pang 2, and Nikola Kasabov 2 1 Graduate School of Science and Technology,

More information

Table of Contents. Multivariate methods. Introduction II. Introduction I

Table of Contents. Multivariate methods. Introduction II. Introduction I Table of Contents Introduction Antti Penttilä Department of Physics University of Helsinki Exactum summer school, 04 Construction of multinormal distribution Test of multinormality with 3 Interpretation

More information

STA 4273H: Statistical Machine Learning

STA 4273H: Statistical Machine Learning STA 4273H: Statistical Machine Learning Russ Salakhutdinov Department of Computer Science! Department of Statistical Sciences! rsalakhu@cs.toronto.edu! h0p://www.cs.utoronto.ca/~rsalakhu/ Lecture 7 Approximate

More information

Computer Vision Group Prof. Daniel Cremers. 9. Gaussian Processes - Regression

Computer Vision Group Prof. Daniel Cremers. 9. Gaussian Processes - Regression Group Prof. Daniel Cremers 9. Gaussian Processes - Regression Repetition: Regularized Regression Before, we solved for w using the pseudoinverse. But: we can kernelize this problem as well! First step:

More information

Dimensionality Reduction

Dimensionality Reduction Lecture 5 1 Outline 1. Overview a) What is? b) Why? 2. Principal Component Analysis (PCA) a) Objectives b) Explaining variability c) SVD 3. Related approaches a) ICA b) Autoencoders 2 Example 1: Sportsball

More information

Bayesian Linear Regression

Bayesian Linear Regression Bayesian Linear Regression Sudipto Banerjee 1 Biostatistics, School of Public Health, University of Minnesota, Minneapolis, Minnesota, U.S.A. September 15, 2010 1 Linear regression models: a Bayesian perspective

More information

Probabilistic Low-Rank Matrix Completion with Adaptive Spectral Regularization Algorithms

Probabilistic Low-Rank Matrix Completion with Adaptive Spectral Regularization Algorithms Probabilistic Low-Rank Matrix Completion with Adaptive Spectral Regularization Algorithms Adrien Todeschini Inria Bordeaux JdS 2014, Rennes Aug. 2014 Joint work with François Caron (Univ. Oxford), Marie

More information

ABSTRACT INTRODUCTION

ABSTRACT INTRODUCTION ABSTRACT Presented in this paper is an approach to fault diagnosis based on a unifying review of linear Gaussian models. The unifying review draws together different algorithms such as PCA, factor analysis,

More information

Package LCAextend. August 29, 2013

Package LCAextend. August 29, 2013 Type Package Package LCAextend August 29, 2013 Title Latent Class Analysis (LCA) with familial dependence in extended pedigrees Version 1.2 Date 2012-03-15 Author Arafat TAYEB ,

More information

Algorithmisches Lernen/Machine Learning

Algorithmisches Lernen/Machine Learning Algorithmisches Lernen/Machine Learning Part 1: Stefan Wermter Introduction Connectionist Learning (e.g. Neural Networks) Decision-Trees, Genetic Algorithms Part 2: Norman Hendrich Support-Vector Machines

More information

PCA and admixture models

PCA and admixture models PCA and admixture models CM226: Machine Learning for Bioinformatics. Fall 2016 Sriram Sankararaman Acknowledgments: Fei Sha, Ameet Talwalkar, Alkes Price PCA and admixture models 1 / 57 Announcements HW1

More information

FACTOR ANALYSIS AND MULTIDIMENSIONAL SCALING

FACTOR ANALYSIS AND MULTIDIMENSIONAL SCALING FACTOR ANALYSIS AND MULTIDIMENSIONAL SCALING Vishwanath Mantha Department for Electrical and Computer Engineering Mississippi State University, Mississippi State, MS 39762 mantha@isip.msstate.edu ABSTRACT

More information

Multilevel Statistical Models: 3 rd edition, 2003 Contents

Multilevel Statistical Models: 3 rd edition, 2003 Contents Multilevel Statistical Models: 3 rd edition, 2003 Contents Preface Acknowledgements Notation Two and three level models. A general classification notation and diagram Glossary Chapter 1 An introduction

More information

Fractional Imputation in Survey Sampling: A Comparative Review

Fractional Imputation in Survey Sampling: A Comparative Review Fractional Imputation in Survey Sampling: A Comparative Review Shu Yang Jae-Kwang Kim Iowa State University Joint Statistical Meetings, August 2015 Outline Introduction Fractional imputation Features Numerical

More information

Error Functions & Linear Regression (1)

Error Functions & Linear Regression (1) Error Functions & Linear Regression (1) John Kelleher & Brian Mac Namee Machine Learning @ DIT Overview 1 Introduction Overview 2 Univariate Linear Regression Linear Regression Analytical Solution Gradient

More information

Introduction to Bayesian Statistics and Markov Chain Monte Carlo Estimation. EPSY 905: Multivariate Analysis Spring 2016 Lecture #10: April 6, 2016

Introduction to Bayesian Statistics and Markov Chain Monte Carlo Estimation. EPSY 905: Multivariate Analysis Spring 2016 Lecture #10: April 6, 2016 Introduction to Bayesian Statistics and Markov Chain Monte Carlo Estimation EPSY 905: Multivariate Analysis Spring 2016 Lecture #10: April 6, 2016 EPSY 905: Intro to Bayesian and MCMC Today s Class An

More information

Machine learning comes from Bayesian decision theory in statistics. There we want to minimize the expected value of the loss function.

Machine learning comes from Bayesian decision theory in statistics. There we want to minimize the expected value of the loss function. Bayesian learning: Machine learning comes from Bayesian decision theory in statistics. There we want to minimize the expected value of the loss function. Let y be the true label and y be the predicted

More information

Chapter 5: Models used in conjunction with sampling. J. Kim, W. Fuller (ISU) Chapter 5: Models used in conjunction with sampling 1 / 70

Chapter 5: Models used in conjunction with sampling. J. Kim, W. Fuller (ISU) Chapter 5: Models used in conjunction with sampling 1 / 70 Chapter 5: Models used in conjunction with sampling J. Kim, W. Fuller (ISU) Chapter 5: Models used in conjunction with sampling 1 / 70 Nonresponse Unit Nonresponse: weight adjustment Item Nonresponse:

More information

Principal Component Analysis (PCA) Principal Component Analysis (PCA)

Principal Component Analysis (PCA) Principal Component Analysis (PCA) Recall: Eigenvectors of the Covariance Matrix Covariance matrices are symmetric. Eigenvectors are orthogonal Eigenvectors are ordered by the magnitude of eigenvalues: λ 1 λ 2 λ p {v 1, v 2,..., v n } Recall:

More information

CMSC858P Supervised Learning Methods

CMSC858P Supervised Learning Methods CMSC858P Supervised Learning Methods Hector Corrada Bravo March, 2010 Introduction Today we discuss the classification setting in detail. Our setting is that we observe for each subject i a set of p predictors

More information