A Quantum-inspired version of the Nearest Mean Classifier

Size: px
Start display at page:

Download "A Quantum-inspired version of the Nearest Mean Classifier"

Transcription

1 A Quantum-inspired version of the Nearest Mean Classifier (with Roberto Giuntini and Ranjith Venkatrama) University of Cagliari November 6th-8th, Verona Quantum Techniques in Machine Learning - QTML 2017

2 A foundational question Can Quantum Theory help classical machine learning (by using classical computer only)? A naive example.

3 Training set, Class, Pattern, Feature Nearest Mean Classifier (NMC) Standard (supervised) Classification scenario Let us consider (as a simple example) two disjoint sets A and B of different objects (say cats and dogs). During the training process, we take n objects from the set A and m objects from the set B. Let C a A and C b B. We can measure two (or more) features of each object a i C a and b i C b (for istance the weight and the lenght of the tail). We say that C a and C b are classes and the objects a i and b i are patterns that are characterized by their features. We write, for example, a i = {x 1, x 2 }, where x 1 and x 2 are the weight and the lenght of the tail of the cat a i, respectively.

4 Training set, Class, Pattern, Feature Nearest Mean Classifier (NMC) Nearest Mean Classifier (NMC) Let us consider the classes C a = {a 1,..., a n } and C b = {b 1,..., b m }, with a i and b i belonging to the training set and an arbitrary pattern c i = {x 1, x 2 } belonging to the test set. The goal is to establish whether is more probably that c i A or c i B. We - only - consider the centroids a and b of C a and C b and the euclidean distances Ed(c i, a ) and Ed(c i, b ). Hence, if Ed(c i, a ) Ed(c i, b ) then (is more probabily that) c i B; otherwise c i A.

5 How to encode a Pattern as a Density operator Normalized Trace Distance All we need in order to provide a Quantum representation of NMC are: a suitable encoding from patterns to quantum objects a quantum counterpart for the Euclidean distance a quantum counterpart for the centroid

6 How to encode a Pattern as a Density operator Normalized Trace Distance Encoding. An Example: Stereographic encoding It is possible to map the pattern a = (x, y) onto the surface of a radius one sphere by the stereographic projection: 2x (x, y) ( x 2 + y 2 + 1, 2y x 2 + y 2 + 1, x 2 + y 2 1 x 2 + y ). By placing the Bloch components: 2x r 1 = x 2 +y 2 +1 ; r 2y 2 = x 2 +y 2 +1 ; r 3 = x 2 +y 2 1 we obtain: x 2 +y 2 +1 ρ a = 1 ( ) 1 + r3 r 1 ir 2 = 2 r 1 + ir 2 1 r 3 1 x 2 + y ( ) x 2 + y 2 x iy. x + iy 1

7 How to encode a Pattern as a Density operator Normalized Trace Distance Example Let us consider the pattern a = {1, 3}. Its corresponding Density pattern ρ a, is: ρ a = 1 11 ( ) i 1 + 3i 1 We call ρ a Density Pattern.

8 Basic Notions How to encode a Pattern as a Density operator Normalized Trace Distance Moon Dataset Figure : Classical Patterns Figure : Density Patterns

9 How to encode a Pattern as a Density operator Normalized Trace Distance Another Example: Projective encoding v (x, y) ( x v, y ) ( x, ȳ) v ψ v = x 0 + ȳ 1 ρ v = ψ v ψ v...and many others.

10 How to encode a Pattern as a Density operator Normalized Trace Distance Distance. Preservation of the Order Let a = {x a, y a } b = {x b, y b } and c = {x c, y c } be three arbitrary patterns and let ρ i be the density pattern associated to the pattern i. If Ed(a, b) Ed(b, c) (where Ed is the Euclidian distance), is it possible to define a Quantum distance such that Qd(ρ a, ρ b ) Qd(ρ b, ρ c )?

11 How to encode a Pattern as a Density operator Normalized Trace Distance Normalized Trace Distance Let us consider two patterns a = {x a, y a } and b = {x b, y b }. ( Let ρ a = ra3 r a1 ir a2 2 r a1 + ir a2 1 r a3 associated to a; similarly for b. ) the density pattern 2 Let place K = (1 ra3 and let we define the normalized )(1 r b3 ) trace distance as: K Td(ρ a, ρ b ), where Td is the usual Trace distance. It is straightforward to show that Ed(a, b) = K Td(ρ a, ρ b ).

12 How to encode a Pattern as a Density operator Normalized Trace Distance At this stage let s keep the definition of centroid unchanged. Classification Hence, given a and b as the centroids of C a and C b respectively, if K Td(ρ x, ρ a ) K Td(ρ x, ρ b ) then x B; otherwise x A. Similarly to the classical case.

13 How to encode a Pattern as a Density operator Normalized Trace Distance In this way we have introduced just a translation of the standard NMC in terms of quantum objects (with the pontential implicit benefits carried out by a Quantum Computer). But this approach turns out to be convenient also on a Classical Computer.

14 Quantum Centroid Comparison Now we introduce a new definition of Quantum Centroid that has not any counterpart in the classical framework. Quantum Centroid Given a dataset {P 1,..., P n }, let us consider the respective set of density patterns {ρ 1,..., ρ n }. The Quantum Centroid is defined as: ρ QC = 1 n n ρ i. i=1 S. Gambs, Quantum classification, arxiv: v2.

15 Quantum Centroid Comparison Observations Some observation: The QC ρ QC is not a pure state and it has not any counterpart in the set of classical pattern in R n ; In contrast to the Classical Centroid, the QC is "sensitive" to the distribution of the patterns.

16 Quantum Centroid Comparison We provide a comparison between the NMC and the "quantum" classification process based on Density Patterns and Quantum Centroids by involving different kinds of standard (artificial) datasets on a Classical Computer. We compare the Error E (E = #CorrectClassifiedPatterns #PatternsOfTheDataset ) and the reliability (in terms of the Cohen s constant k) for both classifiers.

17 Quantum Centroid Comparison Gaussian Dataset Gaussian Dataset: 200 Patterns allocated in two Classes. Figure : Gaussian Dataset Figure : NMC Figure : Quantum Classifier

18 Quantum Centroid Comparison Table : Gaussian Dataset E E1 E2 Pr k TPR FPR NMC QC By randomly dividing the dataset in a training set (80%) and in a test set (20%), the average over 100 experiments gives: NMC Error = ± 6.79; Q Error = ± 6.09.

19 Quantum Centroid Comparison A remark Even if the error of the Quantum Classifier is lower than the Error of the NMC, there are some patterns that are correctly classified by the NMC but not by the Quantum Classifier. Hence, it makes sense to consider a "merging" of the NMC and the Quantum Classifier.

20 Basic Notions Quantum Centroid Comparison Gaussian Dataset Figure : Gaussian Dataset Figure : NMC Figure : Quantum Classifier Figure : NMC & Quantum Classifier

21 Quantum Centroid Comparison Table : Gaussian Dataset E E1 E2 Pr k TPR FPR NMC QC NMC-QC

22 Quantum Centroid Comparison Moon Dataset Moon Dataset: 200 patterns allocated in two Classes. Figure : Moon Dataset Figure : NMC Figure : Quantum Classifier Figure : NMC & Quantum Classifier

23 Quantum Centroid Comparison Table : Moon Dataset E E1 E2 Pr k TPR FPR NMC QC By randomly dividing the dataset in a training set (80%) and in a test set (20%), the average over 100 experiments gives: NMC Error = ± 6.32; Q Error = ± 5.46.

24 Basic Notions Quantum Centroid Comparison Banana Dataset Banana Dataset: 5300 patterns; 2376 belonging to the first Class and 2924 to the second Class. Figure : Banana Dataset Figure : NMC Figure : Quantum Classifier Figure : NMC & Quantum Classifier

25 Quantum Centroid Comparison Table : Banana Dataset E E1 E2 Pr k TPR FPR NMC QC NMC-QC By randomly dividing the dataset in a training set (80%) and in a test set (20%), the average over 100 experiments gives: NMC Error = ± 1.74; Q Error = ± 1.21.

26 Basic Notions Quantum Centroid Comparison 3Gaussian Dataset 3Gaussian Dataset: 450 Patterns allocated in three Classes. Figure : 3Gaussian Dataset Figure : NMC Figure : Quantum Classifier Figure : NMC & Quantum Classifier

27 Quantum Centroid Comparison Here we randomly divide the dataset in a Training set (80% of the patterns) and a Test set (20% of the patterns). We calculate the average over 100 runs for each experiments. A full comparison

28 The Quantum Centroid is not invariant under rescaling the "Quantum" Classifier is not invariant under rescaling! Is it an Embarrasment or is it an Asset? The Error is dependent on both the rescaling of the Patterns and the different encoding. We show how the Error changes by changing the rescaling (it can be shown for two different encodings).

29 Rescaling How the Error of the Quantum Classifier changes by ranging the value of the rescaling. If we know, in principle, the kind of distribution of a given phenomena, we can choose ad hoc the suitable rescaling in order to further improve the accuracy of the classification.

30 Summarizing the procedure 1. Start from classical data 2. Suitable rescaling (if we know in principle the kind of distribution) 3. Encoding (from classical patterns to density patterns) 4. Classification (using a classical computer) 5. Decoding (from density patterns to classical patterns) 6. Merging with the classical result Positive answer to the foundational question.

31 The choise of the "best" Encoding (and/or the best Rescaling) is mostly empirical and it is stritcly dependent on the Dataset. A complete theoretical investigation is still missing!!! A comparison with more performant classifiers (Linear Discriminant Analysis, Quadratic Discriminant Analysis...) also in case of unsupervised or semi-supervised classification could be investigated. (The NMC and the Quantum Classifier are based on the concepts of centroid and distance only but more sofisticated classifiers can be replaced in a quantum framework exploiting superposition principle, entanglement, majorization etc...).

32 Basic Notions From artificial to real datasets As further developments, it will be checked whether the Quantum Classifier could bring some benefit for practical implementations, such as Figure : Handwriting Figure : Fingerprint Recognition Figure : Face Recognition Figure : Biomedical Imaging

33 thank you G. Sergioli, E. Santucci, L. Didaci, J.A. Miskczak, R. Giuntini, A Quantum-inspired version of the Nearest Mean Classifier, Soft Computing (2016). G. Sergioli, G.M. Bosyk, E. Santucci, R. Giuntini, A Quantum-inspired version of the classification problem, International Journal of Theoretical Physics (2017). F. Holik, G. Sergioli, H. Freytes, A. Plastino, Pattern Recognition in non-kolmogorovian Structures, Foundation of Science (2017). E. Santucci, G. Sergioli, Classification Problem in a Quantum Framework, Proceedings of the Vaxjo Conference Quantum and Beyond (2017).

Quantum Minimum Distance Classifier

Quantum Minimum Distance Classifier Article Quantum Minimum Distance Classifier Enrica Santucci 1 University of Cagliari, Piazza D Armi snc - 09123 Cagliari (Italy); enrica.santucci@gmail.com 1 2 3 4 5 6 7 8 9 10 11 12 13 14 Abstract: We

More information

Department of Electrical and Electronic Engineering, University of Cagliari, Cagliari 09123, Italy;

Department of Electrical and Electronic Engineering, University of Cagliari, Cagliari 09123, Italy; entropy Article Quantum Minimum Distance Classifier Enrica Santucci Department of Electrical and Electronic Engineering, University of Cagliari, Cagliari 09123, Italy; enrica.santucci@gmail.com Received:

More information

Quantum machine learning for quantum anomaly detection CQT AND SUTD, SINGAPORE ARXIV:

Quantum machine learning for quantum anomaly detection CQT AND SUTD, SINGAPORE ARXIV: Quantum machine learning for quantum anomaly detection NANA LIU CQT AND SUTD, SINGAPORE ARXIV:1710.07405 TUESDAY 7 TH NOVEMBER 2017 QTML 2017, VERONA Types of anomaly detection algorithms Classification-based

More information

Lecture Notes 1: Vector spaces

Lecture Notes 1: Vector spaces Optimization-based data analysis Fall 2017 Lecture Notes 1: Vector spaces In this chapter we review certain basic concepts of linear algebra, highlighting their application to signal processing. 1 Vector

More information

International Workshop on Applied Mathematics & Quantum Information. Cagliari, November 3-4, 2016

International Workshop on Applied Mathematics & Quantum Information. Cagliari, November 3-4, 2016 International Workshop on Applied Mathematics & Quantum Information Cagliari, November 3-4, 2016 Contents 1 INTRODUCTION AND ACKNOWLEDGMENTS 2 2 INVITED SPEAKERS 3 3 ABSTRACTS OF THE CONTRIBUTIONS 4 Majorization

More information

Hidden Markov Models

Hidden Markov Models 10-601 Introduction to Machine Learning Machine Learning Department School of Computer Science Carnegie Mellon University Hidden Markov Models Matt Gormley Lecture 22 April 2, 2018 1 Reminders Homework

More information

ARTIFICIAL NEURAL NETWORKS گروه مطالعاتي 17 بهار 92

ARTIFICIAL NEURAL NETWORKS گروه مطالعاتي 17 بهار 92 ARTIFICIAL NEURAL NETWORKS گروه مطالعاتي 17 بهار 92 BIOLOGICAL INSPIRATIONS Some numbers The human brain contains about 10 billion nerve cells (neurons) Each neuron is connected to the others through 10000

More information

Classification of handwritten digits using supervised locally linear embedding algorithm and support vector machine

Classification of handwritten digits using supervised locally linear embedding algorithm and support vector machine Classification of handwritten digits using supervised locally linear embedding algorithm and support vector machine Olga Kouropteva, Oleg Okun, Matti Pietikäinen Machine Vision Group, Infotech Oulu and

More information

Generative Learning. INFO-4604, Applied Machine Learning University of Colorado Boulder. November 29, 2018 Prof. Michael Paul

Generative Learning. INFO-4604, Applied Machine Learning University of Colorado Boulder. November 29, 2018 Prof. Michael Paul Generative Learning INFO-4604, Applied Machine Learning University of Colorado Boulder November 29, 2018 Prof. Michael Paul Generative vs Discriminative The classification algorithms we have seen so far

More information

Intro. ANN & Fuzzy Systems. Lecture 15. Pattern Classification (I): Statistical Formulation

Intro. ANN & Fuzzy Systems. Lecture 15. Pattern Classification (I): Statistical Formulation Lecture 15. Pattern Classification (I): Statistical Formulation Outline Statistical Pattern Recognition Maximum Posterior Probability (MAP) Classifier Maximum Likelihood (ML) Classifier K-Nearest Neighbor

More information

arxiv: v2 [quant-ph] 19 Jul 2018

arxiv: v2 [quant-ph] 19 Jul 2018 Multiqubit and multilevel quantum reinforcement learning with quantum technologies F. A. Cárdenas-López,2,*, L. Lamata 3, J. C. Retamal,2, E. Solano 3,4,5 arxiv:709.07848v2 [quant-ph] 9 Jul 208 Departamento

More information

Feature selection and extraction Spectral domain quality estimation Alternatives

Feature selection and extraction Spectral domain quality estimation Alternatives Feature selection and extraction Error estimation Maa-57.3210 Data Classification and Modelling in Remote Sensing Markus Törmä markus.torma@tkk.fi Measurements Preprocessing: Remove random and systematic

More information

Discriminant analysis and supervised classification

Discriminant analysis and supervised classification Discriminant analysis and supervised classification Angela Montanari 1 Linear discriminant analysis Linear discriminant analysis (LDA) also known as Fisher s linear discriminant analysis or as Canonical

More information

Introduction to Machine Learning

Introduction to Machine Learning Introduction to Machine Learning CS4731 Dr. Mihail Fall 2017 Slide content based on books by Bishop and Barber. https://www.microsoft.com/en-us/research/people/cmbishop/ http://web4.cs.ucl.ac.uk/staff/d.barber/pmwiki/pmwiki.php?n=brml.homepage

More information

Lecture 4 Discriminant Analysis, k-nearest Neighbors

Lecture 4 Discriminant Analysis, k-nearest Neighbors Lecture 4 Discriminant Analysis, k-nearest Neighbors Fredrik Lindsten Division of Systems and Control Department of Information Technology Uppsala University. Email: fredrik.lindsten@it.uu.se fredrik.lindsten@it.uu.se

More information

Expectation Maximization (EM)

Expectation Maximization (EM) Expectation Maximization (EM) The Expectation Maximization (EM) algorithm is one approach to unsupervised, semi-supervised, or lightly supervised learning. In this kind of learning either no labels are

More information

Linear classifiers Lecture 3

Linear classifiers Lecture 3 Linear classifiers Lecture 3 David Sontag New York University Slides adapted from Luke Zettlemoyer, Vibhav Gogate, and Carlos Guestrin ML Methodology Data: labeled instances, e.g. emails marked spam/ham

More information

Gaussian Models

Gaussian Models Gaussian Models ddebarr@uw.edu 2016-04-28 Agenda Introduction Gaussian Discriminant Analysis Inference Linear Gaussian Systems The Wishart Distribution Inferring Parameters Introduction Gaussian Density

More information

Linear Classifiers as Pattern Detectors

Linear Classifiers as Pattern Detectors Intelligent Systems: Reasoning and Recognition James L. Crowley ENSIMAG 2 / MoSIG M1 Second Semester 2014/2015 Lesson 16 8 April 2015 Contents Linear Classifiers as Pattern Detectors Notation...2 Linear

More information

Motivating the Covariance Matrix

Motivating the Covariance Matrix Motivating the Covariance Matrix Raúl Rojas Computer Science Department Freie Universität Berlin January 2009 Abstract This note reviews some interesting properties of the covariance matrix and its role

More information

Pattern Recognition. Parameter Estimation of Probability Density Functions

Pattern Recognition. Parameter Estimation of Probability Density Functions Pattern Recognition Parameter Estimation of Probability Density Functions Classification Problem (Review) The classification problem is to assign an arbitrary feature vector x F to one of c classes. The

More information

Machine Learning for Signal Processing Bayes Classification and Regression

Machine Learning for Signal Processing Bayes Classification and Regression Machine Learning for Signal Processing Bayes Classification and Regression Instructor: Bhiksha Raj 11755/18797 1 Recap: KNN A very effective and simple way of performing classification Simple model: For

More information

MULTIVARIATE PATTERN RECOGNITION FOR CHEMOMETRICS. Richard Brereton

MULTIVARIATE PATTERN RECOGNITION FOR CHEMOMETRICS. Richard Brereton MULTIVARIATE PATTERN RECOGNITION FOR CHEMOMETRICS Richard Brereton r.g.brereton@bris.ac.uk Pattern Recognition Book Chemometrics for Pattern Recognition, Wiley, 2009 Pattern Recognition Pattern Recognition

More information

Topics in Natural Language Processing

Topics in Natural Language Processing Topics in Natural Language Processing Shay Cohen Institute for Language, Cognition and Computation University of Edinburgh Lecture 9 Administrativia Next class will be a summary Please email me questions

More information

Clustering VS Classification

Clustering VS Classification MCQ Clustering VS Classification 1. What is the relation between the distance between clusters and the corresponding class discriminability? a. proportional b. inversely-proportional c. no-relation Ans:

More information

SUPERVISED LEARNING: INTRODUCTION TO CLASSIFICATION

SUPERVISED LEARNING: INTRODUCTION TO CLASSIFICATION SUPERVISED LEARNING: INTRODUCTION TO CLASSIFICATION 1 Outline Basic terminology Features Training and validation Model selection Error and loss measures Statistical comparison Evaluation measures 2 Terminology

More information

Principles of Pattern Recognition. C. A. Murthy Machine Intelligence Unit Indian Statistical Institute Kolkata

Principles of Pattern Recognition. C. A. Murthy Machine Intelligence Unit Indian Statistical Institute Kolkata Principles of Pattern Recognition C. A. Murthy Machine Intelligence Unit Indian Statistical Institute Kolkata e-mail: murthy@isical.ac.in Pattern Recognition Measurement Space > Feature Space >Decision

More information

Introduction to Signal Detection and Classification. Phani Chavali

Introduction to Signal Detection and Classification. Phani Chavali Introduction to Signal Detection and Classification Phani Chavali Outline Detection Problem Performance Measures Receiver Operating Characteristics (ROC) F-Test - Test Linear Discriminant Analysis (LDA)

More information

Support Vector Machines. CAP 5610: Machine Learning Instructor: Guo-Jun QI

Support Vector Machines. CAP 5610: Machine Learning Instructor: Guo-Jun QI Support Vector Machines CAP 5610: Machine Learning Instructor: Guo-Jun QI 1 Linear Classifier Naive Bayes Assume each attribute is drawn from Gaussian distribution with the same variance Generative model:

More information

Pattern Recognition and Machine Learning. Learning and Evaluation of Pattern Recognition Processes

Pattern Recognition and Machine Learning. Learning and Evaluation of Pattern Recognition Processes Pattern Recognition and Machine Learning James L. Crowley ENSIMAG 3 - MMIS Fall Semester 2016 Lesson 1 5 October 2016 Learning and Evaluation of Pattern Recognition Processes Outline Notation...2 1. The

More information

Multiscale Autoconvolution Histograms for Affine Invariant Pattern Recognition

Multiscale Autoconvolution Histograms for Affine Invariant Pattern Recognition Multiscale Autoconvolution Histograms for Affine Invariant Pattern Recognition Esa Rahtu Mikko Salo Janne Heikkilä Department of Electrical and Information Engineering P.O. Box 4500, 90014 University of

More information

Advanced Introduction to Machine Learning

Advanced Introduction to Machine Learning 10-715 Advanced Introduction to Machine Learning Homework Due Oct 15, 10.30 am Rules Please follow these guidelines. Failure to do so, will result in loss of credit. 1. Homework is due on the due date

More information

Towards a Ptolemaic Model for OCR

Towards a Ptolemaic Model for OCR Towards a Ptolemaic Model for OCR Sriharsha Veeramachaneni and George Nagy Rensselaer Polytechnic Institute, Troy, NY, USA E-mail: nagy@ecse.rpi.edu Abstract In style-constrained classification often there

More information

Machine Learning. CUNY Graduate Center, Spring Lectures 11-12: Unsupervised Learning 1. Professor Liang Huang.

Machine Learning. CUNY Graduate Center, Spring Lectures 11-12: Unsupervised Learning 1. Professor Liang Huang. Machine Learning CUNY Graduate Center, Spring 2013 Lectures 11-12: Unsupervised Learning 1 (Clustering: k-means, EM, mixture models) Professor Liang Huang huang@cs.qc.cuny.edu http://acl.cs.qc.edu/~lhuang/teaching/machine-learning

More information

CS229 Machine Learning Project: Allocating funds to the right people

CS229 Machine Learning Project: Allocating funds to the right people CS229 Machine Learning Project: Allocating funds to the right people Introduction Dugalic, Adem adugalic@stanford.edu Nguyen, Tram tram@stanford.edu Our project is motivated by a common issue in developing

More information

Hypothesis Evaluation

Hypothesis Evaluation Hypothesis Evaluation Machine Learning Hamid Beigy Sharif University of Technology Fall 1395 Hamid Beigy (Sharif University of Technology) Hypothesis Evaluation Fall 1395 1 / 31 Table of contents 1 Introduction

More information

CS 188: Artificial Intelligence Spring Announcements

CS 188: Artificial Intelligence Spring Announcements CS 188: Artificial Intelligence Spring 2010 Lecture 22: Nearest Neighbors, Kernels 4/18/2011 Pieter Abbeel UC Berkeley Slides adapted from Dan Klein Announcements On-going: contest (optional and FUN!)

More information

Discriminative Direction for Kernel Classifiers

Discriminative Direction for Kernel Classifiers Discriminative Direction for Kernel Classifiers Polina Golland Artificial Intelligence Lab Massachusetts Institute of Technology Cambridge, MA 02139 polina@ai.mit.edu Abstract In many scientific and engineering

More information

An Overview of Outlier Detection Techniques and Applications

An Overview of Outlier Detection Techniques and Applications Machine Learning Rhein-Neckar Meetup An Overview of Outlier Detection Techniques and Applications Ying Gu connygy@gmail.com 28.02.2016 Anomaly/Outlier Detection What are anomalies/outliers? The set of

More information

Linear Classifiers as Pattern Detectors

Linear Classifiers as Pattern Detectors Intelligent Systems: Reasoning and Recognition James L. Crowley ENSIMAG 2 / MoSIG M1 Second Semester 2013/2014 Lesson 18 23 April 2014 Contents Linear Classifiers as Pattern Detectors Notation...2 Linear

More information

Data Preprocessing. Cluster Similarity

Data Preprocessing. Cluster Similarity 1 Cluster Similarity Similarity is most often measured with the help of a distance function. The smaller the distance, the more similar the data objects (points). A function d: M M R is a distance on M

More information

Performance Comparison of K-Means and Expectation Maximization with Gaussian Mixture Models for Clustering EE6540 Final Project

Performance Comparison of K-Means and Expectation Maximization with Gaussian Mixture Models for Clustering EE6540 Final Project Performance Comparison of K-Means and Expectation Maximization with Gaussian Mixture Models for Clustering EE6540 Final Project Devin Cornell & Sushruth Sastry May 2015 1 Abstract In this article, we explore

More information

Supervised locally linear embedding

Supervised locally linear embedding Supervised locally linear embedding Dick de Ridder 1, Olga Kouropteva 2, Oleg Okun 2, Matti Pietikäinen 2 and Robert P.W. Duin 1 1 Pattern Recognition Group, Department of Imaging Science and Technology,

More information

Machine Learning and Data Mining. Bayes Classifiers. Prof. Alexander Ihler

Machine Learning and Data Mining. Bayes Classifiers. Prof. Alexander Ihler + Machine Learning and Data Mining Bayes Classifiers Prof. Alexander Ihler A basic classifier Training data D={x (i),y (i) }, Classifier f(x ; D) Discrete feature vector x f(x ; D) is a con@ngency table

More information

arxiv: v2 [quant-ph] 2 Sep 2008

arxiv: v2 [quant-ph] 2 Sep 2008 Quantum classification arxiv:0809.0444v2 [quant-ph] 2 Sep 2008 Sébastien Gambs Département d informatique et de recherche opérationnelle Université de Montréal C.P. 6128, Succ. Centre-Ville, Montréal (Québec),

More information

Lecture 6: Quantum error correction and quantum capacity

Lecture 6: Quantum error correction and quantum capacity Lecture 6: Quantum error correction and quantum capacity Mark M. Wilde The quantum capacity theorem is one of the most important theorems in quantum hannon theory. It is a fundamentally quantum theorem

More information

Interpreting Deep Classifiers

Interpreting Deep Classifiers Ruprecht-Karls-University Heidelberg Faculty of Mathematics and Computer Science Seminar: Explainable Machine Learning Interpreting Deep Classifiers by Visual Distillation of Dark Knowledge Author: Daniela

More information

ISyE 6416: Computational Statistics Spring Lecture 5: Discriminant analysis and classification

ISyE 6416: Computational Statistics Spring Lecture 5: Discriminant analysis and classification ISyE 6416: Computational Statistics Spring 2017 Lecture 5: Discriminant analysis and classification Prof. Yao Xie H. Milton Stewart School of Industrial and Systems Engineering Georgia Institute of Technology

More information

Announcements. CS 188: Artificial Intelligence Spring Classification. Today. Classification overview. Case-Based Reasoning

Announcements. CS 188: Artificial Intelligence Spring Classification. Today. Classification overview. Case-Based Reasoning CS 188: Artificial Intelligence Spring 21 Lecture 22: Nearest Neighbors, Kernels 4/18/211 Pieter Abbeel UC Berkeley Slides adapted from Dan Klein Announcements On-going: contest (optional and FUN!) Remaining

More information

Face Recognition Using Laplacianfaces He et al. (IEEE Trans PAMI, 2005) presented by Hassan A. Kingravi

Face Recognition Using Laplacianfaces He et al. (IEEE Trans PAMI, 2005) presented by Hassan A. Kingravi Face Recognition Using Laplacianfaces He et al. (IEEE Trans PAMI, 2005) presented by Hassan A. Kingravi Overview Introduction Linear Methods for Dimensionality Reduction Nonlinear Methods and Manifold

More information

Data Mining. Practical Machine Learning Tools and Techniques. Slides for Chapter 4 of Data Mining by I. H. Witten, E. Frank and M. A.

Data Mining. Practical Machine Learning Tools and Techniques. Slides for Chapter 4 of Data Mining by I. H. Witten, E. Frank and M. A. Data Mining Practical Machine Learning Tools and Techniques Slides for Chapter of Data Mining by I. H. Witten, E. Frank and M. A. Hall Statistical modeling Opposite of R: use all the attributes Two assumptions:

More information

Linear & nonlinear classifiers

Linear & nonlinear classifiers Linear & nonlinear classifiers Machine Learning Hamid Beigy Sharif University of Technology Fall 1394 Hamid Beigy (Sharif University of Technology) Linear & nonlinear classifiers Fall 1394 1 / 34 Table

More information

9/26/17. Ridge regression. What our model needs to do. Ridge Regression: L2 penalty. Ridge coefficients. Ridge coefficients

9/26/17. Ridge regression. What our model needs to do. Ridge Regression: L2 penalty. Ridge coefficients. Ridge coefficients What our model needs to do regression Usually, we are not just trying to explain observed data We want to uncover meaningful trends And predict future observations Our questions then are Is β" a good estimate

More information

Linear, threshold units. Linear Discriminant Functions and Support Vector Machines. Biometrics CSE 190 Lecture 11. X i : inputs W i : weights

Linear, threshold units. Linear Discriminant Functions and Support Vector Machines. Biometrics CSE 190 Lecture 11. X i : inputs W i : weights Linear Discriminant Functions and Support Vector Machines Linear, threshold units CSE19, Winter 11 Biometrics CSE 19 Lecture 11 1 X i : inputs W i : weights θ : threshold 3 4 5 1 6 7 Courtesy of University

More information

Pattern recognition. "To understand is to perceive patterns" Sir Isaiah Berlin, Russian philosopher

Pattern recognition. To understand is to perceive patterns Sir Isaiah Berlin, Russian philosopher Pattern recognition "To understand is to perceive patterns" Sir Isaiah Berlin, Russian philosopher The more relevant patterns at your disposal, the better your decisions will be. This is hopeful news to

More information

Hidden Markov Models

Hidden Markov Models 10-601 Introduction to Machine Learning Machine Learning Department School of Computer Science Carnegie Mellon University Hidden Markov Models Matt Gormley Lecture 19 Nov. 5, 2018 1 Reminders Homework

More information

Part I. Linear Discriminant Analysis. Discriminant analysis. Discriminant analysis

Part I. Linear Discriminant Analysis. Discriminant analysis. Discriminant analysis Week 5 Based in part on slides from textbook, slides of Susan Holmes Part I Linear Discriminant Analysis October 29, 2012 1 / 1 2 / 1 Nearest centroid rule Suppose we break down our data matrix as by the

More information

Machine Learning and Deep Learning! Vincent Lepetit!

Machine Learning and Deep Learning! Vincent Lepetit! Machine Learning and Deep Learning!! Vincent Lepetit! 1! What is Machine Learning?! 2! Hand-Written Digit Recognition! 2 9 3! Hand-Written Digit Recognition! Formalization! 0 1 x = @ A Images are 28x28

More information

Beyond the Point Cloud: From Transductive to Semi-Supervised Learning

Beyond the Point Cloud: From Transductive to Semi-Supervised Learning Beyond the Point Cloud: From Transductive to Semi-Supervised Learning Vikas Sindhwani, Partha Niyogi, Mikhail Belkin Andrew B. Goldberg goldberg@cs.wisc.edu Department of Computer Sciences University of

More information

Linear Discriminant Analysis Based in part on slides from textbook, slides of Susan Holmes. November 9, Statistics 202: Data Mining

Linear Discriminant Analysis Based in part on slides from textbook, slides of Susan Holmes. November 9, Statistics 202: Data Mining Linear Discriminant Analysis Based in part on slides from textbook, slides of Susan Holmes November 9, 2012 1 / 1 Nearest centroid rule Suppose we break down our data matrix as by the labels yielding (X

More information

Time Series Classification

Time Series Classification Distance Measures Classifiers DTW vs. ED Further Work Questions August 31, 2017 Distance Measures Classifiers DTW vs. ED Further Work Questions Outline 1 2 Distance Measures 3 Classifiers 4 DTW vs. ED

More information

Introduction to Machine Learning. PCA and Spectral Clustering. Introduction to Machine Learning, Slides: Eran Halperin

Introduction to Machine Learning. PCA and Spectral Clustering. Introduction to Machine Learning, Slides: Eran Halperin 1 Introduction to Machine Learning PCA and Spectral Clustering Introduction to Machine Learning, 2013-14 Slides: Eran Halperin Singular Value Decomposition (SVD) The singular value decomposition (SVD)

More information

Machine Learning And Applications: Supervised Learning-SVM

Machine Learning And Applications: Supervised Learning-SVM Machine Learning And Applications: Supervised Learning-SVM Raphaël Bournhonesque École Normale Supérieure de Lyon, Lyon, France raphael.bournhonesque@ens-lyon.fr 1 Supervised vs unsupervised learning Machine

More information

Performance Evaluation and Hypothesis Testing

Performance Evaluation and Hypothesis Testing Performance Evaluation and Hypothesis Testing 1 Motivation Evaluating the performance of learning systems is important because: Learning systems are usually designed to predict the class of future unlabeled

More information

Notation. Pattern Recognition II. Michal Haindl. Outline - PR Basic Concepts. Pattern Recognition Notions

Notation. Pattern Recognition II. Michal Haindl. Outline - PR Basic Concepts. Pattern Recognition Notions Notation S pattern space X feature vector X = [x 1,...,x l ] l = dim{x} number of features X feature space K number of classes ω i class indicator Ω = {ω 1,...,ω K } g(x) discriminant function H decision

More information

Classifier Complexity and Support Vector Classifiers

Classifier Complexity and Support Vector Classifiers Classifier Complexity and Support Vector Classifiers Feature 2 6 4 2 0 2 4 6 8 RBF kernel 10 10 8 6 4 2 0 2 4 6 Feature 1 David M.J. Tax Pattern Recognition Laboratory Delft University of Technology D.M.J.Tax@tudelft.nl

More information

L11: Pattern recognition principles

L11: Pattern recognition principles L11: Pattern recognition principles Bayesian decision theory Statistical classifiers Dimensionality reduction Clustering This lecture is partly based on [Huang, Acero and Hon, 2001, ch. 4] Introduction

More information

Kaviti Sai Saurab. October 23,2014. IIT Kanpur. Kaviti Sai Saurab (UCLA) Quantum methods in ML October 23, / 8

Kaviti Sai Saurab. October 23,2014. IIT Kanpur. Kaviti Sai Saurab (UCLA) Quantum methods in ML October 23, / 8 A review of Quantum Algorithms for Nearest-Neighbor Methods for Supervised and Unsupervised Learning Nathan Wiebe,Ashish Kapoor,Krysta M. Svore Jan.2014 Kaviti Sai Saurab IIT Kanpur October 23,2014 Kaviti

More information

day month year documentname/initials 1

day month year documentname/initials 1 ECE471-571 Pattern Recognition Lecture 13 Decision Tree Hairong Qi, Gonzalez Family Professor Electrical Engineering and Computer Science University of Tennessee, Knoxville http://www.eecs.utk.edu/faculty/qi

More information

Linear models: the perceptron and closest centroid algorithms. D = {(x i,y i )} n i=1. x i 2 R d 9/3/13. Preliminaries. Chapter 1, 7.

Linear models: the perceptron and closest centroid algorithms. D = {(x i,y i )} n i=1. x i 2 R d 9/3/13. Preliminaries. Chapter 1, 7. Preliminaries Linear models: the perceptron and closest centroid algorithms Chapter 1, 7 Definition: The Euclidean dot product beteen to vectors is the expression d T x = i x i The dot product is also

More information

An introduction to clustering techniques

An introduction to clustering techniques - ABSTRACT Cluster analysis has been used in a wide variety of fields, such as marketing, social science, biology, pattern recognition etc. It is used to identify homogenous groups of cases to better understand

More information

c 4, < y 2, 1 0, otherwise,

c 4, < y 2, 1 0, otherwise, Fundamentals of Big Data Analytics Univ.-Prof. Dr. rer. nat. Rudolf Mathar Problem. Probability theory: The outcome of an experiment is described by three events A, B and C. The probabilities Pr(A) =,

More information

Contents Lecture 4. Lecture 4 Linear Discriminant Analysis. Summary of Lecture 3 (II/II) Summary of Lecture 3 (I/II)

Contents Lecture 4. Lecture 4 Linear Discriminant Analysis. Summary of Lecture 3 (II/II) Summary of Lecture 3 (I/II) Contents Lecture Lecture Linear Discriminant Analysis Fredrik Lindsten Division of Systems and Control Department of Information Technology Uppsala University Email: fredriklindsten@ituuse Summary of lecture

More information

AN INTRODUCTION TO NEURAL NETWORKS. Scott Kuindersma November 12, 2009

AN INTRODUCTION TO NEURAL NETWORKS. Scott Kuindersma November 12, 2009 AN INTRODUCTION TO NEURAL NETWORKS Scott Kuindersma November 12, 2009 SUPERVISED LEARNING We are given some training data: We must learn a function If y is discrete, we call it classification If it is

More information

Kernel Methods. Barnabás Póczos

Kernel Methods. Barnabás Póczos Kernel Methods Barnabás Póczos Outline Quick Introduction Feature space Perceptron in the feature space Kernels Mercer s theorem Finite domain Arbitrary domain Kernel families Constructing new kernels

More information

Pointwise Exact Bootstrap Distributions of Cost Curves

Pointwise Exact Bootstrap Distributions of Cost Curves Pointwise Exact Bootstrap Distributions of Cost Curves Charles Dugas and David Gadoury University of Montréal 25th ICML Helsinki July 2008 Dugas, Gadoury (U Montréal) Cost curves July 8, 2008 1 / 24 Outline

More information

SVM TRADE-OFF BETWEEN MAXIMIZE THE MARGIN AND MINIMIZE THE VARIABLES USED FOR REGRESSION

SVM TRADE-OFF BETWEEN MAXIMIZE THE MARGIN AND MINIMIZE THE VARIABLES USED FOR REGRESSION International Journal of Pure and Applied Mathematics Volume 87 No. 6 2013, 741-750 ISSN: 1311-8080 (printed version); ISSN: 1314-3395 (on-line version) url: http://www.ijpam.eu doi: http://dx.doi.org/10.12732/ijpam.v87i6.2

More information

Jae-Bong Lee 1 and Bernard A. Megrey 2. International Symposium on Climate Change Effects on Fish and Fisheries

Jae-Bong Lee 1 and Bernard A. Megrey 2. International Symposium on Climate Change Effects on Fish and Fisheries International Symposium on Climate Change Effects on Fish and Fisheries On the utility of self-organizing maps (SOM) and k-means clustering to characterize and compare low frequency spatial and temporal

More information

Machine Learning. Regression-Based Classification & Gaussian Discriminant Analysis. Manfred Huber

Machine Learning. Regression-Based Classification & Gaussian Discriminant Analysis. Manfred Huber Machine Learning Regression-Based Classification & Gaussian Discriminant Analysis Manfred Huber 2015 1 Logistic Regression Linear regression provides a nice representation and an efficient solution to

More information

Introduction to Machine Learning Midterm Exam

Introduction to Machine Learning Midterm Exam 10-701 Introduction to Machine Learning Midterm Exam Instructors: Eric Xing, Ziv Bar-Joseph 17 November, 2015 There are 11 questions, for a total of 100 points. This exam is open book, open notes, but

More information

Statistical Machine Learning

Statistical Machine Learning Statistical Machine Learning Christoph Lampert Spring Semester 2015/2016 // Lecture 12 1 / 36 Unsupervised Learning Dimensionality Reduction 2 / 36 Dimensionality Reduction Given: data X = {x 1,..., x

More information

Mixture of Gaussians Models

Mixture of Gaussians Models Mixture of Gaussians Models Outline Inference, Learning, and Maximum Likelihood Why Mixtures? Why Gaussians? Building up to the Mixture of Gaussians Single Gaussians Fully-Observed Mixtures Hidden Mixtures

More information

Introduction to Machine Learning Spring 2018 Note 18

Introduction to Machine Learning Spring 2018 Note 18 CS 189 Introduction to Machine Learning Spring 2018 Note 18 1 Gaussian Discriminant Analysis Recall the idea of generative models: we classify an arbitrary datapoint x with the class label that maximizes

More information

ECE662: Pattern Recognition and Decision Making Processes: HW TWO

ECE662: Pattern Recognition and Decision Making Processes: HW TWO ECE662: Pattern Recognition and Decision Making Processes: HW TWO Purdue University Department of Electrical and Computer Engineering West Lafayette, INDIANA, USA Abstract. In this report experiments are

More information

On Classification: An Empirical Study of Existing Algorithms Based on Two Kaggle Competitions

On Classification: An Empirical Study of Existing Algorithms Based on Two Kaggle Competitions On Classification: An Empirical Study of Existing Algorithms Based on Two Kaggle Competitions Wilson Florero-Salinas (Team Leader), Sha Li (Team Leader) Xiaoyan Chong, Dan Li, Minglu Ma, Abhirupa Sen,

More information

Statistical Pattern Recognition

Statistical Pattern Recognition Statistical Pattern Recognition Feature Extraction Hamid R. Rabiee Jafar Muhammadi, Alireza Ghasemi, Payam Siyari Spring 2014 http://ce.sharif.edu/courses/92-93/2/ce725-2/ Agenda Dimensionality Reduction

More information

Classification Methods II: Linear and Quadratic Discrimminant Analysis

Classification Methods II: Linear and Quadratic Discrimminant Analysis Classification Methods II: Linear and Quadratic Discrimminant Analysis Rebecca C. Steorts, Duke University STA 325, Chapter 4 ISL Agenda Linear Discrimminant Analysis (LDA) Classification Recall that linear

More information

April 18, :46 WSPC/INSTRUCTION FILE ForVaxjorev15.2dc ABSTRACT QUANTUM COMPUTING MACHINES AND QUANTUM COMPUTATIONAL LOGICS

April 18, :46 WSPC/INSTRUCTION FILE ForVaxjorev15.2dc ABSTRACT QUANTUM COMPUTING MACHINES AND QUANTUM COMPUTATIONAL LOGICS International Journal of Quantum Information c World Scientific Publishing Company ABSTRACT QUANTUM COMPUTING MACHINES AND QUANTUM COMPUTATIONAL LOGICS MARIA LUISA DALLA CHIARA Dipartimento di Lettere

More information

Perceptron Revisited: Linear Separators. Support Vector Machines

Perceptron Revisited: Linear Separators. Support Vector Machines Support Vector Machines Perceptron Revisited: Linear Separators Binary classification can be viewed as the task of separating classes in feature space: w T x + b > 0 w T x + b = 0 w T x + b < 0 Department

More information

A Bias Correction for the Minimum Error Rate in Cross-validation

A Bias Correction for the Minimum Error Rate in Cross-validation A Bias Correction for the Minimum Error Rate in Cross-validation Ryan J. Tibshirani Robert Tibshirani Abstract Tuning parameters in supervised learning problems are often estimated by cross-validation.

More information

CS 188: Artificial Intelligence Spring Announcements

CS 188: Artificial Intelligence Spring Announcements CS 188: Artificial Intelligence Spring 2010 Lecture 24: Perceptrons and More! 4/22/2010 Pieter Abbeel UC Berkeley Slides adapted from Dan Klein Announcements W7 due tonight [this is your last written for

More information

MIDTERM: CS 6375 INSTRUCTOR: VIBHAV GOGATE October,

MIDTERM: CS 6375 INSTRUCTOR: VIBHAV GOGATE October, MIDTERM: CS 6375 INSTRUCTOR: VIBHAV GOGATE October, 23 2013 The exam is closed book. You are allowed a one-page cheat sheet. Answer the questions in the spaces provided on the question sheets. If you run

More information

Quantum Cloning WOOTTERS-ZUREK CLONER

Quantum Cloning WOOTTERS-ZUREK CLONER Quantum Cloning Quantum cloning has been a topic of considerable interest for many years. It turns out to be quantum limit for copying an input state and is closely related to linear amplification when

More information

Machine Learning (CS 567) Lecture 5

Machine Learning (CS 567) Lecture 5 Machine Learning (CS 567) Lecture 5 Time: T-Th 5:00pm - 6:20pm Location: GFS 118 Instructor: Sofus A. Macskassy (macskass@usc.edu) Office: SAL 216 Office hours: by appointment Teaching assistant: Cheol

More information

Representing fuzzy structures in quantum computation with mixed states

Representing fuzzy structures in quantum computation with mixed states Representing fuzzy structures in quantum computation with mixed states Hector Freytes Fellow of CONICET-Argentina Email: hfreytes@gmailcom Antonio Arico Department of Mathematics Viale Merello 9, 913 Caglari

More information

Statistical learning theory, Support vector machines, and Bioinformatics

Statistical learning theory, Support vector machines, and Bioinformatics 1 Statistical learning theory, Support vector machines, and Bioinformatics Jean-Philippe.Vert@mines.org Ecole des Mines de Paris Computational Biology group ENS Paris, november 25, 2003. 2 Overview 1.

More information

CS 340 Lec. 18: Multivariate Gaussian Distributions and Linear Discriminant Analysis

CS 340 Lec. 18: Multivariate Gaussian Distributions and Linear Discriminant Analysis CS 3 Lec. 18: Multivariate Gaussian Distributions and Linear Discriminant Analysis AD March 11 AD ( March 11 1 / 17 Multivariate Gaussian Consider data { x i } N i=1 where xi R D and we assume they are

More information

Selection of Classifiers based on Multiple Classifier Behaviour

Selection of Classifiers based on Multiple Classifier Behaviour Selection of Classifiers based on Multiple Classifier Behaviour Giorgio Giacinto, Fabio Roli, and Giorgio Fumera Dept. of Electrical and Electronic Eng. - University of Cagliari Piazza d Armi, 09123 Cagliari,

More information

Learning Kernel Parameters by using Class Separability Measure

Learning Kernel Parameters by using Class Separability Measure Learning Kernel Parameters by using Class Separability Measure Lei Wang, Kap Luk Chan School of Electrical and Electronic Engineering Nanyang Technological University Singapore, 3979 E-mail: P 3733@ntu.edu.sg,eklchan@ntu.edu.sg

More information

Clustering K-means. Clustering images. Machine Learning CSE546 Carlos Guestrin University of Washington. November 4, 2014.

Clustering K-means. Clustering images. Machine Learning CSE546 Carlos Guestrin University of Washington. November 4, 2014. Clustering K-means Machine Learning CSE546 Carlos Guestrin University of Washington November 4, 2014 1 Clustering images Set of Images [Goldberger et al.] 2 1 K-means Randomly initialize k centers µ (0)

More information