A Quantum-inspired version of the Nearest Mean Classifier
|
|
- Wilfrid Kennedy
- 5 years ago
- Views:
Transcription
1 A Quantum-inspired version of the Nearest Mean Classifier (with Roberto Giuntini and Ranjith Venkatrama) University of Cagliari November 6th-8th, Verona Quantum Techniques in Machine Learning - QTML 2017
2 A foundational question Can Quantum Theory help classical machine learning (by using classical computer only)? A naive example.
3 Training set, Class, Pattern, Feature Nearest Mean Classifier (NMC) Standard (supervised) Classification scenario Let us consider (as a simple example) two disjoint sets A and B of different objects (say cats and dogs). During the training process, we take n objects from the set A and m objects from the set B. Let C a A and C b B. We can measure two (or more) features of each object a i C a and b i C b (for istance the weight and the lenght of the tail). We say that C a and C b are classes and the objects a i and b i are patterns that are characterized by their features. We write, for example, a i = {x 1, x 2 }, where x 1 and x 2 are the weight and the lenght of the tail of the cat a i, respectively.
4 Training set, Class, Pattern, Feature Nearest Mean Classifier (NMC) Nearest Mean Classifier (NMC) Let us consider the classes C a = {a 1,..., a n } and C b = {b 1,..., b m }, with a i and b i belonging to the training set and an arbitrary pattern c i = {x 1, x 2 } belonging to the test set. The goal is to establish whether is more probably that c i A or c i B. We - only - consider the centroids a and b of C a and C b and the euclidean distances Ed(c i, a ) and Ed(c i, b ). Hence, if Ed(c i, a ) Ed(c i, b ) then (is more probabily that) c i B; otherwise c i A.
5 How to encode a Pattern as a Density operator Normalized Trace Distance All we need in order to provide a Quantum representation of NMC are: a suitable encoding from patterns to quantum objects a quantum counterpart for the Euclidean distance a quantum counterpart for the centroid
6 How to encode a Pattern as a Density operator Normalized Trace Distance Encoding. An Example: Stereographic encoding It is possible to map the pattern a = (x, y) onto the surface of a radius one sphere by the stereographic projection: 2x (x, y) ( x 2 + y 2 + 1, 2y x 2 + y 2 + 1, x 2 + y 2 1 x 2 + y ). By placing the Bloch components: 2x r 1 = x 2 +y 2 +1 ; r 2y 2 = x 2 +y 2 +1 ; r 3 = x 2 +y 2 1 we obtain: x 2 +y 2 +1 ρ a = 1 ( ) 1 + r3 r 1 ir 2 = 2 r 1 + ir 2 1 r 3 1 x 2 + y ( ) x 2 + y 2 x iy. x + iy 1
7 How to encode a Pattern as a Density operator Normalized Trace Distance Example Let us consider the pattern a = {1, 3}. Its corresponding Density pattern ρ a, is: ρ a = 1 11 ( ) i 1 + 3i 1 We call ρ a Density Pattern.
8 Basic Notions How to encode a Pattern as a Density operator Normalized Trace Distance Moon Dataset Figure : Classical Patterns Figure : Density Patterns
9 How to encode a Pattern as a Density operator Normalized Trace Distance Another Example: Projective encoding v (x, y) ( x v, y ) ( x, ȳ) v ψ v = x 0 + ȳ 1 ρ v = ψ v ψ v...and many others.
10 How to encode a Pattern as a Density operator Normalized Trace Distance Distance. Preservation of the Order Let a = {x a, y a } b = {x b, y b } and c = {x c, y c } be three arbitrary patterns and let ρ i be the density pattern associated to the pattern i. If Ed(a, b) Ed(b, c) (where Ed is the Euclidian distance), is it possible to define a Quantum distance such that Qd(ρ a, ρ b ) Qd(ρ b, ρ c )?
11 How to encode a Pattern as a Density operator Normalized Trace Distance Normalized Trace Distance Let us consider two patterns a = {x a, y a } and b = {x b, y b }. ( Let ρ a = ra3 r a1 ir a2 2 r a1 + ir a2 1 r a3 associated to a; similarly for b. ) the density pattern 2 Let place K = (1 ra3 and let we define the normalized )(1 r b3 ) trace distance as: K Td(ρ a, ρ b ), where Td is the usual Trace distance. It is straightforward to show that Ed(a, b) = K Td(ρ a, ρ b ).
12 How to encode a Pattern as a Density operator Normalized Trace Distance At this stage let s keep the definition of centroid unchanged. Classification Hence, given a and b as the centroids of C a and C b respectively, if K Td(ρ x, ρ a ) K Td(ρ x, ρ b ) then x B; otherwise x A. Similarly to the classical case.
13 How to encode a Pattern as a Density operator Normalized Trace Distance In this way we have introduced just a translation of the standard NMC in terms of quantum objects (with the pontential implicit benefits carried out by a Quantum Computer). But this approach turns out to be convenient also on a Classical Computer.
14 Quantum Centroid Comparison Now we introduce a new definition of Quantum Centroid that has not any counterpart in the classical framework. Quantum Centroid Given a dataset {P 1,..., P n }, let us consider the respective set of density patterns {ρ 1,..., ρ n }. The Quantum Centroid is defined as: ρ QC = 1 n n ρ i. i=1 S. Gambs, Quantum classification, arxiv: v2.
15 Quantum Centroid Comparison Observations Some observation: The QC ρ QC is not a pure state and it has not any counterpart in the set of classical pattern in R n ; In contrast to the Classical Centroid, the QC is "sensitive" to the distribution of the patterns.
16 Quantum Centroid Comparison We provide a comparison between the NMC and the "quantum" classification process based on Density Patterns and Quantum Centroids by involving different kinds of standard (artificial) datasets on a Classical Computer. We compare the Error E (E = #CorrectClassifiedPatterns #PatternsOfTheDataset ) and the reliability (in terms of the Cohen s constant k) for both classifiers.
17 Quantum Centroid Comparison Gaussian Dataset Gaussian Dataset: 200 Patterns allocated in two Classes. Figure : Gaussian Dataset Figure : NMC Figure : Quantum Classifier
18 Quantum Centroid Comparison Table : Gaussian Dataset E E1 E2 Pr k TPR FPR NMC QC By randomly dividing the dataset in a training set (80%) and in a test set (20%), the average over 100 experiments gives: NMC Error = ± 6.79; Q Error = ± 6.09.
19 Quantum Centroid Comparison A remark Even if the error of the Quantum Classifier is lower than the Error of the NMC, there are some patterns that are correctly classified by the NMC but not by the Quantum Classifier. Hence, it makes sense to consider a "merging" of the NMC and the Quantum Classifier.
20 Basic Notions Quantum Centroid Comparison Gaussian Dataset Figure : Gaussian Dataset Figure : NMC Figure : Quantum Classifier Figure : NMC & Quantum Classifier
21 Quantum Centroid Comparison Table : Gaussian Dataset E E1 E2 Pr k TPR FPR NMC QC NMC-QC
22 Quantum Centroid Comparison Moon Dataset Moon Dataset: 200 patterns allocated in two Classes. Figure : Moon Dataset Figure : NMC Figure : Quantum Classifier Figure : NMC & Quantum Classifier
23 Quantum Centroid Comparison Table : Moon Dataset E E1 E2 Pr k TPR FPR NMC QC By randomly dividing the dataset in a training set (80%) and in a test set (20%), the average over 100 experiments gives: NMC Error = ± 6.32; Q Error = ± 5.46.
24 Basic Notions Quantum Centroid Comparison Banana Dataset Banana Dataset: 5300 patterns; 2376 belonging to the first Class and 2924 to the second Class. Figure : Banana Dataset Figure : NMC Figure : Quantum Classifier Figure : NMC & Quantum Classifier
25 Quantum Centroid Comparison Table : Banana Dataset E E1 E2 Pr k TPR FPR NMC QC NMC-QC By randomly dividing the dataset in a training set (80%) and in a test set (20%), the average over 100 experiments gives: NMC Error = ± 1.74; Q Error = ± 1.21.
26 Basic Notions Quantum Centroid Comparison 3Gaussian Dataset 3Gaussian Dataset: 450 Patterns allocated in three Classes. Figure : 3Gaussian Dataset Figure : NMC Figure : Quantum Classifier Figure : NMC & Quantum Classifier
27 Quantum Centroid Comparison Here we randomly divide the dataset in a Training set (80% of the patterns) and a Test set (20% of the patterns). We calculate the average over 100 runs for each experiments. A full comparison
28 The Quantum Centroid is not invariant under rescaling the "Quantum" Classifier is not invariant under rescaling! Is it an Embarrasment or is it an Asset? The Error is dependent on both the rescaling of the Patterns and the different encoding. We show how the Error changes by changing the rescaling (it can be shown for two different encodings).
29 Rescaling How the Error of the Quantum Classifier changes by ranging the value of the rescaling. If we know, in principle, the kind of distribution of a given phenomena, we can choose ad hoc the suitable rescaling in order to further improve the accuracy of the classification.
30 Summarizing the procedure 1. Start from classical data 2. Suitable rescaling (if we know in principle the kind of distribution) 3. Encoding (from classical patterns to density patterns) 4. Classification (using a classical computer) 5. Decoding (from density patterns to classical patterns) 6. Merging with the classical result Positive answer to the foundational question.
31 The choise of the "best" Encoding (and/or the best Rescaling) is mostly empirical and it is stritcly dependent on the Dataset. A complete theoretical investigation is still missing!!! A comparison with more performant classifiers (Linear Discriminant Analysis, Quadratic Discriminant Analysis...) also in case of unsupervised or semi-supervised classification could be investigated. (The NMC and the Quantum Classifier are based on the concepts of centroid and distance only but more sofisticated classifiers can be replaced in a quantum framework exploiting superposition principle, entanglement, majorization etc...).
32 Basic Notions From artificial to real datasets As further developments, it will be checked whether the Quantum Classifier could bring some benefit for practical implementations, such as Figure : Handwriting Figure : Fingerprint Recognition Figure : Face Recognition Figure : Biomedical Imaging
33 thank you G. Sergioli, E. Santucci, L. Didaci, J.A. Miskczak, R. Giuntini, A Quantum-inspired version of the Nearest Mean Classifier, Soft Computing (2016). G. Sergioli, G.M. Bosyk, E. Santucci, R. Giuntini, A Quantum-inspired version of the classification problem, International Journal of Theoretical Physics (2017). F. Holik, G. Sergioli, H. Freytes, A. Plastino, Pattern Recognition in non-kolmogorovian Structures, Foundation of Science (2017). E. Santucci, G. Sergioli, Classification Problem in a Quantum Framework, Proceedings of the Vaxjo Conference Quantum and Beyond (2017).
Quantum Minimum Distance Classifier
Article Quantum Minimum Distance Classifier Enrica Santucci 1 University of Cagliari, Piazza D Armi snc - 09123 Cagliari (Italy); enrica.santucci@gmail.com 1 2 3 4 5 6 7 8 9 10 11 12 13 14 Abstract: We
More informationDepartment of Electrical and Electronic Engineering, University of Cagliari, Cagliari 09123, Italy;
entropy Article Quantum Minimum Distance Classifier Enrica Santucci Department of Electrical and Electronic Engineering, University of Cagliari, Cagliari 09123, Italy; enrica.santucci@gmail.com Received:
More informationQuantum machine learning for quantum anomaly detection CQT AND SUTD, SINGAPORE ARXIV:
Quantum machine learning for quantum anomaly detection NANA LIU CQT AND SUTD, SINGAPORE ARXIV:1710.07405 TUESDAY 7 TH NOVEMBER 2017 QTML 2017, VERONA Types of anomaly detection algorithms Classification-based
More informationLecture Notes 1: Vector spaces
Optimization-based data analysis Fall 2017 Lecture Notes 1: Vector spaces In this chapter we review certain basic concepts of linear algebra, highlighting their application to signal processing. 1 Vector
More informationInternational Workshop on Applied Mathematics & Quantum Information. Cagliari, November 3-4, 2016
International Workshop on Applied Mathematics & Quantum Information Cagliari, November 3-4, 2016 Contents 1 INTRODUCTION AND ACKNOWLEDGMENTS 2 2 INVITED SPEAKERS 3 3 ABSTRACTS OF THE CONTRIBUTIONS 4 Majorization
More informationHidden Markov Models
10-601 Introduction to Machine Learning Machine Learning Department School of Computer Science Carnegie Mellon University Hidden Markov Models Matt Gormley Lecture 22 April 2, 2018 1 Reminders Homework
More informationARTIFICIAL NEURAL NETWORKS گروه مطالعاتي 17 بهار 92
ARTIFICIAL NEURAL NETWORKS گروه مطالعاتي 17 بهار 92 BIOLOGICAL INSPIRATIONS Some numbers The human brain contains about 10 billion nerve cells (neurons) Each neuron is connected to the others through 10000
More informationClassification of handwritten digits using supervised locally linear embedding algorithm and support vector machine
Classification of handwritten digits using supervised locally linear embedding algorithm and support vector machine Olga Kouropteva, Oleg Okun, Matti Pietikäinen Machine Vision Group, Infotech Oulu and
More informationGenerative Learning. INFO-4604, Applied Machine Learning University of Colorado Boulder. November 29, 2018 Prof. Michael Paul
Generative Learning INFO-4604, Applied Machine Learning University of Colorado Boulder November 29, 2018 Prof. Michael Paul Generative vs Discriminative The classification algorithms we have seen so far
More informationIntro. ANN & Fuzzy Systems. Lecture 15. Pattern Classification (I): Statistical Formulation
Lecture 15. Pattern Classification (I): Statistical Formulation Outline Statistical Pattern Recognition Maximum Posterior Probability (MAP) Classifier Maximum Likelihood (ML) Classifier K-Nearest Neighbor
More informationarxiv: v2 [quant-ph] 19 Jul 2018
Multiqubit and multilevel quantum reinforcement learning with quantum technologies F. A. Cárdenas-López,2,*, L. Lamata 3, J. C. Retamal,2, E. Solano 3,4,5 arxiv:709.07848v2 [quant-ph] 9 Jul 208 Departamento
More informationFeature selection and extraction Spectral domain quality estimation Alternatives
Feature selection and extraction Error estimation Maa-57.3210 Data Classification and Modelling in Remote Sensing Markus Törmä markus.torma@tkk.fi Measurements Preprocessing: Remove random and systematic
More informationDiscriminant analysis and supervised classification
Discriminant analysis and supervised classification Angela Montanari 1 Linear discriminant analysis Linear discriminant analysis (LDA) also known as Fisher s linear discriminant analysis or as Canonical
More informationIntroduction to Machine Learning
Introduction to Machine Learning CS4731 Dr. Mihail Fall 2017 Slide content based on books by Bishop and Barber. https://www.microsoft.com/en-us/research/people/cmbishop/ http://web4.cs.ucl.ac.uk/staff/d.barber/pmwiki/pmwiki.php?n=brml.homepage
More informationLecture 4 Discriminant Analysis, k-nearest Neighbors
Lecture 4 Discriminant Analysis, k-nearest Neighbors Fredrik Lindsten Division of Systems and Control Department of Information Technology Uppsala University. Email: fredrik.lindsten@it.uu.se fredrik.lindsten@it.uu.se
More informationExpectation Maximization (EM)
Expectation Maximization (EM) The Expectation Maximization (EM) algorithm is one approach to unsupervised, semi-supervised, or lightly supervised learning. In this kind of learning either no labels are
More informationLinear classifiers Lecture 3
Linear classifiers Lecture 3 David Sontag New York University Slides adapted from Luke Zettlemoyer, Vibhav Gogate, and Carlos Guestrin ML Methodology Data: labeled instances, e.g. emails marked spam/ham
More informationGaussian Models
Gaussian Models ddebarr@uw.edu 2016-04-28 Agenda Introduction Gaussian Discriminant Analysis Inference Linear Gaussian Systems The Wishart Distribution Inferring Parameters Introduction Gaussian Density
More informationLinear Classifiers as Pattern Detectors
Intelligent Systems: Reasoning and Recognition James L. Crowley ENSIMAG 2 / MoSIG M1 Second Semester 2014/2015 Lesson 16 8 April 2015 Contents Linear Classifiers as Pattern Detectors Notation...2 Linear
More informationMotivating the Covariance Matrix
Motivating the Covariance Matrix Raúl Rojas Computer Science Department Freie Universität Berlin January 2009 Abstract This note reviews some interesting properties of the covariance matrix and its role
More informationPattern Recognition. Parameter Estimation of Probability Density Functions
Pattern Recognition Parameter Estimation of Probability Density Functions Classification Problem (Review) The classification problem is to assign an arbitrary feature vector x F to one of c classes. The
More informationMachine Learning for Signal Processing Bayes Classification and Regression
Machine Learning for Signal Processing Bayes Classification and Regression Instructor: Bhiksha Raj 11755/18797 1 Recap: KNN A very effective and simple way of performing classification Simple model: For
More informationMULTIVARIATE PATTERN RECOGNITION FOR CHEMOMETRICS. Richard Brereton
MULTIVARIATE PATTERN RECOGNITION FOR CHEMOMETRICS Richard Brereton r.g.brereton@bris.ac.uk Pattern Recognition Book Chemometrics for Pattern Recognition, Wiley, 2009 Pattern Recognition Pattern Recognition
More informationTopics in Natural Language Processing
Topics in Natural Language Processing Shay Cohen Institute for Language, Cognition and Computation University of Edinburgh Lecture 9 Administrativia Next class will be a summary Please email me questions
More informationClustering VS Classification
MCQ Clustering VS Classification 1. What is the relation between the distance between clusters and the corresponding class discriminability? a. proportional b. inversely-proportional c. no-relation Ans:
More informationSUPERVISED LEARNING: INTRODUCTION TO CLASSIFICATION
SUPERVISED LEARNING: INTRODUCTION TO CLASSIFICATION 1 Outline Basic terminology Features Training and validation Model selection Error and loss measures Statistical comparison Evaluation measures 2 Terminology
More informationPrinciples of Pattern Recognition. C. A. Murthy Machine Intelligence Unit Indian Statistical Institute Kolkata
Principles of Pattern Recognition C. A. Murthy Machine Intelligence Unit Indian Statistical Institute Kolkata e-mail: murthy@isical.ac.in Pattern Recognition Measurement Space > Feature Space >Decision
More informationIntroduction to Signal Detection and Classification. Phani Chavali
Introduction to Signal Detection and Classification Phani Chavali Outline Detection Problem Performance Measures Receiver Operating Characteristics (ROC) F-Test - Test Linear Discriminant Analysis (LDA)
More informationSupport Vector Machines. CAP 5610: Machine Learning Instructor: Guo-Jun QI
Support Vector Machines CAP 5610: Machine Learning Instructor: Guo-Jun QI 1 Linear Classifier Naive Bayes Assume each attribute is drawn from Gaussian distribution with the same variance Generative model:
More informationPattern Recognition and Machine Learning. Learning and Evaluation of Pattern Recognition Processes
Pattern Recognition and Machine Learning James L. Crowley ENSIMAG 3 - MMIS Fall Semester 2016 Lesson 1 5 October 2016 Learning and Evaluation of Pattern Recognition Processes Outline Notation...2 1. The
More informationMultiscale Autoconvolution Histograms for Affine Invariant Pattern Recognition
Multiscale Autoconvolution Histograms for Affine Invariant Pattern Recognition Esa Rahtu Mikko Salo Janne Heikkilä Department of Electrical and Information Engineering P.O. Box 4500, 90014 University of
More informationAdvanced Introduction to Machine Learning
10-715 Advanced Introduction to Machine Learning Homework Due Oct 15, 10.30 am Rules Please follow these guidelines. Failure to do so, will result in loss of credit. 1. Homework is due on the due date
More informationTowards a Ptolemaic Model for OCR
Towards a Ptolemaic Model for OCR Sriharsha Veeramachaneni and George Nagy Rensselaer Polytechnic Institute, Troy, NY, USA E-mail: nagy@ecse.rpi.edu Abstract In style-constrained classification often there
More informationMachine Learning. CUNY Graduate Center, Spring Lectures 11-12: Unsupervised Learning 1. Professor Liang Huang.
Machine Learning CUNY Graduate Center, Spring 2013 Lectures 11-12: Unsupervised Learning 1 (Clustering: k-means, EM, mixture models) Professor Liang Huang huang@cs.qc.cuny.edu http://acl.cs.qc.edu/~lhuang/teaching/machine-learning
More informationCS229 Machine Learning Project: Allocating funds to the right people
CS229 Machine Learning Project: Allocating funds to the right people Introduction Dugalic, Adem adugalic@stanford.edu Nguyen, Tram tram@stanford.edu Our project is motivated by a common issue in developing
More informationHypothesis Evaluation
Hypothesis Evaluation Machine Learning Hamid Beigy Sharif University of Technology Fall 1395 Hamid Beigy (Sharif University of Technology) Hypothesis Evaluation Fall 1395 1 / 31 Table of contents 1 Introduction
More informationCS 188: Artificial Intelligence Spring Announcements
CS 188: Artificial Intelligence Spring 2010 Lecture 22: Nearest Neighbors, Kernels 4/18/2011 Pieter Abbeel UC Berkeley Slides adapted from Dan Klein Announcements On-going: contest (optional and FUN!)
More informationDiscriminative Direction for Kernel Classifiers
Discriminative Direction for Kernel Classifiers Polina Golland Artificial Intelligence Lab Massachusetts Institute of Technology Cambridge, MA 02139 polina@ai.mit.edu Abstract In many scientific and engineering
More informationAn Overview of Outlier Detection Techniques and Applications
Machine Learning Rhein-Neckar Meetup An Overview of Outlier Detection Techniques and Applications Ying Gu connygy@gmail.com 28.02.2016 Anomaly/Outlier Detection What are anomalies/outliers? The set of
More informationLinear Classifiers as Pattern Detectors
Intelligent Systems: Reasoning and Recognition James L. Crowley ENSIMAG 2 / MoSIG M1 Second Semester 2013/2014 Lesson 18 23 April 2014 Contents Linear Classifiers as Pattern Detectors Notation...2 Linear
More informationData Preprocessing. Cluster Similarity
1 Cluster Similarity Similarity is most often measured with the help of a distance function. The smaller the distance, the more similar the data objects (points). A function d: M M R is a distance on M
More informationPerformance Comparison of K-Means and Expectation Maximization with Gaussian Mixture Models for Clustering EE6540 Final Project
Performance Comparison of K-Means and Expectation Maximization with Gaussian Mixture Models for Clustering EE6540 Final Project Devin Cornell & Sushruth Sastry May 2015 1 Abstract In this article, we explore
More informationSupervised locally linear embedding
Supervised locally linear embedding Dick de Ridder 1, Olga Kouropteva 2, Oleg Okun 2, Matti Pietikäinen 2 and Robert P.W. Duin 1 1 Pattern Recognition Group, Department of Imaging Science and Technology,
More informationMachine Learning and Data Mining. Bayes Classifiers. Prof. Alexander Ihler
+ Machine Learning and Data Mining Bayes Classifiers Prof. Alexander Ihler A basic classifier Training data D={x (i),y (i) }, Classifier f(x ; D) Discrete feature vector x f(x ; D) is a con@ngency table
More informationarxiv: v2 [quant-ph] 2 Sep 2008
Quantum classification arxiv:0809.0444v2 [quant-ph] 2 Sep 2008 Sébastien Gambs Département d informatique et de recherche opérationnelle Université de Montréal C.P. 6128, Succ. Centre-Ville, Montréal (Québec),
More informationLecture 6: Quantum error correction and quantum capacity
Lecture 6: Quantum error correction and quantum capacity Mark M. Wilde The quantum capacity theorem is one of the most important theorems in quantum hannon theory. It is a fundamentally quantum theorem
More informationInterpreting Deep Classifiers
Ruprecht-Karls-University Heidelberg Faculty of Mathematics and Computer Science Seminar: Explainable Machine Learning Interpreting Deep Classifiers by Visual Distillation of Dark Knowledge Author: Daniela
More informationISyE 6416: Computational Statistics Spring Lecture 5: Discriminant analysis and classification
ISyE 6416: Computational Statistics Spring 2017 Lecture 5: Discriminant analysis and classification Prof. Yao Xie H. Milton Stewart School of Industrial and Systems Engineering Georgia Institute of Technology
More informationAnnouncements. CS 188: Artificial Intelligence Spring Classification. Today. Classification overview. Case-Based Reasoning
CS 188: Artificial Intelligence Spring 21 Lecture 22: Nearest Neighbors, Kernels 4/18/211 Pieter Abbeel UC Berkeley Slides adapted from Dan Klein Announcements On-going: contest (optional and FUN!) Remaining
More informationFace Recognition Using Laplacianfaces He et al. (IEEE Trans PAMI, 2005) presented by Hassan A. Kingravi
Face Recognition Using Laplacianfaces He et al. (IEEE Trans PAMI, 2005) presented by Hassan A. Kingravi Overview Introduction Linear Methods for Dimensionality Reduction Nonlinear Methods and Manifold
More informationData Mining. Practical Machine Learning Tools and Techniques. Slides for Chapter 4 of Data Mining by I. H. Witten, E. Frank and M. A.
Data Mining Practical Machine Learning Tools and Techniques Slides for Chapter of Data Mining by I. H. Witten, E. Frank and M. A. Hall Statistical modeling Opposite of R: use all the attributes Two assumptions:
More informationLinear & nonlinear classifiers
Linear & nonlinear classifiers Machine Learning Hamid Beigy Sharif University of Technology Fall 1394 Hamid Beigy (Sharif University of Technology) Linear & nonlinear classifiers Fall 1394 1 / 34 Table
More information9/26/17. Ridge regression. What our model needs to do. Ridge Regression: L2 penalty. Ridge coefficients. Ridge coefficients
What our model needs to do regression Usually, we are not just trying to explain observed data We want to uncover meaningful trends And predict future observations Our questions then are Is β" a good estimate
More informationLinear, threshold units. Linear Discriminant Functions and Support Vector Machines. Biometrics CSE 190 Lecture 11. X i : inputs W i : weights
Linear Discriminant Functions and Support Vector Machines Linear, threshold units CSE19, Winter 11 Biometrics CSE 19 Lecture 11 1 X i : inputs W i : weights θ : threshold 3 4 5 1 6 7 Courtesy of University
More informationPattern recognition. "To understand is to perceive patterns" Sir Isaiah Berlin, Russian philosopher
Pattern recognition "To understand is to perceive patterns" Sir Isaiah Berlin, Russian philosopher The more relevant patterns at your disposal, the better your decisions will be. This is hopeful news to
More informationHidden Markov Models
10-601 Introduction to Machine Learning Machine Learning Department School of Computer Science Carnegie Mellon University Hidden Markov Models Matt Gormley Lecture 19 Nov. 5, 2018 1 Reminders Homework
More informationPart I. Linear Discriminant Analysis. Discriminant analysis. Discriminant analysis
Week 5 Based in part on slides from textbook, slides of Susan Holmes Part I Linear Discriminant Analysis October 29, 2012 1 / 1 2 / 1 Nearest centroid rule Suppose we break down our data matrix as by the
More informationMachine Learning and Deep Learning! Vincent Lepetit!
Machine Learning and Deep Learning!! Vincent Lepetit! 1! What is Machine Learning?! 2! Hand-Written Digit Recognition! 2 9 3! Hand-Written Digit Recognition! Formalization! 0 1 x = @ A Images are 28x28
More informationBeyond the Point Cloud: From Transductive to Semi-Supervised Learning
Beyond the Point Cloud: From Transductive to Semi-Supervised Learning Vikas Sindhwani, Partha Niyogi, Mikhail Belkin Andrew B. Goldberg goldberg@cs.wisc.edu Department of Computer Sciences University of
More informationLinear Discriminant Analysis Based in part on slides from textbook, slides of Susan Holmes. November 9, Statistics 202: Data Mining
Linear Discriminant Analysis Based in part on slides from textbook, slides of Susan Holmes November 9, 2012 1 / 1 Nearest centroid rule Suppose we break down our data matrix as by the labels yielding (X
More informationTime Series Classification
Distance Measures Classifiers DTW vs. ED Further Work Questions August 31, 2017 Distance Measures Classifiers DTW vs. ED Further Work Questions Outline 1 2 Distance Measures 3 Classifiers 4 DTW vs. ED
More informationIntroduction to Machine Learning. PCA and Spectral Clustering. Introduction to Machine Learning, Slides: Eran Halperin
1 Introduction to Machine Learning PCA and Spectral Clustering Introduction to Machine Learning, 2013-14 Slides: Eran Halperin Singular Value Decomposition (SVD) The singular value decomposition (SVD)
More informationMachine Learning And Applications: Supervised Learning-SVM
Machine Learning And Applications: Supervised Learning-SVM Raphaël Bournhonesque École Normale Supérieure de Lyon, Lyon, France raphael.bournhonesque@ens-lyon.fr 1 Supervised vs unsupervised learning Machine
More informationPerformance Evaluation and Hypothesis Testing
Performance Evaluation and Hypothesis Testing 1 Motivation Evaluating the performance of learning systems is important because: Learning systems are usually designed to predict the class of future unlabeled
More informationNotation. Pattern Recognition II. Michal Haindl. Outline - PR Basic Concepts. Pattern Recognition Notions
Notation S pattern space X feature vector X = [x 1,...,x l ] l = dim{x} number of features X feature space K number of classes ω i class indicator Ω = {ω 1,...,ω K } g(x) discriminant function H decision
More informationClassifier Complexity and Support Vector Classifiers
Classifier Complexity and Support Vector Classifiers Feature 2 6 4 2 0 2 4 6 8 RBF kernel 10 10 8 6 4 2 0 2 4 6 Feature 1 David M.J. Tax Pattern Recognition Laboratory Delft University of Technology D.M.J.Tax@tudelft.nl
More informationL11: Pattern recognition principles
L11: Pattern recognition principles Bayesian decision theory Statistical classifiers Dimensionality reduction Clustering This lecture is partly based on [Huang, Acero and Hon, 2001, ch. 4] Introduction
More informationKaviti Sai Saurab. October 23,2014. IIT Kanpur. Kaviti Sai Saurab (UCLA) Quantum methods in ML October 23, / 8
A review of Quantum Algorithms for Nearest-Neighbor Methods for Supervised and Unsupervised Learning Nathan Wiebe,Ashish Kapoor,Krysta M. Svore Jan.2014 Kaviti Sai Saurab IIT Kanpur October 23,2014 Kaviti
More informationday month year documentname/initials 1
ECE471-571 Pattern Recognition Lecture 13 Decision Tree Hairong Qi, Gonzalez Family Professor Electrical Engineering and Computer Science University of Tennessee, Knoxville http://www.eecs.utk.edu/faculty/qi
More informationLinear models: the perceptron and closest centroid algorithms. D = {(x i,y i )} n i=1. x i 2 R d 9/3/13. Preliminaries. Chapter 1, 7.
Preliminaries Linear models: the perceptron and closest centroid algorithms Chapter 1, 7 Definition: The Euclidean dot product beteen to vectors is the expression d T x = i x i The dot product is also
More informationAn introduction to clustering techniques
- ABSTRACT Cluster analysis has been used in a wide variety of fields, such as marketing, social science, biology, pattern recognition etc. It is used to identify homogenous groups of cases to better understand
More informationc 4, < y 2, 1 0, otherwise,
Fundamentals of Big Data Analytics Univ.-Prof. Dr. rer. nat. Rudolf Mathar Problem. Probability theory: The outcome of an experiment is described by three events A, B and C. The probabilities Pr(A) =,
More informationContents Lecture 4. Lecture 4 Linear Discriminant Analysis. Summary of Lecture 3 (II/II) Summary of Lecture 3 (I/II)
Contents Lecture Lecture Linear Discriminant Analysis Fredrik Lindsten Division of Systems and Control Department of Information Technology Uppsala University Email: fredriklindsten@ituuse Summary of lecture
More informationAN INTRODUCTION TO NEURAL NETWORKS. Scott Kuindersma November 12, 2009
AN INTRODUCTION TO NEURAL NETWORKS Scott Kuindersma November 12, 2009 SUPERVISED LEARNING We are given some training data: We must learn a function If y is discrete, we call it classification If it is
More informationKernel Methods. Barnabás Póczos
Kernel Methods Barnabás Póczos Outline Quick Introduction Feature space Perceptron in the feature space Kernels Mercer s theorem Finite domain Arbitrary domain Kernel families Constructing new kernels
More informationPointwise Exact Bootstrap Distributions of Cost Curves
Pointwise Exact Bootstrap Distributions of Cost Curves Charles Dugas and David Gadoury University of Montréal 25th ICML Helsinki July 2008 Dugas, Gadoury (U Montréal) Cost curves July 8, 2008 1 / 24 Outline
More informationSVM TRADE-OFF BETWEEN MAXIMIZE THE MARGIN AND MINIMIZE THE VARIABLES USED FOR REGRESSION
International Journal of Pure and Applied Mathematics Volume 87 No. 6 2013, 741-750 ISSN: 1311-8080 (printed version); ISSN: 1314-3395 (on-line version) url: http://www.ijpam.eu doi: http://dx.doi.org/10.12732/ijpam.v87i6.2
More informationJae-Bong Lee 1 and Bernard A. Megrey 2. International Symposium on Climate Change Effects on Fish and Fisheries
International Symposium on Climate Change Effects on Fish and Fisheries On the utility of self-organizing maps (SOM) and k-means clustering to characterize and compare low frequency spatial and temporal
More informationMachine Learning. Regression-Based Classification & Gaussian Discriminant Analysis. Manfred Huber
Machine Learning Regression-Based Classification & Gaussian Discriminant Analysis Manfred Huber 2015 1 Logistic Regression Linear regression provides a nice representation and an efficient solution to
More informationIntroduction to Machine Learning Midterm Exam
10-701 Introduction to Machine Learning Midterm Exam Instructors: Eric Xing, Ziv Bar-Joseph 17 November, 2015 There are 11 questions, for a total of 100 points. This exam is open book, open notes, but
More informationStatistical Machine Learning
Statistical Machine Learning Christoph Lampert Spring Semester 2015/2016 // Lecture 12 1 / 36 Unsupervised Learning Dimensionality Reduction 2 / 36 Dimensionality Reduction Given: data X = {x 1,..., x
More informationMixture of Gaussians Models
Mixture of Gaussians Models Outline Inference, Learning, and Maximum Likelihood Why Mixtures? Why Gaussians? Building up to the Mixture of Gaussians Single Gaussians Fully-Observed Mixtures Hidden Mixtures
More informationIntroduction to Machine Learning Spring 2018 Note 18
CS 189 Introduction to Machine Learning Spring 2018 Note 18 1 Gaussian Discriminant Analysis Recall the idea of generative models: we classify an arbitrary datapoint x with the class label that maximizes
More informationECE662: Pattern Recognition and Decision Making Processes: HW TWO
ECE662: Pattern Recognition and Decision Making Processes: HW TWO Purdue University Department of Electrical and Computer Engineering West Lafayette, INDIANA, USA Abstract. In this report experiments are
More informationOn Classification: An Empirical Study of Existing Algorithms Based on Two Kaggle Competitions
On Classification: An Empirical Study of Existing Algorithms Based on Two Kaggle Competitions Wilson Florero-Salinas (Team Leader), Sha Li (Team Leader) Xiaoyan Chong, Dan Li, Minglu Ma, Abhirupa Sen,
More informationStatistical Pattern Recognition
Statistical Pattern Recognition Feature Extraction Hamid R. Rabiee Jafar Muhammadi, Alireza Ghasemi, Payam Siyari Spring 2014 http://ce.sharif.edu/courses/92-93/2/ce725-2/ Agenda Dimensionality Reduction
More informationClassification Methods II: Linear and Quadratic Discrimminant Analysis
Classification Methods II: Linear and Quadratic Discrimminant Analysis Rebecca C. Steorts, Duke University STA 325, Chapter 4 ISL Agenda Linear Discrimminant Analysis (LDA) Classification Recall that linear
More informationApril 18, :46 WSPC/INSTRUCTION FILE ForVaxjorev15.2dc ABSTRACT QUANTUM COMPUTING MACHINES AND QUANTUM COMPUTATIONAL LOGICS
International Journal of Quantum Information c World Scientific Publishing Company ABSTRACT QUANTUM COMPUTING MACHINES AND QUANTUM COMPUTATIONAL LOGICS MARIA LUISA DALLA CHIARA Dipartimento di Lettere
More informationPerceptron Revisited: Linear Separators. Support Vector Machines
Support Vector Machines Perceptron Revisited: Linear Separators Binary classification can be viewed as the task of separating classes in feature space: w T x + b > 0 w T x + b = 0 w T x + b < 0 Department
More informationA Bias Correction for the Minimum Error Rate in Cross-validation
A Bias Correction for the Minimum Error Rate in Cross-validation Ryan J. Tibshirani Robert Tibshirani Abstract Tuning parameters in supervised learning problems are often estimated by cross-validation.
More informationCS 188: Artificial Intelligence Spring Announcements
CS 188: Artificial Intelligence Spring 2010 Lecture 24: Perceptrons and More! 4/22/2010 Pieter Abbeel UC Berkeley Slides adapted from Dan Klein Announcements W7 due tonight [this is your last written for
More informationMIDTERM: CS 6375 INSTRUCTOR: VIBHAV GOGATE October,
MIDTERM: CS 6375 INSTRUCTOR: VIBHAV GOGATE October, 23 2013 The exam is closed book. You are allowed a one-page cheat sheet. Answer the questions in the spaces provided on the question sheets. If you run
More informationQuantum Cloning WOOTTERS-ZUREK CLONER
Quantum Cloning Quantum cloning has been a topic of considerable interest for many years. It turns out to be quantum limit for copying an input state and is closely related to linear amplification when
More informationMachine Learning (CS 567) Lecture 5
Machine Learning (CS 567) Lecture 5 Time: T-Th 5:00pm - 6:20pm Location: GFS 118 Instructor: Sofus A. Macskassy (macskass@usc.edu) Office: SAL 216 Office hours: by appointment Teaching assistant: Cheol
More informationRepresenting fuzzy structures in quantum computation with mixed states
Representing fuzzy structures in quantum computation with mixed states Hector Freytes Fellow of CONICET-Argentina Email: hfreytes@gmailcom Antonio Arico Department of Mathematics Viale Merello 9, 913 Caglari
More informationStatistical learning theory, Support vector machines, and Bioinformatics
1 Statistical learning theory, Support vector machines, and Bioinformatics Jean-Philippe.Vert@mines.org Ecole des Mines de Paris Computational Biology group ENS Paris, november 25, 2003. 2 Overview 1.
More informationCS 340 Lec. 18: Multivariate Gaussian Distributions and Linear Discriminant Analysis
CS 3 Lec. 18: Multivariate Gaussian Distributions and Linear Discriminant Analysis AD March 11 AD ( March 11 1 / 17 Multivariate Gaussian Consider data { x i } N i=1 where xi R D and we assume they are
More informationSelection of Classifiers based on Multiple Classifier Behaviour
Selection of Classifiers based on Multiple Classifier Behaviour Giorgio Giacinto, Fabio Roli, and Giorgio Fumera Dept. of Electrical and Electronic Eng. - University of Cagliari Piazza d Armi, 09123 Cagliari,
More informationLearning Kernel Parameters by using Class Separability Measure
Learning Kernel Parameters by using Class Separability Measure Lei Wang, Kap Luk Chan School of Electrical and Electronic Engineering Nanyang Technological University Singapore, 3979 E-mail: P 3733@ntu.edu.sg,eklchan@ntu.edu.sg
More informationClustering K-means. Clustering images. Machine Learning CSE546 Carlos Guestrin University of Washington. November 4, 2014.
Clustering K-means Machine Learning CSE546 Carlos Guestrin University of Washington November 4, 2014 1 Clustering images Set of Images [Goldberger et al.] 2 1 K-means Randomly initialize k centers µ (0)
More information