Linear & Non-Linear Discriminant Analysis! Hugh R. Wilson

Similar documents
What is Principal Component Analysis?

Statistical Pattern Recognition

PCA & ICA. CE-717: Machine Learning Sharif University of Technology Spring Soleymani

ROBERTO BATTITI, MAURO BRUNATO. The LION Way: Machine Learning plus Intelligent Optimization. LIONlab, University of Trento, Italy, Apr 2015

CSC 411 Lecture 12: Principal Component Analysis

Machine Learning. CUNY Graduate Center, Spring Lectures 11-12: Unsupervised Learning 1. Professor Liang Huang.

Machine Learning 11. week

Face Recognition Using Laplacianfaces He et al. (IEEE Trans PAMI, 2005) presented by Hassan A. Kingravi

CS4495/6495 Introduction to Computer Vision. 8B-L2 Principle Component Analysis (and its use in Computer Vision)

Dimension Reduction (PCA, ICA, CCA, FLD,

ECE 521. Lecture 11 (not on midterm material) 13 February K-means clustering, Dimensionality reduction

December 20, MAA704, Multivariate analysis. Christopher Engström. Multivariate. analysis. Principal component analysis

Data Mining. Dimensionality reduction. Hamid Beigy. Sharif University of Technology. Fall 1395

Generative classifiers: The Gaussian classifier. Ata Kaban School of Computer Science University of Birmingham

PRINCIPAL COMPONENTS ANALYSIS

Machine Learning (CSE 446): Unsupervised Learning: K-means and Principal Component Analysis

Introduction to Machine Learning

Your Project Proposals

Principal Component Analysis

STATISTICS 407 METHODS OF MULTIVARIATE ANALYSIS TOPICS

Introduction to machine learning and pattern recognition Lecture 2 Coryn Bailer-Jones

20 Unsupervised Learning and Principal Components Analysis (PCA)

Feature Extraction Techniques

Dimensionality Reduction Using PCA/LDA. Hongyu Li School of Software Engineering TongJi University Fall, 2014

L11: Pattern recognition principles

PCA, Kernel PCA, ICA

DIMENSION REDUCTION AND CLUSTER ANALYSIS

Subspace Methods for Visual Learning and Recognition

Machine Learning 2nd Edition

Subspace Analysis for Facial Image Recognition: A Comparative Study. Yongbin Zhang, Lixin Lang and Onur Hamsici

Dimension Reduction Methods

1 Principal Components Analysis

Supervised locally linear embedding

LEC 3: Fisher Discriminant Analysis (FDA)

Principal Components Analysis (PCA)

PCA and admixture models

Data Preprocessing Tasks

Discriminant analysis and supervised classification

Linear Dimensionality Reduction

Supervised Learning: Linear Methods (1/2) Applied Multivariate Statistics Spring 2012

PCA and LDA. Man-Wai MAK

Principal Component Analysis!! Lecture 11!

Dimensionality Reduction

Classification 2: Linear discriminant analysis (continued); logistic regression

Multivariate Statistics Fundamentals Part 1: Rotation-based Techniques

Principal Component Analysis and Linear Discriminant Analysis

CSC321 Lecture 20: Autoencoders

Computation. For QDA we need to calculate: Lets first consider the case that

Classification techniques focus on Discriminant Analysis

STA 414/2104: Lecture 8

Example: Face Detection

Computer Vision Group Prof. Daniel Cremers. 3. Regression

Lecture 7: Con3nuous Latent Variable Models

Revision: Chapter 1-6. Applied Multivariate Statistics Spring 2012

Motivating the Covariance Matrix

Face Recognition. Face Recognition. Subspace-Based Face Recognition Algorithms. Application of Face Recognition

PCA and LDA. Man-Wai MAK

Linear Algebra Methods for Data Mining

Machine Learning (Spring 2012) Principal Component Analysis

Fisher s Linear Discriminant Analysis

Covariance and Correlation Matrix

Lecture 13. Principal Component Analysis. Brett Bernstein. April 25, CDS at NYU. Brett Bernstein (CDS at NYU) Lecture 13 April 25, / 26

STATISTICAL LEARNING SYSTEMS

Eigenimages. Digital Image Processing: Bernd Girod, 2013 Stanford University -- Eigenimages 1

Classification: Linear Discriminant Analysis

Learning Eigenfunctions: Links with Spectral Clustering and Kernel PCA

Kernel methods for comparing distributions, measuring dependence

Image Analysis. PCA and Eigenfaces

Machine Learning. Dimensionality reduction. Hamid Beigy. Sharif University of Technology. Fall 1395

Machine learning comes from Bayesian decision theory in statistics. There we want to minimize the expected value of the loss function.

Deriving Principal Component Analysis (PCA)

Unsupervised Learning: K- Means & PCA

Dimension Reduction Techniques. Presented by Jie (Jerry) Yu

Lecture: Face Recognition

Principal Component Analysis

CS 6375 Machine Learning

Principle Components Analysis (PCA) Relationship Between a Linear Combination of Variables and Axes Rotation for PCA

Announcements (repeat) Principal Components Analysis

MACHINE LEARNING. Methods for feature extraction and reduction of dimensionality: Probabilistic PCA and kernel PCA

Dimensionality Reduction: PCA. Nicholas Ruozzi University of Texas at Dallas

Multivariate Analysis

Chemometrics: Classification of spectra

Robot Image Credit: Viktoriya Sukhanova 123RF.com. Dimensionality Reduction

Eigenfaces. Face Recognition Using Principal Components Analysis

Machine Learning - MT & 14. PCA and MDS

5. Discriminant analysis

Clustering VS Classification

STA 414/2104: Machine Learning

Linear Classification. CSE 6363 Machine Learning Vassilis Athitsos Computer Science and Engineering Department University of Texas at Arlington

7. Variable extraction and dimensionality reduction

Lecture 13 Visual recognition

Linear Algebra. Introduction. Marek Petrik 3/23/2017. Many slides adapted from Linear Algebra Lectures by Martin Scharlemann

CS229 Final Project. Wentao Zhang Shaochuan Xu

Introduction to Graphical Models

Principal Component Analysis -- PCA (also called Karhunen-Loeve transformation)

Principal Component Analysis

ECE 661: Homework 10 Fall 2014

Recognition Using Class Specific Linear Projection. Magali Segal Stolrasky Nadav Ben Jakov April, 2015

The role of dimensionality reduction in classification

A Least Squares Formulation for Canonical Correlation Analysis

Transcription:

Linear & Non-Linear Discriminant Analysis! Hugh R. Wilson PCA Review! Supervised learning! Fisher linear discriminant analysis! Nonlinear discriminant analysis! Research example! Multiple Classes! Unsupervised learning

Reading Monday: Principal components: Kutz, pp. 387 (bottom) - 393! Today: Linear discriminant analysis: Kutz, pp. 442-445

PCA Summary PCA reveals data structure determined by covariance Calculation for M measurements, N samples (N > M): Subtract means from measurements Data covariance matrix CV [Vector, EV] = eig(cv) Dimensionality reduction (e.g. retain 95% of variance) Can be implemented by brain learning networks

Nonlinear PCA Nonlinear axes (a manifold) Problem: not know shape of data in multi-dimensions Similar problem in Discriminant Analysis 1.8 1.6 1.4 1.2 1 0.8 0.6 0.4 0.2 0-0.2-1.5-1 -0.5 0 0.5 1

Supervised Category Learning Issue: learn to discriminate category to which each individual belongs based on measurements! M measurements on each individual! Each individual labeled by category (e.g. dog vs. cat)! Supervised: learning algorithm knows the correct category in each case! Simplest example: Fisher linear discriminant analysis! Two categories (can be generalized)! Use result to categorize new data (cross validation)

Fisher Linear Discriminant Analysis Maximize ratio of covariance between classes to covariance within classes by projection onto vector V! Covariance Between: CovBet! Covariance Within: CovWin! CovWin*V = λ CovBet*V (generalized eigenvalue problem)! Solution: V = eig(inv(covwin)*covbet))! V = vector for maximum class separation! Project onto vector & find point for maximum discrimination! Matlab demo

Fisher LDA Matlab Code Calculate between and within covariances! Eigenvalue problem! Maximize ratio of CovBetween/CovBetween CovBetween = cov([dog; Cat]); CovDog = cov(dog); CovCat = cov(cat); CovWithin = (CovDog + CovCat)/2; (average category cov) [FLD, FV] = eig(inv(covwithin)*covbetween); (eigenvector)

Extension to Non-Linear Class Boundaries Expand dimensionality of measurement space! Include nonlinear functions of individual measurements! x 2, y 2, xy, (x 2 + y 2 ), other nonlinear combinations! Problem: don t use too many or overfit data! Matlab demo 5 4.5 4 3.5 3 2.5 2 1.5 1 0.5-2 -1 0 1 2 3

Non-Linear Rationale Eigenvalue problem: linear algebra! How can linear algebra deal with nonlinear boundaries?! Increase dimensionality of data space with nonlinear measurement combinations x 2, xy, etc.! Discriminant Analysis is still linear in finding the weighting of each of those dimensions! Solution: Ax + By + C x 2 + Dxy + E y 2! Linear in A, B, C, D, E

Discriminant Analysis of Child Faces Does face shape reveal whether child is male or female?! Used 45 child faces! PCA to reduce geometry from 28 to 9 dimensions (94% var)! Non-linear discriminant analysis (included each PC 2 )! Predicted gender of 23 new child faces (cross-validation)! Results: 80% correct on training faces! 78% correct on novel test faces

Extension to more Classes Example: 3 classes A, B, C! Do multiple 2-class discriminant analyses:! A versus (B + C)! B versus (A + C)! C versus (A + B)! Combine results! Effective for both linear & non-linear discriminants! Generalizes to many classes

Unsupervised Learning: Cluster Analysis Given data without category labels! Must learn number of categories & category membership! Analysis: find clusters of nearest neighbours 6 4 2 0-2 -4-6 -5-4 -3-2 -1 0 1 2 3 4

Summary PCA reveals data structure determined by eigenvalues of covariance matrix Fisher LDA (Linear Discriminant Analysis) reveals best axis for data projection to separate two classes Eigenvalue problem for matrix (CovBet)/(CovWin) Generalizes to multiple classes Non-linear Discriminant Analysis: add nonlinear combinations of measurements (extra dimensions) Dimensionality reduction by PCA often used first to simplify Discriminant Analysis (child face example) Computer simulations of brain networks can compute PCA, LDA & NLDA