Principal Component Analysis -- PCA (also called Karhunen-Loeve transformation)

Similar documents
Focus was on solving matrix inversion problems Now we look at other properties of matrices Useful when A represents a transformations.

CS4495/6495 Introduction to Computer Vision. 8B-L2 Principle Component Analysis (and its use in Computer Vision)

Image Analysis. PCA and Eigenfaces

Principal Component Analysis (PCA)

CS 4495 Computer Vision Principle Component Analysis

What is Principal Component Analysis?

1 Principal Components Analysis

Face detection and recognition. Detection Recognition Sally

Lecture 24: Principal Component Analysis. Aykut Erdem May 2016 Hacettepe University

Eigenfaces. Face Recognition Using Principal Components Analysis

Dimensionality Reduction: PCA. Nicholas Ruozzi University of Texas at Dallas

Face Recognition. Face Recognition. Subspace-Based Face Recognition Algorithms. Application of Face Recognition

Robot Image Credit: Viktoriya Sukhanova 123RF.com. Dimensionality Reduction

Principal Component Analysis

7. Variable extraction and dimensionality reduction

Machine Learning 2nd Edition

Advanced Introduction to Machine Learning CMU-10715

Lecture: Face Recognition

Principal Component Analysis

PCA & ICA. CE-717: Machine Learning Sharif University of Technology Spring Soleymani

Reconnaissance d objetsd et vision artificielle

2D Image Processing Face Detection and Recognition

Deriving Principal Component Analysis (PCA)

EECS490: Digital Image Processing. Lecture #26

Introduction to Machine Learning

PCA FACE RECOGNITION

Principal Component Analysis

Face Detection and Recognition

Vectors To begin, let us describe an element of the state space as a point with numerical coordinates, that is x 1. x 2. x =

Recognition Using Class Specific Linear Projection. Magali Segal Stolrasky Nadav Ben Jakov April, 2015

Dimensionality Reduction Using PCA/LDA. Hongyu Li School of Software Engineering TongJi University Fall, 2014

UVA CS 6316/4501 Fall 2016 Machine Learning. Lecture 18: Principal Component Analysis (PCA) Dr. Yanjun Qi. University of Virginia

Face Recognition and Biometric Systems

Linear Subspace Models

Eigenimaging for Facial Recognition

Universität Potsdam Institut für Informatik Lehrstuhl Maschinelles Lernen PCA. Tobias Scheffer

PCA, Kernel PCA, ICA

Maximum variance formulation

Example: Face Detection

Pattern Recognition 2

Expectation Maximization

Unsupervised Learning: K- Means & PCA

Lecture 17: Face Recogni2on

Dimensionality reduction

Machine Learning - MT & 14. PCA and MDS

Principal Component Analysis (PCA) CSC411/2515 Tutorial

Signal Analysis. Principal Component Analysis

Keywords Eigenface, face recognition, kernel principal component analysis, machine learning. II. LITERATURE REVIEW & OVERVIEW OF PROPOSED METHODOLOGY

Introduction to Machine Learning. PCA and Spectral Clustering. Introduction to Machine Learning, Slides: Eran Halperin

System 1 (last lecture) : limited to rigidly structured shapes. System 2 : recognition of a class of varying shapes. Need to:

Covariance and Correlation Matrix

Covariance and Principal Components

Principal Components Analysis (PCA)

Lecture: Face Recognition and Feature Reduction

Image Analysis & Retrieval. Lec 14. Eigenface and Fisherface

PRINCIPAL COMPONENTS ANALYSIS

Singular Value Decomposition. 1 Singular Value Decomposition and the Four Fundamental Subspaces

Structure in Data. A major objective in data analysis is to identify interesting features or structure in the data.

ECE 521. Lecture 11 (not on midterm material) 13 February K-means clustering, Dimensionality reduction

COS 429: COMPUTER VISON Face Recognition

CSC 411 Lecture 12: Principal Component Analysis

Lecture: Face Recognition and Feature Reduction

CS168: The Modern Algorithmic Toolbox Lecture #8: How PCA Works

STA 414/2104: Lecture 8

PRINCIPAL COMPONENT ANALYSIS

20 Unsupervised Learning and Principal Components Analysis (PCA)

Principal Component Analysis

Image Analysis & Retrieval Lec 14 - Eigenface & Fisherface

CS168: The Modern Algorithmic Toolbox Lecture #7: Understanding Principal Component Analysis (PCA)

Karhunen-Loève Transform KLT. JanKees van der Poel D.Sc. Student, Mechanical Engineering

Announcements (repeat) Principal Components Analysis

Lecture 17: Face Recogni2on

Principal Component Analysis (PCA) Principal Component Analysis (PCA)

Lecture 13. Principal Component Analysis. Brett Bernstein. April 25, CDS at NYU. Brett Bernstein (CDS at NYU) Lecture 13 April 25, / 26

MATH 829: Introduction to Data Mining and Analysis Principal component analysis

Lecture 7: Con3nuous Latent Variable Models

Lecture 13 Visual recognition

Statistical Machine Learning

Statistical Pattern Recognition

Principal Component Analysis (PCA)

Principal Components Analysis. Sargur Srihari University at Buffalo

Neuroscience Introduction

Machine Learning. Data visualization and dimensionality reduction. Eric Xing. Lecture 7, August 13, Eric Xing Eric CMU,

SPECTRAL CLUSTERING AND KERNEL PRINCIPAL COMPONENT ANALYSIS ARE PURSUING GOOD PROJECTIONS

L11: Pattern recognition principles

The Mathematics of Facial Recognition

Face Recognition Using Multi-viewpoint Patterns for Robot Vision

Unsupervised Machine Learning and Data Mining. DS 5230 / DS Fall Lecture 7. Jan-Willem van de Meent

COMP 408/508. Computer Vision Fall 2017 PCA for Recognition

Principal Component Analysis

12.2 Dimensionality Reduction

Tutorial on Principal Component Analysis

Computation. For QDA we need to calculate: Lets first consider the case that

Artificial Intelligence Module 2. Feature Selection. Andrea Torsello

15 Singular Value Decomposition

Solutions to Review Problems for Chapter 6 ( ), 7.1

Course 495: Advanced Statistical Machine Learning/Pattern Recognition

Unsupervised Learning

Dimensionality Reduction

PCA Review. CS 510 February 25 th, 2013

Transcription:

Principal Component Analysis -- PCA (also called Karhunen-Loeve transformation) PCA transforms the original input space into a lower dimensional space, by constructing dimensions that are linear combinations of the given features; The objective is to consider independent dimensions along which data have largest variance (i.e., greatest variability);

The space of all face images When viewed as vectors of pixel values, face images are extremely high-dimensional 100x100 image = 10,000 dimensions However, relatively few 10,000-dimensional vectors correspond to valid face images We want to effectively model the subspace of face images

Geometric view Given set of datapoints in D dimensional space find some transformation which will transform the points to lower dimensional space. U d is D x d matrix with d orthonormal column vectors y i x i = x 0 + U d y i are the new coordinates of in d-dimensional space Derivation on the board see handout for more details x i

Statistical view Given multivariate random variable x and set of sample points x, find d uncorrelated linear i components of x such that variance of the components is maximized Such that Derivation on the board y i = u i T x u i T u i =1 and Var(y 1 ) Var(y 2 )

Principal Component Analysis -- PCA PCA enables transforms a number of possibly correlated variables into a smaller number of uncorrelated variables called principal components; The first principal component accounts for as much of the variability in the data as possible; Each succeeding component (orthogonal to the previous ones) accounts for as much of the remaining variability as possible.

Principal Component Analysis PCA is the most commonly used dimension reduction technique. (Also called the Karhunen-Loeve transform). Data samples Compute the mean x 1,,x N Computer the covariance: x = 1 N N x i i=1 Σ x = 1 N N i=1 (x i x )(x i x ) T

Principal Component Analysis Compute the eigenvalues and eigenvectors Solve Σ x x = λx Order them by magnitude: of the matrix Σ x PCA reduces the dimension by keeping direction such that

Principal Component Analysis For many datasets, most of the eigenvalues are negligible and can be discarded. The eigenvalue measures the variation In the direction of corresponding eigenvector Example:

Principal Component Analysis How to get uncorrelated components which Capture most of the variance Project the data onto the selected eigenvectors: y i = e i T (x i x ) If we consider first M eigenvectors we get new lower dimensional representation [y 1,,y M ] Proportion covered by first M eigenvalues

PCA Example The images of an object under different lighting lie in a low-dimensional space. The original images are 256x 256. But the data lies mostly in 3-5 dimensions. First we show the PCA for a face under a range of lighting conditions. The PCA components have simple interpretations. Then we plot as a function of M for several objects under a range of lighting.

Limitations of PCA PCA is not effective for some datasets. For example, if the data is a set of strings (1,0,0,0, ), (0,1,0,0 ),,(0,0,0,,1) then the eigenvalues do not fall off as PCA requires.

Principal Component Analysis -- PCA Statistical view of PCA PCA finds n linearly transformed components so that they explain the maximum amount of variance See hand out/blackboard how to compute the largest principal component We can define PCA in an intuitive way using a recursive formulation:

Simple illustration of PCA First principal component of a two-dimensional data set.

Simple illustration of PCA Second principal component of a two-dimensional data set.

Determining the number of components Plot the eigenvalues each eigenvalue is related to the amount of variation explained by the corresponding axis (eigenvector); If the points on the graph tend to level out (show an elbow shape), these eigenvalues are usually close enough to zero that they can be ignored. In general: Limit the variance accounted for.

Critical information lies in low dimensional subspaces F S A typical eigenvalue spectrum and its division into two orthogonal subspaces

Dimensionality Reduction Applications Need to analyze large amounts multivariate data. Human Faces. Speech Waveforms. Global Climate patterns. Gene Distributions. Difficult to visualize data in dimensions just greater than three. Discover compact representations of high dimensional data. Visualization. Compression. Better Recognition. Probably meaningful dimensions.

Eigenfaces: Key idea Assume that most face images lie on a low-dimensional subspace determined by the first k (k<d) directions of maximum variance Use PCA to determine the vectors or eigenfaces u 1, u k that span that subspace Represent all face images in the dataset as linear combinations of eigenfaces M. Turk and A. Pentland, Face Recognition using Eigenfaces, CVPR 1991

Eigenfaces example Training images x 1,,x N

Eigenfaces example Top eigenvectors: u 1, u k Mean: μ

Eigenfaces example Principal component (eigenvector) u k μ + 3σ k u k μ 3σ k u k

Eigenfaces example Representation (w i1,,w ik ) = (u 1T (x i µ),, u kt (x i µ)) Reconstruction = + ^ x = µ + w 1 u 1 +w 2 u 2 +w 3 u 3 +w 4 u 4 +

Recognition with eigenfaces Process labeled training images: Find mean µ and covariance matrix Σ Find k principal components (eigenvectors of Σ) u 1, u k Project each training image x i onto subspace spanned by principal components: (w i1,,w ik ) = (u 1T (x i µ),, u kt (x i µ)) Given novel image x: Project onto subspace: (w 1,,w k ) = (u 1T (x µ),, u k T (x µ)) Optional: check reconstruction error x x ^ to determine whether image is really a face Classify as closest training face in k-dimensional subspace M. Turk and A. Pentland, Face Recognition using Eigenfaces, CVPR 1991

Eigenfaces (Turk & Pentland, 1991)