Subspace Analysis for Facial Image Recognition: A Comparative Study. Yongbin Zhang, Lixin Lang and Onur Hamsici

Similar documents
Face Recognition. Face Recognition. Subspace-Based Face Recognition Algorithms. Application of Face Recognition

PCA & ICA. CE-717: Machine Learning Sharif University of Technology Spring Soleymani

Uncorrelated Multilinear Principal Component Analysis through Successive Variance Maximization

Principal Component Analysis (PCA)

PCA FACE RECOGNITION

Linear & Non-Linear Discriminant Analysis! Hugh R. Wilson

Support Vector Machines. Introduction to Data Mining, 2 nd Edition by Tan, Steinbach, Karpatne, Kumar

Kernel Methods. Konstantin Tretyakov MTAT Machine Learning

Face recognition Computer Vision Spring 2018, Lecture 21

Face Recognition Using Laplacianfaces He et al. (IEEE Trans PAMI, 2005) presented by Hassan A. Kingravi

Kernel Methods. Konstantin Tretyakov MTAT Machine Learning

CS4495/6495 Introduction to Computer Vision. 8C-L3 Support Vector Machines

2D Image Processing Face Detection and Recognition

Discriminant Uncorrelated Neighborhood Preserving Projections

Machine Learning 11. week

High Dimensional Discriminant Analysis

A Least Squares Formulation for Canonical Correlation Analysis

Face detection and recognition. Detection Recognition Sally

Dimension Reduction (PCA, ICA, CCA, FLD,

Kernel PCA: keep walking... in informative directions. Johan Van Horebeek, Victor Muñiz, Rogelio Ramos CIMAT, Guanajuato, GTO

Bayes Optimal Kernel Discriminant Analysis

Lecture: Face Recognition

Dimensionality Reduction:

Example: Face Detection

Kernel Principal Component Analysis

Machine Learning. CUNY Graduate Center, Spring Lectures 11-12: Unsupervised Learning 1. Professor Liang Huang.

Lecture 24: Principal Component Analysis. Aykut Erdem May 2016 Hacettepe University

Stefanos Zafeiriou, Anastasios Tefas, and Ioannis Pitas

Kernel Discriminant Analysis for Regression Problems

Face Recognition in Subspaces

When Fisher meets Fukunaga-Koontz: A New Look at Linear Discriminants

Deriving Principal Component Analysis (PCA)

Principal Component Analysis and Linear Discriminant Analysis

PCA, Kernel PCA, ICA

Principal Component Analysis

Unsupervised Learning of Hierarchical Models. in collaboration with Josh Susskind and Vlad Mnih

Statistical Pattern Recognition

Lecture 13 Visual recognition

Machine Learning. Dimensionality reduction. Hamid Beigy. Sharif University of Technology. Fall 1395

Comparative Analysis of ICA Based Features

PCA and LDA. Man-Wai MAK

COS 429: COMPUTER VISON Face Recognition

Convergence of Eigenspaces in Kernel Principal Component Analysis

6.036 midterm review. Wednesday, March 18, 15

Data Mining. Dimensionality reduction. Hamid Beigy. Sharif University of Technology. Fall 1395

Jeff Howbert Introduction to Machine Learning Winter

Nonlinear Dimensionality Reduction. Jose A. Costa

Machine Learning (BSMC-GA 4439) Wenke Liu

Subspace Methods for Visual Learning and Recognition

Support Vector Machines II. CAP 5610: Machine Learning Instructor: Guo-Jun QI

Unsupervised Learning: K- Means & PCA

A Unified Bayesian Framework for Face Recognition

Machine Learning. B. Unsupervised Learning B.2 Dimensionality Reduction. Lars Schmidt-Thieme, Nicolas Schilling

Kernel Sliced Inverse Regression With Applications to Classification

MULTI-VARIATE/MODALITY IMAGE ANALYSIS

Principal Component Analysis (PCA)

Classification: The rest of the story

STA 414/2104: Lecture 8

Machine Learning 2nd Edition

Recognition Using Class Specific Linear Projection. Magali Segal Stolrasky Nadav Ben Jakov April, 2015

Fisher s Linear Discriminant Analysis

PATTERN RECOGNITION AND MACHINE LEARNING

Image Analysis & Retrieval. Lec 14. Eigenface and Fisherface

Pattern Recognition 2018 Support Vector Machines

Introduction to Machine Learning

Support Vector Machines for Classification: A Statistical Portrait

December 20, MAA704, Multivariate analysis. Christopher Engström. Multivariate. analysis. Principal component analysis

High Dimensional Discriminant Analysis

Statistical learning theory, Support vector machines, and Bioinformatics

Support Vector Machine (SVM) and Kernel Methods

Principal Component Analysis

COMP 551 Applied Machine Learning Lecture 13: Dimension reduction and feature selection

ECE 521. Lecture 11 (not on midterm material) 13 February K-means clustering, Dimensionality reduction

Kernel Methods. Barnabás Póczos

Face Detection and Recognition

Department of Computer Science and Engineering

PCA and LDA. Man-Wai MAK

Support Vector Machine (continued)

Advanced Introduction to Machine Learning CMU-10715

Aruna Bhat Research Scholar, Department of Electrical Engineering, IIT Delhi, India

Midterm Review CS 6375: Machine Learning. Vibhav Gogate The University of Texas at Dallas

ROBERTO BATTITI, MAURO BRUNATO. The LION Way: Machine Learning plus Intelligent Optimization. LIONlab, University of Trento, Italy, Apr 2015

Image Analysis & Retrieval Lec 14 - Eigenface & Fisherface

Methods for sparse analysis of high-dimensional data, II

Introduction to machine learning and pattern recognition Lecture 2 Coryn Bailer-Jones

BANA 7046 Data Mining I Lecture 6. Other Data Mining Algorithms 1

ADAPTIVE QUASICONFORMAL KERNEL FISHER DISCRIMINANT ANALYSIS VIA WEIGHTED MAXIMUM MARGIN CRITERION. Received November 2011; revised March 2012

Robot Image Credit: Viktoriya Sukhanova 123RF.com. Dimensionality Reduction

Linear Dimensionality Reduction

Individual Stable Space: An Approach to Face Recognition Under Uncontrolled Conditions

The role of dimensionality reduction in classification

Classification. The goal: map from input X to a label Y. Y has a discrete set of possible values. We focused on binary Y (values 0 or 1).

Pattern Recognition 2

Likelihood Ratio in a SVM Framework: Fusing Linear and Non-Linear Face Classifiers

CITS 4402 Computer Vision

A PRACTICAL APPLICATION OF KERNEL BASED FUZZY DISCRIMINANT ANALYSIS

Constructing Optimal Subspaces for Pattern Classification. Avinash Kak Purdue University. November 16, :40pm. An RVL Tutorial Presentation

Lecture 16: Small Sample Size Problems (Covariance Estimation) Many thanks to Carlos Thomaz who authored the original version of these slides

CMU-Q Lecture 24:

COMP 408/508. Computer Vision Fall 2017 PCA for Recognition

Transcription:

Subspace Analysis for Facial Image Recognition: A Comparative Study Yongbin Zhang, Lixin Lang and Onur Hamsici

Outline 1. Subspace Analysis: Linear vs Kernel 2. Appearance-based Facial Image Recognition. 3. Databases : FERET and JAFFE. 4. Experimental Results. 5. Concluding Remarks.

Subspace Approaches Linear and Nonlinear Methods Linear Subspace Unsupervised PCA ICA Nonlinear Subspace Kernel PCA Kernel ICA Kernel LDA Supervised LDA

Subspace Approaches: Linear Subspace Principle Component Analysis (PCA) Basic Idea Assume samples X = [x 1, x 2,, x p ] T Y i = f it X, f it f i = 1 f 1 = argmax_f 1 Var(Y 1 ) f 2 = argmax_f 2 Var(Y 2 ), Cov(Y 1, Y 2 )=0 Data redundancy ---- measured by correlation

Subspace Approaches: Linear Subspace : PCA PCA Algorithm Subspace basis C f = i f i? i where C is Sample Covariance, and? 1 =? 2 = =? p Projection to subspace Y = F T ( X - E(X) ) where F = [f 1 f 2,, f m ], m = p

Subspace Approaches: Linear Subspace : PCA PCA Illustration ϕ 2 x 2 ϕ 1 x 1

Subspace Approaches: Linear Subspace Independent Component Analysis (ICA) Mixing Process Separating Process s A X W u Source Observation Estimation Observation : X = As Estimation : u = Wx = WAs, estimation of s Courtesy of Bruce A Draper et al.

Subspace Approaches: Linear Subspace : ICA ICA Algorithm No closed-form solution. Minimizing mutual information Maximizing Non-Gaussianity Kurtosis: E{x 4 ) - 3(E{x 2 }) 2 Kurtosis is 0 for Gaussian random variables InfoMax Principle

Subspace Approaches: Linear Subspace Linear Discrimination Analysis (LDA) Basic Idea: w = arg max S S b w w w T T S S : between - class scatter matrix : within - class scatter matrix b w w w Explicit knowledge of class information

Subspace Approaches: Linear Subspace : LDA LDA Algorithm Generalized Eigenvector and Eigenvalue S b φ = i λ i S W φ i

Subspace Approaches: Linear Subspace : LDA LDA Illustration

Subspace Approaches: Nonlinear Subspace Kernel Version PCA, ICA and LDA are linear in nature. Sample data are nonlinear in general. Map to a nonlinear space and perform PCA, ICA and LDA PCA - KPCA ICA - KICA LDA - KLDA (KDA)

Subspace Approaches: Nonlinear Subspace : Kernel Version Illustration Courtesy of Schelkolpf et.al

Subspace Approaches: Nonlinear Subspace : Kernel Version Kernel Trick Compute dot production without mapping function Kernel Functions Polynomial : k(x, y) = (x y) d Gaussian : k(x, y) = exp - x-y 2s 2 2

Face Identification Appearance-Based Technique Feature: the intensity of all the pixels in the original image(or image after filtering). - Pros: Simple, No explicit model needed. - Cons: Two many variables, most of which dependent.

Face Identification : Appearance-Based Technique Why Dimension Reduction is Important? Curse of Dimensionality. To achieve generalization, we need at least ten times as many training samples per class as the feature dimension. Reduce Computational Complexity and Storage. Techniques: PCA, LDA,ICA vs Kernel Version

Face Identification Subspace Approach : Face Training Images Test Image

Face Identification Subspace Approach : Expression Test Image

Face Identification Data Base : Facial Images FERET(Face Recognition Technology) 195 subjects, each with images of 9 different poses. We will test identity recognition across poses

Face Identification Data Base : Expression Images JAFFE(Japanese female expression) 213 images, 7 expression(neutral + 6 basic emotions)

Face Identification Image Pretreat Face Processing: Detection, Warping, Alignment. Detection Warp Crop Align SVM-based detector Heisel et.al 2001.

Face Identification Experimental Results: Identity Recognition Leave-one-pose out test. Linear Subspace

Face Identification Experimental Results: Identity Recognition KPCA with different degree of polynomials Fractional Power: d<1(liu et.al. 2004)

Face Identification Experimental Results: Identity Recognition KLDA with different degree of polynomials

Face Identification Experimental Results: Facial Expression Hard label is not suitable! We will rate the images and compare with what human does. c h v, v R = c h v v

Face Identification Experimental Results: Facial Expression Average: PCA : 89%, ICA: 93%

Face Identification Experimental Results: Facial Expression Average: 88%~ 89%

Face Identification Conclusion and discussion Identity: KLDA performs the best. LDA>PCA KPCA>ICA Expression: ICA performs the best. Others performs about the same. Training Time: ICA>KLDA>KPCA LDA>PCA

Face Identification Conclusion and discussion Why ICA is poor in identity, While good in expression? Overfit in case 1. Good fit in case 2 Why d=0.5 performs the best for KPCA in identity recognition? Future Work: Data might be fitted well by sub-gaussian. Adaptively chose kernels from data!