Non-parametric Classification of Facial Features

Similar documents
Reconnaissance d objetsd et vision artificielle

Eigenface-based facial recognition

Face Recognition. Face Recognition. Subspace-Based Face Recognition Algorithms. Application of Face Recognition

COS 429: COMPUTER VISON Face Recognition

Pattern Recognition 2

CITS 4402 Computer Vision

Keywords Eigenface, face recognition, kernel principal component analysis, machine learning. II. LITERATURE REVIEW & OVERVIEW OF PROPOSED METHODOLOGY

Example: Face Detection

Principal Component Analysis -- PCA (also called Karhunen-Loeve transformation)

Lecture: Face Recognition

Lecture 24: Principal Component Analysis. Aykut Erdem May 2016 Hacettepe University

Face Detection and Recognition

Dimensionality Reduction Using PCA/LDA. Hongyu Li School of Software Engineering TongJi University Fall, 2014

Recognition Using Class Specific Linear Projection. Magali Segal Stolrasky Nadav Ben Jakov April, 2015

Principal Component Analysis

EEL 851: Biometrics. An Overview of Statistical Pattern Recognition EEL 851 1

Deriving Principal Component Analysis (PCA)

PCA FACE RECOGNITION

Two-Layered Face Detection System using Evolutionary Algorithm

Advanced Introduction to Machine Learning CMU-10715

PATTERN CLASSIFICATION

Nonparametric Methods Lecture 5

Face detection and recognition. Detection Recognition Sally

Discriminant Uncorrelated Neighborhood Preserving Projections

Lecture 13 Visual recognition

Eigenimaging for Facial Recognition

Robot Image Credit: Viktoriya Sukhanova 123RF.com. Dimensionality Reduction

Eigenfaces. Face Recognition Using Principal Components Analysis

BACKPROPAGATION. Neural network training optimization problem. Deriving backpropagation

p(d θ ) l(θ ) 1.2 x x x

CS4495/6495 Introduction to Computer Vision. 8C-L3 Support Vector Machines

Study on Classification Methods Based on Three Different Learning Criteria. Jae Kyu Suhr

Course 495: Advanced Statistical Machine Learning/Pattern Recognition

STUDY ON METHODS FOR COMPUTER-AIDED TOOTH SHADE DETERMINATION

Pattern Recognition Problem. Pattern Recognition Problems. Pattern Recognition Problems. Pattern Recognition: OCR. Pattern Recognition Books

2D Image Processing Face Detection and Recognition

When Fisher meets Fukunaga-Koontz: A New Look at Linear Discriminants

Image Analysis & Retrieval. Lec 14. Eigenface and Fisherface

Lecture 17: Face Recogni2on

Image Region Selection and Ensemble for Face Recognition

Comparative Assessment of Independent Component. Component Analysis (ICA) for Face Recognition.

Learning with multiple models. Boosting.

p(x ω i 0.4 ω 2 ω

CS4495/6495 Introduction to Computer Vision. 8B-L2 Principle Component Analysis (and its use in Computer Vision)

Dimension Reduction and Component Analysis

Dimension Reduction and Component Analysis Lecture 4

W vs. QCD Jet Tagging at the Large Hadron Collider

Feature Extraction with Weighted Samples Based on Independent Component Analysis

Statistical and Learning Techniques in Computer Vision Lecture 2: Maximum Likelihood and Bayesian Estimation Jens Rittscher and Chuck Stewart

Instance-based Learning CE-717: Machine Learning Sharif University of Technology. M. Soleymani Fall 2016

Symmetric Two Dimensional Linear Discriminant Analysis (2DLDA)

Image Analysis & Retrieval Lec 14 - Eigenface & Fisherface

Parametric Techniques

Classification: The rest of the story

ECE 661: Homework 10 Fall 2014

Holdout and Cross-Validation Methods Overfitting Avoidance

A Unified Bayesian Framework for Face Recognition

L11: Pattern recognition principles

WHEN an object is represented using a linear combination

Face recognition Computer Vision Spring 2018, Lecture 21

Multivariate statistical methods and data mining in particle physics

Comparing Robustness of Pairwise and Multiclass Neural-Network Systems for Face Recognition

arxiv: v3 [cs.lg] 18 Mar 2013

Machine Learning 2nd Edition

Learning Kernel Parameters by using Class Separability Measure

Lecture 16: Small Sample Size Problems (Covariance Estimation) Many thanks to Carlos Thomaz who authored the original version of these slides

Face Recognition Using Eigenfaces

Machine learning for pervasive systems Classification in high-dimensional spaces

Parametric Techniques Lecture 3

ISSN: Jurnal. Ilmiah. eknik. omputer. Kelompok Kellmuan Teknlk Komputer Sekolah Teknlk Elektro a Informatlka Instltut Teknologl Bandung

Image Analysis. PCA and Eigenfaces

Eigenfaces and Fisherfaces

Worst-Case Analysis of the Perceptron and Exponentiated Update Algorithms

Time Series Classification

Drift Reduction For Metal-Oxide Sensor Arrays Using Canonical Correlation Regression And Partial Least Squares

EM in High-Dimensional Spaces

The Mathematics of Facial Recognition

Clustering VS Classification

20 Unsupervised Learning and Principal Components Analysis (PCA)

A Novel Activity Detection Method

Aruna Bhat Research Scholar, Department of Electrical Engineering, IIT Delhi, India

Learning Methods for Linear Detectors

SPECTRAL CLUSTERING AND KERNEL PRINCIPAL COMPONENT ANALYSIS ARE PURSUING GOOD PROJECTIONS

Lecture 17: Face Recogni2on

Final project STK4030/ Statistical learning: Advanced regression and classification Fall 2015

Principal Component Analysis and Singular Value Decomposition. Volker Tresp, Clemens Otte Summer 2014

Automatic Identity Verification Using Face Images

Discriminative Direction for Kernel Classifiers

Algorithm-Independent Learning Issues

Midterm: CS 6375 Spring 2015 Solutions

Orthogonal Laplacianfaces for Face Recognition

Pattern Classification

Unsupervised Learning: K- Means & PCA

Problem Session #5. EE368/CS232 Digital Image Processing

ABSTRACT INTRODUCTION

Multilayer Neural Networks. (sometimes called Multilayer Perceptrons or MLPs)

Learning Linear Detectors

Simultaneous and Orthogonal Decomposition of Data using Multimodal Discriminant Analysis

Principal Component Analysis (PCA)

Face Recognition Using Multi-viewpoint Patterns for Robot Vision

Transcription:

Non-parametric Classification of Facial Features Hyun Sung Chang Department of Electrical Engineering and Computer Science Massachusetts Institute of Technology Problem statement In this project, I attempted to classify facial images based on various external characteristics, such as gender, expression, and accessories they are taking on. Rather than extracting any particular parameters describing faces, e.g., the distances among eyes, nose, and mouse, I used grey-scale face images themselves, fitted to 128x128 window, as the inputs. Dataset The set used for this project together with detailed description is available here at the course website. The set consists of 2,000 training face images (facer, 1,997 of them labeled) and 2,000 test face images (faces, 1,996 of them labeled). Because the image size is 128x128, each image can be considered as a point in a huge dimensional space. The dimensionality reduction has been conducted using principal component analysis (PCA) on 100 sample faces, all from the training set, so each image can be represented by 99 eigenface coefficients, as well as the mean face. The composition of set is shown in Table 1. For example, notice that, in terms of expression, funny faces were significantly fewer than the other two classes and that few people wore glasses or bandana. One interesting thing is that no bandana image was included in the samples used to generate the eigenfaces.

Table 1. Dataset composition gender expression male female serious smiling funny Eigenface generating 61/100 39/100 Eigenface generating 45/100 51/100 4/100 Training (facer) Testing (faces) 1,150/1,997 847/1,997 Training (facer) 1,277/1,996 719/1,996 Testing (faces) 917/1,997 1,043/1,997 37/1,997 1,097/1,996 836/1,996 63/1,996 Eigenface generating Training (facer) Testing (faces) glasses bandana on off on off 4/100 96/100 Eigenface 0/100 100/100 generating 59/1,997 1,938/1,997 Training 13/1,997 1,984/1,997 (facer) 8/1,996 1,988/1,996 Testing 8/1,996 1,988/1,996 (faces) Objective of this project The objective of this project lies in two aspects: 1) to practice with meaningful classification problem using the methods learned from the class; 2) to look into inherent limitations of PCA approach. Eigenface representation Let be eigenfaces and be the sample faces used to generate the set of eigenfaces. The PCA finds so that can be well represented by their linear combinations. Let be an arbitrary face and be its eigenface representation, that is,.

Note that is just a linear combination of, which implies that the sensitivity issue should be aroused. For example, because there was no bandana image in, we may see that and may be significantly different from each other if is a facial image of the person who is wearing a bandana. The approximation error between and can be measured in terms of peak-signal-tonoise-ratio (PSNR) defined by where, is the number of pixels. Figure 1 shows a bandana image example. The PSNR is as low as 14.47dB. Note that, in eigenface representation, other regions than the bandana pixels were also severely distorted as a result of making best efforts to compensate for the bandana region. This may lead to the classification errors even also for the other criteria, not only for the bandana. Figure 1. Bandana image example and its eigenface approximation. Figure 2 shows the actual PSNR distribution for the training set and the test set. The images in the test set show somewhat low PSNR; for some particular samples, the PSNR was significantly low. This low PSNR may contribute to the classification error.

Figure 2. PSNR distribution of each face image in training set (facer) and in test set (faces). (left: facer, right: faces) Figure 3 illustrates how the discriminant value and the PSNR for gender classification (+ = male, - = female) when a linear discriminant was used. Note that a majority of male samples of high PSNR were correctly classified. For the female samples, such a correlation was not that noticeable; instead, the discriminant value and the PSNR looked rather uncorrelated. I think this may be because of the inherent ill-performance of our classifier against female images (See Table 2)

Figure 3. Plot of discriminant value versus PSNR for male and female face samples. (left: male, right: female) Classification practice For this part of experiment, the following classification schemes were tested: k-nn method (k-nn) Linear discriminant (LD) Neural network with one-hidden layer (NN-2) Neural network with two-hidden layers (NN-3) Also the performance may have to be compared to random guess schemes. RG-1: choose the class whose prior probability is maximum RG-2: choose the class randomly according to their prior probabilities Table 2 through Table 5 show the classification results. In most cases, LD and NN-2 showed best performance than the other two schemes and also than the two RG schemes. Nearly all classifiers failed to detect the glasses, bandana, and also funny expression which are all characterized as extreme minority, i.e., whose prior probability is very low. Table 2. Comparison of k-nn ( ), LD, NN-2 ( ), NN-3 ( ), RG- 1, RG-2 for gender classification. k-nn detect miss LD detect miss male 823 454 male 1,026 251 female 402 317 female 375 344 NN-2 detect miss NN-3 detect miss male 1,008 269 male 763 514 female 378 341 female 544 175 RG-1 detect miss RG-2 detect miss male 1,277 0 male 753 542 female 0 719 female 305 414 Table 3. Comparison Comparison of k-nn ( ), LD, NN-2 ( ), NN-3 ( ), RG-1, RG-2 for expression classification.

k-nn detect miss LD detect miss serious 586 511 serious 936 161 smiling 468 368 smiling 623 213 funny 0 63 funny 0 63 NN-2 detect miss NN-3 detect miss serious 932 165 serious 963 134 smiling 617 219 smiling 593 243 funny 0 63 funny 0 63 RG-1 detect miss RG-2 detect miss serious 0 1,097 serious 504 593 smiling 836 0 smiling 437 399 funny 0 63 funny 1 62 Table 4. Comparison of k-nn ( ), LD, NN-2 ( ), NN-3 ( ), RG- 1, RG-2 for glasses detection. k-nn detect miss LD detect miss off 1,988 0 off 1,986 2 NN-2 detect miss NN-3 detect miss on 2 6 on 0 8 off 1,962 16 off 1,958 30 RG-1 detect miss RG-2 detect miss off 1,988 0 off 1,988 0 Table 5. Comparison of k-nn ( ), LD, NN-2 ( ), NN-3 ( ), RG- 1, RG-2 for bandana detection. k-nn detect miss LD detect miss off 1,988 0 off 1,988 0 NN-2 detect miss NN-3 detect miss off 1,988 0 off 1,986 2

RG-1 detect miss RG-2 detect miss off 1,988 0 off 1,988 0 From this experiment, I concluded that 1) Samples from minority class (with very low prior probability) tend to be missclassified with any classifier; 2) Eigenface approach is good for identity recognition purpose, robust to noise and partial loss of, but not as good for classification purpose dealing with extraneous face samples, i.e., not used for the eigenface generation. Remarks After Monday presentation, I applied AdaBoost on LD and Parzen window for each classification and obtained preliminary results, but the classification performance was not improved so much. Particularly, I am looking into the working details for Parzen window since my preliminary results were far from those in [2]. Mostly due to limited time, multilinear analysis method has not been attempted. Future direction of study should include the analytical and experimental study of multi-linear analysis method. References [1] Richard O. Duda, Peter E. Hart, and David G. Stork, Pattern Classification. New York, NY: John Wiley & Sons, 2001. [2] Tiwari, Face recognition: eigenfaces with 99 PCA coefficients, MAS.622J/1.126J Project Presentation, MIT, Fall 2004. [3] W. S. Yambor, Analysis of PCA-based and Fisher discriminant-based image recognition algorithms, Master's thesis, Dept. of Comp. Sci., Colorado State Univ., July 2000. [4] M. Turk and A. Pentland, Eigenfaces for recognition, J. Cogn. Neurosci., vol. 3, no. 1, pp. 71-86, 1991. [5] H. A. Rowely, S. Baluja, and T. Kanade, Neural network based face detection, IEEE Trans. Pattern Anal. Machine Intell., vol. 20, no. 1, pp. 23-38, Jan. 1998. [pdf] [6] Face recognition homepage. [Online]. Available: http://www.face-rec.org