Comparing Robustness of Pairwise and Multiclass Neural-Network Systems for Face Recognition

Similar documents
Multi-Class Linear Dimension Reduction by. Weighted Pairwise Fisher Criteria

Classification of Ordinal Data Using Neural Networks

COMP-4360 Machine Learning Neural Networks

Iterative Laplacian Score for Feature Selection

ARTIFICIAL NEURAL NETWORKS گروه مطالعاتي 17 بهار 92

Shankar Shivappa University of California, San Diego April 26, CSE 254 Seminar in learning algorithms

Lecture 7 Artificial neural networks: Supervised learning

HYPERGRAPH BASED SEMI-SUPERVISED LEARNING ALGORITHMS APPLIED TO SPEECH RECOGNITION PROBLEM: A NOVEL APPROACH

Applied Statistics. Multivariate Analysis - part II. Troels C. Petersen (NBI) Statistics is merely a quantization of common sense 1

Artificial Neural Network

A summary of Deep Learning without Poor Local Minima

Holdout and Cross-Validation Methods Overfitting Avoidance

Machine learning for pervasive systems Classification in high-dimensional spaces

A Unified Bayesian Framework for Face Recognition

Constrained Optimization and Support Vector Machines

CS 188: Artificial Intelligence. Outline

Linking non-binned spike train kernels to several existing spike train metrics

Object Recognition Using a Neural Network and Invariant Zernike Features

Artificial Neural Networks

Artificial Neural Networks

Learning Gaussian Process Models from Uncertain Data

Neural Networks, Computation Graphs. CMSC 470 Marine Carpuat

Lecture 4: Feed Forward Neural Networks

BANA 7046 Data Mining I Lecture 6. Other Data Mining Algorithms 1

Neural Networks and the Back-propagation Algorithm

EEE 241: Linear Systems

Discriminant Uncorrelated Neighborhood Preserving Projections

Exploiting Sparse Non-Linear Structure in Astronomical Data

Unit 8: Introduction to neural networks. Perceptrons

Non-parametric Classification of Facial Features

Artificial Neural Networks

Learning and Memory in Neural Networks

Multiple Similarities Based Kernel Subspace Learning for Image Classification

Lecture 4: Perceptrons and Multilayer Perceptrons

Kernel Sliced Inverse Regression With Applications to Classification

Robust Multiple Estimator Systems for the Analysis of Biophysical Parameters From Remotely Sensed Data

Application of Artificial Neural Networks in Evaluation and Identification of Electrical Loss in Transformers According to the Energy Consumption

Machine learning comes from Bayesian decision theory in statistics. There we want to minimize the expected value of the loss function.

Cheng Soon Ong & Christian Walder. Canberra February June 2018

Artificial Neural Networks. Q550: Models in Cognitive Science Lecture 5

Hu and Zernike Moments for Sign Language Recognition

Machine Learning for Large-Scale Data Analysis and Decision Making A. Neural Networks Week #6

ABSTRACT INTRODUCTION

100 inference steps doesn't seem like enough. Many neuron-like threshold switching units. Many weighted interconnections among units

Face Recognition Using Laplacianfaces He et al. (IEEE Trans PAMI, 2005) presented by Hassan A. Kingravi

Linear vs Non-linear classifier. CS789: Machine Learning and Neural Network. Introduction

VC-dimension of a context-dependent perceptron

Artificial Neural Networks

Real Estate Price Prediction with Regression and Classification CS 229 Autumn 2016 Project Final Report

A Novel Rejection Measurement in Handwritten Numeral Recognition Based on Linear Discriminant Analysis

OVERLAPPING ANIMAL SOUND CLASSIFICATION USING SPARSE REPRESENTATION

Discriminative Fields for Modeling Spatial Dependencies in Natural Images

Discriminative Direction for Kernel Classifiers

Invariant Pattern Recognition using Dual-tree Complex Wavelets and Fourier Features

Learning Binary Classifiers for Multi-Class Problem

Determining the Optimal Decision Delay Parameter for a Linear Equalizer

Artificial Neural Network

ECE662: Pattern Recognition and Decision Making Processes: HW TWO

Power Supply Quality Analysis Using S-Transform and SVM Classifier

Introduction to Deep Learning

Hierarchical Boosting and Filter Generation

STUDY ON METHODS FOR COMPUTER-AIDED TOOTH SHADE DETERMINATION

EEL 851: Biometrics. An Overview of Statistical Pattern Recognition EEL 851 1

ARTIFICIAL NEURAL NETWORK WITH HYBRID TAGUCHI-GENETIC ALGORITHM FOR NONLINEAR MIMO MODEL OF MACHINING PROCESSES

Unsupervised Clustering of Human Pose Using Spectral Embedding

Search for Inspiral GW Signals Associated with Short GRBs using Artificial Neural Networks

Support Vector Machine. Industrial AI Lab.

Analysis of Multilayer Neural Network Modeling and Long Short-Term Memory

Iterative face image feature extraction with Generalized Hebbian Algorithm and a Sanger-like BCM rule

Support Vector Machine (SVM) and Kernel Methods

Support Vector Machine (SVM) and Kernel Methods

BAYESIAN CLASSIFICATION OF HIGH DIMENSIONAL DATA WITH GAUSSIAN PROCESS USING DIFFERENT KERNELS

Analysis of Multiclass Support Vector Machines

A Tutorial on Support Vector Machine

Prediction of gas emission quantity using artificial neural networks

Artificial Neural Networks Examination, June 2004

CSC242: Intro to AI. Lecture 21

Convolutional Associative Memory: FIR Filter Model of Synapse

Approximate Inference Part 1 of 2

Neural Networks DWML, /25

Mining Classification Knowledge

Machine Learning. Ensemble Methods. Manfred Huber

Robust Inverse Covariance Estimation under Noisy Measurements

Approximate Inference Part 1 of 2

Speaker Representation and Verification Part II. by Vasileios Vasilakakis

An artificial neural networks (ANNs) model is a functional abstraction of the

Non-negative Matrix Factorization on Kernels

Application of hopfield network in improvement of fingerprint recognition process Mahmoud Alborzi 1, Abbas Toloie- Eshlaghy 1 and Dena Bazazian 2

Intelligent Modular Neural Network for Dynamic System Parameter Estimation

In the Name of God. Lectures 15&16: Radial Basis Function Networks

Artificial Intelligence

Principal Component Analysis (PCA)

Analysis and visualization of protein-protein interactions. Olga Vitek Assistant Professor Statistics and Computer Science

Artificial Neural Networks Francesco DI MAIO, Ph.D., Politecnico di Milano Department of Energy - Nuclear Division IEEE - Italian Reliability Chapter

Convolutional networks. Sebastian Seung

Study on Classification Methods Based on Three Different Learning Criteria. Jae Kyu Suhr

Rapid Object Recognition from Discriminative Regions of Interest

Introduction to Support Vector Machines

ARTIFICIAL INTELLIGENCE. Artificial Neural Networks

Machine Learning Techniques

Transcription:

Comparing Robustness of Pairwise and Multiclass Neural-Network Systems for Face Recognition J. Uglov, V. Schetinin, C. Maple Computing and Information System Department, University of Bedfordshire, Luton, UK Abstract. Noise, corruptions and variations in face images can seriously hurt the performance of face recognition systems. To make such systems robust, multiclass neuralnetwork classifiers capable of learning from noisy data have been suggested. However on large face data sets such systems cannot provide the robustness at a high level. In this paper we explore a pairwise neural-network system as an alternative approach to improving the robustness of face recognition. In our experiments this approach is shown to outperform the multiclass neural-network system in terms of the predictive accuracy on the face images corrupted by noise. 1. Introduction Performance of face recognition systems is achieved at a high level when such systems are robust to noise, corruptions and variations in face images [1]. To make face recognition systems robust, multiclass artificial neural networks (ANNs) capable of learning from noisy data have been suggested [1]. However on large face data sets such neural-network systems cannot provide the robustness at a high level [1] - [3]. To overcome this problem pairwise classification systems have been proposed, see e.g. [3], [4]. In this paper we explore a pairwise neural-network system as an alternative approach to improving the robustness. In our experiments this approach is shown to outperform the multiclass neural-network system in terms of the predictive accuracy on the face image data described in [5]. In section 2 we briefly describe face image representation and noise problems, and then in section 3 we describe a pairwise neural-network system proposed for face recognition. Section 4 describes our experiments and finally section 5 concludes the paper. 2. Face Image Representation and Noise Problems Following to [1] [3], we use the principal component analysis (PCA) to represent face images as m-dimensional vectors of components. The PCA is the common technique for data representation in face recognition systems. The first two principal components which make the most important contribution in face recognition can be used to visualise the scatter of patterns of different classes (faces). Therefore the use of such a visualisation allows us to observe how the noise can corrupt the boundaries of classes. For example, Fig. 1 shows two graphs depicting the examples of four classes whose centres of gravity are visually distinct. The left side plot depicts the examples taken from the original data while the right side plot depicts these examples containing

noise components drawn from a Gaussian density function with zero mean and the standard deviation alpha = 0.5. Fig. 1. An example of scattering the samples drawn from the four classes for alpha = 0 (the left side) and alpha = 0.5 (the right side) in a plane of the first two principal components p 1 and p 2. From this plot we can observe that the noise components corrupt the boundary of the given classes, and therefore the performance of a face recognition system can be affected. From these plots we can also observe that the boundaries between pairs of the classes can remain almost the same. This inspire us to exploit such a classification scheme to implement a pairwise neural-network system for face recognition. 3. A Pairwise Neural-Network System The idea behind the pairwise classification is to use two-class ANNs learning to classify all possible pairs of classes. Therefore for C classes the pairwise system should include C*(C 1)/2 ANNs learnt to solve two-class problems. For example, for classes 1, 2, and 3 depicted in Fig. 2, the number of two-class ANNs is equal to 3. In this figure the lines f 1/2, f 1/3 and f 2/3 are the dividing hyperplanes learnt by the ANNs. We can simply assume these functions are given the positive values for examples of the classes standing first in the lower indexes (1, 1, and 2) and the negative values for the classes standing second in there (2, 3, and 3).

f 1/3 1 g 1 = f 1/2 + f 1/3 f 1/2 3 2 g 3 = f 1/3 f 2/3 g 2 = f 2/3 f 1/2 f 2/3 x Fig. 2. Hyperplanes f 1/2, f 1/3 and f 2/3 learnt to divide the following pairs of classes: 1 versus 2, 1 versus 3, and 2 versus 3. Now we can combine hyperplanes f1/2, f1/3 and f2/3 to build up the new dividing hyperplanes g1, g2, and g3. The first hyperplane g1 combines the functions f1/2 and f1/3 so that g 1 = f 1/2 + f 1/3. These functions are taken with weights of 1.0 because both functions f 1/2 and f 1/3 give the positive output values on the examples of class 1. Likewise, the second and third hyperplanes are as follows: g2 = f2/3 f1/2 and g3 = f1/3 f2/3. In practice each of hyperplanes g1,, gc, can be implemented as a two-layer feedforward ANN with a given number of hidden neurons fully connected to the input nodes. Then we can introduce the output neuron summing all outputs of the ANNs to make a final decision. For example, the pairwise neural-network system depicted in Fig 3 consists of three neural networks performing the functions f1/2, f1/3, and f2/3. The three output neurons g1, g2, and g 3 are connected to these networks with weights equal to (+1, +1), ( 1, +1) and ( 1, 1), respectively. f 1/2 +1 x 1 +1 g 1 C 1 C 2 x 2 f 1/3 +1 g 2 x m g 3 C 3 f 2/3 Fig. 3. An example of pairwise neural-network system for C = 3 classes. In general, a pairwise neural-network system consists of C(C 1)/2 neural networks, performing functions f 1/2,, f i/j,, f C - 1/C, and C output neurons g 1,, g C, where i < j = 2,, C. We can see that the weights of output neuron g i connected to the hidden neurons fi/k and fk/i should be equal to + 1 and 1, respectively.

4. Experiments The goal of our experiments is to compare the robustness of the proposed pairwise and standard multiclass neural-network systems on the Cambridge ORL face image data set [5] (in a full paper, the experiments will run on different face image data sets). To estimate the robustness we add noise components to the data and then estimate the performance on the test data within 5 fold cross-validation. The performances of the pairwise and multiclass systems are listed in Table 1 and shown in Fig. 4. Table 1. Performance of the pairwise (P) and multiclass (M) neural-network systems over alpha. The performances are represented by the means and 2 intervals. alpha 0.0 0.1 0.3 0.5 0.7 0.9 1.1 1.3 P, mean 0.972 0.966 0.953 0.920 0.859 0.772 0.659 0.556 P, 2 ±0.004 ±0.013 ±0.017 ±0.013 ±0.018 ±0.030 ±0.028 ±0.031 M, mean 0.952 0.951 0.932 0.898 0.802 0.678 0.557 0.419 M, 2 ±0.017 ±0.016 ±0.025 ±0.016 ±0.015 ±0.052 ±0.036 ±0.050 From this table we can see that for alpha ranging between 0.0 and 1.3 the proposed pairwise system significantly outperforms the multiclass systems. For alpha = 0.0 the improvement in the performance is 2.0% while for alpha = 1.1 the improvement becomes 10.2%. Fig. 4. Performance of the pairwise and multiclass neural-network systems over alpha. Solid lines and bars are the mean and 2 intervals, respectively.

5. Conclusion We have proposed a pairwise neural-network system for face recognition in order to reduce the negative effect of noise and corruptions in face images. Within such a classification scheme we expect that the improvement in the performance can be achieved on the base of our observation that boundaries between pairs of classes remain almost the same while a noise level increases. We have compared the performances of the proposed pairwise and standard multiclass neural-network systems on the face dataset [5]. Evaluating the mean values and standard deviations of the performances under different levels of noise in the data, we have found that the proposed pairwise system is superior to the multiclass neural-network system. Thus we conclude that the proposed pairwise system is capable of decreasing the negative effect of noise and corruptions in face images. Clearly this is a very desirable property for face recognition systems when the robustness is of crucial importance. 6. References 1. S.Y. Kung, M.W. Mak and S.H. Lin. Biometric Authentication: A Machine Learning Approach. Pearson Education, 2005 2. C. Liu and H. Wechler. Robust coding scheme for indexing and retrieval from large face database. IEEE Trans Image Processing, 9(1), 13237, 2000 3. A.S. Tolba, A.H. El-Baz and A.A. El-Harby. Face Recognition: A Literature Review. IJSP. 2(2), 8803, 2005 4. T. Hastie and R. Tibshirani. Classification by pairwise coupling. Advances in NIPS, 10, 507-513, 1998 5. E.S. Samaria. Face recognition using hidden Markov models. PhD thesis. University of Cambridge, 1994