Score Normalization in Multimodal Biometric Systems

Similar documents
Likelihood Ratio Based Biometric Score Fusion

Biometric Fusion: Does Modeling Correlation Really Matter?

Analytical Study of Biometrics Normalization and Fusion Techniques For Designing a Multimodal System

Dynamic Linear Combination of Two-Class Classifiers

Semi-parametric Distributional Models for Biometric Fusion

A Confidence-Based Late Fusion Framework For Audio-Visual Biometric Identification

System Design and Performance Assessment: A Biometric Menagerie Perspective. Norman Poh Lecturer Dept of Computing

BIOMETRIC verification systems are used to verify the

I D I A P R E S E A R C H R E P O R T. Samy Bengio a. November submitted for publication

Likelihood Ratio in a SVM Framework: Fusing Linear and Non-Linear Face Classifiers

1 Biometric Authentication: A Copula Based Approach

Multimodal Biometric Fusion Joint Typist (Keystroke) and Speaker Verification

Theoretical Statistical Correlation for Biometric Identification Performance

Longitudinal Study of Fingerprint Recognition

Constrained Optimization and Support Vector Machines

Fusion of Dependent and Independent Biometric Information Sources

Likelihood Ratio-Based Biometric Verification

Meta-Recognition, Machine Learning and the Open Set Problem

Nearest Neighbourhood Classifiers in Biometric Fusion

Research Article Multimodal Personal Verification Using Likelihood Ratio for the Match Score Fusion

Time Series Classification

Fingerprint Individuality

Revisiting Doddington s Zoo: A Systematic Method to Assess User-dependent Variabilities

Biometric scores fusion based on total error rate minimization

Learning Classification with Auxiliary Probabilistic Information Quang Nguyen Hamed Valizadegan Milos Hauskrecht

Error Rates. Error vs Threshold. ROC Curve. Biometrics: A Pattern Recognition System. Pattern classification. Biometrics CSE 190 Lecture 3

Biometric Data Fusion Based on Subjective Logic 1

Biometrics: Introduction and Examples. Raymond Veldhuis

Role of Assembling Invariant Moments and SVM in Fingerprint Recognition

When enough is enough: early stopping of biometrics error rate testing

EEL 851: Biometrics. An Overview of Statistical Pattern Recognition EEL 851 1

Has this Person Been Encountered Before? : Modeling an Anonymous Identification System

If you wish to cite this paper, please use the following reference:

Classification CE-717: Machine Learning Sharif University of Technology. M. Soleymani Fall 2012

Support Vector Machines: Kernels

A Unified Bayesian Framework for Face Recognition

Probabilistic modeling. The slides are closely adapted from Subhransu Maji s slides

Design of Robust Classifiers for Adversarial Environments

probability of k samples out of J fall in R.

SVM based on personal identification system using Electrocardiograms

Score Level Fusion of Multimodal Biometrics based on Entropy Function

Study on Classification Methods Based on Three Different Learning Criteria. Jae Kyu Suhr

CSC 411: Lecture 04: Logistic Regression

PATTERN RECOGNITION AND MACHINE LEARNING CHAPTER 2: PROBABILITY DISTRIBUTIONS

Bayesian Decision Theory

σ(a) = a N (x; 0, 1 2 ) dx. σ(a) = Φ(a) =

Introduction to Support Vector Machines

Automated Statistical Recognition of Partial Discharges in Insulation Systems.

Ch 4. Linear Models for Classification

Machine Learning! in just a few minutes. Jan Peters Gerhard Neumann

Middlesex University Research Repository

Hybrid Multi-Biometric Person Authentication System

Support Vector Machine (SVM) & Kernel CE-717: Machine Learning Sharif University of Technology. M. Soleymani Fall 2012

Cheng Soon Ong & Christian Walder. Canberra February June 2018

A COMPLEXITY FRAMEWORK FOR COMBINATION OF CLASSIFIERS IN VERIFICATION AND IDENTIFICATION SYSTEMS

Robust Speaker Identification

Pattern Recognition and Machine Learning. Bishop Chapter 6: Kernel Methods

Classification 1: Linear regression of indicators, linear discriminant analysis

Large-scale Image Annotation by Efficient and Robust Kernel Metric Learning

Lecture 3: Multiclass Classification

Probabilistic classification CE-717: Machine Learning Sharif University of Technology. M. Soleymani Fall 2016

Application of hopfield network in improvement of fingerprint recognition process Mahmoud Alborzi 1, Abbas Toloie- Eshlaghy 1 and Dena Bazazian 2

Lecture 16: Small Sample Size Problems (Covariance Estimation) Many thanks to Carlos Thomaz who authored the original version of these slides

Artificial Neural Networks Examination, June 2004

Support Vector Machine (SVM) and Kernel Methods

This is an accepted version of a paper published in Elsevier Information Fusion. If you wish to cite this paper, please use the following reference:

Algorithm-Independent Learning Issues

Comparing Robustness of Pairwise and Multiclass Neural-Network Systems for Face Recognition

HMM and IOHMM Modeling of EEG Rhythms for Asynchronous BCI Systems

EXTRACTING BIOMETRIC BINARY STRINGS WITH MINIMAL AREA UNDER THE FRR CURVE FOR THE HAMMING DISTANCE CLASSIFIER

Machine Learning Ensemble Learning I Hamid R. Rabiee Jafar Muhammadi, Alireza Ghasemi Spring /

Class 4: Classification. Quaid Morris February 11 th, 2011 ML4Bio

Parametric Models. Dr. Shuang LIANG. School of Software Engineering TongJi University Fall, 2012

ECE521 week 3: 23/26 January 2017

Independent Component Analysis and Unsupervised Learning

Analysis of 2x2 Cross-Over Designs using T-Tests

Securing BioEncoded IrisCodes Against Correlation Attacks

Assignment No A-05 Aim. Pre-requisite. Objective. Problem Statement. Hardware / Software Used

Random Sampling LDA for Face Recognition

Palmprint based Verification System Robust to Occlusion using Low-order Zernike Moments of Sub-images

Classifier Evaluation. Learning Curve cleval testc. The Apparent Classification Error. Error Estimation by Test Set. Classifier

Curve Fitting Re-visited, Bishop1.2.5

Linear Models for Classification

Non-parametric Classification of Facial Features

Revision: Chapter 1-6. Applied Multivariate Statistics Spring 2012

Optimal Power Allocation for Distributed BLUE Estimation with Linear Spatial Collaboration

Independent Component Analysis and Unsupervised Learning. Jen-Tzung Chien

Principles of Pattern Recognition. C. A. Murthy Machine Intelligence Unit Indian Statistical Institute Kolkata

ECE 661: Homework 10 Fall 2014

OBJECT DETECTION AND RECOGNITION IN DIGITAL IMAGES

Machine Learning Basics Lecture 7: Multiclass Classification. Princeton University COS 495 Instructor: Yingyu Liang

Graphical Object Models for Detection and Tracking

Optimal Treatment Regimes for Survival Endpoints from a Classification Perspective. Anastasios (Butch) Tsiatis and Xiaofei Bai

Robust scale estimation with extensions

CSE 417T: Introduction to Machine Learning. Lecture 11: Review. Henry Chai 10/02/18

Linear Models for Classification

Classifier s Complexity Control while Training Multilayer Perceptrons

Unsupervised Anomaly Detection for High Dimensional Data

Classification and Pattern Recognition

Transcription:

Score Normalization in Multimodal Biometric Systems Karthik Nandakumar and Anil K. Jain Michigan State University, East Lansing, MI Arun A. Ross West Virginia University, Morgantown, WV http://biometrics.cse.mse.edu

Multimodal Biometric Systems Multiple sources of biometric information are integrated to enhance matching performance Increases population coverage by reducing failure to enroll rate Anti-spoofing; difficult to spoof multiple traits simultaneously Fingerprint Face Hand geometry Iris

Fusion in Multimodal Biometrics Biometric Fusion Prior to matching After matching Sensor Level Raw Data Feature Level Feature Sets Dynamic Classifier Selection Classifier Fusion Confidence Level Matching Scores Rank Level Class Ranks Abstract Level Class Labels

Fusion at the Matching Score Level Fusion at the matching score level offers the best tradeoff in terms of information content and ease in fusion Classification Approach Combination Approach Experiments indicate that the combination approach performs better than the classification approach* *A. Ross, A.K. Jain, Information Fusion in Biometrics, Pattern Recognition Letters, Sep. 2003

Fusion Rules* Problem: Classify input pattern Z into one of m possible classes (c 1,,c m ) based on evidence provided by R classifiers Let x i be the feature vector for the i th classifier derived from Z; x i s are independent Assign Z c j, if g(c j ) g(c k ), 1 k m, k j Product rule: g(c r ) = Sum rule: g(c r ) = R i= 1 R i= 1 P( c r x i ) P( c r x i ) Max rule: g(c r ) = max P( c i r x i ) Min rule: g(c r ) = min P( c i r x i ) *Kittler et al., On Combining Classifiers, IEEE PAMI, March 1998

Computing the Posteriori Probability To make use of the fusion rules, we need to compute P(c j x i ) At the matching score level, we have only the score s ij (not x i ) Verlinde et al. proposed that f is a monotonic function s ij = f(p(c j x i )) + η(x i ) η is the error introduced by the biometric system due to problems in acquisition and feature extraction processes To estimate η, the biometric system should output a matching score along with a confidence measure on that score indicating the quality of the input feature vector x i If η is known, we can estimate P(c j x i ) from s ij using nonparametric density estimation methods (e.g., Parzen window)

Score Normalization If η is unknown, the errors in the estimation of P(c j x i ) will be very large; hence, it is better to combine the scores directly Combination of scores has the following problems: Non-homogeneous scores: distance or similarity Ranges may be different; e.g., [0,100] or [0,1000] Distributions may be different Modify the location and scale parameters of score distributions of individual matchers to transform the scores into a common domain Robustness: Should not be affected by the outliers Efficiency: Estimated parameters of the score distribution should be close to the true values

Score Normalization Techniques Min-max normalization: Given matching scores {s k }, k=1,2,..,n the normalized scores are given by: s ' = s min{ s } max{ s } min{ s } k k k Decimal scaling: Used when scores of different matchers differ by a logarithmic factor; e.g., one matcher has scores in the range [0,1] and the other matcher has scores in the range [0, 1000] s s ' = n 10, n = log max{ s } 10 k

Z-score: Score Normalization Techniques s s ' = µ σ Median and Median Absolute Deviation (MAD): ( s median) s ' = MAD MAD = median({ s } median ) Double Sigmoid function: s ' = 1 s t 1+ exp 2 r r = r 1, if s < t r = r 2, otherwise k

Score Normalization Techniques Tanh estimators: s ( s µ ) GH ' = 0.5 tanh 0.01 + 1, σ GH where µ GH and σ GH are the mean and standard deviation estimates of the genuine score distribution as given by Hampel estimators* Min-max, Z-score, and Tanh normalization schemes are efficient Median, Double Sigmoid, and Tanh methods are robust *Hampel et al., Robust Statistics: The Approach Based on Influence Functions, 1986

Experimental Setup Face Hand-geometry Fingerprint Database of 100 users with three modalities (5 samples/user)

Experimental Setup Feature Extraction and Matching d f Normalization S f Eigenfaces Fusion Rule Decision Feature Extraction and Matching s p Normalization s p Minutiae Feature Extraction and Matching d h Normalization s h 14-D hand feature vector Fusion Rules Sum rule: mean(s f, s p, s h ) Min rule: min(s f, s p, s h ) Max rule: max(s f, s p, s h )

Distribution of Matching Scores Face Fingerprint Hand-geometry

Performance of Individual Modalities

Normalization followed by Sum Rule

Sensitivity of Min-max to Outliers

Sensitivity of Z-score to Outliers

Sensitivity of Tanh to Outliers

Summary The effects of different score normalization techniques in a multimodal biometric system have been studied Min-max, Z-score, and Tanh normalization techniques followed by sum rule resulted in the best recognition performance Min-max and Z-score methods are efficient but sensitive to outliers in the training data; Tanh is both efficient and robust If the parameters of the matching score distribution are known, simple methods like Min-max and Z-score would suffice If the parameters are to be estimated from noisy training data, one should choose robust methods like Tanh normalization If the system can provide both score and some confidence measure on that score, non-parametric density estimation methods can be used to estimate the posteriori probability

Efficiency of a Normalization Scheme Before Normalization After Sigmoid Normalization After Min-max Normalization After Tanh Normalization