IBM Research Report. The Relation between the ROC Curve and the CMC

Similar documents
IBM Research Report. Meertens Number and Its Variations

Predicting large population data cumulative match characteristic performance from small population data

IBM Research Report. Conditions for Separability in Generalized Laplacian Matrices and Diagonally Dominant Matrices as Density Matrices

IBM Research Report. On the Linear Relaxation of the p-median Problem I: Oriented Graphs

ISCA Archive

IBM Research Report. Research Division Almaden - Austin - Beijing - Haifa - India - T. J. Watson - Tokyo - Zurich

EXTRACTING BIOMETRIC BINARY STRINGS WITH MINIMAL AREA UNDER THE FRR CURVE FOR THE HAMMING DISTANCE CLASSIFIER

IBM Research Report. Decision Support in Multi-Vendor Services Partner Negotiation

IBM Research Report. Quantum-Based Simulation Analysis of Scaling in Ultra-Thin Body Device Structures

IBM Research Report. On a Connection between Facility Location and Perfect Graphs. Mourad Baïou University of Clermont

Error Rates. Error vs Threshold. ROC Curve. Biometrics: A Pattern Recognition System. Pattern classification. Biometrics CSE 190 Lecture 3

Biometrics: Introduction and Examples. Raymond Veldhuis

Has this Person Been Encountered Before? : Modeling an Anonymous Identification System

Yorktown Heights, New York * San Jose, California * Zurich, Switzerland

BIOMETRIC verification systems are used to verify the

IBM Research Report. Minimum Up/Down Polytopes of the Unit Commitment Problem with Start-Up Costs

Semi-parametric Distributional Models for Biometric Fusion

ebay/google short course: Problem set 2

g(y)dy = f(x)dx. y B If we write x = x(y) in the second integral, then the change of variable technique gives g(y)dy = f{x(y)} dx dy dy.

Likelihood Ratio-Based Biometric Verification

IBM Research Report. A Feasibility Pump for Mixed Integer Nonlinear Programs

Orientation Map Based Palmprint Recognition

IBM Research Report. Gust Speed Forecasting Using Weather Model Outputs and Meteorological Observations

Fingerprint Individuality

Estimation and sample size calculations for correlated binary error rates of biometric identification devices

IBM Research Report. Accurate 3D Capacitance of Parallel Plates from 2D Analytical Superposition

Likelihood Ratio Based Biometric Score Fusion

Score calibration for optimal biometric identification

PIN-based cancelable biometrics

If you wish to cite this paper, please use the following reference:

ORF 245 Fundamentals of Statistics Joint Distributions

Chapter 4 Multiple Random Variables

Probability Methods in Civil Engineering Prof. Dr. Rajib Maity Department of Civil Engineering Indian Institute of Technology, Kharagpur

PhysicsAndMathsTutor.com

How to Deal with Multiple-Targets in Speaker Identification Systems?

Expression arrays, normalization, and error models

HST.582J / 6.555J / J Biomedical Signal and Image Processing Spring 2007

Part Two. Diagnostic Test

Estimation of Relative Operating Characteristics of Text Independent Speaker Verification

9 - Markov processes and Burt & Allison 1963 AGEC

Central Limit Theorem and the Law of Large Numbers Class 6, Jeremy Orloff and Jonathan Bloom

Aruna Bhat Research Scholar, Department of Electrical Engineering, IIT Delhi, India

Political Science 236 Hypothesis Testing: Review and Bootstrapping

System Design and Performance Assessment: A Biometric Menagerie Perspective. Norman Poh Lecturer Dept of Computing

CH 24 IDENTITIES. [Each product is 35] Ch 24 Identities. Introduction

Expectation Maximization Deconvolution Algorithm

NON-STANDARD SPESIMENS ULTRASOUND DEFECTOSCOPY OF PIPES

Bivariate Distributions. Discrete Bivariate Distribution Example

Equations and Inequalities

EE 302 Division 1. Homework 6 Solutions.

Derivatives 2: The Derivative at a Point

8.2 Graphing More Complicated Rational Functions

Likelihood Ratio in a SVM Framework: Fusing Linear and Non-Linear Face Classifiers

Biometric Hash based on Statistical Features of Online Signatures

A3. Statistical Inference

MORE CURVE SKETCHING

Palmprint Verification with Moments

Ch. 5 Joint Probability Distributions and Random Samples

Preface.

IBM Research Report. Hilbert's Nullstellensatz and an Algorithm for Proving Combinatorial Infeasibility

CS/EE 181a 2008/09 Lecture 4

Particle Methods as Message Passing

THE MOMENT-PRESERVATION METHOD OF SUMMARIZING DISTRIBUTIONS. Stanley L. Sclove

AP Statistics Cumulative AP Exam Study Guide

2 3 x = 6 4. (x 1) 6

4.5 Rational functions.

Securing BioEncoded IrisCodes Against Correlation Attacks

Score Normalization in Multimodal Biometric Systems

On the Individuality of Fingerprints: Models and Methods

hypothesis a claim about the value of some parameter (like p)

ECE 661: Homework 10 Fall 2014

Signal Detection Basics - CFAR

Topic 17: Simple Hypotheses

6664/01 Edexcel GCE Core Mathematics C2 Bronze Level B3

6.867 Machine learning

Practice Problems Section Problems

Bayesian Decision Theory Tutorial Visual Recognition Tutorial 1

Introduction to Probability Theory for Graduate Economics Fall 2008

A TWO STAGE DYNAMIC TRILATERAL FILTER REMOVING IMPULSE PLUS GAUSSIAN NOISE

Machine Learning 4771

MA 1128: Lecture 08 03/02/2018. Linear Equations from Graphs And Linear Inequalities

Markscheme May 2016 Calculus Higher level Paper 3

IBM Research Report. Application of Three-Parameter Lognormal Distribution in EM Data Analysis

Chapter 3 Single Random Variables and Probability Distributions (Part 1)

Hypothesis tests

Nonlinear Oscillations and Chaos

Solutions of Linear Equations

IBM Research Report. Stochasic Unit Committment Problem. Julio Goez Lehigh University. James Luedtke University of Wisconsin

Dynamic Linear Combination of Two-Class Classifiers

A Contrario Detection of False Matches in Iris Recognition

2. A Basic Statistical Toolbox

Math 320-1: Midterm 2 Practice Solutions Northwestern University, Fall 2014

Theory of Computation

Modelling using ARMA processes

22. RADICALS. x add 5. multiply by 7

First Variation of a Functional

Chapter 2. Random Variable. Define single random variables in terms of their PDF and CDF, and calculate moments such as the mean and variance.

irst we need to know that there are many ways to indicate multiplication; for example the product of 5 and 7 can be written in a variety of ways:

review session gov 2000 gov 2000 () review session 1 / 38

UNIT 4A MATHEMATICAL MODELING OF INVERSE, LOGARITHMIC, AND TRIGONOMETRIC FUNCTIONS Lesson 2: Modeling Logarithmic Functions

Transcription:

RC23885 (W0602-49) February 7, 2006 Mathematics IBM Research Report The Relation between the ROC Curve and the CMC R. M. Bolle, J. H. Connell, S. Pankanti, N. K. Ratha, A. W. Senior IBM Research Division Thomas J. Watson Research Center P.O. Bo 704 Yorktown Heights, NY 0598 Research Division Almaden - Austin - Beijing - Haifa - India - T. J. Watson - Tokyo - Zurich LIMITED DISTRIBUTION NOTICE: This report has been submitted for publication outside of IBM and will probably be copyrighted if accepted for publication. Ithas been issued as a Research Report for early dissemination of its contents. In view of the transfer of copyright to the outside publisher, its distribution outside of IBM prior to publication should be limited to peer communications and specific requests. After outside publication, requests should be filled only by reprints or legally obtained copies of the article (e.g., payment of royalties). Copies may be requested from IBM T. J. Watson Research Center, P. O. Bo 28, Yorktown Heights, NY 0598 USA (email: reports@us.ibm.com). Some reports are available on the internet at http://domino.watson.ibm.com/library/cyberdig.nsf/home.

The Relation Between the ROC Curve and the CMC R.M. Bolle, J.H. Connell, S. Pankanti, N.K. Ratha and A.W. Senior IBM Thomas J. Watson Research Center, Hawthorne, NY 0532 Eploratory Computer Vision Group {bolle, jconnell, sharat, ratha, aws}@us.ibm.com Abstract The Cumulative Match Curve (CMC) is used as a measure of : m identification system performance. It judges the ranking capabilities of an identification system. The Receiver Operating Characteristic curve (ROC curve) of a verification system, on the other hand, epresses the quality of a : matcher. The ROC plots the False Accept Rate (FAR) of a : matcher versus the False Reject Rate (FRR) of the matcher. We show that the CMC is also related to the FAR and FRR of a : matcher, i.e., the matcher that is used to rank the candidates by sorting the scores. This has as a consequence that when a : matcher is used for identification, that is, for sorting match scores from high to low, the CMC does not offer any additional information beyond the FAR and FRR curves. The CMC is just another way of displaying the data and can be computed from the FAR and FRR. Introduction Performance evaluation of : biometric verification systems is achieved through estimating the False Accept Rate (FAR) and the False Reject Rate (FRR) and using these estimates to construct a ROC (Receiver Operating Characteristic) curve that epresses the tradeoff between the FAR and FRR []. The ROC is a well-accepted measure to epress the performance of : matchers. How to measure the performance of identification systems is perhaps less clear. One measure is the Cumulative Match Curve (CMC) [2, 3, 4], which epresses the performance of biometric identification systems (: m search engines) that return ranked lists of candidates. Specifically, when an identification system always returns the identities associated with the K highest-scoring biometric samples from an enrollment database (gallery). To estimate the CMC, the match scores between a query sample and the m biometric samples in the database are sorted. The lower the rank of the genuine matching biometric in the enrollment database, the better the ; m identification system. The matcher is tested on multiple galleries and typically it is assumed that the same : matcher can be used for all galleries. The database (gallery) is an integral part of a : m search engine because without an enrollment database there is no identification system. However, the testing methodology of identification systems does not allow for tuning the system to the database. Moreover, we show that there is a relationship between the ROC and the CMC associated with a : matcher (see also []). That is, given the ROC or given the FAR/FRR of a : matcher, the CMC can be computed and epresses how good this particular : matcher is at sorting galleries with respect to input query samples. This paper is organized as follows. The net section introduces the CMC and Section 3 shows how to estimate a CMC given data samples. Section 4 introduces the closed world assumption under which the CMC is often estimated. In Section 5 the relation between the ROC and CMC is derived. Section 6 gives some empirical results and Section 7 ends with some conclusions. 2 TheCMC Assume that we have a (large) set of biometric samples B i with associated ground truth ID(B i ). Key to measuring a CMC curve associated with a : matcher is the assembly of two subsets of samples [4]:. A gallery set G. There are m biometric samples in gallery set G {B,B 2,..., B m }; these are m biometric identifiers of different subjects. The gallery set G can be thought of as the enrolled database of biometric identifiers. 2. A probe set denoted as Q is a set of n unknown samples (B,B 2,..., B n) or {B l,l,..., n}, associated with the n subjects. The probe or query set Q can be from any set of individuals. However, usually probe identities are pre-

sumedtobeinthegalleryg. The probe set may contain more than one biometric sample of a given person and need not contain a sample of each subject in G. Given a query biometric B l Q and a biometric B i G, the output of a biometric matcher is a similarity score s(b l,b i). In order to estimate the Cumulative Match Curve, each probe biometric is matched to every gallery biometric and a total of n m similarity scores is computed, S l {s(b l,b ),s(b l,b 2 ),..., s(b l,b m )}, l,..., n. We have n sets S l of m similarity scores each. The scores s(b l,b i),i,.., m for each probe biometric B l, l,..., n are ordered as s(b l,b () ) s(b l,b (2) )... s(b l,b (m) ) () and probe B l is assigned the rank k l k if the matching sample from G is B (k). Hence the rank estimate k l of probe B l is k if the matching gallery biometric is in the k-th location of the sorted list of (), or equivalently, if the matching identity is in the k-th location of the associated list (or vector) of biometric identifiers, C m (B l; G) (B (),B (2),..., B (k),..., B (m) ). (2) We denote this list as C m (B l ; G) because the sorted list depends on input query B l and on the gallery G. 3 Estimating the CMC We have set K of n rank estimates {k l ; l,..., n} (with k l m), one estimate for each probe biometric B l ; each rank is defined only if the correct identity is in the ordered list of gallery biometrics C m (B l ; G) of (2). Before going into the Cumulative Match Curve (CMC), let us first define the discrete rank probabilities P (k),k,..., m, of a biometric search engine. These probabilities associated with a search engine are simply the probabilities, summing to, that the identity associated with a probe has rank k. TheP (k) are the true frequencies of occurrence, or discrete probabilities, of ranks k m. Basically, P (k) could be any discrete Rank Probability Mass (RPM) function. However, a Rank Probability Mass function with low average rank, is preferred. Given the probe and gallery data, the P (k) associated with a biometric matcher, are estimated by ˆP (k) n (# k l k) (3) n (# k l K k), k,..., m. The probability ˆP (k) that a matching biometric has rank k is estimated as the fraction of probe biometrics B l for which k l k; ˆP () is an estimate of the probability that the rank of any probe is, ˆP (2) is an estimate of the probability that the rank is 2, and so on. The Cumulative Match Curve estimates the distribution of the ranks k l,l,..., n of probes {B,B 2,..., B n}; CMC(k) is the fraction of probe biometrics B l that have rank k l k. That is, CMC(k) n (# k l k) (4) n (# k l K k) n n (k l k). l Hence by definition, the true CMC is the sum of the P (k) CMC(k) k P (r); k,..., m, (5) r (see Figure ). The CMC estimate is the distribution of the estimated ranks, denoted as k l and estimates the probability Prob(k l k). The random variable k l takes on discrete values, 2,..., m, determined by gallery G size m. Figure shows a Cumulative Match Curve CMC(k); a higher CMC() and a steeper slope of the curve at k indicate a better : matcher to implement : m search through sorting. CMC(k) k Figure. The CMC (left) is a weighted sum of step functions (k l k) at k k l as on the right To estimate (4) probes can be constructed in two ways: one of which is with the closed universe assumption. Every probe biometric identity ID(B l ) is some identity ID(B i) in the gallery G of size m. 4 Closed universe Each probe biometric B l has a corresponding gallery biometric B i and the correct match is always somewhere in the return vector of (2). This means that for each probe l,..., n, a rank k l, k l m, is defined and can be determined from reordered list C m (B l; G) (B (),B (2),..., B (m) ). (6) k k

To identify an unknown sample B l from the samples B i,i,..., m in gallery G, let the decision rule be to choose the highest-scoring candidate in vector C m (B l ; G) of (6) as the correct answer: Decide ID(B l)id(d () ), (7) the top sample from reordered vector C m (B l ; G) (B (),B (2),..., B (m) ). Selecting the top gallery biometric C (B l ; G) B () G as the correct match may appear to be a good identification strategy, because identifier B l is in gallery G and may be believed to be the most likely candidate on the top of the ranked list. The estimated rank probability ˆP (k ) of (3) is an estimate of the probability of correct identification, under this assumption. A subject cannot be falsely rejected and cannot be falsely accepted; he or she can only be misidentified. Alternatively, to make a decision, one can construct a short vector (candidate list) C K (B l ; G) of identities of the K top-scoring gallery biometrics B i. This list of identities associated with the top K gallery biometrics, C K (B l ; G) (B (),B (2),..., B (K) ) T comprises the K most likely identities. Now, suppose the gallery G is some most wanted list and the decision rule is ID(B l ) is Most wanted, if B l C K(B l ; G), or, ID(B l ) is Most wanted if rank of B l K. (8) That is, a correct decision is deemed to have been made if the true identity of unknown sample B l is in the candidate list of length K. This is the negative identification problem: Is subject ID(B l ) em not on some list? Of course, the closed world assumes that every subject ID(B l ) that is authenticated is on this wanted list. This assumption will be relaed in Section 5.3. As is the case when we only consider the top candidate on the list (7), also for K>,asampleB l cannot be falsely rejected and cannot be falsely accepted, he or she can only be mistakenly left off the candidate list. Again, only identification errors can be made, with probability Prob(B l is misidentified; K) ˆP F (K). (9) This error probability depends first of all on candidate list size K and is lower for larger K. It further depends on the gallery size m and is higher for larger m. A third influence is the quality of the match engine, or the distributions of genuine and imposter scores. The probability of correct identification then is Prob(B l is identified; K) CMC(K) ˆP F (K). The Cumulative Match Curve is an estimate of the probability of correct identification as a function of candidate list size K identification. The discrete function CMC(K) converges to when K approaches m. The faster the convergence, the more likely it is that the genuine matches are in short candidate lists for a given biometric matcher. 5 The CMC, FAR/FRR relation Even though the decision rule of (8) does not use any thresholds, the match engine produces scores s(b l,b i), and we can relate the CMC to the False Accept and False Reject rates of the underlying : biometric matcher used in the : m search engine, by realizing that for every search the genuine score of the correct template can be interpreted as a virtual threshold on the imposter scores s(b l,b i), ID(B l ) ID(B i) from the rest of the database. (Note that this relation between the CMC and the FAR/FRR only holds when the : matcher is used for sorting the gallery scores.) Each probe Q is a biometric B l with a corresponding biometric B i in gallery G. Matching probe biometric B l to a gallery of size m is equivalent to drawing of m random variables:. One genuine score X s(b l,b i) for B i Gwith B l and B i from the same identity. 2. A set of (m ) imposter scores {Y,..., Y m }, i.e., {s(b l,b i),i,..., (l ), (l +),..., m}, where the B i are in the gallery but not including biometric B i matching to B l. To look at this, we turn to the rank probabilities of (3) first to simplify the analysis ˆP (k) Prob(rank B l is k), k,..., m. These can be determined using joint probabilities p(, y,..., y m ) of the m random variables {s(b l,b i),i,..., m}. 5. Gallery size m 2 Pick a gallery size of m 2for the moment; this can be depicted in the y plane. Any probe biometric B l with l will result in a pair of random variables (X, Y ), with X s(b l,b ) a genuine score and Y s(b l,b 2) an imposter score. The genuine score is from probability density function (PDF) f(); the imposter scores are from PDF g(y), each associated with the : matcher. There eist only two rank probabilities P () Prob(rank B l is ) Prob(X Y ), P (2) Prob(rank B l is 2) Prob(Y > X). Let the joint probability density for a probe (X, Y ) be p(, y) f() g(y), assuming independence of the genuine and imposter score (see Figure 2). The rank probabil-

y - FAR() g(y) p(, y) g() f() f() Figure 2. The probability of misidentification when gallery size is m 2 - FRR(y) g(y) f(y) ity P () is the integral of p(, y) over the area >yand is the probability of correct identification. Rank probability P (2) P (), the probability of misidentification, is the integral of p(, y) over the area <y. These two rank probabilities can be written down. First look at the probability of correct identification, P (). In Figure 2, this is the probability that genuine score X>Y. This is the probability mass for >y P () Prob(X > Y) p(, y)dyd 0 y0 0 y0 f() g(y) y0 y >y g(y)dy d f()d dy f()[ FAR()] d g(y)[ FRR(y)] dy (0) The interpretation of (0) is shown in Figure 3. Given any genuine score X, the probability that y<is the probability that y is not a False Accept when adjusting threshold t o to declare score a match, i.e., is [ FAR()]. Conversely, given any y, the probability that > y is [ FRR(y)], the probability that is not a False Reject at threshold y. What is the probability that the correct identity has rank 2, i.e., what is P (2)? From Figure 2 it is seen that the probability that the imposter similarity score Y is larger than X is the integral of p(, y) over the shaded region y> P (2) P () Prob(X <Y) p(, y)dyd f()far()d <y y0 0 g(y)frr(y)dy. () Interpretations of these probabilities are as above. The first integral of () is shown in Figure 4a. Given some genuine Figure 3. For gallery size m 2, the probability of correct identification is the integral of f()[ FAR()] and g(y)[ FRR(y)] as in (0) score X, the probability that an imposter score y is larger than is FAR(), the probability of a False Accept at threshold. The shaded area in the detail of Figure 4b is the probability of incorrect identification; this is the integral f()far() over all possible genuine scores. The second integral of () is an integral of possible values for imposter scores y. For any y the probability that genuine score is less than y is the probability of a False Reject at threshold y; the integral over g(y)frr(y) P (2). 5.2 Gallery size m>2 When m>2, a probe Q is the drawing of one random genuine score X and the drawing of (m ) random imposter scores Y i,i,..., m. Let us first eamine rank probability k, i.e., P (). This is Prob(rank X is ), hence the probability that X> Y,X > Y 2,..., X > Y m. Again assuming independence, the joint probability density of (, y,y 2,..) is p(, y,..., y m )f() m i y g(y i ). The rank probability P () is the integral of p(, y,y 2,..), as in Figure 2, but over a hyper region this time: P () y < y p(, y m <,..., y m )ddy dy m ( 0 f() m ) i y0 g(y i)dy i d. (2)

FAR () g() f() This is given by P (k) ( ) m k y y k > y k y m < p(,..., y m )d dy m, FAR () a f() accounting for the number of ways (k ) imposter scores can be selected from (m ). Using (0) and (), this becomes ( m ) ( k ) k 0 i y g(y i)dy i f() ( m ) y0 g(y i)dy i d. ik f() FAR () b Figure 4. For gallery size m 2, the probability of misidentification is the integral of f()far() Using (0), this becomes P () 0 f()[ FAR()] m d. (3) Figure 5 compares P () for m 2of (0) with the P () for m>2 of (3). The probability [ FAR()] m [ FAR()] for large m and Prob(rank X is ) obviously becomes less and less for larger m. - FAR() [ - FAR()] m - f() Figure 5. The probability P () that genuine score X is assigned rank goes down for increasing m By etending the above, we are now ready to construct P (k), the probability that the some correct identity ends up in the k-th position of the return vector C m. The probability that rank X is k is the probability that (k ) of the (m ) imposter scores Y,..., Y m are greater than X. And we have P ( (k) m ) k 0 [FAR()]k f()[ FAR()] m k d. (4) The interpretation of (4) is as above. Given some genuine score X with probability density f(), there are (k ) imposter scores Y l greater than with probability [FAR()] k. There are an additional (m k) imposter scores Y l that are less than with probability [ FAR()] k. An approimate interpretation of (4) is the following. If we assume that the genuine score PDF f() is relatively narrow and can be well approimated by an impulse function δ( ˆ), with ˆ the epected match score of B l, (4) becomes ( ) m P (k) [FAR(ˆ)] k [ FAR(ˆ)] m k. (5) k This is the probability that (k ) imposter scores Y are greater than ˆ, and (m k) imposter scores Y are less than ˆ. Remember, the Cumulative Match Curve CMC(K) of (7) is an estimate of the probability of correct identification as a function of candidate vector C K size K. AsampleB is correctly identified if the corresponding identity in the gallery has rank K or less, and following (5), we get for the true CMC CMC(K) K k P (k) K k ( m k ) 0 [FAR()]k f()[ FAR()] m k d. (6) When using (5), this becomes approimately CMC(K) K ( m ) k k [FAR(ˆ)] k [ FAR(ˆ)] m k, K,..., m. (7) The probability that the true rank k of the correct identity K is the sum of the true individual probabilities that rank is k from k to k K.

One thing about Epression (7) between the CMC of a : m search engine and the FAR/FRR of a : match engine should be noted again. Given the way the CMC is computed by sorting s(b l,b i) as in () is only one particular way of implementing : m search based an a : search engine. 5.3 Open world Suppose there are an additional n unknown biometrics of unknown samples B in the probe. Then the CMC is weighted by the prior probability that B G n known m. n known + n unknown m + n unknown While the CMC is a useful characterization of a system in controlled evaluations, it is only a partial characterization of the performance of a system. 6 Eperiment Figure 6 shows a plot of an empirical CMC versus its theoretical form. The theoretical form was estimated by first computing the FAR and FRR distributions from the matcher scores and then using Epression (6) to estimate CMC(K) for K ranging from to the size of the gallery. Given the matcher scores, the empirical form was estimated by calculating the CMC directly from its definition and without first computing the FAR or FRR distribution. CMC(K) 0.99 0.98 0.97 0.96 0.95 about two percent and it is empirical evidence supporting the integrity of our theoretical derivation. 7 Conclusions Performance evaluation of biometric systems is an important topic. The generally accepted method for epressing the performance of a : verification system is an ROC curve. Typically, the performance of a : m identification system is epressed by a Cumulative Match Curve (CMC), which is really a discrete function. In this paper we have shown that the CMC is directly related to the ROC, the tradeoff between the FAR and FRR as a function of the operating threshold. That is, given the characteristics of a : matcher in terms of genuine and imposter score distributions or estimates thereof, the CMC can be constructed. Hence, as it is used today, the CMC is more a performance measure of a : matcher (verification system) than a performance measure of a : m identification system. When computed with the probe-gallery approach, the CMC merely epresses the sorting capabilities of a : matcher. A better way of testing : m identification systems is to make the galleries public some time before the test and sequester the probe biometric samples till the time of the test. It is then up to identification system designer how to use a : matcher in the identification task if a : matcher is used at all. In our eperiment, we have used a very simple identification scheme based on a : matcher. Note that in many applications, the identification process may involve more than a matcher (see, for eample, [], Chapter 6). In such situations, the discrepancy between empirical CMC (from identification procedure) and the theoretical CMC (from theoretical analysis of : FAR and FRR using Epression (6) can be an important metric for assessing the additional value provided by identification process. References cmcformal cmcempirical 0.94 0 00 Candidate List Size, K Figure 6. Theoretical versus empirical form of the CMC. It is easy to see that as candidate list size, K increases, the error decreases and when K reaches the gallery size, both empirical and theoretical estimates (correctly) reach. For this very same reason, note that the difference in empirical and theoretical estimates decreases as the candidate list size increases. Further, it can be observed that the maimum discrepancy between theoretical and empirical estimates is [] R. Bolle, J. Connell, S. Pankanti, N. Ratha, and A. Senior. Guide to Biometrics. Springer, New York, 2004. [2] P. Grother and P. Philips. Models of large population recognition performance. In Proc. CVPR, volume 2, pages 68 75, 2004. [3] P. Phillips, P. Grother, R. Micheals, D. Blackburn, T. Elham, and J. M. Bone. FRVT 2002: Facial recognition vendor test. Technical report, DoD, Apr. 2003. [4] P. Phillips, H. Moon, S. Rizvi, and P. Rauss. The FERET evaluation methodology for face-recognition algorithms. IEEE PAMI, 22(0):090 04, Oct. 2000.