Technical Report. Results from 200 billion iris cross-comparisons. John Daugman. Number 635. June Computer Laboratory

Similar documents
A Contrario Detection of False Matches in Iris Recognition

Score calibration for optimal biometric identification

DEMODULATION BY COMPLEX-VALUED WAVELETS FOR STOCHASTIC PATTERN RECOGNITION

Technical Report. Biometric decision landscapes. John Daugman. Number 482. January Computer Laboratory UCAM-CL-TR-482 ISSN

independent. The mere failure of a test of statistical independence can thereby

BLAST: Target frequencies and information content Dannie Durand

Name: Firas Rassoul-Agha

Fingerprint Individuality

INTERNATIONAL JOURNAL OF ENGINEERING SCIENCES & RESEARCH TECHNOLOGY

Biometrics: Introduction and Examples. Raymond Veldhuis

When enough is enough: early stopping of biometrics error rate testing

CS 1538: Introduction to Simulation Homework 1

B.N.Bandodkar College of Science, Thane. Random-Number Generation. Mrs M.J.Gholba

Probabilities and distributions

Discrete Random Variables

Probability and Probability Distributions. Dr. Mohammed Alahmed

0460 GEOGRAPHY. 0460/43 Paper 4 (Alternative to Coursework), maximum raw mark 60

CSE 103 Homework 8: Solutions November 30, var(x) = np(1 p) = P r( X ) 0.95 P r( X ) 0.

Preliminary Statistics Lecture 2: Probability Theory (Outline) prelimsoas.webs.com

Statistical Methods in Particle Physics

CHAPTER 2 Estimating Probabilities

STAT2201. Analysis of Engineering & Scientific Data. Unit 3

Random Variable. Discrete Random Variable. Continuous Random Variable. Discrete Random Variable. Discrete Probability Distribution

Chapte The McGraw-Hill Companies, Inc. All rights reserved.

New Directions in Computer Science

Introductory Probability

QUALIFICATION SPECIFICATION

BNAD 276 Lecture 5 Discrete Probability Distributions Exercises 1 11

Bits. Chapter 1. Information can be learned through observation, experiment, or measurement.

Probability Dr. Manjula Gunarathna 1

Probability Rules. MATH 130, Elements of Statistics I. J. Robert Buchanan. Fall Department of Mathematics

3. Tests in the Bernoulli Model

Lecture Notes. This lecture introduces the idea of a random variable. This name is a misnomer, since a random variable is actually a function.

5.3 Conditional Probability and Independence

Machine Learning

ECE521 Lecture7. Logistic Regression

Discrete Distributions

Bernoulli Trials, Binomial and Cumulative Distributions

Human-Oriented Robotics. Probability Refresher. Kai Arras Social Robotics Lab, University of Freiburg Winter term 2014/2015

Corners, Blobs & Descriptors. With slides from S. Lazebnik & S. Seitz, D. Lowe, A. Efros

Are Symmetrical Patterns More Successful?

1. Discrete Distributions

ACM 116: Lecture 2. Agenda. Independence. Bayes rule. Discrete random variables Bernoulli distribution Binomial distribution

Probability Distributions.

Probability (10A) Young Won Lim 6/12/17

ECE 302: Probabilistic Methods in Electrical Engineering

Chapter 3. Chapter 3 sections

1/18/2011. Chapter 6: Probability. Introduction to Probability. Probability Definition

The probability of an event is viewed as a numerical measure of the chance that the event will occur.

Maximum-Likelihood fitting

Deep Learning for Computer Vision

LEGAL DISCLAIMER. APG Coin (APG) White Paper (hereinafter 'the White Paper', 'the Document') is presented for informational purposes only

Descriptive Data Summarization

PROD. TYPE: COM ARTICLE IN PRESS. The importance of being random: statistical principles of iris recognition. John Daugman

Statistics 100A Homework 5 Solutions

Topic 3: Introduction to Probability

I. Introduction to probability, 1

IMBALANCED DATA. Phishing. Admin 9/30/13. Assignment 3: - how did it go? - do the experiments help? Assignment 4. Course feedback

exp{ (x i) 2 i=1 n i=1 (x i a) 2 (x i ) 2 = exp{ i=1 n i=1 n 2ax i a 2 i=1

Gaussian Mixture Distance for Information Retrieval

Test Sample and Size. Synonyms. Definition. Main Body Text. Michael E. Schuckers 1. Sample Size; Crew designs

Discrete Binary Distributions

Midterm #1 - Solutions

Some Material on the Statistics Curriculum

Assignment 1 (Sol.) Introduction to Machine Learning Prof. B. Ravindran. 1. Which of the following tasks can be best solved using Clustering.

Discrete Distributions

6.3 Bernoulli Trials Example Consider the following random experiments

Expected Value - Revisited

02 Background Minimum background on probability. Random process

Statistical Experiment A statistical experiment is any process by which measurements are obtained.

GEOG 508 GEOGRAPHIC INFORMATION SYSTEMS I KANSAS STATE UNIVERSITY DEPARTMENT OF GEOGRAPHY FALL SEMESTER, 2002

Page Max. Possible Points Total 100

Machine Learning for Large-Scale Data Analysis and Decision Making A. Week #1

Discrete Probability Distributions

SRI Briefing Note Series No.8 Communicating uncertainty in seasonal and interannual climate forecasts in Europe: organisational needs and preferences

MATH 10 INTRODUCTORY STATISTICS

KCC White Paper: The 100 Year Hurricane. Could it happen this year? Are insurers prepared? KAREN CLARK & COMPANY. June 2014

Randomized Algorithms

Background to Statistics

The Decision List Machine

Chapter Three. Hypothesis Testing

City of Hermosa Beach Beach Access and Parking Study. Submitted by. 600 Wilshire Blvd., Suite 1050 Los Angeles, CA

Management Programme. MS-08: Quantitative Analysis for Managerial Applications

Unit 4 Probability. Dr Mahmoud Alhussami

SCIENCE OCR LEVEL 2 CAMBRIDGE TECHNICAL. Cambridge TECHNICALS CHEMISTRY OF PRODUCTION CERTIFICATE/DIPLOMA IN D/505/3209 LEVEL 2 UNIT 10

Face Recognition Using Eigenfaces

Binomial and Poisson Probability Distributions

Statistics for scientists and engineers

Noise Robust Isolated Words Recognition Problem Solving Based on Simultaneous Perturbation Stochastic Approximation Algorithm

Discrete Mathematics and Probability Theory Spring 2016 Rao and Walrand Note 14

CS 160: Lecture 16. Quantitative Studies. Outline. Random variables and trials. Random variables. Qualitative vs. Quantitative Studies

Linear Models: Comparing Variables. Stony Brook University CSE545, Fall 2017

Statistical Concepts. Distributions of Data

Grundlagen der Künstlichen Intelligenz

Math 180B Problem Set 3

7.1 What is it and why should we care?

An-Najah National University Faculty of Engineering Industrial Engineering Department. Course : Quantitative Methods (65211)

Discrete probability distributions

Section 2.4 Bernoulli Trials

Transcription:

Technical Report UCAM-CL-TR-635 ISSN 1476-2986 Number 635 Computer Laboratory Results from 200 billion iris cross-comparisons John Daugman June 2005 15 JJ Thomson Avenue Cambridge CB3 0FD United Kingdom phone +44 1223 763500 http://www.cl.cam.ac.uk/

c 2005 John Daugman Technical reports published by the University of Cambridge Computer Laboratory are freely available via the Internet: http://www.cl.cam.ac.uk/techreports/ ISSN 1476-2986

Results from 200 billion iris cross-comparisons John Daugman Abstract Statistical results are presented for biometric recognition of persons by their iris patterns, based on 200 billion cross-comparisons between different eyes. The database consisted of 632,500 iris images acquired in the Middle East, in a national border-crossing protection programme that uses the Daugman algorithms for iris recognition. A total of 152 different nationalities were represented in this database. The set of exhaustive cross-comparisons between all possible pairings of irises in the database shows that with reasonable acceptance thresholds, the False Match rate is less than 1 in 200 billion. Recommendations are given for the numerical decision threshold policy that would enable reliable identification performance on a national scale in the UK. 1 Introduction In 2001 the United Arab Emirates (UAE) Ministry of Interior launched a national bordercrossing security programme that is today deployed at all 17 of the UAE s air, land, and sea ports. It is based upon mathematical analysis of the random patterns visible in the iris of a person s eye (iris recognition), using algorithms developed by the author of this report and cameras made by LG. All deployments of iris recognition worldwide use these same Daugman algorithms, and most use the same LG cameras; but a valid criticism that has often been raised in discussions of other possible national-scale deployments is the absence of large-scale statistical test reports. For example, although the algorithms have been deployed successfully at Schiphol Airport since 2001 for biometrically controlled entry into The Netherlands without passport inspection, no statistical analysis of performance has been published. Several smaller-scale tests of the iris recognition algorithms have consistently shown them to produce no False Matches (a bibliography of reports is available at: http://www.cl.cam.ac.uk/users/jgd1000/iristests.pdf), but those tests involved relatively small databases, and the iris images were often acquired in laboratory rather than real-world conditions. In the UAE border-crossing deployment, nearly 2 trillion (2 million-million) iris comparisons have been performed to date, as all foreign nationals visiting the Emirates have their irises compared against all the IrisCodes (mathematical descriptions of registered iris patterns) stored in a central database. Some 40,000 persons have thereby been caught trying to re-enter the UAE with false travel documents since this deployment began. The Abu Dhabi Directorate of Police report that so far there have been no False Matches; yet it was desired to exploit the large enrollment database to understand better the statistical powers of iris recognition technology. For example, by computing the similarities between 3

all possible pairings of different irises in the database, much could be learned about the robustness of the algorithms against making any False Matches when there are such vast numbers of opportunities for making False Matches. This would help to illuminate the technology s potential for even larger-scale national deployments. UAE Interior Minister Sheikh Saif Bin Zayyed therefore graciously consented to make the enrollment database of IrisCodes available to the University of Cambridge for detailed analysis. This report is the result of that analysis. The iris recognition algorithms have been described previously in detailed aspects in Daugman(1993, 2001, 2003, 2004) and will not be reviewed here. 2 Cross-comparison results Statistical analysis of the decision-making power of iris identification requires examining the distribution of similarity scores obtained from comparing different irises, and likewise the scores obtained from comparing different images of same irises. Ideally these two distributions should be well-separated, with rapidly decaying tails, since it is the overlap of the distributions tails that determines error rates. Count 0 3 Billion 6 Billion 9 Billion 12 Billion 200 Billion Iris Cross-Comparisons, 0 Rotations, UAE Database All bits agree 200,027,808,750 pair comparisons among 632,500 different irises mean = 0.499, stnd.dev. = 0.034 solid curve: binomial distribution All bits disagree 0.0 0.1 0.2 0.3 0.4 0.5 0.6 0.7 0.8 0.9 1.0 Hamming Distance The above distribution shows all cross-comparison similarity scores obtained from making all possible pair comparisons amongst 632,500 different irises. N different objects can generate a total of N (N 1)/2 different pairings, which for this database means 4

200,027,808,750, or about 200 billion different pair comparisons. The similarity metric is Hamming Distance, which is simply the fraction of bits that disagree between any two IrisCodes. Thus if two IrisCodes happened to agree in every bit, their Hamming Distance would be 0 and they would score a count at the extreme left in the above distribution. If they disagreed in every bit, they would score a count at 1.0 on the extreme right. The vast majority of IrisCodes from different eyes disagreed in about 50% of their bits, as expected since the bits are equiprobable and uncorrelated between different eyes. Very few pairings of IrisCodes could disagree in fewer than 35% or more than 65% of their bits. The smallest and largest Hamming Distances found in this set of 200 billion simple raw comparisons of different IrisCodes were 0.265 and 0.750 respectively. The solid curve that fits the data extremely closely in the above plot is a binomial probability density function. This theoretical form was chosen because comparisons between bits from different IrisCodes are Bernoulli trials, or conceptually coin tosses, and Bernoulli trials generate binomial distributions. If one tossed a coin whose probability of heads is p (typically but not necessarily p = 0.5) in a series of N independent tosses and counted the number m of heads outcomes, and if one tallied this fraction x = m/n in a large number of such repeated runs of N tosses, then the resulting distribution of x would follow the solid curve in the figure above and be described by this equation: N! f(x) = m!(n m)! pm (1 p) (N m) (1) The analogy between tossing coins and comparing bits between different IrisCodes is deep but imperfect, because any given IrisCode has internal correlations arising from iris features, especially in the radial direction. Their effect is to reduce the value of N to a number significantly smaller than the number of bits that are actually compared between two IrisCodes; N becomes the number of effectively independent bit comparisons. The value of p is very close to 0.5 (empirically 0.499 for this database), because the states of each bit are equiprobable a priori, and so any pair of bits from different IrisCodes are equally likely to agree or disagree. Different people expose different amounts of iris between their eyelids, and the amount visible depends also on circumstances. Therefore the number of bits available for comparison between two different IrisCodes is quite variable. A close match (say a Hamming Distance of HD = 0.100) based on only few compared bits is much less indicative of identity than an apparently poorer match (say HD = 0.200) based on a large number of compared bits. In the coin-tossing analogy, obtaining very imbalanced outcomes just by chance is much more likely in a short series of tosses than in a long series. This requires a renormalisation of any observed raw Hamming Distance score HD raw into one HD norm whose deviation from 0.5 has been re-scaled for statistical significance, based on the number of bits n that were actually compared between the two IrisCodes: n HD norm = 0.5 (0.5 HD raw ) (2) 911 The parameters in the above equation influence the standard deviation of the distribution of normalised Hamming Distance scores, and they give the distribution a stable form which permits a stable decision criterion to be used. When IrisCodes are compared in a search for a match, it cannot be known precisely what was the amount of head tilt, camera tilt, or eye rotation when the IrisCodes were 5

obtained. Therefore it is necessary to make comparisons over a reasonable range of relative tilts (rotations) between every pair of IrisCodes, keeping the best match as their similarity score. This generates a new distribution that is skewed towards lower Hamming Distances even between unrelated irises. 200 Billion Iris Cross-Comparisons, 7 Rotations, UAE Database Count 0 6 Billion 12 Billion 18 Billion All bits agree 200,027,808,750 pair comparisons among 632,500 different irises (best match after 7 rotations each) mean = 0.456, stnd.dev. = 0.0214 solid curve: binomial distribution extreme-value All bits disagree 0.0 0.1 0.2 0.3 0.4 0.5 0.6 0.7 0.8 0.9 1.0 Hamming Distance The new distribution after k rotations of IrisCodes in the search process still has a simple analytic form that can be derived theoretically. Let f 0 (x) be the raw density distribution obtained for the HD scores between different irises after comparing them only in a single relative orientation; for example, f 0 (x) might be the binomial defined in Eqn (1). Then F 0 (x) = x 0 f 0(x)dx, the cumulative of f 0 (x) from 0 to x, becomes the probability of getting a False Match in such a test when using HD acceptance criterion x. Clearly, then, the probability of not making a False Match when using decision criterion x is 1 F 0 (x) after a single test, and it is [1 F 0 (x)] k after carrying out k such tests independently at k different relative orientations. It follows that the probability of a False Match after a best of k test of agreement, when using HD criterion x, regardless of the actual form of the raw unrotated distribution f 0 (x), is F k (x) = 1 [1 F 0 (x)] k, and the expected density f k (x) associated with this cumulative is: f k (x) = d dx F k(x) = kf 0 (x) [1 F 0 (x)] k 1 (3) Equation (3) is the solid curve in the above figure, fitting the distribution of the same set of 200 billion IrisCode comparisons after 7 relative rotations of each pair. 6

3 Observed False Match Rates The cumulative scores under the left tail of the preceding distribution, up to various Hamming Distance thresholds, reveal the False Match Rates among the 200 billion iris comparisons if the acceptance decision policy used those thresholds. These rates are provided in the following Table. No such matches were found with Hamming Distances below about 0.260, but the Table has been extended down to 0.220 using Equation (3) for extreme value samples of the binomial (plotted as the solid curve in the preceding figure) to extrapolate the theoretically expected False Match Rates for such decision policies. HD Criterion Observed False Match Rate 0.220 0 (theor: 1 in 5 10 15 ) 0.225 0 (theor: 1 in 1 10 15 ) 0.230 0 (theor: 1 in 3 10 14 ) 0.235 0 (theor: 1 in 9 10 13 ) 0.240 0 (theor: 1 in 3 10 13 ) 0.245 0 (theor: 1 in 8 10 12 ) 0.250 0 (theor: 1 in 2 10 12 ) 0.255 0 (theor: 1 in 7 10 11 ) 0.262 1 in 200 billion 0.267 1 in 50 billion 0.272 1 in 13 billion 0.277 1 in 2.7 billion 0.282 1 in 284 million 0.287 1 in 96 million 0.292 1 in 40 million 0.297 1 in 18 million 0.302 1 in 8 million 0.307 1 in 4 million 0.312 1 in 2 million 0.317 1 in 1 million The US Department of Homeland Security recently sponsored independent testing of the same Daugman algorithms. In a total of 1,707,061,393 (1.7 billion) cross-comparisons between different irises, the smallest Hamming Distance observed was in the range of 0.280, consistent with the above Table. (Source: International Biometric Group, Independent Testing of Iris Recognition Technology, May 2005.) In the UK with a national population of about 60 million, an all-against-all comparison of IrisCodes to detect any multiple identities could be reliably performed using a decision criterion of about 0.220 without making False Matches. Employing such a decision policy in off-line exhaustive comparisons would allow the detection of multiple identities even when up to about 22% of the bits in their IrisCodes disagreed, due to corrupting factors in image acquisition such as poor focus, motion blur, limited cooperation, and so forth. In everyday transactions in which an identity is first asserted, matches could be accepted with a Hamming Distance as high as perhaps 0.33, allowing 33% of the bits to be incorrect, thereby tolerating unfavourable image capture conditions. 7

4 Camera properties and imaging conditions Different cameras deliver iris images of differing quality and resolution, in part depending on the sophistication of their optics (e.g. autofocus and autozoom), and in part because they employ different wavelengths of the near-infrared spectrum. The following two graphs from earlier studies (Daugman 2003, 2004) illustrate that although the distribution of similarity scores for different-eye comparisons (shown in black) is stable and similar to that presented in this document for 200 billion UAE iris comparisons, the distribution of similarity scores for same-eye comparisons (shown in white on the left of each graph) is dependent on camera quality and imaging conditions. Decision Environment for Iris Recognition: Sub-optimal Camera Decision Environment for Iris Recognition: Optimal Camera same different same different mean = 0.110 stnd.dev. = 0.065 mean = 0.458 stnd.dev. = 0.0197 mean = 0.019 stnd.dev. = 0.039 mean = 0.456 stnd.dev. = 0.020 Count d = 7.3 2.3 million comparisons d = 14.1 482,600 comparisons 0.0 0.1 0.2 0.3 0.4 0.5 0.6 0.7 0.8 0.9 1.0 Hamming Distance 0.0 0.1 0.2 0.3 0.4 0.5 0.6 0.7 0.8 0.9 1.0 Hamming Distance 5 Conclusion and recommendations Iris recognition can be reliably used on a national basis in an Identity Cards scheme, including the capability for exhaustive IrisCode comparisons to detect multiple identities, if the decision policy employs a threshold criterion of about 0.22 Hamming Distance for off-line exhaustive comparisons and a criterion near 0.33 for everyday one-to-one matches. References Daugman, J.G. (1993) High confidence visual recognition of persons by a test of statistical independence. IEEE Transactions on Pattern Analysis and Machine Intelligence, 15 (11): 1148 1161. Daugman J.G. (2001) Statistical richness of visual phase information: Update on recognising persons by their iris patterns. International Journal of Computer Vision, 45 (1): 25 38. Daugman J.G. and Downing C.J. (2001) Epigenetic randomness, complexity, and singularity of human iris patterns. Proceedings of the Royal Society (London): B. Biological Sciences, 268: 1737 1740. Daugman J.G. (2003) The importance of being random: Statistical principles of iris recognition. Pattern Recognition, 36: 279 291. Daugman J.G. (2004) How iris recognition works. IEEE Transactions on Circuits and Systems for Video Technology, 14 (1): 21 30. 8