A Novel Rejection Measurement in Handwritten Numeral Recognition Based on Linear Discriminant Analysis

Similar documents
Intelligent Handwritten Digit Recognition using Artificial Neural Network

SYMBOL RECOGNITION IN HANDWRITTEN MATHEMATI- CAL FORMULAS

Pairwise Neural Network Classifiers with Probabilistic Outputs

Research on Feature Extraction Method for Handwritten Chinese Character Recognition Based on Kernel Independent Component Analysis

Invariant Pattern Recognition using Dual-tree Complex Wavelets and Fourier Features

Pattern Recognition and Machine Learning. Perceptrons and Support Vector machines

A New Hybrid System for Recognition of Handwritten-Script

Introduction to Signal Detection and Classification. Phani Chavali

Reconnaissance d objetsd et vision artificielle

Pattern Recognition Approaches to Solving Combinatorial Problems in Free Groups

A PERTURBATION-BASED APPROACH FOR MULTI-CLASSIFIER SYSTEM DESIGN

AN INVESTIGATION ON EFFICIENT FEATURE EXTRACTION APPROACHES FOR ARABIC LETTER RECOGNITION

Optical Character Recognition of Jutakshars within Devanagari Script

EEL 851: Biometrics. An Overview of Statistical Pattern Recognition EEL 851 1

6.867 Machine Learning

L11: Pattern recognition principles

Regularized Discriminant Analysis and Reduced-Rank LDA

Multi-Class Linear Dimension Reduction by. Weighted Pairwise Fisher Criteria

Feature Extraction Using Zernike Moments

ARTIFICIAL NEURAL NETWORKS گروه مطالعاتي 17 بهار 92

Selection of Classifiers based on Multiple Classifier Behaviour

Support Vector Machines (SVM) in bioinformatics. Day 1: Introduction to SVM

Pattern Recognition and Machine Learning

Myoelectrical signal classification based on S transform and two-directional 2DPCA

FARSI CHARACTER RECOGNITION USING NEW HYBRID FEATURE EXTRACTION METHODS

ECE 521. Lecture 11 (not on midterm material) 13 February K-means clustering, Dimensionality reduction

Gaussian Process Style Transfer Mapping for Historical Chinese Character Recognition

Linear Classification and SVM. Dr. Xin Zhang

Machine learning comes from Bayesian decision theory in statistics. There we want to minimize the expected value of the loss function.

SUPPORT VECTOR MACHINE

STA 414/2104: Lecture 8

ECE 661: Homework 10 Fall 2014

PCA and LDA. Man-Wai MAK

Comparing Robustness of Pairwise and Multiclass Neural-Network Systems for Face Recognition

Support Vector Machine & Its Applications

Support Vector Machines for Classification: A Statistical Portrait

Partial discriminative training for classification of overlapping classes in document analysis

Classifier Complexity and Support Vector Classifiers

The Perceptron. Volker Tresp Summer 2014

Adaptive Boosting of Neural Networks for Character Recognition

The Perceptron. Volker Tresp Summer 2016

PCA and LDA. Man-Wai MAK

Machine Learning Linear Classification. Prof. Matteo Matteucci

Weight Quantization for Multi-layer Perceptrons Using Soft Weight Sharing

Support Vector Machine (SVM) & Kernel CE-717: Machine Learning Sharif University of Technology. M. Soleymani Fall 2012

Dimensionality Reduction Using PCA/LDA. Hongyu Li School of Software Engineering TongJi University Fall, 2014

ECE662: Pattern Recognition and Decision Making Processes: HW TWO

A New OCR System Similar to ASR System

Multiple Similarities Based Kernel Subspace Learning for Image Classification

Engineering Part IIB: Module 4F10 Statistical Pattern Processing Lecture 6: Multi-Layer Perceptrons I

Ch 4. Linear Models for Classification

A Statistical Analysis of Fukunaga Koontz Transform

Object Recognition Using a Neural Network and Invariant Zernike Features

Incorporating Invariances in Nonlinear Support Vector Machines

SGN (4 cr) Chapter 5

Style-conscious quadratic field classifier

PCA & ICA. CE-717: Machine Learning Sharif University of Technology Spring Soleymani

EXAM IN STATISTICAL MACHINE LEARNING STATISTISK MASKININLÄRNING

Classification with Perceptrons. Reading:

A Hierarchical Representation for the Reference Database of On-Line Chinese Character Recognition

STUDY ON METHODS FOR COMPUTER-AIDED TOOTH SHADE DETERMINATION

Artificial Neural Networks Examination, June 2005

Discriminative Direction for Kernel Classifiers

Towards a Ptolemaic Model for OCR

Linear, threshold units. Linear Discriminant Functions and Support Vector Machines. Biometrics CSE 190 Lecture 11. X i : inputs W i : weights

Contents Lecture 4. Lecture 4 Linear Discriminant Analysis. Summary of Lecture 3 (II/II) Summary of Lecture 3 (I/II)

CSE 417T: Introduction to Machine Learning. Final Review. Henry Chai 12/4/18

Intro. ANN & Fuzzy Systems. Lecture 15. Pattern Classification (I): Statistical Formulation

Small sample size generalization

Neural Networks. Prof. Dr. Rudolf Kruse. Computational Intelligence Group Faculty for Computer Science

LINEAR CLASSIFICATION, PERCEPTRON, LOGISTIC REGRESSION, SVC, NAÏVE BAYES. Supervised Learning

Data Mining. Linear & nonlinear classifiers. Hamid Beigy. Sharif University of Technology. Fall 1396

Support Vector Machines. CAP 5610: Machine Learning Instructor: Guo-Jun QI

Spectral and Spatial Methods for the Classification of Urban Remote Sensing Data

Machine Learning for Signal Processing Bayes Classification and Regression

Iterative Laplacian Score for Feature Selection

COMS 4771 Lecture Course overview 2. Maximum likelihood estimation (review of some statistics)

Constrained Optimization and Support Vector Machines

Environmental Sound Classification in Realistic Situations

CS4495/6495 Introduction to Computer Vision. 8C-L3 Support Vector Machines

Machine Learning 2010

Lecture 4 Discriminant Analysis, k-nearest Neighbors

Problem Set 2. MAS 622J/1.126J: Pattern Recognition and Analysis. Due: 5:00 p.m. on September 30

Notes on Discriminant Functions and Optimal Classification

Learning Kernel Parameters by using Class Separability Measure

Introduction to Machine Learning

Machine Learning Practice Page 2 of 2 10/28/13

Intelligent Systems: Reasoning and Recognition. Perceptrons and Support Vector Machines

Classifier Selection. Nicholas Ver Hoeve Craig Martek Ben Gardner

Issues and Techniques in Pattern Classification

SUPERVISED LEARNING: INTRODUCTION TO CLASSIFICATION

Role of Assembling Invariant Moments and SVM in Fingerprint Recognition

UNIVERSITY of PENNSYLVANIA CIS 520: Machine Learning Final, Fall 2013

Feature selection and extraction Spectral domain quality estimation Alternatives

Lecture 13 Visual recognition

The Perceptron. Volker Tresp Summer 2018

Linear & nonlinear classifiers

Pattern Recognition. Parameter Estimation of Probability Density Functions

Introduction to Support Vector Machines

Final Examination CS540-2: Introduction to Artificial Intelligence

Transcription:

009 0th International Conference on Document Analysis and Recognition A Novel Rejection easurement in Handwritten Numeral Recognition Based on Linear Discriminant Analysis Chun Lei He Louisa Lam Ching Y Suen CENPARI (Centre for Pattern Recognition and achine Intelligence) Computer Science and Software Engineering Department, Concordia University ontreal, Quebec, H3G 8, Canada Emails: {cl_he, llam, suen} @ cenparmiconcordiaca Abstract his paper presents a Linear Discriminant Analysis based easurement (LDA) on the output from classifiers as a criterion to reject the patterns which cannot be classified with high reliability his is important in applications (such as in processing of financial documents) where errors can be very costly and therefore less tolerable than rejections o implement the rejection, which can be considered to be a two-class problem of accepting the classification result or otherwise, Linear Discriminant Analysis (LDA) is used to determine the rejection threshold at a new approach LDA is designed to take into consideration the confidence values of the classifier outputs & the relations between them, and it is an improvement over traditional rejection measurements such as First Rank easurement (FR) and First wo Ranks easurement (FR) Experiments are conducted on the CENPARI Arabic Isolated Numerals Database he results show that LDA is more effective, and it can achieve a higher reliability while achieving a high recognition rate Introduction In Optical Character Recognition (OCR) application, current research methods have achieved recognition rates higher than 99% on some Latin numeral databases such as NIS [, ] and CENPARI Database [3] However, even these low error rates can be costly in some applications It is the general expectation that OCR machines should achieve a high recognition rate as well as high reliability, which is defined by: Number of rejected samples Rejection rate = 00% otal number of test samples Recognition rate Reliability = 00% 00% - Rejection rate Achieving high recognition and reliability requires methods capable of assigning generally higher confidences to correct recognition results than to incorrect ones his confidence scoring method may consist of implementing a simple function of appropriate parameters drawn directly from the recognition process, or it may be a learning task in which a classifier is trained to use an array of parameters to distinguish correct recognitions from misclassifications [4] It seems that the Bayes decision rule embodies a rejection rule, namely, the decision can be based on the maximum confidence value provided this maximum exceeds a certain threshold value However, this approach did not perform satisfactorily when experiments were performed on the CENPARI Arabic Isolated Numerals Database he distribution of incorrectly classified samples is not Gaussian in shape, but remains flat throughout a range of confidence values his is the case while the correctly classified samples do follow a Gaussian distribution he results on the training set are shown in Fig In this paper, we modify the LDA method [5] and apply it to the measurement level outputs so that samples with low confidence can also have a Gaussian distribution separate from that of the correctly classified data LDA is a supervised classification method widely used to find the linear combination of features for separating two or more classes he main idea of LDA is to project high-dimensional data onto a line and perform classification in this one-dimensional space It provides a linear projection of the data with the outcome of maximum between-class variance and minimum within-class variance By finding the feature 978-0-7695-375-/09 $500 009 IEEE DOI 009/ICDAR00989 45

space that can best discriminate an object from others, these discriminative methods have been successfully used in pattern classification applications including Chinese character recognition [6], face recognition [7], etc good outputs should be easily separated into two classes: the confidence value of the first rank and others In the following discussion, we assume that the classifier which outputs the probabilities of the patterns for each class would be analogous to the case when the classifier outputs the distances First Rank easurement (FR) & First wo Ranks easurement (FR) Figure Distribution of the output on the raining set Offline handwritten character recognition of languages such as English, Chinese, and Japanese has been researched extensively for over twenty years However, Arabic handwriting recognition research has started only in recent years even though Arabic is one of the most widely spoken languages in the world [8] Currently, Alamri et al [9] designed the CENPARI Arabic database which includes isolated Indian/Arabic numerals, numeral strings, Arabic isolated letters, and Arabic words Experiments reported in this paper are conducted on the isolated numerals in this database Detecting samples recognized with low confidence for rejection and thus achieving a high reliability in handwritten numeral recognition is our objective in this research We define a novel rejection measurement LDA easurement (LDA), and we compare it to other rejection measurements, such as First Rank easurement (FR) and First wo Ranks easurement (FR) in Section hen, we describe the experiments on the CENPARI Arabic Isolated Numerals Database and compare the results obtained from using the three measurements (Section 3) Finally, in Section 4, we conclude that the LDA is more effective than the other two measurements Rejection easurements In considering the outputs of classifiers for the rejection option as a two-class problem, (acceptable or rejected classification), the outputs at the measurement level can be considered as features for the rejection option In an output vector, whose components may represent distances or probabilities, we expect the confidence value (measure) of the first rank (most likely class) to be far distant from the confidence values or measures of the other classes In other words, Generally, rejection strategies can be directly applied to the outputs with a probability estimation In an -class problem, suppose P ( = { p (, p (,, p is the classification output vector of the given pattern x, with probabilities p i ( in descending order he decision function would be based on sgn( Φ ( x ) ), where is a threshold derived from the training data, and Φ ( x ) = p( If Φ (, the classifier rejects the pattern and does not assign it to a class (it might instead be passed to a human operator) his has the consequence that on the remaining patterns, a lower error rate can be achieved his method uses the First Rank easurement (FR) [0] Under this method, the frequency distribution according to confidence values of samples in the training set is considered and the threshold is determined However, FR cannot distinguish between reliable and unreliable patterns with the probability distribution of erroneous samples shown in Fig o overcome this deficiency of FR, we have designed First wo Rank easurement (FR) [], which uses the difference between the probabilities p ( and p ( x ) of the first two ranks as a condition of rejection In FR, the measurement function is Φ ( = p ( p(, where can be any distance measurement, and the decision function is based on sgn( Φ ( ), where is a threshold derived from the training set However, FR cannot solve the problem in some cases For example, if p ( x ) p ( x ) is relatively large compared to, but the difference between p ( x ) p 3 ( x ) is much larger, this pattern may still be accepted, when this pattern should have been rejected since the top two classes are close together in terms of relative distance LDA easurement (LDA) 45

o consider the relative difference between the measurements in the first two ranks and all other measurements, LDA is defined and applied Since rejection in classification can be considered as a twoclass problem (acceptance or rejection), we apply LDA [5] for two classes, to implement rejection LDA approaches the problem by assuming that the conditional probability density functions of the two classes are both normally distributed here are n = n + n observations with d features in the n training set, where { x i } i = arise from class n ω and { x i} arise from class ω Gaussian-based discrimination assumes two normal distributions: ( x ω) ~ N ( μ, ) and ( x ω ) ~ N ( μ, ) In LDA, the projection axis (discriminant vector) w for discriminating between two classes is estimated to maximize the Fisher criterion: J ( w) = tr (( w S ww) ( w S B w)) where tr( ) denotes the trace of a matrix, S B and S w denote the between-class scatter matrix and withinclass scatter matrix respectively, and w is the optimal discriminant vector For the two classes ω and ω, with a priori probabilities p and p (it is often assumed that p = p = 0 5 ), the within-class and between-class scatter matrices can be written as S w = p + p =, S B = ( μ μ )( μ μ ) where is the average variance of the two classes It can be shown that the maximum separation occurs when: w = Sw ( μ μ ) = ( p + p ) ( μ μ ) = ( μ μ) o use this principle to the outputs for the rejection option as a one-dimensional application, we define the () two sets G ( = { p, and () G ( x ) = { p(, p3(,, p hen, μ = p(, μ = p ( ), i x = hus, in LDA, = ( p ( μ) = 0, ( p i ( μ ), and = p( pi ( w = Φ 3( = ( ) hen the decision function would be based on sgn( Φ 3( 3 ), where 3 is a threshold derived from the training set, and all values are scaled to [0, ] In summary, when compared to FR and FR, LDA should be more reliable and informative since it compares the relative difference of the measures in the first two ranks with all other measures 3 Experiments and Results he CENPARI Arabic Isolated Numerals Database is used for the experiments It contains 8,585, 6,99, and 6,99 samples in the raining, Validation, and est sets, respectively Since validation was not implemented in this experiment, the raining and Validation sets were combined to be the raining set here are 0 classes (0-9), and five samples of each numeral are shown in Figure Figure Samples from CENPARI Arabic Isolated Numerals Database In the recognition process, the standard procedures of image pre-processing, feature extraction, and classification were implemented In image preprocessing, we perform noise removal, grayscale normalization, size normalization, and binarization of the grayscale images Gradient features were extracted from pseudo gray-scale images [] he Robert filter, which uses the masks and, was applied to 0 0 0 0 calculate the gradient strengths and directions of pixels he direction of the gradient was quantized to 3 levels with an interval of π / 6 he normalized character image was divided into 8 (9 9) blocks After 453

extracting the strengths and directions in each image, the spatial resolution was reduced from 9 9 to 5 5 by down sampling every two horizontal and every two vertical blocks with a 5 5 Gaussian filter Similarly, the directional resolution was reduced from 3 to 6 levels by down sampling with a weight vector[ 4 6 4], to produce a feature vector of size 0 4 400 (5 5 6) oreover, the transformation y = x was applied to make the distribution of the features Gaussian-like he feature set size was reduced to 400 by principal component analysis (KL transform) Finally, we scaled the feature vectors by a constant factor so that the values of feature components range from 0 to Support Vector achines [3] was chosen as a classifier SVs with different kernel functions can transform a non-linear separable problem into a linear separable one by projecting data into the feature space, and then SVs can find the optimal separating hyperplane Radial Basis Function (RBF) was chosen as the kernel in this research he recognition rate on the test set is 9848% (able ), which is significantly higher than the performance (9360%) in [9] on the same database, and the number of errors was 94 (5%), most of which are also unrecognizable by human beings able Performance with different thresholds in LDA compared with [9] hreshold 0 005 035 [9] Recog Rate (%) 9848 940 9006 9360 Error Rate (%) 53 07 0 640 Reliability (%) 9848 9970 9987 9360 he experimental results on the test set are shown in Figure 3 he solid lines represent the distributions of errors, and the dotted lines represent the distribution of correctly recognized samples he distributions based on FR are shown in Figure 3(a) It is similar to the distributions of the raining data Although the correct samples display a Gaussian distribution, the errors are distributed almost evenly for confidence values (measurements) ranging from 04 to, so the graph is too flat to separate correctly and incorrectly classified samples based on FR When compared to FR, FR is more discriminating, as the range of measurements in FR is wider than FR However, the distribution of errors in FR is flat as well LDA is more discriminating than FR and FR his is because the errors plus correctly classified samples with low confidence values are assigned small measurements In LDA, most incorrectly classified samples (78/94) retain very low measurements (less than 005) hus, LDA enables the rejection of samples with low reliability with the thresholds obtained from the training set (a) (b) (c) Figure 3 Distributions of the three measurements on the est Set: (a) FR (b) FR (c) LDA he performances using different thresholds on the three measurements are shown in Figure 4 As illustrated, when the threshold 3 is set at 005, the reliability increases from 9848% to 9970% with LDA, while the reliabilities with FR and FR are 985% and 9848% respectively It shows that LDA is more effective in increasing reliability 454

Reliability 0995 099 0985 098 Reliablity from different rejection measurements in LibSV 0 0 0 03 04 05 06 07 08 09 hreshold FR FR LDA Figure 4 Reliability with different thresholds used on the three measurements When the reliability of LDA is 9970%, there are 6 errors, all of which are shown in Figure 5 As can be seen, these errors are reasonable since even human beings would find it difficult to recognize these samples, or to distinguish between samples of and 3 written in the same styles in Arabic 0 3 3 3 3 3 3 3 3 3 3 3 3 4 5 0 8 Figure 5 Incorrectly classified images his method has also been trained and tested on the NIS database, and the resulting reliabilities display a pattern very similar to that shown in Figure 4 In particular, when the threshold is set at 0, reliabilities of 9897%, 9909% and 9988% are achieved on the test set for FR, FR and LDA, with recognition rates of 9897%, 9884% and 966% respectively 4 Conclusion he rejection option is very useful in preventing misclassifications, especially in applications which require a high reliability We designed a novel rejection criterion using the LDA easurement (LDA), which relies on the principle of LDA and considers relationships among the probabilities in each output vector We trained this rejection measurement on the training set and tested it on the test set of different databases At the same time, we compared LDA with other measurements such as FR and FR he results indicate that LDA achieved a higher reliability than the other measurements when a small threshold was set In the future, we can apply this rejection method to train multi-classifier systems oreover, we can also apply this methodology to semi-supervised learning, so that we could reject the data with unreliable classification results produced by supervised learning 5 References [] F Lauer, C Y Suen, and G Bloch, A trainable feature extractor for handwritten digit recognition, Pattern Recognition, 40(6), 007, pp 86 84 [] P Zhang, D Bui, and C Y Suen, A novel cascade ensemble classifier system with a high recognition performance on handwritten digits, Pattern Recognition, 40(), 006, pp 345 349 [3] C L Liu, K Nakashima, H Sako, and H Fujisawa, Handwritten digit recognition: Investigation of normalization and feature extraction techniques, Pattern Recognition, 37(), 004, pp 65 79 [4] J F Pitrelli and P Perrone, Confidence-scoring post-processing for off-line handwritten-character recognition verification, Proc of 7 th Int Conference on Document Analysis and Recognition (ICDAR 03), I, 003, pp 78 8 [5] R A Fisher, he use of multiple measurements in taxonomic problems, Annals of Eugenics, 7, 936, pp 79 88 [6] -F Gao and C-L Liu, High accuracy handwritten Chinese character recognition using LDA-based compound distances, Pattern Recognition, 4, 008, pp 344 345 [7] F ang and H ao, Fast linear discriminant analysis using binary bases, Pattern Recognition Letters, 8 (6), 007, pp 09 8 [8] A El Sagheer, N suruta, and R-I aniguchi, Arabic lip-reading system: A combination of Hypercolumn Neural Network odel with Hidden arkov odel, Artificial Intelligence and Soft Computing ASC, arbella, Spain, 004, 9 93 [9] H Alamri, J Sadri, C Y Suen, and N Nobile, A Novel Comprehensive Database for Arabic Off-Line Handwriting Recognition, Proc of the th Int Conference on Frontiers in Handwriting Recognition (ICFHR 008), ontreal, Canada, 008, pp 664 669 [0] J X Dong, A Krzyzak, and CY Suen, Fast SV training algorithm with decomposition on very large datasets, IEEE rans Pattern Analysis and achine Intelligence, 7(4), 005, pp 603 68 [] C Y Suen, C Nadal, R Legault, A ai, and L Lam, Computer recognition of unconstrained handwritten numerals, Proc IEEE, 80(7), 99, pp 6 80 [] Shi, Y Fujisawa, Wakabayashi, and F Kimura, Handwritten numeral recognition using gradient and curvature of gray scale image, Pattern Recognition, 35(0), 00, pp 05 059 [3] V Vapnik and A Lerner, Pattern recognition using generalized portrait method, Automation and Remote Control, 4, 963, pp 774 780 455