3. System Overview and Approach
3.1 Preprocessing of Different Biometric Features
Nine different biometric features such as color-face, color-iris, color-eye, right and left hand fingerprints, hand-writing, palm-print, gait and wrist-vein were used in different classifiers individually. The advantage to keep color patterns were, the classifiers take the color of the patterns as additional features. Thus the classifiers can identify different very similar same features depending on the color. As for example, in case of color-face patterns, the classifier is able to recognize different very similar faces as different faces depending on skin color. All the biometric patterns of training and test dataset must be preprocessed before learning and identification.
There were 8 different preprocessing steps which were applicable on all 8 different biometric features excluding Iris pattern. Also all 8 steps of preprocessing were not required for these 8 different biometric features. Each step is described below for different biometric patterns.
(i) RGB to gray scale image conversion: This step was applicable only on Handwriting patterns. Here all the handwriting patterns were converted from RGB to gray-scale.
(ii) Removal of noise: The patterns of training and test dataset may be noisy/blurred. In this step, all the noises have been removed from different noisy biometric patterns.
(iii) De-blurring the patterns: In this process, for color-face, color-eye, palm-print and wrist-vein patterns Lucy-Richardson method was used and for handwriting and fingerprints of right and left hands, Wiener Filter was utilized to de-blur the blurred patterns and obtain sharp patterns.
(iv) Background elimination: Then the backgrounds of face, right and left hand fingerprints, and handwriting were removed. Gaussian model was utilized to remove the background of facial patterns. Here a sum-of-Gaussians model (2 Gaussians) was utilized for the color at each pixel results in a simple way of separating the background.
(v) Image compression: The patterns were compressed by exchanging block of pixels with the mode value of the pixel intensities where mode is the intensity value which has occurred more frequently in the block. Following are the steps for image compression:
- Compute the number of pixels of the pattern.
- Calculate the block size in such a way that compressed pattern is of a particular size, say x×y pixels.
- In each block calculate the mode value of the pixel intensities.
- Exchange each block by its corresponding mode intensity to get corresponding compressed pattern of size x×y pixels.
(vi) Image normalization: In this step, all the patterns of particular biometric features are normalized into equal and lower dimensions.
(vii) Conversion of gray-scale patterns into binary patterns: Right and left hand fingerprints, handwriting, palm-print patterns were converted into corresponding binary patterns.
(viii) Conversion of RGB/binary patterns into 1D matrix: In this last step, all different biometric patterns such as color-face, color-iris, color-eye, wrist-vein, right and left hand fingerprints, handwriting, palm-print and gait patterns were converted into corresponding 1D matrix files. These sets were the input to the clustering algorithms of corresponding individual classifiers.
• Preprocessing to extract Color-Iris patterns
The necessary steps to extract the color-iris patterns from color-eye patterns are given below:
(i) Compression of eye images: At the first step, all color-eye patterns were compressed.
(ii) Iris boundary localization: For the detection of the iris boundary, the radial-suppression edge detection algorithm [21] was utilized which was similar to Canny edge detection algorithm. A non-separable wavelet transform was utilized in the radial-suppression edge detection algorithm to extract the wavelet transform modulus of the iris image and then radial non maxima suppression was utilized to keep the annular edges and concurrently eliminate the radial edges. Then deduce the final binary edge map by eliminating the isolated edges utilizing an edge thresholding.
Then for the detection of the final iris boundaries, circular Hough Transformation [22] was utilized and also deducts their radius and center. The Hough transform [22] is explained, as in (1), for the circular boundary and a set of recovered edge points xj, yj (with j=1,..., n).
where,
with
For each edge point[TeX:] $$\left( x _ { j } , y _ { j } \right) , g \left( x _ { i } , y _ { j } , x _ { 0 } , y _ { c } , r \right) = 0$$ for each parameter0 xcycr that implies a circle through that point. The triplet maximizing H corresponds to the largest number of edge points that represents the contour of interest.
(iii) Extraction of the iris: Then, the excess portions other than the irises, as for example eyelids, eyelashes and eyebrows were eliminated for the extraction of the iris.
(iv) (iv) Conversion of RGB images into 1D matrix: Finally RGB-iris patterns were converted into 1D matrix files. This set was the input to the SOM network of SOM based modified RBFN classifier.
3.2. Theoretical Approach of the Present System
Five different classifiers such as OCA based RBFN, Modified OCA based RBFN, SOM based RBFN, combination of Malsburg learning and BPN, HBC based RBFN were used to build this multiclassification system. Among of these 5 classifiers, first 4 classifiers were used twice in this multimodal system and overall 9 different biometric features were identified by 9 classifiers. Every biometric feature was trained and tested with 5 different classifiers and finally that classifier was selected for this system which gave best performance in terms of accuracy for corresponding biometric. This system comprised of three super-classifiers. In first super-classifier, color-face, color-iris and color-eye patterns were identified separately using Modified OCA based RBFN, SOM based RBFN and a Malsburg learning and BPN combination respectively and then super-classifier1conclude the person’s identification based on programming based boosting method. In second super-classifier, OCA based RBFN, HBC based RBFN and a Malsburg learning and BPN combination performed identification of right and left hand fingerprints and handwriting respectively and super-classifier2 combine these three different classifiers conclusion depending on programing based boosting logic and conclude the decision. Similarly in third super-classifier, palm-print, gait(silhouettes) and wrist-vein patterns were identified by Modified OCA based RBFN, SOM based RBFN and OCA based RBFN respectively and super-classifier3 conclude the person’s identification using the logic of programing based boosting method. Finally mega-superclassifier integrates all these three super-classifiers decision based on again programing based boosting method to conclude the final identification/authentication of the person (Fig. 1).
For the above mentioned single classifiers in case of modified RBFNs, the networks [23-27]] individually consists of three layers: (1) an input layer for different biometric pattern representation, (2) a hidden (clustering) layer comprising ‘basis units’, and (3) an output (classification) layer. The outputs of different clustering algorithms (mean ‘μ’, standard deviation ‘σ’ and corresponding approximated normal distribution output functions) are applied in ‘basis units’ of corresponding RBFN. Thus, for the above-mentioned single classifiers, OCA, MOCA, SOM and HBC are the first phase of learning for the corresponding classifier and using BP learning the optimal weights are obtained, which is the second phase of learning.
In this multiple classification system, the time complexities of 5 individual single classifiers algorithms are as follows. The time complexity of modified RBFN including OCA and BP learning is O(n*(k*d+p)), the time complexity of modified RBFN including MOCA and BP learning is O(n*(m*k+p)), the time complexity of modified RBFN including HBC and BP learning is O(n*(n+p)), the time complexity of Malsburg learning and Back propagation Network combination is O(n*(n+p)) and the time complexity of modified RBFN including SOM and BP learning is O(n*p+K2). Here, “n” is total patterns in the pattern set, “k” is total clusters formed, “d” is the dimensionality of each and every pattern, “p” is total iteration to optimize the weight, “m” is total iteration where optimal solution is reached and “K” is total map units.
3.2.1 Classifiers of super-classifier1
In the first classifier of super-classifier1, MOCA [27,28] was applied to make groups of the input face pattern set which were used as input of the RBFN. It forms clusters of various expressions of faces of each person and view (person-view). Then BP Learning of RBFN classifies the “person-view” into “person”. Here total input nodes of RBFN was equivalents to the total training face patterns, whereas the output nodes sets to the number of classes and the hidden units were equivalent to the total clusters formed by MOCA.
Block diagram of the present system for testing identification.
In the second classifier of super-classifier1, a Self-organizing network [23] with Kohonen’s learning was utilized to create clusters of the preprocessed input iris pattern set which were used as input of the RBFN. The SOM can form the two dimensional feature map (for this system 15×15 feature map) from which directly the number of clusters can be evaluated. It creates clusters of various expressions of eyes (irises) of every person and of two eyes (left and right). The BP Learning of RBFN combines the irises of two eyes (left and right) of each person into “person iris”.
In the third classifier of super-classifier1, a competitive network (Malsburg Learning) [29] was applied to form clusters of the preprocessed input eye data set. It makes groups of various expressions of eyes of each person’s two eyes (left and right). Then the BPN [29,30] classifier was used which classifies the two eyes (left and right) of each person into “person eye”. From the output layer of BP learning the optimal weights can be obtained. Here total input nodes of BPN was equivalents to the total training eye patterns, the total hidden layer nodes were sets to the total clusters formed by Malsburg learning network and the total output nodes sets to the number of classes.
3.2.2 Classifiers of super-classifier2
In the first classifier of super-classifier 2, OCA [24,25, 31-33] was used to form clusters of the preprocessed input right-hand fingerprints set (thumb, second, third and fourth finger–as per standard dataset CASIA version 5). These clustering outputs were applied as input of the RBFN. Here OCA formed groups of various qualities of fingerprints of each person and individual fingers (person-finger). The BP Learning of RBFN combines the different fingers’ fingerprints of a person into “person fingerprint”.
In the second classifier of super-classifier2, HBC [26] algorithm was used to form groups of the preprocessed input left-hand fingerprints set (thumb, second, third and fourth finger–as per standard dataset CASIA version 5). Similar to the previous classifier, these clustering outputs were applied as input of the RBFN. Here HBC also formed groups of various qualities of fingerprints of each person and individual fingers (person-finger). Then the BP learning of RBFN classifier was used to classify the different fingers’ fingerprints of a person into “person fingerprint” like previous classifier.
In the third classifier of super-classifier2, again the Malsburg learning and BPN combination was utilized for handwriting identification. Here, Malsburg learning network formed groups of various qualities of handwritings (name and surname) of each person. Then the BPN was utilized which classifies the “person name-person surname” into “person name”.
3.2.3 Classifiers of super-classifier3
In the first classifier of super-classifier3, again MOCA was applied to form groups of the preprocessed input palm-print (left and right hand) set which were used as input to the RBFN. Then BP Learning of RBFN combines the palm prints of two hands (left and right) of each person into “person-palm print”.
In the second classifier of super-classifier3, SOM based RBFN was utilized for gait recognition. Silhouettes were taken here as pattern set. We divided the continuous sequences of gait in few time slots and take the alternative slots of gait patterns to train the classifier. The other instances were taken to test the classifier. Here SOM network was used to form the group of gait patterns of same time slots for each person. Finally BP learning of RBFN, classifies the different time slots of gait patterns of each person as “person-gait”.
In the third classifier of super classifier3, OCA based RBFN was utilized for wrist-vein identification. Here OCA is applied to form groups of the input wrist-vein (left and right hand) set which were used as input of the RBFN. Finally BP Learning of RBFN classifies the wrist-veins of two hands (left and right) of each person into “person-wrist vein”.
3.2.4 Programming based boosting
The present multi-classifier used programming based boosting [34,35] in super-classifiers and megasuper- classifier, i.e. super-classifiers and mega-super-classifier concluded the final person identification depending on programming based boosting method incorporating the conclusions of three distinct classifiers or super-classifiers. In this method, the weight of the vote of each classifier or super-classifier is pre-assigned or customized previously. The separate links’ weights from the distinct classifiers or super-classifiers into the integrator are ‘programmed’. These weights are the measurements in respect of normalized accuracy of the distinct classifiers or super-classifiers.
3.2.5 Identification learning with different biometric patterns
To learn the different classifiers with 9 different biometric features (color-face, color-eye/color-iris, one common training database for these two biometrics, right and left hand fingerprints, handwriting, palm-print, gait (silhouettes), and wrist-vein); 8 different training databases were used. Every database comprises of various biometric patterns of different persons.
In color-face database, for each person’s facial patterns, 6 various expressions and also 3 separate angular views, i.e., frontal, 90° left side and 90° right side view were included. In color-iris/eye database, for each person’s iris/eye patterns, 8 various expressions of left and right eyes were taken separately. In the right and left hand fingerprints distinct databases, for each person, 3 various qualities of fingerprints (hard-press, medium-press and soft-press) as well as four different fingers’ fingerprints (thumb, second, third and fourth finger–as per standard database CASIA version 5) were incorporated. The handwriting database comprises of 6 various qualities of handwritings (name and surname separately) for every person. In the palm-print database, both right and left hands’ four different qualities of palm-prints were taken for each person. In the gait database, patterns of different time slots were taken for every person. Finally in the wrist-vein database, both right and left hands’ four various qualities of wrist-vein patterns were taken for each person (Fig. 2).
Samples of some training and test patterns of various biometric features for single classifiers.
All the preprocessed different biometric patterns were fed as input to the corresponding separate classifiers. When the classifiers learned all the different training patterns (9 different biometric features) for all different people, the classifiers were ready for recognition of people through these learned patterns, which were called as of ‘known’ persons. The biometric patterns that were not included and learned during training process of the classifiers are called as of ‘unknown’ persons (Fig. 3).
Block diagram for learning identification of individual single classifiers.
3.2.6 Identification Testing with Different Biometric Patterns
The test databases of different biometrics contained different people’s (similar as training database) patterns (color-face, color-eye/color-iris, right and left hand fingerprints, handwriting, palm-print, gait (silhouettes) and wrist-vein) of different qualities/expressions/instances. These test patterns were entirely non-identical from training databases (Fig. 2).
For the performance evaluation of three different super-classifiers and mega-super-classifier, the test databases for testing contained pattern sets of various people (similar as training database). Every test set of super-classifier1 contained one color-face and color-eye. Each test set of super-classifier2 contained one right-hand fingerprint, left-hand fingerprint and handwriting and each test set of superclassifier3 contained one palm-print, gait (silhouette) and wrist-vein pattern. In the test database of mega-super-classifier, each test set comprised of one color-face, color-eye, right-hand fingerprint, lefthand fingerprint, handwriting, palm-print, gait and wrist-vein pattern. The patterns of every test set were also of several qualities/expressions/instances which were totally non-identical from training databases (Fig. 4). The test sets for 9 single classifiers, three super-classifiers and mega-super-classifier also contained several unknown patterns of various qualities/expressions/instances.
A sample of test pattern set (person3) of mega-super-classifier for person identification.
The test patterns of different biometrics were fed as input to 9 different preprocessors. The preprocessed patterns were taken as inputs to the formerly trained networks of 9 single classifiers. After completion of the training of different classifiers, high output values were obtained for known patterns and low output values were obtained for unknown patterns. A threshold value was necessary to differentiate between known and unknown biometric patterns. The threshold was set as the mean of the minimum output value from known patterns and maximum output value from unknown patterns. This threshold value was different for different biometric patterns. For the given biometric pattern if the respective output value is above threshold then it is considered as known pattern. The BP networks produce different output activation in different overall output units. The probability of belongingness of the given test pattern into the different classes can be obtained from the normalized activation of each output unit. Here the test pattern is considered to belong to a class for which the normalized activation itself presents the probability of belongingness of that input test pattern into the specific class. Then three individual super-classifiers conclude the identifications of the person depending on programming based boosting method incorporating the conclusions of corresponding three distinct single classifiers. Finally the mega-super-classifier determines the final identification of the person depending on again programming based boosting method incorporating the conclusions of three different super-classifiers. Finally we compute the probability of belongingness of the given test pattern set for that corresponding class decided by super-classifiers and mega-super-classifier by considering the minimum value of probability among three different classifiers/super-classifiers.
If for a test pattern set two or more classifiers produce contrary outputs, then the decision obtained by the classifier/ super-classifier with higher weighted link have to be accepted with minimum probability. So, the proposed algorithm of super-classifier/mega-super-classifier performs well also for such contrary conditions (Fig. 1 and Algorithm 1).
Algorithm 1:
Algorithm for Person Identification with Super-classifiers and Mega-super-classifier
Time complexity of Algorithm 1: In the above mentioned algorithm from step 1 to step 5, i.e. in case of single classifiers, complexity is O(n). From step 9 to step 14, for the super-classifiers again the time complexity is O(n) and finally for mega-super-classifier the complexity is O(n). Hence, the total time complexity of the algorithm is O(n).
4. Result and Performance Analysis
Nine different biometric traits were taken from 8 different standard databases to make the training and test databases of this system. It was not possible for us to collect all the variety of biometric patterns from a single standard database. Hence it was presumed that, different biometric patterns of various standard databases were of identical specific people without losing any generality to estimate this multiclassifier’s performance.
We utilized training and test databases for color-face samples from FEI database (http://fei.edu.br/~cet/facedatabase.html), eyes/irises from UTIRIS database (http://utiris.wordpress.com/), right and left hand fingerprints from CASIA Fingerprint Image Database Version 5.0 (http://biometrics.idealtest.org/ dbDetailForUser.do?id=7), handwritings from IAM handwriting database (http://www.iam.unibe.ch/ fki/databases/iam-handwriting-database/download-the-iam-handwriting-database), palm-prints from CASIA Palm print Image Database (http://biometrics.idealtest.org/dbDetailForUser.do?id=5), gaits (silhouettes) from CASIA Gait Database (http://www.cbsr.ia.ac.cn/english/Gait%20Databases.asp) and Wrist-veins from CIE Biometrics (http://biometrics.put.poznan.pl/vein-dataset/).
4.1. Performance Evaluation Metrics of the Classifiers
Holdout method [26,27,29] was utilized for performance evaluation of the individual classifiers, super-classifiers and mega-super-classifier. Confusion matrices have been implemented in case of individual single classifiers, super-classifiers and also for mega-super-classifier.
A Confusion matrix [26-29,34-36]] is a table that is used to describe the performance of a classifier. Figuring a confusion matrix can provide a superior idea of what the classification model is getting right and what kinds of errors it is producing (Fig. 5).
Confusion matrix (2 class).
From the aforesaid binary confusion matrix (Fig. 5) containing only two classes (say P and Q), the accuracy, precision, recall and F-score [26,27,29] are defined as follows:
When holdout method is used to estimate the classifiers’ performance, such kinds of samples are taken for testing which are excluded in training database. When accuracy metric is applied for a classifiers’ performance estimation, the whole performance is reflected irrespective of the distinct performance for each class. That is why accuracy metric is more appropriate for the classifiers’ performance evaluation through a specific numeric value. Precision, recall and F-score metrics are utilized to explain the performance of each class.
4.2. Experimental Results
The present system was developed to learn on a computer with Intel Core 2 Duo E8400, 3.00 GHz processor with 4 GB RAM and Windows 7 32-bit Operating System. The multi-classification system was implemented using MATLAB R2008b.
Some significant part of experimental result which manages contrary case (each classifier is recognizing different person) is given below:
Graphical representation of / vs. threshold of MOCA for color-face (a) and palm-print (b).
Now if general majority voting logic is considered then super-classifier2 decide the given pattern set as of ‘unknown person’ as three separate classifiers produce three different results of identification. But whenever programing based boosting algorithm is applied, the super-classifier2 decides the given pattern set as of person 1 with probability 0.39164. Here, the respective weights of the links of three single classifiers are 0.2899, 0.3445, and 0.3656. The weight of the link associated to third classifier is highest and the minimum graded probability is gained from first classifier. Therefore the maximum weighted conclusion with minimum probability (the safest acceptable probability) is decided as final result.
From Fig. 6, graphical demonstration of [Sb, Sw] vs. Threshold of MOCA explains the maximizing of [Sb, Sw] through T1 to T2. For training color-face patterns, the value of T1 and T2 are 2500 and 7000, and for palm-print 4000 and 4170, respectively. The thresholds were incremented by 500 for color-face and 10 for palm-print. From T1 through T2 desired numbers of clusters are obtained. The mean value between T1 and T2 was the final particular threshold to obtain the perfect clusters without any misclassification.
Confusion matrix for super-classifier1
Confusion matrix for super-classifier2
Confusion matrix for super-classifier3
Confusion matrix for super-classifier4
Tables 1–4 displays the confusion matrices of three different super-classifiers and also mega-superclassifier. Table 5 displays the accuracies of 9 different classifiers for 9 different biometric traits, three different super-classifiers and mega-super-classifier. The accuracies for single classifiers are ≥90% except right hand fingerprint patterns. In this system OCA based RBFN was utilized for right hand fingerprint and wrist-vein identification. OCA may not provide perfect clusters in every case and as a consequence the accuracy for right hand finger-prints is comparatively low. Conventional OCA considers one specific intra cluster similarity or threshold to form the clusters. But in this algorithm inter cluster distances is not considerable. So the possibility for misclassification is there among the group of patterns. MOCA use both intra cluster distance and inter cluster distance to create perfect clusters and avoid misclassification among the clusters. Thus MOCA based RBFN classifiers give higher accuracy for different biometrics. The accuracies for three different super-classifiers are ≥95% and for mega-super-classifier is 98.89%. Therefore, it is evident that the mega-super-classifier is effective for person recognition than the single classifiers and also super-classifiers utilizing single or three biometric traits independently.
Accuracy of the classifiers (Holdout method) Classifiers
Performance measurement of the classifiers (1st super-classifier)
Performance measurement of the classifiers (2nd super-classifier)
Performance measurement of the classifiers (3rd super-classifier)
Performance measurement of the Mega-Super-classifier
From Tables 6–9, precision, recall and F-score metrics explain the performance of every class with holdout method for all the classifiers. Similarly in Table 10, the present multi-classifier displays overall low recognition time (<1 second). The limitation or disadvantage of this multi-classifier is that, it took quite high time for training. But training is a single time process whereas recognition can be done for multiple times. Once the system is completely trained, then it can perform the recognition several times for different inputs as per users’ demand. Hence utilizing multiple classifiers, accurate as well as reliable identification can be obtained with minimum identification time at the cost of training time.
Learning time of the biometric features (unit: second)
Table 11 displays a comparative analysis of the developed multi-classifier, using different parameters with other multimodal systems mentioned in section 2 [14-20]. Compare to other multimodal systems, our developed system effectively deals with 9 different biometric features to give a very secure and reliable person authentication system with higher accuracy as well as low identification time. Therefore, the proposed approach shows improvement in case of both accuracy and identification time as compared to methods mentioned in the Section 2.
Comparative study with other multimodal systems