UBSegNet: Unified Biometric Region of Interest Segmentation Network

UBSegNet: Unified Biometric Region of Interest Segmentation Network

Ranjeet Ranjan Jha, Daksh Thapar, Aditya Nigam School of Computing and Electrical Engineering,
Indian Institute of Technology Mandi
Mandi, India - 175005
Email: d16044,s16007@students.iitmandi.ac.in, aditya@iitmandi.ac.in
   Shreyas Malakarjun Patil Department of Electrical Engineering,
Indian Institute of Technology, Jodhpur
Jodhpur, Rajasthan, India
Email: patil.3@iitj.ac.in
Abstract

Digital human identity management, can now be seen as a social necessity, as it is essentially required in almost every public sector such as, financial inclusions, security, banking, social networking . Hence, in today’s rampantly emerging world with so many adversarial entities, relying on a single biometric trait is being too optimistic. In this paper, we have proposed a novel end-to-end, Unified Biometric ROI Segmentation Network (), for extracting region of interest from five different biometric traits face, iris, palm, knuckle and 4-slap fingerprint. The architecture of the proposed consists of two stages: (i) Trait classification and (ii) Trait localization. For these stages, we have used a state of the art region based convolutional neural network (RCNN), comprising of three major parts namely convolutional layers, region proposal network (RPN) along with classification and regression heads. The model has been evaluated over various huge publicly available biometric databases. To the best of our knowledge this is the first unified architecture proposed, segmenting multiple biometric traits. It has been tested over around images ( images per trait) and produces very good results. Our work on unified biometric segmentation, opens up the vast opportunities in the field of multiple biometric traits based authentication systems.

Multi-Modal Biometrics, CNN, RCNN;

I Introduction

In the prevailing world, biometric traits used for identification of an individual, is the best possible user friendly, effective and efficient human identity management solution. Over the last few decades, many different biometric traits have been developed and explored extensively, such as face [1], iris [2], fingerprint [3], knuckle [4], palm [5], ear [6] . Due to the uniqueness of these biometric traits they have achieved similar performance with an improved user experience, as compared to other existing token/knowledge based alternatives and people have started using them in many places regarding their security concerns. Traditionally in many places where security is concerned, password and PIN based methods had been used in the past. In password based authentication systems, every person needs to remember his or her password. It has proved to be an efficient method (only up-to few passwords) but often leads to people forgetting their passwords. Also they are pretty easy to crack, which results in huge losses to individuals in terms of privacy as well as finance. Because of these limiting nature of password related security, biometric identification and authentication methods are preferred over traditional passwords and PIN based methods. The physical nature of the physiological biometric based identification, is the key to its secure trait, therefore it has grown to be a vast area of research in recent times.

Any biometric based authentication system contains the following six modules, (i) Data acquisition, (ii) Region of Interest (ROI) extraction, (iii) Quality estimation, (iv) Data pre-processing (v) Feature extraction and finally (vi) Matching and fusion.

Motivation : The ROI extraction is a very early step and one can easily infer that its performance plays a pivotal role in the overall system performance as all subsequent modules has to work over the region currently extracted. Several state of the art methods are available for the segmentation of any individual biometric trait. Since uni-modal systems performance got limited due to several external and environmental factors, researchers have moved to multi-modal biometric system, which uses combination of several traits for identification. Therefore a robust and efficient multi-modal biometric ROI extraction algorithm is necessarily required. Few example traits utilized are shown in Figure 1.

Fig. 1: Examples of the Biometric Traits used

Related Work : Huge amount of work has already been done in order to segment individual biometric ROI. But to best of our knowledge no multi-modal ROI extraction has been reported till now.

In  [3], four slap fingerprint segmentation has been performed by clustering and averaging pixel intensity of various non-overlapping boxes to localize the four fingerprints. Several iris segmentation methods has been discussed in  [7]. They mostly utilize circular hough transform or its variants for segmenting the inner circle. For segmenting the outer circle they applies the circular integro-differential operator. The face detection algorithm proposed by Viola and Jones [8], has been extensively used till date that uses a novel representation of images known as integral image and formulates a classifier using cascaded AdaBoost classifiers. Similarly, there are several knuckleprint ROI extraction algorithms such as the one presented in [9] and [10]. They have used curvature gabor filters to estimate the middle knuckle line and middle knuckle point. Finally, an area in accordance with the image size around the point has been considered. In [11], palm region has been extracting using similarity constraints such as thresholding and few spatial information.

Contribution : The major contribution of our work is to provide a single fully automated network for biometric ROI extraction that can be trained and perform well for a vast variety of available biometric traits. In this paper, we have proposed an unified biometric ROI segmentation network (), which can take input as any face, iris, palm, knuckle or four slap fingerprints image and provides the actual region of interest in that image as shown in Fig. 7. Also, it can classify the extracted ROI into different biometric traits, so that it can later be matched with the appropriate gallery sample. To the best of our knowledge, this is the first unified deep learning architecture utilized to classify and localize any type of biometric sample.

Justification for Model Selection : We have propose a method for region extraction using deep convolutional neural network. In traditional object detection and classification approaches, first image is passed through a region proposal algorithm or network (), that returns multiple prospective candidate bounding boxes. These boxes are later given to a classifier for classification as an input. One example of such an approach is the Region based Convolutional Neural Network (RCNN) [12], in which the region proposals are fed to a CNN classifier, but such models have very high time complexities with a good accuracy. Object Detection has also been performed as a single regression problem, as attempted in the case of YOLO [13]. Even though it is faster than RCNN, the issue with YOLO is that it several times it gives inaccurate results. Also its training requires more efforts and we have observed that in the case of biometric samples, its features are not very well transferable. Hence, we have selected a faster version of which gives much better accuracy as well as transferable features as compared with YOLO, named as Faster RCNN [14].

Rest of the paper has been organized as follows: Section 2, explains in detail the proposed Deep CNN architecture. Section 3, discuss the databases and testing protocol. Section 4, provides the experimental results and Section 5 concludes the paper.

Fig. 2: Overview of the proposed network. There are two heads, (a) Classification and (b) Bounding Box heads.

Ii Proposed UBSegNet

In this section, we have discussed all the techniques that we have utilized in order to train the proposed , so as to obtain satisfactory results in the case of a multi-modal segmentation network. An overview for the proposed network has been shown in Figure 2.

Ii-a UBSegNet : The Network Architecture

The proposed , mainly consisting of these components, a) A set of shared convolutional layers for extracting the discriminative features, b) A region proposal network for regenerating candidate bounding boxes, c) ROI ( Region of Interest) Pooling layer and finally d) A regression and a classification heads as shown in Figure 2.

(a) Shared layers : The Convolutional layers are shared between regression and classification heads and the region proposal network. Generally, these layers can be take from any popular networks (pre-trained over ImageNet) such as the ResNet or VGG/AlexNet. In the proposed network after some experimentation (over available public biometric databases), we have inferred that utilizing few layers of ResNet [15] is the best option. Such empirical analysis enable us to extract many discriminative features. Since over dataset is totally different from the one over which ResNet got initially trained, we have carefully pruned the network from the end up-to a point where one can achieve desired accuracy with lesser layers and very fast. It has been observed that a layered network (subset of full ResNet) gives similar results, for our dataset with much smaller time and space complexities.

(b) Region Proposal Network : The Region Proposal Network (RPN) proposes various regions, over which the regression and classification heads are applied. The RPN takes a matrix, as an input, from the feature map obtained from previously defined shared layers. It then considers several anchor boxes of different scales and aspect ratios (as shown in Fig. 3), so as to select the best fit anchor box for every ground truth bounding box. Later on, it selects the coordinates of these anchor boxes by regressing them the ground truth bounding box. The similarity between the anchor and the bounding boxes are measured using Intersection over Union (). For each of the bounding box, at-least one anchor has to be chosen. These anchor boxes are further pruned to one (in case if they are more) per bounding box, using non-maximum suppression(). The network, gives an output of and values, signifying the coordinates of the anchor boxes ( values per box) and the probability ( one value per box) of box existence respectively.

Fig. 3: Region Proposal Network Implementation

In Figure 3, we have shown the implementation details of the region proposal network. It takes an input of size and uses a convolution layer to learn filter of size . The regression and probability values are computed by learning and convolution filters of size of depth respectively, over the obtained feature maps. The anchor boxes are chosen to be scale and shape invariant. Three aspect ratios has been considered for anchor box at three different scales as shown in Figure 4. Hence, a total of nine anchor boxes has been considered for each window under consideration.

Fig. 4: Anchor Boxes

(c) ROI Pooling Layer : The ROI pooling takes as input, an arbitrary sized matrix (basically a region proposed by ) and converts it to a fixed size vector (), and apply max pooling over such a re-sized grid. It is easy to back propagate through this layer as it is just a max pooling applied over to different regions of a feature map.

(d) Classification and Regression heads : Finally, the network turns into two heads predicting the class scores and bounding box coordinates. The multiple regions obtained after ROI pooling are finally fed to a network consisting of a few convolutional layers and a few fully connected layers to predict the class scores and the regression parameters (bounding boxes coordinates). The layer specific details of the proposed has been shown in Table I.

Network Component Layer Specifications
Type, # Filters Filter Size Batch Nor Activation Function
Shared Layers First ResNet Layers
Region Conv2D, 512 3*3 Yes Relu
Proposal Per Anchor 4 1*1*512 For BB Regression
Network Per Anchor 1 1*1*512 For Classification Score
ROI Pooling Resize ever block to 14*14 and Maxpooling
Classification Conv2D, 512 1*1 Yes Relu
Head Conv2D, 512 3*3 Yes Relu
Conv2D, 2048 1*1 Yes Relu
Regression Conv2D, 2048 1*1 Yes Relu
TABLE I: Architecture of , as shown in Figs. 2, 3

Ii-B Training

(a) Ground truth generation: The ground truth with respect to face, iris, palm, knuckle and 4-slap fingerprint has been generated using [8], [7], [5], [9], [3] respectively. The RPN, outputs 4K and K values corresponding to K anchor boxes for each input. We have used as a similarity measure between the anchor boxes and the bounding boxes provided as ground truth. The anchor boxes with the maximum IOU while compared with the ground truth are given high probabilities, termed “positive”. It is ensured that each of the bounding boxes has to have at-least one positive anchor box corresponding to it.

(b) Training RPN network : Initially, we have trained the region proposal network along with the shared layers using the above computed ground truths for the RPN. We are training from scratch rather than considering pre-trained weights in order to make out the trained model as problem specific as possible. One has to notice, that RPN along with the shared layers has to be trained as an end-to-end network so as to achieve good performance.

(c) Training classification and regression heads : In the next step, we have to train the classification and the regression heads using the obtained region proposals. This also has to be carried out in end-to-end fashion through ROI pooling layer and shared convolutional layers.

(d) Fine Tuning RPN : Once we have trained the shared layers for RPN and both heads (as in Steps (b),(c)), the best possible and discriminative features have been learned at shared layers attaining the maximum accuracy. But the problem is, that RPN is trained as end-to-end in Step (b), along with shared layers. Hence, we have fine tuned the RPN layers keeping the shared layers frozen, in order to learn the anchor box prediction and their probabilities.

(e) Fine tuning the Classification and Regression heads : Similarly, the classification and regression heads has to be fine tuned in order to take a different feature map as an input, keeping the weights of shared layers frozen, to get satisfactory results.

(f) Losses : In order to train such a deep network, we have considered four kinds of losses, (i) RPN regression loss, (ii) RPN classification loss, (iii) Final regression loss and (iv) final classification loss. During every epoch, it first trains the RPN network followed by training the final regression and classification heads. The losses used for RPN classification as well as trait classification are “binary cross-entropy” and “categorical cross-entropy” respectively. The mean squared error (MSE), loss function has been used for regression of both region proposal network (RPN) as well as bounding boxes.

Size invariant network : Our network can take input of any size (size invariant), mainly because of this tweaked implementation of region proposal network (RPN) and the classification and regression heads. The region proposal network has been implemented as convolutional layers and the ground truth is corresponding to the image size, making our RPN size invariant. In the case of classification and regression heads, the ROI pooling layer serves this purpose, as the pooling layers take in any arbitrary sized region of interest (ROI) and pools it into a fixed sized output as discussed above. This fixed sized output has been fed to a network consisting of convolutional and fully connected layers, making the classification and regression heads size invariant too.

All the network hyper-parameters, have been selected “emperically” by maximizing the system performance over a validation set.

Iii Testing Strategy Used

We have used over images for training while around images has been used for testing, per trait, in-order to generate the proposed . The trained has been tested by evaluating intersection over union (), that we have used to obtain the accuracy of our proposed network. It is the most widely used evaluating parameter, to check the efficiency of any algorithm/network, for object localization. Iterative thresholding has been applied to over each of the traits individually, as well over all the traits to determine the individual trait as well as overall performance analysis.

(a) Accuracy Vs IOU Graph : To visualize system performance, we have plotted a graph, showing accuracy at each threshold for each trait as well as overall, as shown in Fig.  5. The IOU ranges from to . Where , indicates that the boxes do not match at all and indicates that the boxes are perfectly matched. When the threshold is high the number of images (in %) having IOU more than the threshold will be less, where as it is at threshold.

Such a graph can be plotted as follows: Compute IOU of predicted and ground truth boxes. The predicted boxes having the same predicted predicted as that of the ground truth along with their respective distance from ground truth has been considered to match the boxes in images containing multiple boxes ( four slap images). Generate a histogram over IOU values at an step of so as to get a smooth curve. Normalize it, so as to get a probability distribution function (PDF) and compute its cumulative distribution function (CDF).

The accuracy at each IOU threshold () can be defined as : and can be computed using the Eq. (1), as shown in Fig. 5.

(1)
Fig. 5: Steps involved for generating Accuracy Vs IOU Graph

(b) Precision and Recall : In addition to the accuracy values, precision and recall has also been calculated for the proposed network validation as defined in Eqs. (2), (3).

(2)
(3)

Precision and recall are calculated so as to validate our approach, while calculating accuracy we only consider the true predicted boxes and not all the predicted boxes, hence the need for precision. Similarly the intersection over union values calculated are with respect to the ground truth bounding boxes, but it may so happen that all the ground truth boxes are not considered while calculating accuracy, therefore we take into account this detail while computing the recall values.

Biometric Traits Accuracy Precision Recall
Overlap IOU Threshold Overlap IOU Threshold Overlap IOU Threshold
0.35 0.5 0.65 0.35 0.5 0.65 0.35 0.5 0.65
Knuckle 99.14 93.72 78.3 98.94 93.48 77.97 98.77 93.38 77.00
Iris 96.78 90.24 87.42 94.84 88.43 85.67 96.78 90.24 87.42
Face 98.46 98.34 98.27 97.46 97.34 97.27 97.91 97.77 97.70
Palm 99.66 99.52 98.33 98.14 98.00 96.83 97.80 97.66 96.50
4 Slap Fingerprint 99.74 96.94 84.38 99.24 96.46 83.95 98.70 95.93 83.49
All Traits 99.00 96.62 90.53 98.51 96.14 90.08 97.97 95.61 89.58
TABLE II: The Accuracy, Precision and Recall Values at different Overlap (IOU) thresholds

(c) Four Slap Fingerprint Testing : Testing for all the other traits has been performed by taking the bounding box predicted with the maximum probability and finding the IOU for it, the ground truth. But for 4-slap fingerprint, the testing mechanism has to be varied as there are multiple boxes in the ground truth. Hence, we have used a method using -axis projection and distances to address this issue.

We first take the projection of the bounding boxes on -axis, and compute the overlap between each of the predicted bounding boxes. If the overlap between any two boxes, is greater than a empirically selected threshold, then drop those boxes which are below to any other reported block. This is done as multiple fingerprint bounding boxes have been reported by , but we know that there must not be two fingerprint one below other. Basically we have dropped the lower fingerprint in these cases. After getting all predicted bounding boxes, we have computed the distance each of them the ground truth boxes and then considering the box with minimum distance as the “corresponding” box. Finally, we have computed the IOU between, “corresponding” predicted and ground truth boxes.

Iv Experimental Results and Discussion

The experiments in accordance with the testing strategy and parameters as mentioned above has been conducted in two fold, (i) Individual biometric traits in which images has been used for testing. The Accuracy vs. IOU threshold graph is shown in Fig. 6, where different colours have been used to plot the graphs for different traits. In Fig. 6, the combined Accuracy Vs. Overlap Threshold graph validates the effectiveness of our proposed . Table II, shows the values obtained for Accuracy, Precision and Recall for the all experiments performed. One can observe that the network produces high accuracy even up-to overlap IOU threshold for almost all the traits. Slight accuracy drop has been observed when overlap IOU threshold becomes more than , especially for knuckle. It may be as the curvature like features are evenly distributed and most of the previous approaches [10] tried to obtain the centre line or the center point. Network may not be able to capture such a symmetry. Even though it is performing better than the existing systems.

Network was performing excellently for face and palm traits. The prime reason behind this is that these traits contain features that are easily distinguishable from the others under their respective region of interests. In terms of precision and recall very similar trends have been observed. From Table II, Fig. 6 one can infer that the proposed network has been performing very well across all the traits. Some network predictions are shown in Figure 7.

Fig. 6: Accuracy Vs. Overlap Threshold Graph when tested for all the traits individually and combined

Comparative Performance Analysis : To best of our knowledge, this is the first ever proposed multi-modal biometrics deep learning segmentation network. Hence, we have not compared our results with any other method. Although, one can compare it with the existing techniques, such as [8], [7], [5], [9], [3], but such comparison may not be justified due to two reasons : (i) They have been tested only over single trait (we have performed multi-class classification) and (ii) None of them have used deep learning. Still we have observed that the proposed network performs better that previous individual trait techniques.

Fig. 7: ROI obtained for various traits using UBSegNet. Colored box shows our predicted box along with its probability

Network Limitations : We have tried to train on ear data. It became a very challenging task due to these three reasons : (1) Less amount to annotated data, (2) unavailability of any good segmentation algorithm, (3) It has features very similar face and started our network started to confuse it with face. Due to its same texture as that of face and similar shape the region proposal was quite effective but the classification accuracy was not satisfactory. Hence we inferred that, if two traits are “similar” or “subset” of each other, network will not be trained optimally. In future, we will incorporate ear as well, extending it to a six class biometric segmentation problem.

V Conclusion

In this paper, we have proposed a novel end-to-end network for extracting ROI from any biometric trait (face, iris, palm, knuckle and 4-slap fingerprint). To best of our knowledge this is the first unified architecture proposed, segmenting multiple biometric traits. It has been tested over around images and produces very good results. Our work on unified biometric segmentation, opens up the vast opportunities in the field of multi-biometric authentication systems.

References

  • [1] R. Ajmera, A. Nigam, and P. Gupta, “3d face recognition using kinect,” in Proceedings of the 2014 Indian Conference on Computer Vision Graphics and Image Processing.   ACM, 2014, p. 76.
  • [2] A. Nigam, V. Krishna, A. Bendale, and P. Gupta, “Iris recognition using block local binary patterns and relational measures,” in IEEE International Joint Conference on Biometrics (IJCB), 2014.   IEEE, 2014, pp. 1–6.
  • [3] N. Singh, A. Nigam, P. Gupta, and P. Gupta, “Four slap fingerprint segmentation,” in International Conference on Intelligent Computing.   Springer, 2012, pp. 664–671.
  • [4] A. Nigam and P. Gupta, “Multimodal personal authentication system fusing palmprint and knuckleprint,” in International Conference on Intelligent Computing.   Springer, Berlin, Heidelberg, 2013, pp. 188–193.
  • [5] G. Badrinath and P. Gupta, “Palmprint based recognition system using phase-difference information,” Future Generation Computer Systems, vol. 28, no. 1, pp. 287 – 305, 2012.
  • [6] A. Nigam and P. Gupta, “Robust ear recognition using gradient ordinal relationship pattern,” in Asian Conference on Computer Vision.   Springer, Cham, 2014, pp. 617–632.
  • [7] A. Bendale, A. Nigam, S. Prakash, and P. Gupta, “Iris segmentation using improved hough transform.” in ICIC (3).   Springer, 2012, pp. 408–415.
  • [8] P. Viola and M. J. Jones, “Robust real-time face detection,” International journal of computer vision, vol. 57, no. 2, pp. 137–154, 2004.
  • [9] G. Jaswal, A. Nigam, and R. Nath, “Deepknuckle: revealing the human identity,” Multimedia Tools and Applications, 2017.
  • [10] A. Nigam and P. Gupta, “Finger-knuckle-print roi extraction using curvature gabor filter for human authentication.” in VISIGRAPP (3: VISAPP), 2016, pp. 366–373.
  • [11] S. Shekhar, B. S. Kumar, and S. Ramesh, “Robust approach for palm (roi) extraction in palmprint recognition system,” in 2012 IEEE International Conference on Engineering Education: Innovative Practices and Future Trends (AICERA), July 2012, pp. 1–6.
  • [12] R. Girshick, J. Donahue, T. Darrell, and J. Malik, “Rich feature hierarchies for accurate object detection and semantic segmentation,” in The IEEE Conference on Computer Vision and Pattern Recognition (CVPR), June 2014.
  • [13] J. Redmon, S. Divvala, R. Girshick, and A. Farhadi, “You only look once: Unified, real-time object detection,” in Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, 2016, pp. 779–788.
  • [14] S. Ren, K. He, R. Girshick, and J. Sun, “Faster r-cnn: Towards real-time object detection with region proposal networks,” IEEE transactions on pattern analysis and machine intelligence, vol. 39, no. 6, pp. 1137–1149, 2017.
  • [15] K. He, X. Zhang, S. Ren, and J. Sun, “Deep residual learning for image recognition,” in Proceedings of the IEEE conference on computer vision and pattern recognition, 2016, pp. 770–778.
Comments 0
Request Comment
You are adding the first comment!
How to quickly get a good reply:
  • Give credit where it’s due by listing out the positive aspects of a paper before getting into which changes should be made.
  • Be specific in your critique, and provide supporting evidence with appropriate references to substantiate general statements.
  • Your comment should inspire ideas to flow and help the author improves the paper.

The better we are at sharing our knowledge with each other, the faster we move forward.
""
The feedback must be of minimum 40 characters and the title a minimum of 5 characters
   
Add comment
Cancel
Loading ...
276256
This is a comment super asjknd jkasnjk adsnkj
Upvote
Downvote
""
The feedback must be of minumum 40 characters
The feedback must be of minumum 40 characters
Submit
Cancel

You are asking your first question!
How to quickly get a good answer:
  • Keep your question short and to the point
  • Check for grammar or spelling errors.
  • Phrase it like a question
Test
Test description