4.6 Article

ViT-Cap: A Novel Vision Transformer-Based Capsule Network Model for Finger Vein Recognition

Journal

APPLIED SCIENCES-BASEL
Volume 12, Issue 20, Pages -

Publisher

MDPI
DOI: 10.3390/app122010364

Keywords

finger vein; biometrics; computer vision; deep learning

Funding

  1. Key R&D Project of Jilin Provincial Science and Technology Development Plan in 2020 [20200401103GX]

Ask authors/readers for more resources

In this study, a new model combining the vision transformer architecture with the capsule network (ViT-Cap) was proposed for finger vein recognition. The model explores finger vein image information based on global and local attention and selectively focuses on important finger vein feature information. Experimental results showed that the proposed model achieved better recognition accuracy compared to the original vision transformer, capsule network, and other advanced finger vein recognition algorithms. Moreover, the model achieved state-of-the-art performance in terms of equal error rate (EER), particularly on the FV-USM datasets, demonstrating its effectiveness and reliability in finger vein recognition.
Finger vein recognition has been widely studied due to its advantages, such as high security, convenience, and living body recognition. At present, the performance of the most advanced finger vein recognition methods largely depends on the quality of finger vein images. However, when collecting finger vein images, due to the possible deviation of finger position, ambient lighting and other factors, the quality of the captured images is often relatively low, which directly affects the performance of finger vein recognition. In this study, we proposed a new model for finger vein recognition that combined the vision transformer architecture with the capsule network (ViT-Cap). The model can explore finger vein image information based on global and local attention and selectively focus on the important finger vein feature information. First, we split-finger vein images into patches and then linearly embedded each of the patches. Second, the resulting vector sequence was fed into a transformer encoder to extract the finger vein features. Third, the feature vectors generated by the vision transformer module were fed into the capsule module for further training. We tested the proposed method on four publicly available finger vein databases. Experimental results showed that the average recognition accuracy of the algorithm based on the proposed model was above 96%, which was better than the original vision transformer, capsule network, and other advanced finger vein recognition algorithms. Moreover, the equal error rate (EER) of our model achieved state-of-the-art performance, especially reaching less than 0.3% under the test of FV-USM datasets which proved the effectiveness and reliability of the proposed model in finger vein recognition.

Authors

I am an author on this paper
Click your name to claim this paper and add it to your profile.

Reviews

Primary Rating

4.6
Not enough ratings

Secondary Ratings

Novelty
-
Significance
-
Scientific rigor
-
Rate this paper

Recommended

No Data Available
No Data Available