4.6 Article

ViT-Cap: A Novel Vision Transformer-Based Capsule Network Model for Finger Vein Recognition

期刊

APPLIED SCIENCES-BASEL
卷 12, 期 20, 页码 -

出版社

MDPI
DOI: 10.3390/app122010364

关键词

finger vein; biometrics; computer vision; deep learning

资金

  1. Key R&D Project of Jilin Provincial Science and Technology Development Plan in 2020 [20200401103GX]

向作者/读者索取更多资源

In this study, a new model combining the vision transformer architecture with the capsule network (ViT-Cap) was proposed for finger vein recognition. The model explores finger vein image information based on global and local attention and selectively focuses on important finger vein feature information. Experimental results showed that the proposed model achieved better recognition accuracy compared to the original vision transformer, capsule network, and other advanced finger vein recognition algorithms. Moreover, the model achieved state-of-the-art performance in terms of equal error rate (EER), particularly on the FV-USM datasets, demonstrating its effectiveness and reliability in finger vein recognition.
Finger vein recognition has been widely studied due to its advantages, such as high security, convenience, and living body recognition. At present, the performance of the most advanced finger vein recognition methods largely depends on the quality of finger vein images. However, when collecting finger vein images, due to the possible deviation of finger position, ambient lighting and other factors, the quality of the captured images is often relatively low, which directly affects the performance of finger vein recognition. In this study, we proposed a new model for finger vein recognition that combined the vision transformer architecture with the capsule network (ViT-Cap). The model can explore finger vein image information based on global and local attention and selectively focus on the important finger vein feature information. First, we split-finger vein images into patches and then linearly embedded each of the patches. Second, the resulting vector sequence was fed into a transformer encoder to extract the finger vein features. Third, the feature vectors generated by the vision transformer module were fed into the capsule module for further training. We tested the proposed method on four publicly available finger vein databases. Experimental results showed that the average recognition accuracy of the algorithm based on the proposed model was above 96%, which was better than the original vision transformer, capsule network, and other advanced finger vein recognition algorithms. Moreover, the equal error rate (EER) of our model achieved state-of-the-art performance, especially reaching less than 0.3% under the test of FV-USM datasets which proved the effectiveness and reliability of the proposed model in finger vein recognition.

作者

我是这篇论文的作者
点击您的名字以认领此论文并将其添加到您的个人资料中。

评论

主要评分

4.6
评分不足

次要评分

新颖性
-
重要性
-
科学严谨性
-
评价这篇论文

推荐

暂无数据
暂无数据