指关节指纹分类:利用曼巴视觉低复杂性和高精度

Chiron Bang , Ali Salem Altaher , Ahmed Altaher , Hawraa Moamin , Thamer Alshammari , Basmh Alkanjr , Hasan Altaher , Mohammed G. Al-Jassani , Hanqi Zhuang
{"title":"指关节指纹分类:利用曼巴视觉低复杂性和高精度","authors":"Chiron Bang ,&nbsp;Ali Salem Altaher ,&nbsp;Ahmed Altaher ,&nbsp;Hawraa Moamin ,&nbsp;Thamer Alshammari ,&nbsp;Basmh Alkanjr ,&nbsp;Hasan Altaher ,&nbsp;Mohammed G. Al-Jassani ,&nbsp;Hanqi Zhuang","doi":"10.1016/j.fraope.2025.100261","DOIUrl":null,"url":null,"abstract":"<div><div>Although FKP has been recognized as a viable alternative to other biometric modalities, its adoption remains in the early stages due to its accuracy being lower than that of more competitive options. This research aims to bridge this performance gap by presenting an advancement in the classification of Finger Knuckle Prints (FKPs) using the Vision Mamba (ViM) model. The experimental study, conducted on a dataset from Hong Kong Polytechnic University containing 7,920 images from 165 individuals, evaluated the ViM model’s performance against several pretrained classification models. ViM achieved an impressive accuracy of 99.1%, outperforming other models such as AlexNet (96.2%), SCNN (98.3%), and EfficientNet (98.0%), highlighting its superior capability in FKP classification. With around 7 million parameters, ViM balances complexity and performance, engineered specifically to capture fine-grained FKP features, such as texture and line patterns. Its use of weight decay mitigates overfitting, and it demonstrates resilience in occlusion scenarios by maintaining performance despite missing FKP components. Spatial attention mechanisms further enhance classification accuracy by prioritizing the most informative regions. Seventeen pretrained deep neural networks were evaluated for their effectiveness in classifying FKPs, with experimental results consistently demonstrating the superior performance of the ViM model. ViM exemplifies an advanced deep learning approach for biometric applications, combining high precision with efficient resource usage. Its versatile design – incorporating bidirectional SSMs for global context modeling and position embeddings for spatial awareness – extends its applicability to visual tasks beyond FKP identification. However, users should take its complexity and resource requirements into account for practical implementation.</div></div>","PeriodicalId":100554,"journal":{"name":"Franklin Open","volume":"11 ","pages":"Article 100261"},"PeriodicalIF":0.0000,"publicationDate":"2025-04-23","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":"{\"title\":\"Finger Knuckle Print classification: Leveraging vision Mamba for low complexity and high accuracy\",\"authors\":\"Chiron Bang ,&nbsp;Ali Salem Altaher ,&nbsp;Ahmed Altaher ,&nbsp;Hawraa Moamin ,&nbsp;Thamer Alshammari ,&nbsp;Basmh Alkanjr ,&nbsp;Hasan Altaher ,&nbsp;Mohammed G. Al-Jassani ,&nbsp;Hanqi Zhuang\",\"doi\":\"10.1016/j.fraope.2025.100261\",\"DOIUrl\":null,\"url\":null,\"abstract\":\"<div><div>Although FKP has been recognized as a viable alternative to other biometric modalities, its adoption remains in the early stages due to its accuracy being lower than that of more competitive options. This research aims to bridge this performance gap by presenting an advancement in the classification of Finger Knuckle Prints (FKPs) using the Vision Mamba (ViM) model. The experimental study, conducted on a dataset from Hong Kong Polytechnic University containing 7,920 images from 165 individuals, evaluated the ViM model’s performance against several pretrained classification models. ViM achieved an impressive accuracy of 99.1%, outperforming other models such as AlexNet (96.2%), SCNN (98.3%), and EfficientNet (98.0%), highlighting its superior capability in FKP classification. With around 7 million parameters, ViM balances complexity and performance, engineered specifically to capture fine-grained FKP features, such as texture and line patterns. Its use of weight decay mitigates overfitting, and it demonstrates resilience in occlusion scenarios by maintaining performance despite missing FKP components. Spatial attention mechanisms further enhance classification accuracy by prioritizing the most informative regions. Seventeen pretrained deep neural networks were evaluated for their effectiveness in classifying FKPs, with experimental results consistently demonstrating the superior performance of the ViM model. ViM exemplifies an advanced deep learning approach for biometric applications, combining high precision with efficient resource usage. Its versatile design – incorporating bidirectional SSMs for global context modeling and position embeddings for spatial awareness – extends its applicability to visual tasks beyond FKP identification. However, users should take its complexity and resource requirements into account for practical implementation.</div></div>\",\"PeriodicalId\":100554,\"journal\":{\"name\":\"Franklin Open\",\"volume\":\"11 \",\"pages\":\"Article 100261\"},\"PeriodicalIF\":0.0000,\"publicationDate\":\"2025-04-23\",\"publicationTypes\":\"Journal Article\",\"fieldsOfStudy\":null,\"isOpenAccess\":false,\"openAccessPdf\":\"\",\"citationCount\":\"0\",\"resultStr\":null,\"platform\":\"Semanticscholar\",\"paperid\":null,\"PeriodicalName\":\"Franklin Open\",\"FirstCategoryId\":\"1085\",\"ListUrlMain\":\"https://www.sciencedirect.com/science/article/pii/S2773186325000519\",\"RegionNum\":0,\"RegionCategory\":null,\"ArticlePicture\":[],\"TitleCN\":null,\"AbstractTextCN\":null,\"PMCID\":null,\"EPubDate\":\"\",\"PubModel\":\"\",\"JCR\":\"\",\"JCRName\":\"\",\"Score\":null,\"Total\":0}","platform":"Semanticscholar","paperid":null,"PeriodicalName":"Franklin Open","FirstCategoryId":"1085","ListUrlMain":"https://www.sciencedirect.com/science/article/pii/S2773186325000519","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
引用次数: 0

摘要

虽然FKP已被公认为其他生物识别模式的可行替代方案,但由于其准确性低于更具竞争力的选择,其采用仍处于早期阶段。本研究旨在通过使用视觉曼巴(ViM)模型在手指指关节指纹(FKPs)分类方面取得进展,从而弥合这一性能差距。实验研究在香港理工大学的数据集上进行,该数据集包含来自165个人的7920张图像,并将ViM模型与几个预训练的分类模型进行了比较。ViM达到了令人印象深刻的99.1%的准确率,优于AlexNet(96.2%)、SCNN(98.3%)和effentnet(98.0%)等其他模型,突出了其在FKP分类方面的卓越能力。ViM拥有大约700万个参数,平衡了复杂性和性能,专门用于捕获细粒度的FKP特征,如纹理和线条图案。它使用的权重衰减减轻了过拟合,并通过保持性能,尽管缺少FKP组件在遮挡情况下显示弹性。空间注意机制通过优先考虑信息量最大的区域,进一步提高了分类精度。17个预训练的深度神经网络对fkp分类的有效性进行了评估,实验结果一致证明了ViM模型的优越性能。ViM为生物识别应用提供了一种先进的深度学习方法,将高精度与高效的资源利用相结合。它的多功能设计——结合了用于全局上下文建模的双向ssm和用于空间感知的位置嵌入——扩展了它在FKP识别之外的视觉任务中的适用性。但是,在实际实现时,用户应该考虑到它的复杂性和资源需求。
本文章由计算机程序翻译,如有差异,请以英文原文为准。

Finger Knuckle Print classification: Leveraging vision Mamba for low complexity and high accuracy

Finger Knuckle Print classification: Leveraging vision Mamba for low complexity and high accuracy
Although FKP has been recognized as a viable alternative to other biometric modalities, its adoption remains in the early stages due to its accuracy being lower than that of more competitive options. This research aims to bridge this performance gap by presenting an advancement in the classification of Finger Knuckle Prints (FKPs) using the Vision Mamba (ViM) model. The experimental study, conducted on a dataset from Hong Kong Polytechnic University containing 7,920 images from 165 individuals, evaluated the ViM model’s performance against several pretrained classification models. ViM achieved an impressive accuracy of 99.1%, outperforming other models such as AlexNet (96.2%), SCNN (98.3%), and EfficientNet (98.0%), highlighting its superior capability in FKP classification. With around 7 million parameters, ViM balances complexity and performance, engineered specifically to capture fine-grained FKP features, such as texture and line patterns. Its use of weight decay mitigates overfitting, and it demonstrates resilience in occlusion scenarios by maintaining performance despite missing FKP components. Spatial attention mechanisms further enhance classification accuracy by prioritizing the most informative regions. Seventeen pretrained deep neural networks were evaluated for their effectiveness in classifying FKPs, with experimental results consistently demonstrating the superior performance of the ViM model. ViM exemplifies an advanced deep learning approach for biometric applications, combining high precision with efficient resource usage. Its versatile design – incorporating bidirectional SSMs for global context modeling and position embeddings for spatial awareness – extends its applicability to visual tasks beyond FKP identification. However, users should take its complexity and resource requirements into account for practical implementation.
求助全文
通过发布文献求助,成功后即可免费获取论文全文。 去求助
来源期刊
自引率
0.00%
发文量
0
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
确定
请完成安全验证×
copy
已复制链接
快去分享给好友吧!
我知道了
右上角分享
点击右上角分享
0
联系我们:info@booksci.cn Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。 Copyright © 2023 布克学术 All rights reserved.
京ICP备2023020795号-1
ghs 京公网安备 11010802042870号
Book学术文献互助
Book学术文献互助群
群 号:604180095
Book学术官方微信