IEEE transactions on biometrics, behavior, and identity science最新文献

筛选
英文 中文
eDifFIQA: Towards Efficient Face Image Quality Assessment Based on Denoising Diffusion Probabilistic Models eDifFIQA:基于去噪扩散概率模型的高效人脸图像质量评估
IEEE transactions on biometrics, behavior, and identity science Pub Date : 2024-03-12 DOI: 10.1109/TBIOM.2024.3376236
Žiga Babnik;Peter Peer;Vitomir Štruc
{"title":"eDifFIQA: Towards Efficient Face Image Quality Assessment Based on Denoising Diffusion Probabilistic Models","authors":"Žiga Babnik;Peter Peer;Vitomir Štruc","doi":"10.1109/TBIOM.2024.3376236","DOIUrl":"https://doi.org/10.1109/TBIOM.2024.3376236","url":null,"abstract":"State-of-the-art Face Recognition (FR) models perform well in constrained scenarios, but frequently fail in difficult real-world scenarios, when no quality guarantees can be made for face samples. For this reason, Face Image Quality Assessment (FIQA) techniques are often used by FR systems, to provide quality estimates of captured face samples. The quality estimate provided by FIQA techniques can be used by the FR system to reject samples of low-quality, in turn improving the performance of the system and reducing the number of critical false-match errors. However, despite steady improvements, ensuring a good trade-off between the performance and computational complexity of FIQA methods across diverse face samples remains challenging. In this paper, we present DifFIQA, a powerful unsupervised approach for quality assessment based on the popular denoising diffusion probabilistic models (DDPMs) and the extended (eDifFIQA) approach. The main idea of the base DifFIQA approach is to utilize the forward and backward processes of DDPMs to perturb facial images and quantify the impact of these perturbations on the corresponding image embeddings for quality prediction. Because of the iterative nature of DDPMs the base DifFIQA approach is extremely computationally expensive. Using eDifFIQA we are able to improve on both the performance and computational complexity of the base DifFIQA approach, by employing label optimized knowledge distillation. In this process, quality information inferred by DifFIQA is distilled into a quality-regression model. During the distillation process we use an additional source of quality information hidden in the relative position of the embedding to further improve the predictive capabilities of the underlying regression model. By choosing different feature extraction backbone models as the basis for the quality-regression eDifFIQA model, we are able to control the trade-off between the predictive capabilities and computational complexity of the final model. We evaluate three eDifFIQA variants of varying sizes in comprehensive experiments on 7 diverse datasets containing static-images and a separate video-based dataset, with 4 target CNN-based FR models and 2 target Transformer-based FR models and against 10 state-of-the-art FIQA techniques, as well as against the initial DifFIQA baseline and a simple regression-based predictor DifFIQA(R), distilled from DifFIQA without any additional optimization. The results show that the proposed label optimized knowledge distillation improves on the performance and computationally complexity of the base DifFIQA approach, and is able to achieve state-of-the-art performance in several distinct experimental scenarios. Furthermore, we also show that the distilled model can be used directly for face recognition and leads to highly competitive results.","PeriodicalId":73307,"journal":{"name":"IEEE transactions on biometrics, behavior, and identity science","volume":"6 4","pages":"458-474"},"PeriodicalIF":0.0,"publicationDate":"2024-03-12","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"142713790","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Rediscovering Minutiae Matching Through One Shot Learning’s Siamese Framework in Poor Quality Footprint Images 在质量较差的足迹图像中通过单次学习的连体框架重新发现细节匹配
IEEE transactions on biometrics, behavior, and identity science Pub Date : 2024-03-10 DOI: 10.1109/TBIOM.2024.3399402
Riti Kushwaha;Gaurav Singal;Neeta Nain
{"title":"Rediscovering Minutiae Matching Through One Shot Learning’s Siamese Framework in Poor Quality Footprint Images","authors":"Riti Kushwaha;Gaurav Singal;Neeta Nain","doi":"10.1109/TBIOM.2024.3399402","DOIUrl":"https://doi.org/10.1109/TBIOM.2024.3399402","url":null,"abstract":"Footprint biometrics is one of the emerging techniques, which can be utilized in different security systems. A human footprint has unique traits which is sufficient to recognize any person. Existing work evaluates the shape features and texture features but very few authors have explored minutiae features, hence this article provides a study based on minutiae features. The current State-of-the-art methods utilize machine learning techniques, which suffer from low accuracy in case of poor-quality of data. These machine learning techniques provide approx 97% accuracy while using good quality images but are not able to perform well when we use poor quality images. We have proposed a minutiae matching system based on deep learning techniques which is able to handle samples with adequate noise. We have used Convolution Neural Network for the feature extraction. It uses two different ridge flow estimation methods, i.e., ConvNet-based and dictionary-based. Furthermore, fingerprint-matching metrics are used for footprint feature evaluation. We initially employed a contrastive-based loss function, resulting in an accuracy of 56%. Subsequently, we adapted our approach by implementing a distance-based loss function, which improved the accuracy to 66%.","PeriodicalId":73307,"journal":{"name":"IEEE transactions on biometrics, behavior, and identity science","volume":"6 3","pages":"398-408"},"PeriodicalIF":0.0,"publicationDate":"2024-03-10","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"141725603","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
SFace2: Synthetic-Based Face Recognition With w-Space Identity-Driven Sampling SFace2:基于合成的人脸识别与 w 空间身份驱动采样
IEEE transactions on biometrics, behavior, and identity science Pub Date : 2024-02-29 DOI: 10.1109/TBIOM.2024.3371502
Fadi Boutros;Marco Huber;Anh Thi Luu;Patrick Siebke;Naser Damer
{"title":"SFace2: Synthetic-Based Face Recognition With w-Space Identity-Driven Sampling","authors":"Fadi Boutros;Marco Huber;Anh Thi Luu;Patrick Siebke;Naser Damer","doi":"10.1109/TBIOM.2024.3371502","DOIUrl":"https://doi.org/10.1109/TBIOM.2024.3371502","url":null,"abstract":"The use of synthetic data for training neural networks has recently received increased attention, especially in the area of face recognition. This was mainly motivated by the increase of privacy, ethical, and legal concerns of using privacy-sensitive authentic data to train face recognition models. Many authentic datasets such as MS-Celeb-1M or VGGFace2 that have been widely used to train state-of-the-art deep face recognition models are retracted and officially no longer maintained or provided by official sources as they often have been collected without explicit consent. Toward this end, we first propose a synthetic face generation approach, SFace which utilizes a class-conditional generative adversarial network to generate class-labeled synthetic face images. To evaluate the privacy aspect of using such synthetic data in face recognition development, we provide an extensive evaluation of the identity relation between the generated synthetic dataset and the original authentic dataset used to train the generative model. The investigation proved that the associated identity of the authentic dataset to the one with the same class label in the synthetic dataset is hardly possible, strengthening the possibility for privacy-aware face recognition training. We then propose three different learning strategies to train the face recognition model on our privacy-friendly dataset, SFace, and report the results on five authentic benchmarks, demonstrating its high potential. Noticing the relatively low (in comparison to authentic data) identity discrimination in SFace, we started by analysing the w-space of the class-conditional generator, finding identity information that is highly correlated to that in the embedding space. Based on this finding, we proposed an approach that performs the sampling in the w-space driven to generate data with higher identity discrimination, the SFace2. Our experiments showed the disentanglement of the latent w-space and the benefit of training face recognition models on the more identity-discriminated synthetic dataset SFace2.","PeriodicalId":73307,"journal":{"name":"IEEE transactions on biometrics, behavior, and identity science","volume":"6 3","pages":"290-303"},"PeriodicalIF":0.0,"publicationDate":"2024-02-29","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"141725586","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
SSPRA: A Robust Approach to Continuous Authentication Amidst Real-World Adversarial Challenges SSPRA:应对真实世界对抗性挑战的持续验证稳健方法
IEEE transactions on biometrics, behavior, and identity science Pub Date : 2024-02-27 DOI: 10.1109/TBIOM.2024.3369590
Frank Chen;Jingyu Xin;Vir V. Phoha
{"title":"SSPRA: A Robust Approach to Continuous Authentication Amidst Real-World Adversarial Challenges","authors":"Frank Chen;Jingyu Xin;Vir V. Phoha","doi":"10.1109/TBIOM.2024.3369590","DOIUrl":"https://doi.org/10.1109/TBIOM.2024.3369590","url":null,"abstract":"In real-world deployment, continuous authentication for mobile devices faces challenges such as intermittent data streams, variable data quality, and varying modality reliability. To address these challenges, we introduce a framework based on Markov process, named State-Space Perturbation-Resistant Approach (SSPRA). SSPRA integrates a two-level multi-modality fusion mechanism and dual state transition machines (STMs). This two-level fusion integrates probabilities from available modalities at each inspection (vertical-level) and evolves state probabilities over time (horizontal-level), thereby enhancing decision accuracy. It effectively manages modality disruptions and adjusts to variations in modality reliability. The dual STMs trigger appropriate responses upon detecting suspicious data, managing data fluctuations and extending operational duration, thus improving user experience. In our simulations, covering standard operations and adversarial scenarios like zero to non-zero-effort (ZE/NZE) attacks, modality disconnections, and data fluctuations, SSPRA consistently outperformed all baselines, including Sim’s HMM and three state-of-the-art deep-learning models. Notably, in adversarial attack scenarios, SSPRA achieved substantial reductions in False Alarm Rate (FAR) - 36.31%, 36.58%, and 8.26% - and improvements in True Alarm Rate (TAR) - 33.15%, 33.75%, and 5.1% compared to the DeepSense, Siamese-structured network, and UMSNet models, respectively. Furthermore, it outperformed all baselines in modality disconnection and fluctuation scenarios, underscoring SSPRA’s potential in addressing real-world challenges in mobile device authentication.","PeriodicalId":73307,"journal":{"name":"IEEE transactions on biometrics, behavior, and identity science","volume":"6 2","pages":"245-260"},"PeriodicalIF":0.0,"publicationDate":"2024-02-27","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"140345490","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
From Modalities to Styles: Rethinking the Domain Gap in Heterogeneous Face Recognition 从模式到风格:反思异构人脸识别中的领域差距
IEEE transactions on biometrics, behavior, and identity science Pub Date : 2024-02-13 DOI: 10.1109/TBIOM.2024.3365350
Anjith George;Sébastien Marcel
{"title":"From Modalities to Styles: Rethinking the Domain Gap in Heterogeneous Face Recognition","authors":"Anjith George;Sébastien Marcel","doi":"10.1109/TBIOM.2024.3365350","DOIUrl":"10.1109/TBIOM.2024.3365350","url":null,"abstract":"Heterogeneous Face Recognition (HFR) focuses on matching faces from different domains, for instance, thermal to visible images, making Face Recognition (FR) systems more versatile for challenging scenarios. However, the domain gap between these domains and the limited large-scale datasets in the target HFR modalities make it challenging to develop robust HFR models from scratch. In our work, we view different modalities as distinct styles and propose a method to modulate feature maps of the target modality to address the domain gap. We present a new Conditional Adaptive Instance Modulation (CAIM) module that seamlessly fits into existing FR networks, turning them into HFR-ready systems. The CAIM block modulates intermediate feature maps, efficiently adapting to the style of the source modality and bridging the domain gap. Our method enables end-to-end training using a small set of paired samples. We extensively evaluate the proposed approach on various challenging HFR benchmarks, showing that it outperforms state-of-the-art methods. The source code and protocols for reproducing the findings will be made publicly available.","PeriodicalId":73307,"journal":{"name":"IEEE transactions on biometrics, behavior, and identity science","volume":"6 4","pages":"475-485"},"PeriodicalIF":0.0,"publicationDate":"2024-02-13","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"140673423","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
GSCL: Generative Self-Supervised Contrastive Learning for Vein-Based Biometric Verification GSCL:基于静脉的生物识别验证的生成式自我监督对比学习
IEEE transactions on biometrics, behavior, and identity science Pub Date : 2024-02-08 DOI: 10.1109/TBIOM.2024.3364021
Wei-Feng Ou;Lai-Man Po;Xiu-Feng Huang;Wing-Yin Yu;Yu-Zhi Zhao
{"title":"GSCL: Generative Self-Supervised Contrastive Learning for Vein-Based Biometric Verification","authors":"Wei-Feng Ou;Lai-Man Po;Xiu-Feng Huang;Wing-Yin Yu;Yu-Zhi Zhao","doi":"10.1109/TBIOM.2024.3364021","DOIUrl":"https://doi.org/10.1109/TBIOM.2024.3364021","url":null,"abstract":"Vein-based biometric technology offers secure identity authentication due to the concealed nature of blood vessels. Despite the promising performance of deep learning-based biometric vein recognition, the scarcity of vein data hinders the discriminative power of deep features, thus affecting overall performance. To tackle this problem, this paper presents a generative self-supervised contrastive learning (GSCL) scheme, designed from a data-centric viewpoint to fully mine the potential prior knowledge from limited vein data for improving feature representations. GSCL first utilizes a style-based generator to model vein image distribution and then generate numerous vein image samples. These generated vein images are then leveraged to pretrain the feature extraction network via self-supervised contrastive learning. Subsequently, the network undergoes further fine-tuning using the original training data in a supervised manner. This systematic combination of generative and discriminative modeling allows the network to comprehensively excavate the semantic prior knowledge inherent in vein data, ultimately improving the quality of feature representations. In addition, we investigate a multi-template enrollment method for improving practical verification accuracy. Extensive experiments conducted on public finger vein and palm vein databases, as well as a newly collected finger vein video database, demonstrate the effectiveness of GSCL in improving representation quality.","PeriodicalId":73307,"journal":{"name":"IEEE transactions on biometrics, behavior, and identity science","volume":"6 2","pages":"230-244"},"PeriodicalIF":0.0,"publicationDate":"2024-02-08","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"140345483","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Cascade Transformer Reasoning Embedded by Uncertainty for Occluded Person Re-Identification 基于不确定性的级联变换推理用于被排除人员的再识别
IEEE transactions on biometrics, behavior, and identity science Pub Date : 2024-02-02 DOI: 10.1109/TBIOM.2024.3361677
Hanqing Zheng;Yuxuan Shi;Hefei Ling;Zongyi Li;Runsheng Wang;Zhongyang Li;Ping Li
{"title":"Cascade Transformer Reasoning Embedded by Uncertainty for Occluded Person Re-Identification","authors":"Hanqing Zheng;Yuxuan Shi;Hefei Ling;Zongyi Li;Runsheng Wang;Zhongyang Li;Ping Li","doi":"10.1109/TBIOM.2024.3361677","DOIUrl":"https://doi.org/10.1109/TBIOM.2024.3361677","url":null,"abstract":"Occluded person re-identification is a challenging task due to various noise introduced by occlusion. Previous methods utilize body detectors to exploit more clues which are overdependent on accuracy of detection results. In this paper, we propose a model named Cascade Transformer Reasoning Embedded by Uncertainty Network (CTU) which does not require external information. Self-attention of the transformer models long-range dependency to capture difference between pixels, which helps the model focus on discriminative information of human bodies. However, noise such as occlusion will bring a high level of uncertainty to feature learning and makes self-attention learn undesirable dependency. We invent a novel structure named Uncertainty Embedded Transformer (UT) Layer to involve uncertainty in computing attention weights of self-attention. Introducing uncertainty mechanism helps the network better evaluate the dependency between pixels and focus more on human bodies. Additionally, our proposed transformer layer generates an attention mask through Cascade Attention Module (CA) to guide the next layer to focus more on key areas of the feature map, decomposing feature learning into cascade stages. Extensive experiments over challenging datasets Occluded-DukeMTMC, P-DukeMTMC, etc., verify the effectiveness of our method.","PeriodicalId":73307,"journal":{"name":"IEEE transactions on biometrics, behavior, and identity science","volume":"6 2","pages":"219-229"},"PeriodicalIF":0.0,"publicationDate":"2024-02-02","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"140345463","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Pose Impact Estimation on Face Recognition Using 3-D-Aware Synthetic Data With Application to Quality Assessment 利用三维感知合成数据估计姿势对人脸识别的影响并应用于质量评估
IEEE transactions on biometrics, behavior, and identity science Pub Date : 2024-02-02 DOI: 10.1109/TBIOM.2024.3361657
Marcel Grimmer;Christian Rathgeb;Christoph Busch
{"title":"Pose Impact Estimation on Face Recognition Using 3-D-Aware Synthetic Data With Application to Quality Assessment","authors":"Marcel Grimmer;Christian Rathgeb;Christoph Busch","doi":"10.1109/TBIOM.2024.3361657","DOIUrl":"https://doi.org/10.1109/TBIOM.2024.3361657","url":null,"abstract":"Evaluating the quality of facial images is essential for operating face recognition systems with sufficient accuracy. The recent advances in face quality standardisation (ISO/IEC CD3 29794-5) recommend the usage of component quality measures for breaking down face quality into its individual factors, hence providing valuable feedback for operators to re-capture low-quality images. In light of recent advances in 3D-aware generative adversarial networks, we propose a novel dataset, Syn-YawPitch, comprising 1,000 identities with varying yaw-pitch angle combinations. Utilizing this dataset, we demonstrate that pitch angles beyond 30 degrees have a significant impact on the biometric performance of current face recognition systems. Furthermore, we propose a lightweight and explainable pose quality predictor that adheres to the draft international standard of ISO/IEC CD3 29794–5 and benchmark it against state-of-the-art face image quality assessment algorithms.","PeriodicalId":73307,"journal":{"name":"IEEE transactions on biometrics, behavior, and identity science","volume":"6 2","pages":"209-218"},"PeriodicalIF":0.0,"publicationDate":"2024-02-02","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"140345503","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Implicit Mutual Learning With Dual-Branch Networks for Face Super-Resolution 利用双分支网络进行隐式互学以实现人脸超分辨率
IEEE transactions on biometrics, behavior, and identity science Pub Date : 2024-01-19 DOI: 10.1109/TBIOM.2024.3354333
Kangli Zeng;Zhongyuan Wang;Tao Lu;Jianyu Chen;Zheng He;Zhen Han
{"title":"Implicit Mutual Learning With Dual-Branch Networks for Face Super-Resolution","authors":"Kangli Zeng;Zhongyuan Wang;Tao Lu;Jianyu Chen;Zheng He;Zhen Han","doi":"10.1109/TBIOM.2024.3354333","DOIUrl":"https://doi.org/10.1109/TBIOM.2024.3354333","url":null,"abstract":"Face super-resolution (SR) algorithms have recently made significant progress. However, most existing methods prefer to employ texture and structure information together to promote the generation of high-resolution features, neglecting the mutual encouragement between them, as well as the effective unification of their own low-level and high-level information, thus yielding unsatisfactory results. To address these problems, we propose an implicit mutual learning of dual-branch networks for face super-resolution, which adequately considers both extraction and aggregation of structure and texture information. The proposed approach consists of four essential blocks. First, the deep feature extractor is equipped with a deep feature reinforcement module (DFRM) based on two-stage cross-dimensional attention (TCA), which behaves in the texture enhancement and structure reconstruction branches, respectively. Then, we elaborate two information exchange blocks for two branches, one for the first information exchange block (FIEB) from the texture branch to the structure branch and one for the second information exchange block (SIEB) from the structure branch to the texture branch. These two interaction blocks perform further fusion enhancement of potential features. Finally, a hybrid fusion network (HFNet) based on supervised attention executes adaptive aggregation of the enhanced texture and structure maps. Additionally, we use a joint loss function that modifies the recovery of structure information, diminishes the use of potentially erroneous information, and encourages the generation of realistic face images. Experiments on public datasets show that our method consistently achieves better quantitative and qualitative results than SOTA methods.","PeriodicalId":73307,"journal":{"name":"IEEE transactions on biometrics, behavior, and identity science","volume":"6 2","pages":"182-194"},"PeriodicalIF":0.0,"publicationDate":"2024-01-19","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"140345504","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Learning Joint Local-Global Iris Representations via Spatial Calibration for Generalized Presentation Attack Detection 通过空间校准学习局部-全局联合虹膜表征,用于通用演示攻击检测
IEEE transactions on biometrics, behavior, and identity science Pub Date : 2024-01-17 DOI: 10.1109/TBIOM.2024.3355136
Gaurav Jaswal;Aman Verma;Sumantra Dutta Roy;Raghavendra Ramachandra
{"title":"Learning Joint Local-Global Iris Representations via Spatial Calibration for Generalized Presentation Attack Detection","authors":"Gaurav Jaswal;Aman Verma;Sumantra Dutta Roy;Raghavendra Ramachandra","doi":"10.1109/TBIOM.2024.3355136","DOIUrl":"https://doi.org/10.1109/TBIOM.2024.3355136","url":null,"abstract":"Existing Iris Presentation Attack Detection (IPAD) systems do not generalize well across datasets, sensors and subjects. The main reason for the same is the presence of similarities in bonafide samples and attacks, and intricate iris textures. The proposed DFCANet (Dense Feature Calibration Attention-Assisted Network) uses feature calibration convolution and residual learning to generate domain-specific iris feature representations at local and global scales. DFCANet’s channel attention enables the use of discriminative feature learning across channels. Compared to state-of-the-art methods, DFCANet achieves significant performance gains for the IIITD-CLI, IIITD-WVU, IIIT-CSD, Clarkson-15, Clarkson-17, NDCLD-13, and NDCLD-15 benchmark datasets. Incremental learning in DFCANet overcomes data scarcity issues and cross-domain challenges. This paper also pursues the challenging soft-lens attack scenarios. An additional study conducted over contact lens detection task suggests high domain-specific feature modeling capacities of the proposed network.","PeriodicalId":73307,"journal":{"name":"IEEE transactions on biometrics, behavior, and identity science","volume":"6 2","pages":"195-208"},"PeriodicalIF":0.0,"publicationDate":"2024-01-17","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"140345462","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
0
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
确定
请完成安全验证×
相关产品
×
本文献相关产品
联系我们:info@booksci.cn Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。 Copyright © 2023 布克学术 All rights reserved.
京ICP备2023020795号-1
ghs 京公网安备 11010802042870号
Book学术文献互助
Book学术文献互助群
群 号:481959085
Book学术官方微信