IEEE transactions on biometrics, behavior, and identity science最新文献

筛选
英文 中文
MDL-Net: Multi-Task Learning Network for Face Forgery Detection and Localization Using Dual-Stream Feature Extraction and Reconstruction MDL-Net:基于双流特征提取和重建的人脸伪造检测和定位多任务学习网络
IF 5
IEEE transactions on biometrics, behavior, and identity science Pub Date : 2026-03-01 Epub Date: 2026-01-22 DOI: 10.1109/TBIOM.2026.3656922
Huasong Yi;Qian Jiang;Hongyue Huang;Li Tang;Shaowen Yao;Shin-Jye Lee;Xin Jin
{"title":"MDL-Net: Multi-Task Learning Network for Face Forgery Detection and Localization Using Dual-Stream Feature Extraction and Reconstruction","authors":"Huasong Yi;Qian Jiang;Hongyue Huang;Li Tang;Shaowen Yao;Shin-Jye Lee;Xin Jin","doi":"10.1109/TBIOM.2026.3656922","DOIUrl":"https://doi.org/10.1109/TBIOM.2026.3656922","url":null,"abstract":"Advancement in face forgery technologies has led to the generation of increasingly realistic images, making them difficult for the human eye to distinguish. These technologies pose risks in areas such as face recognition and identity verification, including fraud and social engineering attacks. While face forgery detection algorithms achieve high accuracy within a single dataset, their generalization performance in cross-dataset scenarios remains insufficient. To address this issue, this study introduces the Multi-task Face Forgery Detection and Localization Network (MDL-Net), which improves cross-dataset detection performance through dual-stream feature extraction and reconstructed feature enhancement. First, the face forgery reconstruction module restores and reconstructs images, amplifying subtle features in the tampered regions. Second, leveraging pixel-level residual information, the reconstruction attention module enhances feature representation by comparing input images with their reconstructed counterparts. Finally, the localization module precisely identifies forged regions through pixel-level segmentation. This approach encourages the model to extract forgery-specific features and learn more generalized and robust representations, achieving high detection and localization accuracy. Experiments conducted on the FaceForensics++ and WildDeepfake datasets demonstrate the superior performance and robustness of the proposed framework. Comparative experiments on other benchmark datasets validate the enhanced generalization ability of the method while maintaining high detection accuracy. The code is available at <uri>https://github.com/jinxinhuo/MDL-Net</uri>","PeriodicalId":73307,"journal":{"name":"IEEE transactions on biometrics, behavior, and identity science","volume":"8 3","pages":"398-411"},"PeriodicalIF":5.0,"publicationDate":"2026-03-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"147685342","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
IEEE Transactions on Biometrics, Behavior, and Identity Science Information for Authors IEEE生物识别、行为和身份科学信息作者汇刊
IF 5
IEEE transactions on biometrics, behavior, and identity science Pub Date : 2026-03-01 Epub Date: 2026-04-16 DOI: 10.1109/TBIOM.2026.3678354
{"title":"IEEE Transactions on Biometrics, Behavior, and Identity Science Information for Authors","authors":"","doi":"10.1109/TBIOM.2026.3678354","DOIUrl":"https://doi.org/10.1109/TBIOM.2026.3678354","url":null,"abstract":"","PeriodicalId":73307,"journal":{"name":"IEEE transactions on biometrics, behavior, and identity science","volume":"8 3","pages":"C3-C3"},"PeriodicalIF":5.0,"publicationDate":"2026-03-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://ieeexplore.ieee.org/stamp/stamp.jsp?tp=&arnumber=11481992","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"147685532","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
IEEE Transactions on Biometrics, Behavior, and Identity Science Publication Information IEEE生物计量学、行为与身份科学学报
IF 5
IEEE transactions on biometrics, behavior, and identity science Pub Date : 2026-03-01 Epub Date: 2026-04-16 DOI: 10.1109/TBIOM.2026.3678355
{"title":"IEEE Transactions on Biometrics, Behavior, and Identity Science Publication Information","authors":"","doi":"10.1109/TBIOM.2026.3678355","DOIUrl":"https://doi.org/10.1109/TBIOM.2026.3678355","url":null,"abstract":"","PeriodicalId":73307,"journal":{"name":"IEEE transactions on biometrics, behavior, and identity science","volume":"8 3","pages":"C2-C2"},"PeriodicalIF":5.0,"publicationDate":"2026-03-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://ieeexplore.ieee.org/stamp/stamp.jsp?tp=&arnumber=11481990","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"147685556","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Impact of Sunglasses on One-to-Many Facial Identification Accuracy 墨镜对一对多人脸识别准确率的影响
IF 5
IEEE transactions on biometrics, behavior, and identity science Pub Date : 2026-03-01 Epub Date: 2026-02-23 DOI: 10.1109/TBIOM.2026.3666896
Sicong Tian;Haiyu Wu;Michael C. King;Kevin W. Bowyer
{"title":"Impact of Sunglasses on One-to-Many Facial Identification Accuracy","authors":"Sicong Tian;Haiyu Wu;Michael C. King;Kevin W. Bowyer","doi":"10.1109/TBIOM.2026.3666896","DOIUrl":"https://doi.org/10.1109/TBIOM.2026.3666896","url":null,"abstract":"Probe images used in one-to-many facial identification often deviate from ‘mugshot quality’ in various ways. This paper explores how the accuracy of one-to-many facial identification is degraded by the person in the probe image wearing sunglasses, and how this accuracy degradation can be ameliorated. We show that sunglasses degrade accuracy for mugshot-quality images by an amount similar to strong blur or noticeably lower resolution. Further, we demonstrate that the combination of sunglasses with blur or lower resolution has an additive effect in degrading accuracy. These results have important implications for developing criteria to qualify a probe image for the accuracy that can be expected if it is used for facial identification. We explore two approaches to reduce the accuracy loss. First, we show that by adding synthetic sunglasses to all gallery images, we can recover approximately 27% of the accuracy loss with no change to the face-matching model. Second, we demonstrate that training a sunglasses-aware model using augmented training data can achieve up to 33% higher accuracy for sunglasses-wearing probes while maintaining baseline accuracy for non-sunglasses probes. Moreover, combining both approaches achieves even greater accuracy recovery (up to 43% for males and 35% for females). Additionally, our results show that increasing the representation of sunglasses-wearing images evenly across all identities in training data is more effective than concentrating them on fewer identities. The effectiveness of our sunglasses-aware model is validated through t-SNE visualization and a 16% increase in feature similarity between sunglasses and no-sunglasses images of a given identity. Our investigation further reveals that baseball caps play a role similar to sunglasses in degrading facial identification accuracy. To support replication and further research, info to obtain the face image datasets used is available at <uri>https://cvrl.nd.edu/projects/data/</uri>","PeriodicalId":73307,"journal":{"name":"IEEE transactions on biometrics, behavior, and identity science","volume":"8 3","pages":"461-476"},"PeriodicalIF":5.0,"publicationDate":"2026-03-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"147685321","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Point-Supervised Facial Expression Spotting With Gaussian-Based Instance-Adaptive Intensity Modeling 基于高斯的实例自适应强度建模的点监督面部表情识别
IF 5
IEEE transactions on biometrics, behavior, and identity science Pub Date : 2026-03-01 Epub Date: 2026-01-12 DOI: 10.1109/TBIOM.2026.3651893
Yicheng Deng;Hideaki Hayashi;Hajime Nagahara
{"title":"Point-Supervised Facial Expression Spotting With Gaussian-Based Instance-Adaptive Intensity Modeling","authors":"Yicheng Deng;Hideaki Hayashi;Hajime Nagahara","doi":"10.1109/TBIOM.2026.3651893","DOIUrl":"https://doi.org/10.1109/TBIOM.2026.3651893","url":null,"abstract":"Automatic facial expression spotting, which aims to identify facial expression instances in untrimmed videos, is crucial for facial expression analysis. Existing methods primarily focus on fully-supervised learning and rely on costly, time-consuming temporal boundary annotations. In this paper, we investigate point-supervised facial expression spotting (P-FES), where only a single timestamp annotation per instance is required for training. We propose a unique two-branch framework for P-FES. First, to mitigate the limitation of hard pseudo-labeling, which often confuses neutral and expression frames with various intensities, we propose a Gaussian-based instance-adaptive intensity modeling (GIM) module to model instance-level expression intensity distribution for soft pseudo-labeling. By detecting the pseudo-apex frame around each point label, estimating the duration, and constructing an instance-level Gaussian distribution, GIM assigns soft pseudo-labels to expression frames for more reliable intensity supervision. The GIM module is incorporated into our framework to optimize the class-agnostic expression intensity branch. Second, we design a class-aware apex classification branch that distinguishes macro- and micro-expressions solely based on their pseudo-apex frames. During inference, the two branches work independently: the class-agnostic expression intensity branch generates expression proposals, while the class-aware apex-classification branch is responsible for macro- and micro-expression classification. Furthermore, we introduce an intensity-aware contrastive loss to enhance discriminative feature learning and suppress neutral noise by contrasting neutral frames with expression frames with various intensities. Extensive experiments on the SAMM-LV, CAS(ME)2, and CAS(ME)3 datasets demonstrate the effectiveness of our proposed framework. Code is available at <uri>https://github.com/KinopioIsAllIn/GIM</uri>","PeriodicalId":73307,"journal":{"name":"IEEE transactions on biometrics, behavior, and identity science","volume":"8 3","pages":"378-391"},"PeriodicalIF":5.0,"publicationDate":"2026-03-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://ieeexplore.ieee.org/stamp/stamp.jsp?tp=&arnumber=11339514","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"147685468","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Identity Reflections: Hyperspectral Skin Analysis via Reflectance Distance Learning With Context-Aware Contrastive Transformers 身份反射:高光谱皮肤分析通过反射远程学习与环境感知对比变压器
IF 5
IEEE transactions on biometrics, behavior, and identity science Pub Date : 2026-03-01 Epub Date: 2026-01-23 DOI: 10.1109/TBIOM.2026.3657642
Emanuela Marasco;Bhargavi Janga;Gautham Gali;Raghavendra Ramachandra
{"title":"Identity Reflections: Hyperspectral Skin Analysis via Reflectance Distance Learning With Context-Aware Contrastive Transformers","authors":"Emanuela Marasco;Bhargavi Janga;Gautham Gali;Raghavendra Ramachandra","doi":"10.1109/TBIOM.2026.3657642","DOIUrl":"https://doi.org/10.1109/TBIOM.2026.3657642","url":null,"abstract":"Hyperspectral Imaging (HSI) enables real-time analysis with rich spectral detail, offering a promising path to spoof-resistant AI-based identity verification. Leveraging this spectral information, we present a novel biometric approach that extracts and analyzes skin reflectance patterns from hyperspectral fingertip images. In prior work, hyperspectral analysis has primarily been applied to facial biometrics; however, the associated lighting requirements are often impractical or inconvenient for users in real-world scenarios. From a computational perspective, while deep learning has significantly advanced recognition using conventional imaging, effectively applying these frameworks to the rich and complex signals captured by HSI remains a critical open challenge. Convolutional Neural Networks (CNNs) often have difficulty preserving the sequential structure of spectral information. Vision transformers effectively capture global context through self-attention but often face challenges in modeling fine-grained local features due to the lack of convolutional inductive biases. To address existing limitations, this work introduces a novel integration of Context-Enriched Contrastive Loss (CECL) into two architectures: SpectralFormer, a ViT-based model that employs decoupled spatial–spectral processing to handle spectral signatures extracted from carefully selected spatial regions; and the Swin Transformer, which leverages a unified 3D joint representation. For the first time, SpectralFormer is employed for identity verification, leveraging group-wise spectral embeddings, inspired by ViT’s patch-based processing, across adjacent bands, along with intermediate skip connections to enhance the analysis of complex spectral data. The proposed approach was evaluated on three biometric databases, including the Mason FHSD-2022 hyperspectral fingertip dataset collected at George Mason University with 100 subjects. The code will be publicly available on GitHub.","PeriodicalId":73307,"journal":{"name":"IEEE transactions on biometrics, behavior, and identity science","volume":"8 3","pages":"431-444"},"PeriodicalIF":5.0,"publicationDate":"2026-03-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"147685524","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
MoGA-ETA: Generalized Face Anti-Spoofing With Enhanced Text Guidance and Alignment MoGA-ETA:增强文本引导和对齐的广义人脸反欺骗
IF 5
IEEE transactions on biometrics, behavior, and identity science Pub Date : 2026-03-01 Epub Date: 2026-01-22 DOI: 10.1109/TBIOM.2026.3657113
Liepiao Zhang;Kun Liu;Tianming Xie;Junduan Huang;Zitong Yu;Wenxiong Kang
{"title":"MoGA-ETA: Generalized Face Anti-Spoofing With Enhanced Text Guidance and Alignment","authors":"Liepiao Zhang;Kun Liu;Tianming Xie;Junduan Huang;Zitong Yu;Wenxiong Kang","doi":"10.1109/TBIOM.2026.3657113","DOIUrl":"https://doi.org/10.1109/TBIOM.2026.3657113","url":null,"abstract":"Face anti-spoofing (FAS) methods struggle with cross-domain generalization due to variations in attack types and imaging conditions. While recent vision-language approaches leverage image-text contrastive learning, they face three critical limitations: 1) simplistic text prompts fail to capture nuanced spoofing artifacts like moiré patterns or 3D mask boundaries, 2) suboptimal visual encoder fine-tuning leads to either inadequate adaptation or overfitting, and 3) substantial semantic gaps between image and text features undermine cross-modal alignment. To address these limitations, we propose the Enhanced Text Guidance and Alignment (MoGA-ETA) to enhance language-guided face anti-spoofing. First, we introduce an in-context text prompt enhancement strategy to generate detailed descriptions with fine-grained textual descriptions, capturing both global attributes (e.g., illumination anomalies) and local cues (e.g., Moiré patterns) through the large vision-language model. Second, we design and embed a mixture of gradient adapters into the language-image contrastive pretrained visual encoder during the post-training period. Third, we investigate image-text alignment via a hard positive and negative mining strategy to improve the efficacy of cross-modal contrastive learning. Extensive experiments on leave-one-out cross-domain benchmarks demonstrate that our method can achieve state-of-the-art performance, and ablation studies confirm the necessity of each component. The code is released at <uri>https://github.com/zlpiao/MoGA-ETA</uri>","PeriodicalId":73307,"journal":{"name":"IEEE transactions on biometrics, behavior, and identity science","volume":"8 3","pages":"340-353"},"PeriodicalIF":5.0,"publicationDate":"2026-03-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"147685353","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
FaceXBench: Evaluating Multimodal LLMs on Face Understanding FaceXBench:评估面部理解的多模态法学硕士
IF 5
IEEE transactions on biometrics, behavior, and identity science Pub Date : 2026-03-01 Epub Date: 2026-01-19 DOI: 10.1109/TBIOM.2026.3655668
Kartik Narayan;V. S. Vibashan;Vishal M. Patel
{"title":"FaceXBench: Evaluating Multimodal LLMs on Face Understanding","authors":"Kartik Narayan;V. S. Vibashan;Vishal M. Patel","doi":"10.1109/TBIOM.2026.3655668","DOIUrl":"https://doi.org/10.1109/TBIOM.2026.3655668","url":null,"abstract":"Multimodal Large Language Models (MLLMs) demonstrate impressive problem-solving abilities across a wide range of tasks and domains. However, their capacity for face understanding has not been systematically studied. To address this gap, we introduce FaceXBench, a comprehensive benchmark designed to evaluate MLLMs on complex face understanding tasks. FaceXBench includes 5,000 multimodal multiple-choice questions derived from 25 public datasets and a newly created dataset, FaceXAPI. These questions cover 14 tasks across 6 broad categories, assessing MLLMs’ face understanding abilities in bias and fairness, face authentication, recognition, analysis, localization and tool retrieval. Using FaceXBench, we conduct an extensive evaluation of 26 open-source MLLMs alongside 2 proprietary models, revealing the unique challenges in complex face understanding tasks. We analyze the models across three evaluation settings: zero-shot, in-context task description, and chain-of-thought prompting. Our detailed analysis reveals that current MLLMs, including advanced models like GPT-4o, and GeminiPro 1.5, show significant room for improvement. We believe FaceXBench will be a crucial resource for developing MLLMs equipped to perform sophisticated face understanding.","PeriodicalId":73307,"journal":{"name":"IEEE transactions on biometrics, behavior, and identity science","volume":"8 3","pages":"354-364"},"PeriodicalIF":5.0,"publicationDate":"2026-03-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"147685546","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Face Morphing Attack Generation and Detection: A State-of-the-Art Review 面部变形攻击的生成与检测:最新进展综述
IF 5
IEEE transactions on biometrics, behavior, and identity science Pub Date : 2026-03-01 Epub Date: 2026-01-19 DOI: 10.1109/TBIOM.2026.3655515
Davide Antonutti;Justin Ilyes;Laurenz Ruzicka;Silvia Poletti;Marcel Hasenbalg;Martin Boyer;David Fischinger
{"title":"Face Morphing Attack Generation and Detection: A State-of-the-Art Review","authors":"Davide Antonutti;Justin Ilyes;Laurenz Ruzicka;Silvia Poletti;Marcel Hasenbalg;Martin Boyer;David Fischinger","doi":"10.1109/TBIOM.2026.3655515","DOIUrl":"https://doi.org/10.1109/TBIOM.2026.3655515","url":null,"abstract":"Face morphing attacks pose a significant threat to the security and reliability of biometric identity verification systems, particularly in real-world applications such as passport issuance and border control. In a morphing attack, facial features from two or more individuals are blended into a single image that can deceive face recognition systems, allowing multiple individuals to share a biometric identity. This survey provides a comprehensive overview of the literature on face morphing attack detection (MAD). It begins by introducing the practical implications of morphing attacks and their relevance in operational contexts. The paper then explores a wide range of morphing generation techniques, including both classical landmark-based approaches and modern deep learning-based methods such as GANs and diffusion models. The work proceeds with an extensive review and categorization of existing MAD techniques, grouped into meaningful categories based on their underlying principles, ranging from texture-based and quality-based methods to deep learning and hybrid approaches. By organizing the literature and identifying current trends, strengths, and limitations, this survey offers valuable insight for researchers and practitioners seeking to understand, evaluate, or develop robust solutions against face morphing attacks.","PeriodicalId":73307,"journal":{"name":"IEEE transactions on biometrics, behavior, and identity science","volume":"8 3","pages":"412-430"},"PeriodicalIF":5.0,"publicationDate":"2026-03-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://ieeexplore.ieee.org/stamp/stamp.jsp?tp=&arnumber=11357967","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"147685343","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Advancing Brainwave-Based Biometrics: A Large-Scale, Multi-Session Evaluation 先进的基于脑波的生物识别技术:大规模、多阶段的评估
IF 5
IEEE transactions on biometrics, behavior, and identity science Pub Date : 2026-03-01 Epub Date: 2026-02-23 DOI: 10.1109/TBIOM.2026.3666044
Matin Fallahi;Patricia Arias Cabarcos;Thorsten Strufe
{"title":"Advancing Brainwave-Based Biometrics: A Large-Scale, Multi-Session Evaluation","authors":"Matin Fallahi;Patricia Arias Cabarcos;Thorsten Strufe","doi":"10.1109/TBIOM.2026.3666044","DOIUrl":"https://doi.org/10.1109/TBIOM.2026.3666044","url":null,"abstract":"The field of brainwave-based biometrics has gained attention for its potential to revolutionize user authentication through hands-free interaction, resistance to shoulder surfing, continuous authentication, and revocability. However, current research often relies on single-session or limited-session datasets with fewer than 55 subjects, raising concerns about the generalizability of the findings. To address this gap, we conducted a large-scale study using a public brainwave dataset comprising 345 subjects and over 6,007 sessions (an average of 17 per subject) recorded over five years using three headsets. Our results reveal that deep learning approaches significantly outperform hand-crafted feature extraction methods. We also observe Equal Error Rates (EER) increases over time (e.g., from 6.7% after 1 day to 14.3% after a year). Therefore, it is necessary to reinforce the enrollment set after successful login attempts. Moreover, we demonstrate that fewer brainwave measurement sensors can be used, with an acceptable increase in EER, which is necessary for transitioning from medical-grade to affordable consumer-grade devices. Finally, we compared our results to prior work and existing biometric standards. While our performance is on par with or exceeds previous approaches, it still falls short of industrial benchmarks. Based on the results, we hypothesize that further improvements are possible with larger training sets. To support future research, we have open-sourced our analysis code.","PeriodicalId":73307,"journal":{"name":"IEEE transactions on biometrics, behavior, and identity science","volume":"8 3","pages":"445-460"},"PeriodicalIF":5.0,"publicationDate":"2026-03-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://ieeexplore.ieee.org/stamp/stamp.jsp?tp=&arnumber=11407492","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"147685352","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
0
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
确定
请完成安全验证×
相关产品
×
本文献相关产品
联系我们:info@booksci.cn Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。 Copyright © 2023 布克学术 All rights reserved.
京ICP备2023020795号-1
ghs 京公网安备 11010802042870号
Book学术文献互助
Book学术文献互助群
群 号:604180095
Book学术官方微信
小红书