{"title":"Derivative-Guided Dual-Attention Mechanisms in Patch Transformer for Efficient Automated Recognition of Auditory Brainstem Response Latency.","authors":"Yin Liu, Huanghong Sun, Qiang Li, Kangkang Li, Xinxing Fu, Hao Zhu, Tiecheng Song, Yue Zhao, Tiantian Wang, Chenqiang Gao","doi":"10.1109/TNSRE.2025.3558730","DOIUrl":null,"url":null,"abstract":"<p><p>Accurate recognition of auditory brainstem response (ABR) wave latencies is essential for clinical practice but remains a subjective and time-consuming process. Existing AI approaches face challenges in generalization, complexity, and semantic sparsity due to single sampling-point analysis. This study introduces the Derivative-Guided Patch Dual-Attention Transformer (Patch-DAT), a novel, lightweight, and generalizable deep learning (DL) model for the automated recognition of latencies for waves I, III, and V. Patch-DAT divides the ABR time series into overlapping patches to aggregate semantic information, better capturing local temporal patterns. Meanwhile, leveraging the fact that ABR waves occur at the zero crossing of the first derivative, Patch-DAT incorporates a first derivative-guided dual-attention mechanism to model global dependencies. Trained and validated on large-scale, diverse datasets from two hospitals, Patch-DAT(with a size of 0.36 MB) achieves accuracies of 92.29% and 98.07% at 0.1 ms and 0.2 ms error scales, respectively, on a held-out test set. It also performs well on an independent dataset with accuracies of 88.50% and 95.14%, demonstrating strong generalization across clinical settings. Ablation studies highlight the contributions of the patching strategy and dual-attention mechanisms. Compared to previous state-of-the-art DL models, Patch-DAT shows superior accuracy and reduced complexity, making it a promising solution for object recognition of ABR latencies. Additionally, we systematically investigate how sample size and data heterogeneity affect model generalization, indicating the importance of large, diverse datasets in training robust DL models. Future work will focus on expanding dataset diversity and improving model interpretability to further improve clinical relevance.</p>","PeriodicalId":13419,"journal":{"name":"IEEE Transactions on Neural Systems and Rehabilitation Engineering","volume":"PP ","pages":""},"PeriodicalIF":4.8000,"publicationDate":"2025-04-08","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"IEEE Transactions on Neural Systems and Rehabilitation Engineering","FirstCategoryId":"5","ListUrlMain":"https://doi.org/10.1109/TNSRE.2025.3558730","RegionNum":2,"RegionCategory":"医学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"Q2","JCRName":"ENGINEERING, BIOMEDICAL","Score":null,"Total":0}
引用次数: 0
Abstract
Accurate recognition of auditory brainstem response (ABR) wave latencies is essential for clinical practice but remains a subjective and time-consuming process. Existing AI approaches face challenges in generalization, complexity, and semantic sparsity due to single sampling-point analysis. This study introduces the Derivative-Guided Patch Dual-Attention Transformer (Patch-DAT), a novel, lightweight, and generalizable deep learning (DL) model for the automated recognition of latencies for waves I, III, and V. Patch-DAT divides the ABR time series into overlapping patches to aggregate semantic information, better capturing local temporal patterns. Meanwhile, leveraging the fact that ABR waves occur at the zero crossing of the first derivative, Patch-DAT incorporates a first derivative-guided dual-attention mechanism to model global dependencies. Trained and validated on large-scale, diverse datasets from two hospitals, Patch-DAT(with a size of 0.36 MB) achieves accuracies of 92.29% and 98.07% at 0.1 ms and 0.2 ms error scales, respectively, on a held-out test set. It also performs well on an independent dataset with accuracies of 88.50% and 95.14%, demonstrating strong generalization across clinical settings. Ablation studies highlight the contributions of the patching strategy and dual-attention mechanisms. Compared to previous state-of-the-art DL models, Patch-DAT shows superior accuracy and reduced complexity, making it a promising solution for object recognition of ABR latencies. Additionally, we systematically investigate how sample size and data heterogeneity affect model generalization, indicating the importance of large, diverse datasets in training robust DL models. Future work will focus on expanding dataset diversity and improving model interpretability to further improve clinical relevance.
期刊介绍:
Rehabilitative and neural aspects of biomedical engineering, including functional electrical stimulation, acoustic dynamics, human performance measurement and analysis, nerve stimulation, electromyography, motor control and stimulation; and hardware and software applications for rehabilitation engineering and assistive devices.