IEEE Journal of Biomedical and Health Informatics最新文献

筛选
英文 中文
Paradigm-Shifting Attention-based Hybrid View Learning for Enhanced Mammography Breast Cancer Classification with Multi-Scale and Multi-View Fusion. 范式转移-基于注意力的混合视图学习在多尺度和多视图融合增强乳房x线摄影乳腺癌分类中的应用。
IF 6.7 2区 医学
IEEE Journal of Biomedical and Health Informatics Pub Date : 2025-05-12 DOI: 10.1109/JBHI.2025.3569726
Haoran Zhao, Chengwei Zhang, Jiong Chen, Zhaotong Li, Fei Wang, Song Gao
{"title":"Paradigm-Shifting Attention-based Hybrid View Learning for Enhanced Mammography Breast Cancer Classification with Multi-Scale and Multi-View Fusion.","authors":"Haoran Zhao, Chengwei Zhang, Jiong Chen, Zhaotong Li, Fei Wang, Song Gao","doi":"10.1109/JBHI.2025.3569726","DOIUrl":"10.1109/JBHI.2025.3569726","url":null,"abstract":"<p><p>Breast cancer poses a serious threat to women's health, and its early detection is crucial for enhancing patient survival rates. While deep learning has significantly advanced mammographic image analysis, existing methods struggle to balance between view consistency with input adaptability. Furthermore, current models face challenges in accurately capturing multi-scale features, especially when subtle lesion variations across different scales are involved. To address this challenge, this paper proposes a Hybrid View Learning (HVL) paradigm that unifies traditional Single-View and Multi-View Learning approaches. The core component of this paradigm, our Attention-based Hybrid View Learning (AHVL) framework, incorporates two essential attention mechanisms: Contrastive Switch Attention (CSA) and Selective Pooling Attention (SPA). The CSA mechanism flexibly alternates between self-attention and cross-attention based on data integrity, integrating a pre-trained language model for contrastive learning to enhance model stability. Meanwhile, the SPA module employs multi-scale feature pooling and selection to capture critical features from mammographic images, overcoming the limitations of traditional models that struggle with fine-grained lesion detection. Experimental validation on the INbreast and CBIS-DDSM datasets shows that the AHVL framework outperforms both single-view and multi-view methods, especially under extreme view missing conditions. Even with an 80% missing rate on both datasets, AHVL maintains the highest accuracy and experiences the smallest performance decline in metrics like F1 score and AUC-PR, demonstrating its robustness and stability. This study redefines mammographic image analysis by leveraging attention-based hybrid view processing, setting a new standard for precise and efficient breast cancer diagnosis.</p>","PeriodicalId":13073,"journal":{"name":"IEEE Journal of Biomedical and Health Informatics","volume":"PP ","pages":""},"PeriodicalIF":6.7,"publicationDate":"2025-05-12","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"143998494","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"医学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
CSBNC-PAL: Consistency Semi-supervised Brain Network Classification Framework with Prototypical-Adversarial Learning. 基于原型-对抗学习的一致性半监督脑网络分类框架。
IF 6.7 2区 医学
IEEE Journal of Biomedical and Health Informatics Pub Date : 2025-05-12 DOI: 10.1109/JBHI.2025.3569734
Junzhong Ji, Gan Liu, Xingyu Wang
{"title":"CSBNC-PAL: Consistency Semi-supervised Brain Network Classification Framework with Prototypical-Adversarial Learning.","authors":"Junzhong Ji, Gan Liu, Xingyu Wang","doi":"10.1109/JBHI.2025.3569734","DOIUrl":"10.1109/JBHI.2025.3569734","url":null,"abstract":"<p><p>In recent years, semi-supervised learning (SSL) for functional brain network (FBN) classification has gained considerable attention due to its potential to leverage large amounts of unlabeled data from multisite. However, existing SSL methods often struggle to address the distributional differences across different sites, which limits their ability to extract discriminative features from the unlabeled data, thus hindering classification performance. To overcome this challenge, we propose a novel consistency semi-supervised FBN classification framework with prototypical-adversarial learning, termed CSBNC-PAL. Specifically, we first design a contrastive consistency module (CCM) that utilizes contrastive learning to more effectively exploit unlabeled data and learn preliminary feature representations. Then, we introduce a prototype alignment module (PAM) that computes site-aware prototypes through weighted feature clustering to guide inter-site feature alignment, and achieve inter-site equilibrium feature representations. Finally, we develop an adversarial alignment module (AAM) that employs site-discriminative adversarial training based on a gradient reversal layer to guide intra-site feature alignment, and learn site-invariant features. The three modules above are optimized collectively in an end-to-end manner, ensuring effective learning from both labeled and unlabeled data while alleviating the distribution differences of multisite data. Experiments on the ABIDE I, ABIDE II, and ADHD-200 datasets demonstrate that the CSBNC-PAL outperforms many state-of-the-art SSL methods in FBN classification.</p>","PeriodicalId":13073,"journal":{"name":"IEEE Journal of Biomedical and Health Informatics","volume":"PP ","pages":""},"PeriodicalIF":6.7,"publicationDate":"2025-05-12","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"143981414","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"医学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Benchmarking Radiology Report Generation From Noisy Free-Texts. 从嘈杂的自由文本生成基准放射学报告。
IF 6.7 2区 医学
IEEE Journal of Biomedical and Health Informatics Pub Date : 2025-05-12 DOI: 10.1109/JBHI.2025.3569428
Yujian Yuan, Yanting Zheng, Liangqiong Qu
{"title":"Benchmarking Radiology Report Generation From Noisy Free-Texts.","authors":"Yujian Yuan, Yanting Zheng, Liangqiong Qu","doi":"10.1109/JBHI.2025.3569428","DOIUrl":"10.1109/JBHI.2025.3569428","url":null,"abstract":"<p><p>Automatic radiology report generation can enhance diagnostic efficiency and accuracy. However, clean open-source imaging scan-report pairs are limited in scale and variety. Moreover, the vast amount of radiological texts available online is often too noisy to be directly employed. To address this challenge, we introduce a novel task called Noisy Report Refinement (NRR), which generates radiology reports from noisy free-texts. To achieve this, we propose a report refinement pipeline that leverages large language models (LLMs) enhanced with guided self-critique and report selection strategies. To address the inability of existing radiology report generation metrics in measuring cleanliness, radiological usefulness, and factual correctness across various modalities of reports in NRR task, we introduce a new benchmark, NRRBench, for NRR evaluation. This benchmark includes two online-sourced datasets and four clinically explainable LLM-based metrics: two metrics evaluate the matching rate of radiology entities and modality-specific template attributes respectively, one metric assesses report cleanliness, and a combined metric evaluates overall NRR performance. Experiments demonstrate that guided self-critique and report selection strategies significantly improve the quality of refined reports. Additionally, our proposed metrics show a much higher correlation with noisy rate and error count of reports than radiology report generation metrics in evaluating NRR.</p>","PeriodicalId":13073,"journal":{"name":"IEEE Journal of Biomedical and Health Informatics","volume":"PP ","pages":""},"PeriodicalIF":6.7,"publicationDate":"2025-05-12","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"144019407","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"医学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Facial Anomaly Appraisal Using Discrepancy Optimization-Driven Automatic Inpainting. 基于差异优化的面部异常自动修复方法。
IF 6.7 2区 医学
IEEE Journal of Biomedical and Health Informatics Pub Date : 2025-05-12 DOI: 10.1109/JBHI.2025.3569290
Abdullah Hayajneh, Erchin Serpedin, Mitchell A Stotland
{"title":"Facial Anomaly Appraisal Using Discrepancy Optimization-Driven Automatic Inpainting.","authors":"Abdullah Hayajneh, Erchin Serpedin, Mitchell A Stotland","doi":"10.1109/JBHI.2025.3569290","DOIUrl":"https://doi.org/10.1109/JBHI.2025.3569290","url":null,"abstract":"<p><p>This work presents a novel machine learning and signal processing framework designed to consistently detect, localize, and rate facial anomalies such as cleft lip deformity. The goal of this research is to establish a universal and objective measure of facial abnormalities, capable of sensitively identifying both subtle and significant deformities. The proposed model utilizes an enhanced two-phase automatic inpainting method for face normalization, effectively removing anomalies from the image and replacing them with normal facial content. The framework leverages an efficient knowledge distillation model to estimate the initial heatmap that highlights potential facial anomalies. This heatmap is subsequently converted into a mask for inpainting, which is applied to normalize the original face. A deep convolutional neural network (CNN)-based feature extraction method is then employed to compare the anomalous facial image with its normalized counterpart, enabling robust detection and evaluation of various facial anomalies. This is achieved by obtaining a noise-reduced final heatmap that more accurately scores the level of normality in the face. The normalization protocol delivers results comparable to state-of-the-art methods, while being significantly faster, taking less than one second from image upload to obtaining the face rating. This makes it highly feasible for deployment in mobile applications. Additionally, the proposed method does not require anomalous data for model training, while efficiently detecting and assessing various facial anomalies. We demonstrate that this unique computerized image appraisal system generates facial normality/abnormality scores that closely correlate with human intuition, exhibiting 92% correlation with human scores.</p>","PeriodicalId":13073,"journal":{"name":"IEEE Journal of Biomedical and Health Informatics","volume":"PP ","pages":""},"PeriodicalIF":6.7,"publicationDate":"2025-05-12","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"143994974","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"医学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
GAMMNet: Gating Multi-head Attention in a Multi-modal Deep Network for Sound Based Respiratory Disease Detection. GAMMNet:基于声音的呼吸疾病检测的多模态深度网络中的多头注意门控。
IF 6.7 2区 医学
IEEE Journal of Biomedical and Health Informatics Pub Date : 2025-05-12 DOI: 10.1109/JBHI.2025.3569160
Shaokang Liu, Zhaoji Dai, Zihong Zhuang, Xianwei Zheng, Minfan He, Qing Miao
{"title":"GAMMNet: Gating Multi-head Attention in a Multi-modal Deep Network for Sound Based Respiratory Disease Detection.","authors":"Shaokang Liu, Zhaoji Dai, Zihong Zhuang, Xianwei Zheng, Minfan He, Qing Miao","doi":"10.1109/JBHI.2025.3569160","DOIUrl":"https://doi.org/10.1109/JBHI.2025.3569160","url":null,"abstract":"<p><p>Respiratory diseases present significant challenges to global health due to their high morbidity and mortality rates. Traditional diagnostic methods, such as chest radiographs and blood tests, often lead to unnecessary costs and resource strain, as well as potential risks of cross-contamination during these procedures. In recent years, contactless sensing and intelligent technologies, particularly multi-modal sound-based deep learning methods, have emerged as promising solutions for the early detection of respiratory diseases. While these methods have shown encouraging results, the integration of multi-modal features has not been sufficiently explored, which limits the enhancement of diagnostic accuracy. To address this issue, we introduce GAMMNet, a novel multi-modal neural network designed to enhance the detection of respiratory diseases by leveraging multi-modal sound data collected from contactless recording devices. GAMMNet utilizes a unique gating mechanism that adaptively regulates the influence of each modality on the classification results. Additionally, our model incorporates multi-head attention and linear transformation modules to further enhance classification performance. Our GAMMNet achieves state-of-the-art classification results, compared to existing deep learning based methods, on real-world multi-modal respiratory sound datasets. These findings demonstrate the robustness and effectiveness of GAMMNet in the contactless monitoring and early detection of respiratory diseases.</p>","PeriodicalId":13073,"journal":{"name":"IEEE Journal of Biomedical and Health Informatics","volume":"PP ","pages":""},"PeriodicalIF":6.7,"publicationDate":"2025-05-12","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"143981417","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"医学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Function-structural Interaction with Progressive and Multi-level Feature Fusion for ADHD Classification. 功能结构相互作用与渐进和多层次特征融合在ADHD分类中的应用。
IF 6.7 2区 医学
IEEE Journal of Biomedical and Health Informatics Pub Date : 2025-05-12 DOI: 10.1109/JBHI.2025.3569090
Chunhong Cao, Xingxing Li, Mengyang Wang, Xiaolong Chen, Xieping Gao
{"title":"Function-structural Interaction with Progressive and Multi-level Feature Fusion for ADHD Classification.","authors":"Chunhong Cao, Xingxing Li, Mengyang Wang, Xiaolong Chen, Xieping Gao","doi":"10.1109/JBHI.2025.3569090","DOIUrl":"https://doi.org/10.1109/JBHI.2025.3569090","url":null,"abstract":"<p><p>Individuals with Attention Deficit Hyperactivity Disorder (ADHD) exhibit intricate structural and functional interconnectivity across multiple brain regions. These patients demonstrate abnormal alterations in both respective modal brain regions and co-occurrent brain regions. Furthermore, there exist multi-level relationships between these abnormal brain structures and functions, encompassing hierarchical interactions between function-structural alterations as well as hierarchical progression from local regions to broader brain networks. However, most existing multi-modal ADHD classification approaches independently embed functional and structural data into separate spaces for information integration, often predominately focusing on uni-modal features. This approaches lead to a significant loss of features related to functiona-structural interaction relationships. Additionally, it is crucial for ADHD classification to accurately identify both uni-modal and co-occurrent abnormal alterations in brain regions which have hierarchical progression relationships. This study proposes a function-structural interaction multi-modal network with progressive and multi-level feature fusion (FSIPM) for ADHD classification. The main contributions are threefold: 1) An innovative function-structural interaction method is proposed to facilitate the mutual regulation of information across modalities, thereby relieving modal feature bias caused by integrated fusion. 2) A multi-level refinement framework is designed to promote the identification of both individual and co-occurrent abnormal brain regions. This progressive approach models the function-structural alterations of abnormal brain regions and the hierarchical relationships from local to brain networks, ensuring a deeper understanding of brain abnormalities. 3) Multi-level feature fusion aims to minimize the loss of details caused by consecutive sampling operations during the progressive process of the network, contributing to a more accurate and nuanced representation of ADHD-related brain alterations. Experimental results on the ADHD-200 and ABIDE I datasets demonstrate that FSIPM achieves competitive performance in ADHD classification while revealing uni-modal and co-occurrent altered brain regions that are consistent with clinical findings.</p>","PeriodicalId":13073,"journal":{"name":"IEEE Journal of Biomedical and Health Informatics","volume":"PP ","pages":""},"PeriodicalIF":6.7,"publicationDate":"2025-05-12","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"143984641","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"医学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
TarMGDif: Target-specific Molecular Graphs Generation Based on Diffusion Model. TarMGDif:基于扩散模型的靶向分子图生成。
IF 6.7 2区 医学
IEEE Journal of Biomedical and Health Informatics Pub Date : 2025-05-12 DOI: 10.1109/JBHI.2025.3569105
Shuang Wang, Yunjing Zhang, Dingming Liang, Kaiyu Dong, Tao Song
{"title":"TarMGDif: Target-specific Molecular Graphs Generation Based on Diffusion Model.","authors":"Shuang Wang, Yunjing Zhang, Dingming Liang, Kaiyu Dong, Tao Song","doi":"10.1109/JBHI.2025.3569105","DOIUrl":"https://doi.org/10.1109/JBHI.2025.3569105","url":null,"abstract":"<p><p>Generating drug-like molecules that specifically bind to target proteins remains a resource-intensive challenge. Many studies focus on designing effective networks to accurately extract relevant features from target proteins, which can be challenging. Additionally, most target-specific molecule generation methods based on diffusion models process the 3D information of molecules and proteins, necessitating the maintenance of equivariance at each step. This paper proposes TarMGDif, a novel target-specific molecular graph generation model based on a discrete denoising diffusion framework which could handle graph structure. TarMGDif incorporates a global features embedding network that captures ring features to generate chemically valid rings, while the time step of the diffusion model is also learned through this network. Besides, a novel node-to-edge attention module is proposed to capture dependencies between nodes and edges.Extensive experiments conducted on three datasets demonstrate the advanced performance of TarMGDif. Furthermore, through transfer learning, the model generates molecules specifically targeting the DRD2 protein, with the newly designed molecules exhibiting pharmacological properties similar to known inhibitors. These findings underscore the potential of TarMGDif in facilitating the efficient design of target-specific drug-like molecules.</p>","PeriodicalId":13073,"journal":{"name":"IEEE Journal of Biomedical and Health Informatics","volume":"PP ","pages":""},"PeriodicalIF":6.7,"publicationDate":"2025-05-12","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"144009376","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"医学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
A Medical Multimodal Large Language Model for Pediatric Pneumonia. 儿童肺炎的医学多模态大语言模型。
IF 6.7 2区 医学
IEEE Journal of Biomedical and Health Informatics Pub Date : 2025-05-12 DOI: 10.1109/JBHI.2025.3569361
Weiwei Tian, Xinyu Huang, Tianhao Cheng, Wen He, Jinwu Fang, Rui Feng, Daoying Geng, Xiaobo Zhang
{"title":"A Medical Multimodal Large Language Model for Pediatric Pneumonia.","authors":"Weiwei Tian, Xinyu Huang, Tianhao Cheng, Wen He, Jinwu Fang, Rui Feng, Daoying Geng, Xiaobo Zhang","doi":"10.1109/JBHI.2025.3569361","DOIUrl":"https://doi.org/10.1109/JBHI.2025.3569361","url":null,"abstract":"<p><p>Pediatric pneumonia is the leading cause of death among children under five years worldwide, imposing a substantial burden on affected families. Currently, there are three significant hurdles in diagnosing and treating pediatric pneumonia. Firstly, pediatric pneumonia shares similar symptoms with other respiratory diseases, making rapid and accurate differential diagnosis challenging. Secondly, primary hospitals often lack sufficient medical resources and experienced doctors. Lastly, providing personalized diagnostic reports and treatment recommendations is labor-intensive and time-consuming. To tackle these challenges, we proposed a Medical Multimodal Large Language Model for Pediatric Pneumonia (P2Med-MLLM). It was capable of handling diverse clinical tasks-such as generating free-text medical records and radiology reports-within a unified framework. Specifically, P2Med-MLLM was trained on a large-scale dataset, including real clinical information from 163,999 outpatient and 8,684 inpatient cases. It can process both plain text data (e.g., outpatient and inpatient records) and interleaved image-text pairs (e.g., 2D chest X-ray images, 3D chest Computed Tomography images, and corresponding radiology reports). We designed a three-stage training strategy to enable P2Med-MLLM to comprehend medical knowledge and follow instructions for various clinical decision-support tasks. To rigorously evaluate P2Med-MLLM's performance, we conducted automatic scoring by the large language model and manual scoring by the specialist on the test set of 642 samples, meticulously verified by pediatric pulmonology specialists. The results demonstrated the reliability of automated scoring and the superiority of P2Med-MLLM. This work plays a crucial role in assisting doctors with prompt diagnosis and treatment planning, reducing severe symptom mortality rates, and optimizing the allocation of medical resources.</p>","PeriodicalId":13073,"journal":{"name":"IEEE Journal of Biomedical and Health Informatics","volume":"PP ","pages":""},"PeriodicalIF":6.7,"publicationDate":"2025-05-12","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"143965541","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"医学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
A Finetuning Deep Learning Framework for Pan-species Promoters with Pseudo Time Series Analysis on Time and Frequency Space. 基于时频空间伪时间序列分析的泛物种启动子微调深度学习框架。
IF 6.7 2区 医学
IEEE Journal of Biomedical and Health Informatics Pub Date : 2025-05-09 DOI: 10.1109/JBHI.2025.3568145
Ruimeng Li, Qinke Peng, Haozhou Li, Wentong Sun
{"title":"A Finetuning Deep Learning Framework for Pan-species Promoters with Pseudo Time Series Analysis on Time and Frequency Space.","authors":"Ruimeng Li, Qinke Peng, Haozhou Li, Wentong Sun","doi":"10.1109/JBHI.2025.3568145","DOIUrl":"https://doi.org/10.1109/JBHI.2025.3568145","url":null,"abstract":"<p><p>Promoter identification and classification play crucial roles in unraveling gene mechanisms. Promoters are characterized by specific motifs, such as the TATA-box for eukaryotes and the Pribnow box for prokaryotes, which are known as elements. These constitute the core components, intimately tied to promoter function. However, the heterogeneity of promoters across different species poses a significant challenge to improving identification models. In our study, we introduce ProTriCNN, a deep learning method designed for promoter identification. Based on promoters representation, ProTriCNN treats promoters as pseudo-time series, utilizing this approach to capture the intricate heterogeneity of promoter elements. Furthermore, we introduce TransPro, a ProTriCNN-based Fine-tuning framework to improve identification performance across different species. To better align source species and target species, the TransPro utilizes elements and species evolutionary trees to represent the locality difference between source and target species across various levels and time-frequency space, respectively. Compared to state-of-the-art methods, ProTriCNN demonstrates superior performance across all species, achieving an average accuracy improvement of 2.1% and a 20% enhancement in the Matthews coefficient. TransPro further attains accuracy improvement of the highest 8% and a 25% enhancement in the Matthews coefficient compared to ProTriCNN. The source code and the associated datasets are freely available at https://github.com/Limomo33/promoter.</p>","PeriodicalId":13073,"journal":{"name":"IEEE Journal of Biomedical and Health Informatics","volume":"PP ","pages":""},"PeriodicalIF":6.7,"publicationDate":"2025-05-09","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"144018488","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"医学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Integrating Fine-Tuned LLM with Acoustic Features for Enhanced Detection of Alzheimer's Disease. 整合微调LLM与声学特征,增强阿尔茨海默病的检测。
IF 6.7 2区 医学
IEEE Journal of Biomedical and Health Informatics Pub Date : 2025-05-09 DOI: 10.1109/JBHI.2025.3566615
Filippo Casu, Andrea Lagorio, Pietro Ruiu, Giuseppe A Trunfio, Enrico Grosso
{"title":"Integrating Fine-Tuned LLM with Acoustic Features for Enhanced Detection of Alzheimer's Disease.","authors":"Filippo Casu, Andrea Lagorio, Pietro Ruiu, Giuseppe A Trunfio, Enrico Grosso","doi":"10.1109/JBHI.2025.3566615","DOIUrl":"https://doi.org/10.1109/JBHI.2025.3566615","url":null,"abstract":"<p><p>Dementia represents a global public health concern, with the early detection of Alzheimer's disease, the most prevalent form of dementia, being of paramount importance. Given the limited availability of suitable biomarkers, research has shown that early cognitive impairment can be identified through patients' spoken language. This paper presents a multi-modal system for automatic Alzheimer's disease detection using speech. The system has been trained on spoken recordings of healthy individuals and Alzheimer's patients describing an image, a task requiring linguistic and cognitive skills. Built on fine-tuned advanced Large Language Models, audio feature extractors, and classifiers, the system, after an extensive comparison of single and multi-modal architectures, achieves optimal results with the combination of Mistral-7B, VGGish, and Support Vector Classifier, outperforming previous methods on the ADReSSo 2021 test set.</p>","PeriodicalId":13073,"journal":{"name":"IEEE Journal of Biomedical and Health Informatics","volume":"PP ","pages":""},"PeriodicalIF":6.7,"publicationDate":"2025-05-09","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"144011799","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"医学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
0
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
确定
请完成安全验证×
相关产品
×
本文献相关产品
联系我们:info@booksci.cn Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。 Copyright © 2023 布克学术 All rights reserved.
京ICP备2023020795号-1
ghs 京公网安备 11010802042870号
Book学术文献互助
Book学术文献互助群
群 号:481959085
Book学术官方微信