Computerized Medical Imaging and Graphics最新文献

筛选
英文 中文
TCDE-Net: An unsupervised dual-encoder network for 3D brain medical image registration TCDE-Net:用于三维脑医学图像配准的无监督双编码器网络
IF 5.4 2区 医学
Computerized Medical Imaging and Graphics Pub Date : 2025-03-23 DOI: 10.1016/j.compmedimag.2025.102527
Xin Yang , Dongxue Li , Liwei Deng , Sijuan Huang , Jing Wang
{"title":"TCDE-Net: An unsupervised dual-encoder network for 3D brain medical image registration","authors":"Xin Yang ,&nbsp;Dongxue Li ,&nbsp;Liwei Deng ,&nbsp;Sijuan Huang ,&nbsp;Jing Wang","doi":"10.1016/j.compmedimag.2025.102527","DOIUrl":"10.1016/j.compmedimag.2025.102527","url":null,"abstract":"<div><div>Medical image registration is a critical task in aligning medical images from different time points, modalities, or individuals, essential for accurate diagnosis and treatment planning. Despite significant progress in deep learning-based registration methods, current approaches still face considerable challenges, such as insufficient capture of local details, difficulty in effectively modeling global contextual information, and limited robustness in handling complex deformations. These limitations hinder the precision of high-resolution registration, particularly when dealing with medical images with intricate structures. To address these issues, this paper presents a novel registration network (TCDE-Net), an unsupervised medical image registration method based on a dual-encoder architecture. The dual encoders complement each other in feature extraction, enabling the model to effectively handle large-scale nonlinear deformations and capture intricate local details, thereby enhancing registration accuracy. Additionally, the detail-enhancement attention module aids in restoring fine-grained features, improving the network's capability to address complex deformations such as those at gray-white matter boundaries. Experimental results on the OASIS, IXI, and Hammers-n30r95 3D brain MR dataset demonstrate that this method outperforms commonly used registration techniques across multiple evaluation metrics, achieving superior performance and robustness. Our code is available at <span><span>https://github.com/muzidongxue/TCDE-Net</span><svg><path></path></svg></span>.</div></div>","PeriodicalId":50631,"journal":{"name":"Computerized Medical Imaging and Graphics","volume":"123 ","pages":"Article 102527"},"PeriodicalIF":5.4,"publicationDate":"2025-03-23","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"143697647","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"医学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Class balancing diversity multimodal ensemble for Alzheimer’s disease diagnosis and early detection 类平衡多样性多模态集成在阿尔茨海默病诊断和早期检测中的应用
IF 5.4 2区 医学
Computerized Medical Imaging and Graphics Pub Date : 2025-03-22 DOI: 10.1016/j.compmedimag.2025.102529
Arianna Francesconi , Lazzaro di Biase , Donato Cappetta , Fabio Rebecchi , Paolo Soda , Rosa Sicilia , Valerio Guarrasi , Alzheimer’s Disease Neuroimaging Initiative
{"title":"Class balancing diversity multimodal ensemble for Alzheimer’s disease diagnosis and early detection","authors":"Arianna Francesconi ,&nbsp;Lazzaro di Biase ,&nbsp;Donato Cappetta ,&nbsp;Fabio Rebecchi ,&nbsp;Paolo Soda ,&nbsp;Rosa Sicilia ,&nbsp;Valerio Guarrasi ,&nbsp;Alzheimer’s Disease Neuroimaging Initiative","doi":"10.1016/j.compmedimag.2025.102529","DOIUrl":"10.1016/j.compmedimag.2025.102529","url":null,"abstract":"<div><div>Alzheimer’s disease (AD) poses significant global health challenges due to its increasing prevalence and associated societal costs. Early detection and diagnosis of AD are critical for delaying progression and improving patient outcomes. Traditional diagnostic methods and single-modality data often fall short in identifying early-stage AD and distinguishing it from Mild Cognitive Impairment (MCI). This study addresses these challenges by introducing a novel approach: multImodal enseMble via class BALancing diversity for iMbalancEd Data (IMBALMED). IMBALMED integrates multimodal data from the Alzheimer’s Disease Neuroimaging Initiative database, including clinical assessments, neuroimaging phenotypes, biospecimen, and subject characteristics data. It employs a new ensemble of model classifiers, designed specifically for this framework, which combines eight distinct families of learning paradigms trained with diverse class balancing techniques to overcome class imbalance and enhance model accuracy. We evaluate IMBALMED on two diagnostic tasks (binary and ternary classification) and four binary early detection tasks (at 12, 24, 36, and 48 months), comparing its performance with state-of-the-art algorithms and an unbalanced dataset method. To further validate the proposed model and ensure genuine generalization to real-world scenarios, we conducted an external validation experiment using data from the most recent phase of the ADNI dataset. IMBALMED demonstrates superior diagnostic accuracy and predictive performance in both binary and ternary classification tasks, significantly improving early detection of MCI at a 48-month time point and showing excellent generalizability in the 12-month task during external validation. The method shows improved classification performance and robustness, offering a promising solution for early detection and management of AD.</div></div>","PeriodicalId":50631,"journal":{"name":"Computerized Medical Imaging and Graphics","volume":"123 ","pages":"Article 102529"},"PeriodicalIF":5.4,"publicationDate":"2025-03-22","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"143697586","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"医学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
SpineMamba: Enhancing 3D spinal segmentation in clinical imaging through residual visual Mamba layers and shape priors SpineMamba:通过残余视觉曼巴层和形状先验增强临床成像中的3D脊柱分割
IF 5.4 2区 医学
Computerized Medical Imaging and Graphics Pub Date : 2025-03-22 DOI: 10.1016/j.compmedimag.2025.102531
Zhiqing Zhang , Tianyong Liu , Guojia Fan , Na Li , Bin Li , Yao Pu , Qianjin Feng , Shoujun Zhou
{"title":"SpineMamba: Enhancing 3D spinal segmentation in clinical imaging through residual visual Mamba layers and shape priors","authors":"Zhiqing Zhang ,&nbsp;Tianyong Liu ,&nbsp;Guojia Fan ,&nbsp;Na Li ,&nbsp;Bin Li ,&nbsp;Yao Pu ,&nbsp;Qianjin Feng ,&nbsp;Shoujun Zhou","doi":"10.1016/j.compmedimag.2025.102531","DOIUrl":"10.1016/j.compmedimag.2025.102531","url":null,"abstract":"<div><div>Accurate segmentation of three-dimensional (3D) clinical medical images is critical for the diagnosis and treatment of spinal diseases. However, the complexity of spinal anatomy and the inherent uncertainties of current imaging technologies pose significant challenges for the semantic segmentation of spinal images. Although convolutional neural networks (CNNs) and Transformer-based models have achieved remarkable progress in spinal segmentation, their limitations in modeling long-range dependencies hinder further improvements in segmentation accuracy. To address these challenges, we propose a novel framework, SpineMamba, which incorporates a residual visual Mamba layer capable of effectively capturing and modeling the deep semantic features and long-range spatial dependencies in 3D spinal data. To further enhance the structural semantic understanding of the vertebrae, we also propose a novel spinal shape prior module that captures specific anatomical information about the spine from medical images, significantly enhancing the model’s ability to extract structural semantic information of the vertebrae. Extensive comparative and ablation experiments across three datasets demonstrate that SpineMamba outperforms existing state-of-the-art models. On two computed tomography (CT) datasets, the average Dice similarity coefficients achieved are 94.40±4% and 88.28±3%, respectively, while on a magnetic resonance (MR) dataset, the model achieves a Dice score of 86.95±10%. Notably, SpineMamba surpasses the widely recognized nnU-Net in segmentation accuracy, with a maximum improvement of 3.63 percentage points. These results highlight the precision, robustness, and exceptional generalization capability of SpineMamba.</div></div>","PeriodicalId":50631,"journal":{"name":"Computerized Medical Imaging and Graphics","volume":"123 ","pages":"Article 102531"},"PeriodicalIF":5.4,"publicationDate":"2025-03-22","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"143706291","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"医学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Multi-modal MRI synthesis with conditional latent diffusion models for data augmentation in tumor segmentation 基于条件潜在扩散模型的多模态MRI合成在肿瘤分割中的数据增强
IF 5.4 2区 医学
Computerized Medical Imaging and Graphics Pub Date : 2025-03-21 DOI: 10.1016/j.compmedimag.2025.102532
Aghiles Kebaili , Jérôme Lapuyade-Lahorgue , Pierre Vera , Su Ruan
{"title":"Multi-modal MRI synthesis with conditional latent diffusion models for data augmentation in tumor segmentation","authors":"Aghiles Kebaili ,&nbsp;Jérôme Lapuyade-Lahorgue ,&nbsp;Pierre Vera ,&nbsp;Su Ruan","doi":"10.1016/j.compmedimag.2025.102532","DOIUrl":"10.1016/j.compmedimag.2025.102532","url":null,"abstract":"<div><div>Multimodality is often necessary for improving object segmentation tasks, especially in the case of multilabel tasks, such as tumor segmentation, which is crucial for clinical diagnosis and treatment planning. However, a major challenge in utilizing multimodality with deep learning remains: the limited availability of annotated training data, primarily due to the time-consuming acquisition process and the necessity for expert annotations. Although deep learning has significantly advanced many tasks in medical imaging, conventional augmentation techniques are often insufficient due to the inherent complexity of volumetric medical data. To address this problem, we propose an innovative slice-based latent diffusion architecture for the generation of 3D multi-modal images and their corresponding multi-label masks. Our approach enables the simultaneous generation of the image and mask in a slice-by-slice fashion, leveraging a positional encoding and a Latent Aggregation module to maintain spatial coherence and capture slice sequentiality. This method effectively reduces the computational complexity and memory demands typically associated with diffusion models. Additionally, we condition our architecture on tumor characteristics to generate a diverse array of tumor variations and enhance texture using a refining module that acts like a super-resolution mechanism, mitigating the inherent blurriness caused by data scarcity in the autoencoder. We evaluate the effectiveness of our synthesized volumes using the BRATS2021 dataset to segment the tumor with three tissue labels and compare them with other state-of-the-art diffusion models through a downstream segmentation task, demonstrating the superior performance and efficiency of our method. While our primary application is tumor segmentation, this method can be readily adapted to other modalities. Code is available here : <span><span>https://github.com/Arksyd96/multi-modal-mri-and-mask-synthesis-with-conditional-slice-based-ldm</span><svg><path></path></svg></span>.</div></div>","PeriodicalId":50631,"journal":{"name":"Computerized Medical Imaging and Graphics","volume":"123 ","pages":"Article 102532"},"PeriodicalIF":5.4,"publicationDate":"2025-03-21","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"143687870","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"医学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Deep Guess acceleration for explainable image reconstruction in sparse-view CT 稀疏视图CT中可解释图像重建的深度猜测加速
IF 5.4 2区 医学
Computerized Medical Imaging and Graphics Pub Date : 2025-03-21 DOI: 10.1016/j.compmedimag.2025.102530
Elena Loli Piccolomini , Davide Evangelista , Elena Morotti
{"title":"Deep Guess acceleration for explainable image reconstruction in sparse-view CT","authors":"Elena Loli Piccolomini ,&nbsp;Davide Evangelista ,&nbsp;Elena Morotti","doi":"10.1016/j.compmedimag.2025.102530","DOIUrl":"10.1016/j.compmedimag.2025.102530","url":null,"abstract":"<div><div>Sparse-view Computed Tomography (CT) is an emerging protocol designed to reduce X-ray dose radiation in medical imaging. Reconstructions based on the traditional Filtered Back Projection algorithm suffer from severe artifacts due to sparse data. In contrast, Model-Based Iterative Reconstruction (MBIR) algorithms, though better at mitigating noise through regularization, are too computationally costly for clinical use. This paper introduces a novel technique, denoted as the Deep Guess acceleration scheme, using a trained neural network both to quicken the regularized MBIR and to enhance the reconstruction accuracy. We integrate state-of-the-art deep learning tools to initialize a clever starting guess for a proximal algorithm solving a non-convex model and thus computing a (mathematically) interpretable solution image in a few iterations. Experimental results on real and synthetic CT images demonstrate the Deep Guess effectiveness in (very) sparse tomographic protocols, where it overcomes its mere variational counterpart and many data-driven approaches at the state of the art. We also consider a ground truth-free implementation and test the robustness of the proposed framework to noise.</div></div>","PeriodicalId":50631,"journal":{"name":"Computerized Medical Imaging and Graphics","volume":"123 ","pages":"Article 102530"},"PeriodicalIF":5.4,"publicationDate":"2025-03-21","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"143706292","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"医学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Automatic colon segmentation on T1-FS MR images T1-FS MR图像的自动冒号分割。
IF 5.4 2区 医学
Computerized Medical Imaging and Graphics Pub Date : 2025-03-17 DOI: 10.1016/j.compmedimag.2025.102528
Bernat Orellana , Isabel Navazo , Pere Brunet , Eva Monclús , Álvaro Bendezú , Fernando Azpiroz
{"title":"Automatic colon segmentation on T1-FS MR images","authors":"Bernat Orellana ,&nbsp;Isabel Navazo ,&nbsp;Pere Brunet ,&nbsp;Eva Monclús ,&nbsp;Álvaro Bendezú ,&nbsp;Fernando Azpiroz","doi":"10.1016/j.compmedimag.2025.102528","DOIUrl":"10.1016/j.compmedimag.2025.102528","url":null,"abstract":"<div><div>The volume and distribution of the colonic contents provides valuable insights into the effects of diet on gut microbiotica involving both clinical diagnosis and research. In terms of Magnetic Resonance Imaging modalities, T2-weighted images allow the segmentation of the colon lumen, while fecal and gas contents can be only distinguished on the T1-weighted Fat-Sat modality. However, the manual segmentation of T1-weighted Fat-Sat is challenging, and no automatic segmentation methods are known.</div><div>This paper proposed a non-supervised algorithm providing an accurate T1-weighted Fat-Sat colon segmentation via the registration of an existing colon segmentation in T2-weighted modality.</div><div>The algorithm consists of two phases. It starts with a registration process based on a classical deformable registration method, followed by a novel Iterative Colon Registration process that utilizes a mesh deformation approach. This approach is guided by a probabilistic model that provides the likelihood of the colon boundary, followed by a shape preservation process of the colon segmentation on T2-weighted images. The iterative process converges to achieve an optimal fit for colon segmentation in T1-weighted Fat-Sat images.</div><div>The segmentation algorithm has been tested on multiple datasets (154 scans) and acquisition machines (3) as part of the proof of concept for the proposed methodology. The quantitative evaluation was based on two metrics: the percentage of ground truth labeled feces correctly identified by our proposal (<span><math><mrow><mn>93</mn><mo>±</mo><mn>5</mn><mtext>%</mtext></mrow></math></span>), and the volume variation between the existing colon segmentation in the T2-weighted modality and the colon segmentation computed in T1-weighted Fat-Sat images.</div><div>Quantitative and medical evaluations demonstrated a degree of accuracy, usability, and stability concerning the acquisition hardware, making the algorithm suitable for clinical application and research.</div></div>","PeriodicalId":50631,"journal":{"name":"Computerized Medical Imaging and Graphics","volume":"123 ","pages":"Article 102528"},"PeriodicalIF":5.4,"publicationDate":"2025-03-17","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"143670163","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"医学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
A multimodal framework for assessing the link between pathomics, transcriptomics, and pancreatic cancer mutations 用于评估病理、转录组学和胰腺癌突变之间联系的多模式框架
IF 5.4 2区 医学
Computerized Medical Imaging and Graphics Pub Date : 2025-03-15 DOI: 10.1016/j.compmedimag.2025.102526
Francesco Berloco, Gian Maria Zaccaria, Nicola Altini, Simona Colucci, Vitoantonio Bevilacqua
{"title":"A multimodal framework for assessing the link between pathomics, transcriptomics, and pancreatic cancer mutations","authors":"Francesco Berloco,&nbsp;Gian Maria Zaccaria,&nbsp;Nicola Altini,&nbsp;Simona Colucci,&nbsp;Vitoantonio Bevilacqua","doi":"10.1016/j.compmedimag.2025.102526","DOIUrl":"10.1016/j.compmedimag.2025.102526","url":null,"abstract":"<div><div>In Pancreatic Ductal Adenocarcinoma (PDAC), predicting genetic mutations directly from histopathological images using Deep Learning can provide valuable insights. The combination of several omics can provide further knowledge on mechanisms underlying tumor biology. This study aimed at developing an explainable multimodal pipeline to predict genetic mutations for the <em>KRAS</em>, <em>TP53</em>, <em>SMAD4</em>, and <em>CDKN2A</em> genes, integrating pathomic features with transcriptomics from two independent datasets, the TCGA-PAAD, assumed as training set, and the CPTAC-PDA, as external validation set. Large and small configurations of CLAM (Clustering-constrained Attention Multiple Instance Learning) models were evaluated with three different feature extractors (ResNet50, UNI, and CONCH). RNA-seq data were pre-processed both conventionally and using three autoencoder architectures. The processed transcript panels were input into machine learning (ML) models for mutation classification. Attention maps and SHAP were employed, highlighting significant features from both data modalities. A fusion layer or a voting mechanism combined the outputs from pathomic and transcriptomic models, obtaining a multimodal prediction. Performance comparisons were assessed by Area Under Receiver Operating Characteristic (AUROC) and Precision-Recall (AUPRC) curves. On the validation set, for <em>KRAS</em>, multimodal ML achieved 0.92 of AUROC and 0.98 of AUPRC. For <em>TP53</em>, the multimodal voting model achieved 0.75 of AUROC and 0.85 of AUPRC. For <em>SMAD4</em> and <em>CDKN2A</em>, transcriptomic ML models achieved AUROC of 0.71 and 0.65, while multimodal ML showed AUPRC of 0.39 and 0.37, respectively. This approach demonstrated the potential of combining pathomics with transcriptomics, offering an interpretable framework for predicting key genetic mutations in PDAC.</div></div>","PeriodicalId":50631,"journal":{"name":"Computerized Medical Imaging and Graphics","volume":"123 ","pages":"Article 102526"},"PeriodicalIF":5.4,"publicationDate":"2025-03-15","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"143644207","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"医学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Liver lesion segmentation in ultrasound: A benchmark and a baseline network 超声肝病变分割:一个基准和基线网络
IF 5.4 2区 医学
Computerized Medical Imaging and Graphics Pub Date : 2025-03-14 DOI: 10.1016/j.compmedimag.2025.102523
Jialu Li , Lei Zhu , Guibao Shen , Baoliang Zhao , Ying Hu , Hai Zhang , Weiming Wang , Qiong Wang
{"title":"Liver lesion segmentation in ultrasound: A benchmark and a baseline network","authors":"Jialu Li ,&nbsp;Lei Zhu ,&nbsp;Guibao Shen ,&nbsp;Baoliang Zhao ,&nbsp;Ying Hu ,&nbsp;Hai Zhang ,&nbsp;Weiming Wang ,&nbsp;Qiong Wang","doi":"10.1016/j.compmedimag.2025.102523","DOIUrl":"10.1016/j.compmedimag.2025.102523","url":null,"abstract":"<div><div>Accurate liver lesion segmentation in ultrasound is a challenging task due to high speckle noise, ambiguous lesion boundaries, and inhomogeneous intensity distribution inside the lesion regions. This work first collected and annotated a dataset for liver lesion segmentation in ultrasound. In this paper, we propose a novel convolutional neural network to learn dual self-attentive transformer features for boosting liver lesion segmentation by leveraging the complementary information among non-local features encoded at different layers of the transformer architecture. To do so, we devise a dual self-attention refinement (DSR) module to synergistically utilize self-attention and reverse self-attention mechanisms to extract complementary lesion characteristics between cascaded multi-layer feature maps, assisting the model to produce more accurate segmentation results. Moreover, we propose a False-Positive–Negative loss to enable our network to further suppress the non-liver-lesion noise at shallow transformer layers and enhance more target liver lesion details into CNN features at deep transformer layers. Experimental results show that our network outperforms state-of-the-art methods quantitatively and qualitatively.</div></div>","PeriodicalId":50631,"journal":{"name":"Computerized Medical Imaging and Graphics","volume":"123 ","pages":"Article 102523"},"PeriodicalIF":5.4,"publicationDate":"2025-03-14","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"143644208","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"医学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
CQENet: A segmentation model for nasopharyngeal carcinoma based on confidence quantitative evaluation CQENet:基于置信度定量评价的鼻咽癌分割模型
IF 5.4 2区 医学
Computerized Medical Imaging and Graphics Pub Date : 2025-03-13 DOI: 10.1016/j.compmedimag.2025.102525
Yiqiu Qi , Lijun Wei , Jinzhu Yang , Jiachen Xu , Hongfei Wang , Qi Yu , Guoguang Shen , Yubo Cao
{"title":"CQENet: A segmentation model for nasopharyngeal carcinoma based on confidence quantitative evaluation","authors":"Yiqiu Qi ,&nbsp;Lijun Wei ,&nbsp;Jinzhu Yang ,&nbsp;Jiachen Xu ,&nbsp;Hongfei Wang ,&nbsp;Qi Yu ,&nbsp;Guoguang Shen ,&nbsp;Yubo Cao","doi":"10.1016/j.compmedimag.2025.102525","DOIUrl":"10.1016/j.compmedimag.2025.102525","url":null,"abstract":"<div><div>Accurate segmentation of the tumor regions of nasopharyngeal carcinoma (NPC) is of significant importance for radiotherapy of NPC. However, the precision of existing automatic segmentation methods for NPC remains inadequate, primarily manifested in the difficulty of tumor localization and the challenges in delineating blurred boundaries. Additionally, the black-box nature of deep learning models leads to insufficient quantification of the confidence in the results, preventing users from directly understanding the model’s confidence in its predictions, which severely impacts the clinical application of deep learning models. This paper proposes an automatic segmentation model for NPC based on confidence quantitative evaluation (CQENet). To address the issue of insufficient confidence quantification in NPC segmentation results, we introduce a confidence assessment module (CAM) that enables the model to output not only the segmentation results but also the confidence in those results, aiding users in understanding the uncertainty risks associated with model outputs. To address the difficulty in localizing the position and extent of tumors, we propose a tumor feature adjustment module (FAM) for precise tumor localization and extent determination. To address the challenge of delineating blurred tumor boundaries, we introduce a variance attention mechanism (VAM) to assist in edge delineation during fine segmentation. We conducted experiments on a multicenter NPC dataset, validating that our proposed method is effective and superior to existing state-of-the-art models, possessing considerable clinical application value.</div></div>","PeriodicalId":50631,"journal":{"name":"Computerized Medical Imaging and Graphics","volume":"123 ","pages":"Article 102525"},"PeriodicalIF":5.4,"publicationDate":"2025-03-13","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"143644206","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"医学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
NaMA-Mamba: Foundation model for generalizable nasal disease detection using masked autoencoder with Mamba on endoscopic images NaMA-Mamba:在内窥镜图像上使用蒙面自动编码器与Mamba进行通用鼻疾病检测的基础模型
IF 5.4 2区 医学
Computerized Medical Imaging and Graphics Pub Date : 2025-03-12 DOI: 10.1016/j.compmedimag.2025.102524
Wensheng Wang , Zewen Jin , Xueli Liu , Xinrong Chen
{"title":"NaMA-Mamba: Foundation model for generalizable nasal disease detection using masked autoencoder with Mamba on endoscopic images","authors":"Wensheng Wang ,&nbsp;Zewen Jin ,&nbsp;Xueli Liu ,&nbsp;Xinrong Chen","doi":"10.1016/j.compmedimag.2025.102524","DOIUrl":"10.1016/j.compmedimag.2025.102524","url":null,"abstract":"<div><div>Artificial intelligence (AI) has shown great promise in analyzing nasal endoscopic images for disease detection. However, current AI systems require extensive expert-labeled data for each specific medical condition, limiting their applications. In this work, the challenge is addressed through two key innovations, the creation of the first large-scale pre-training dataset of nasal endoscopic images, and the development of a novel self-learning AI system specifically designed for nasal endoscopy, named NaMA-Mamba. In the proposed NaMA-Mamba model, two key technologies are utilized, which are the nasal endoscopic state space model (NE-SSM) for analyzing sequences of images and an enhanced learning mechanism (CoMAE) for capturing fine details in nasal tissues. These innovations enable the system to learn effectively from unlabeled images while maintaining high accuracy across different diagnostic tasks. In extensive testing, NaMA-Mamba achieved remarkable results using minimal labeled data, matching the performance of traditional systems that require full expert labeling while needing only 1% of the labeled data for tasks such as detecting nasal polyps and identifying nasopharyngeal cancer. These results demonstrate the potential of NaMA-Mamba to significantly improve the efficiency and accessibility of AI-assisted nasal disease diagnosis in clinical practice.</div></div>","PeriodicalId":50631,"journal":{"name":"Computerized Medical Imaging and Graphics","volume":"122 ","pages":"Article 102524"},"PeriodicalIF":5.4,"publicationDate":"2025-03-12","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"143620168","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"医学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
0
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
确定
请完成安全验证×
相关产品
×
本文献相关产品
联系我们:info@booksci.cn Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。 Copyright © 2023 布克学术 All rights reserved.
京ICP备2023020795号-1
ghs 京公网安备 11010802042870号
Book学术文献互助
Book学术文献互助群
群 号:481959085
Book学术官方微信