Self-supervised U-transformer network with mask reconstruction for metal artifact reduction.

IF 3.3 3区 医学 Q2 ENGINEERING, BIOMEDICAL
Fanning Kong, Zaifeng Shi, Huaisheng Cao, Yudong Hao, Qingjie Cao
{"title":"Self-supervised U-transformer network with mask reconstruction for metal artifact reduction.","authors":"Fanning Kong, Zaifeng Shi, Huaisheng Cao, Yudong Hao, Qingjie Cao","doi":"10.1088/1361-6560/adbaae","DOIUrl":null,"url":null,"abstract":"<p><p><i>Objective</i>. Metal artifacts severely damaged human tissue information from the computed tomography (CT) image, posing significant challenges to disease diagnosis. Deep learning has been widely explored for the metal artifact reduction (MAR) task. Nevertheless, paired metal artifact CT datasets suitable for training do not exist in reality. Although the synthetic CT image dataset provides additional training data, the trained networks still generalize poorly to real metal artifact data.<i>Approach.</i>A self-supervised U-shaped transformer network is proposed to focus on model generalizability enhancement in MAR tasks. This framework consists of a self-supervised mask reconstruction pre-text task and a down-stream task. In the pre-text task, the CT images are randomly corrupted by masks. They are recovered with themselves as the label, aiming at acquiring the artifacts and tissue structure of the actual physical situation. Down-stream task fine-tunes MAR target through labeled images. Utilizing the multi-layer long-range feature extraction capabilities of the Transformer efficiently captures features of metal artifacts. The incorporation of the MAR bottleneck allows for the distinction of metal artifact features through cross-channel self-attention.<i>Main result</i>. Experiments demonstrate that the framework maintains strong generalization ability in the MAR task, effectively preserving tissue details while suppressing metal artifacts. The results achieved a peak signal-to-noise ratio of 43.86 dB and a structural similarity index of 0.9863 while ensuring the efficiency of the model inference. In addition, the Dice coefficient and mean intersection over union are improved by 11.70% and 9.51% in the segmentation of the MAR image, respectively.<i>Significance.</i>The combination of unlabeled real-artifact CT images and labeled synthetic-artifact CT images facilitates a self-supervised learning process that positively contributes to model generalizability.</p>","PeriodicalId":20185,"journal":{"name":"Physics in medicine and biology","volume":" ","pages":""},"PeriodicalIF":3.3000,"publicationDate":"2025-03-10","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"Physics in medicine and biology","FirstCategoryId":"5","ListUrlMain":"https://doi.org/10.1088/1361-6560/adbaae","RegionNum":3,"RegionCategory":"医学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"Q2","JCRName":"ENGINEERING, BIOMEDICAL","Score":null,"Total":0}
引用次数: 0

Abstract

Objective. Metal artifacts severely damaged human tissue information from the computed tomography (CT) image, posing significant challenges to disease diagnosis. Deep learning has been widely explored for the metal artifact reduction (MAR) task. Nevertheless, paired metal artifact CT datasets suitable for training do not exist in reality. Although the synthetic CT image dataset provides additional training data, the trained networks still generalize poorly to real metal artifact data.Approach.A self-supervised U-shaped transformer network is proposed to focus on model generalizability enhancement in MAR tasks. This framework consists of a self-supervised mask reconstruction pre-text task and a down-stream task. In the pre-text task, the CT images are randomly corrupted by masks. They are recovered with themselves as the label, aiming at acquiring the artifacts and tissue structure of the actual physical situation. Down-stream task fine-tunes MAR target through labeled images. Utilizing the multi-layer long-range feature extraction capabilities of the Transformer efficiently captures features of metal artifacts. The incorporation of the MAR bottleneck allows for the distinction of metal artifact features through cross-channel self-attention.Main result. Experiments demonstrate that the framework maintains strong generalization ability in the MAR task, effectively preserving tissue details while suppressing metal artifacts. The results achieved a peak signal-to-noise ratio of 43.86 dB and a structural similarity index of 0.9863 while ensuring the efficiency of the model inference. In addition, the Dice coefficient and mean intersection over union are improved by 11.70% and 9.51% in the segmentation of the MAR image, respectively.Significance.The combination of unlabeled real-artifact CT images and labeled synthetic-artifact CT images facilitates a self-supervised learning process that positively contributes to model generalizability.

基于掩模重构的自监督u型变压器网络。
目的:金属伪影严重破坏了CT图像中的人体组织信息,对疾病诊断提出了重大挑战。深度学习(DL)在金属伪影还原(MAR)任务中得到了广泛的探索。然而,现实中并不存在适合训练的配对金属伪影CT数据集。尽管合成CT图像数据集提供了额外的训练数据,但训练后的网络对真实金属伪影数据的泛化能力仍然很差。 ;提出了一种自监督u型变压器网络(SUTransNet),以增强MAR任务中的模型泛化能力。该框架由自监督掩码重构文本前任务和下游任务组成。在pre-text任务中,CT图像被蒙版随机破坏。它们以自身为标签进行复原,目的是获取实际物理状态的人工制品和组织结构。下游任务通过标记图像对MAR目标进行微调。利用Transformer的多层远程特征提取功能,可以有效地捕获金属工件的特征。MAR瓶颈的结合允许通过跨通道自关注来区分金属工件特征。实验表明,该框架在MAR任务中保持了较强的泛化能力,有效地保留了组织细节,同时抑制了金属伪影。在保证模型推理效率的前提下,峰值信噪比(PSNR)为43.86 dB,结构相似性指数(SSIM)为0.9863。此外,在分割MAR图像时,Dice系数和Mean Intersection over Union (MIoU)分别提高了11.70%和9.51% . ;将未标记的真实伪影CT图像与标记的合成伪影CT图像相结合,促进了自监督学习过程,对模型的泛化有积极的贡献。
本文章由计算机程序翻译,如有差异,请以英文原文为准。
求助全文
约1分钟内获得全文 求助全文
来源期刊
Physics in medicine and biology
Physics in medicine and biology 医学-工程:生物医学
CiteScore
6.50
自引率
14.30%
发文量
409
审稿时长
2 months
期刊介绍: The development and application of theoretical, computational and experimental physics to medicine, physiology and biology. Topics covered are: therapy physics (including ionizing and non-ionizing radiation); biomedical imaging (e.g. x-ray, magnetic resonance, ultrasound, optical and nuclear imaging); image-guided interventions; image reconstruction and analysis (including kinetic modelling); artificial intelligence in biomedical physics and analysis; nanoparticles in imaging and therapy; radiobiology; radiation protection and patient dose monitoring; radiation dosimetry
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
确定
请完成安全验证×
copy
已复制链接
快去分享给好友吧!
我知道了
右上角分享
点击右上角分享
0
联系我们:info@booksci.cn Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。 Copyright © 2023 布克学术 All rights reserved.
京ICP备2023020795号-1
ghs 京公网安备 11010802042870号
Book学术文献互助
Book学术文献互助群
群 号:604180095
Book学术官方微信