Mud-Net:用于在计算机断层扫描中同时减少稀疏视图和金属伪影的多域深度展开网络

Baoshun Shi, Ke Jiang, Shaolei Zhang, Qiusheng Lian, Yanwei Qin, Yunsong Zhao
{"title":"Mud-Net:用于在计算机断层扫描中同时减少稀疏视图和金属伪影的多域深度展开网络","authors":"Baoshun Shi, Ke Jiang, Shaolei Zhang, Qiusheng Lian, Yanwei Qin, Yunsong Zhao","doi":"10.1088/2632-2153/ad1b8e","DOIUrl":null,"url":null,"abstract":"\n Sparse-view computed tomography (SVCT) is regarded as a promising technique to accelerate data acquisition and reduce radiation dose. However, in the presence of metallic implants, SVCT inevitably makes the reconstructed CT images suffer from severe metal artifacts and streaking artifacts due to the lack of sufficient projection data. Previous stand-alone SVCT and metal artifact reduction (MAR) methods to solve the problem of simultaneously sparse-view and metal artifact reduction (SVMAR) are plagued by insufficient correction accuracy. To overcome this limitation, we propose a multi-domain deep unrolling network, called Mud-Net, for SVMAR. Specifically, we establish a joint sinogram, image, artifact, and coding domains deep unrolling reconstruction model to recover high-quality CT images from the under-sampled sinograms corrupted by metallic implants. To train this multi-domain network effectively, we embed multi-domain knowledge into the network training process. Comprehensive experiments demonstrate that our method is superior to both existing MAR methods in the full-view MAR task and previous SVCT methods in the SVMAR task.","PeriodicalId":503691,"journal":{"name":"Machine Learning: Science and Technology","volume":"14 4","pages":""},"PeriodicalIF":0.0000,"publicationDate":"2024-01-05","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":"{\"title\":\"Mud-Net: Multi-domain deep unrolling network for simultaneous sparse-view and metal artifact reduction in computed tomography\",\"authors\":\"Baoshun Shi, Ke Jiang, Shaolei Zhang, Qiusheng Lian, Yanwei Qin, Yunsong Zhao\",\"doi\":\"10.1088/2632-2153/ad1b8e\",\"DOIUrl\":null,\"url\":null,\"abstract\":\"\\n Sparse-view computed tomography (SVCT) is regarded as a promising technique to accelerate data acquisition and reduce radiation dose. However, in the presence of metallic implants, SVCT inevitably makes the reconstructed CT images suffer from severe metal artifacts and streaking artifacts due to the lack of sufficient projection data. Previous stand-alone SVCT and metal artifact reduction (MAR) methods to solve the problem of simultaneously sparse-view and metal artifact reduction (SVMAR) are plagued by insufficient correction accuracy. To overcome this limitation, we propose a multi-domain deep unrolling network, called Mud-Net, for SVMAR. Specifically, we establish a joint sinogram, image, artifact, and coding domains deep unrolling reconstruction model to recover high-quality CT images from the under-sampled sinograms corrupted by metallic implants. To train this multi-domain network effectively, we embed multi-domain knowledge into the network training process. Comprehensive experiments demonstrate that our method is superior to both existing MAR methods in the full-view MAR task and previous SVCT methods in the SVMAR task.\",\"PeriodicalId\":503691,\"journal\":{\"name\":\"Machine Learning: Science and Technology\",\"volume\":\"14 4\",\"pages\":\"\"},\"PeriodicalIF\":0.0000,\"publicationDate\":\"2024-01-05\",\"publicationTypes\":\"Journal Article\",\"fieldsOfStudy\":null,\"isOpenAccess\":false,\"openAccessPdf\":\"\",\"citationCount\":\"0\",\"resultStr\":null,\"platform\":\"Semanticscholar\",\"paperid\":null,\"PeriodicalName\":\"Machine Learning: Science and Technology\",\"FirstCategoryId\":\"1085\",\"ListUrlMain\":\"https://doi.org/10.1088/2632-2153/ad1b8e\",\"RegionNum\":0,\"RegionCategory\":null,\"ArticlePicture\":[],\"TitleCN\":null,\"AbstractTextCN\":null,\"PMCID\":null,\"EPubDate\":\"\",\"PubModel\":\"\",\"JCR\":\"\",\"JCRName\":\"\",\"Score\":null,\"Total\":0}","platform":"Semanticscholar","paperid":null,"PeriodicalName":"Machine Learning: Science and Technology","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/10.1088/2632-2153/ad1b8e","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
引用次数: 0

摘要

稀疏视图计算机断层扫描(SVCT)被认为是一种很有前途的技术,可加快数据采集速度并减少辐射剂量。然而,在存在金属植入物的情况下,由于缺乏足够的投影数据,SVCT 不可避免地会使重建的 CT 图像出现严重的金属伪影和条纹伪影。以往独立的 SVCT 和金属伪影还原(MAR)方法在解决同时还原稀疏视图和金属伪影(SVMAR)的问题时,受到校正精度不足的困扰。为了克服这一局限性,我们提出了一种用于 SVMAR 的多域深度展开网络,称为 Mud-Net。具体来说,我们建立了一个联合窦状图、图像、伪影和编码域的深度展开重建模型,以从被金属植入物破坏的低采样窦状图中恢复高质量的 CT 图像。为了有效地训练这个多域网络,我们在网络训练过程中嵌入了多域知识。综合实验证明,我们的方法在全视图 MAR 任务中优于现有的 MAR 方法,在 SVMAR 任务中优于之前的 SVCT 方法。
本文章由计算机程序翻译,如有差异,请以英文原文为准。
Mud-Net: Multi-domain deep unrolling network for simultaneous sparse-view and metal artifact reduction in computed tomography
Sparse-view computed tomography (SVCT) is regarded as a promising technique to accelerate data acquisition and reduce radiation dose. However, in the presence of metallic implants, SVCT inevitably makes the reconstructed CT images suffer from severe metal artifacts and streaking artifacts due to the lack of sufficient projection data. Previous stand-alone SVCT and metal artifact reduction (MAR) methods to solve the problem of simultaneously sparse-view and metal artifact reduction (SVMAR) are plagued by insufficient correction accuracy. To overcome this limitation, we propose a multi-domain deep unrolling network, called Mud-Net, for SVMAR. Specifically, we establish a joint sinogram, image, artifact, and coding domains deep unrolling reconstruction model to recover high-quality CT images from the under-sampled sinograms corrupted by metallic implants. To train this multi-domain network effectively, we embed multi-domain knowledge into the network training process. Comprehensive experiments demonstrate that our method is superior to both existing MAR methods in the full-view MAR task and previous SVCT methods in the SVMAR task.
求助全文
通过发布文献求助,成功后即可免费获取论文全文。 去求助
来源期刊
自引率
0.00%
发文量
0
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
确定
请完成安全验证×
copy
已复制链接
快去分享给好友吧!
我知道了
右上角分享
点击右上角分享
0
联系我们:info@booksci.cn Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。 Copyright © 2023 布克学术 All rights reserved.
京ICP备2023020795号-1
ghs 京公网安备 11010802042870号
Book学术文献互助
Book学术文献互助群
群 号:481959085
Book学术官方微信