Deep unfolding as iterative regularization for imaging inverse problems

IF 2 2区 数学 Q1 MATHEMATICS, APPLIED
Zhuoxu Cui, Qingyong Zhu, Jing Cheng, Bo Zhang, Dong Liang
{"title":"Deep unfolding as iterative regularization for imaging inverse problems","authors":"Zhuoxu Cui, Qingyong Zhu, Jing Cheng, Bo Zhang, Dong Liang","doi":"10.1088/1361-6420/ad1a3c","DOIUrl":null,"url":null,"abstract":"\n Deep unfolding methods have gained significant popularity in the field of inverse problems as they have driven the design of deep neural networks (DNNs) using iterative algorithms. In contrast to general DNNs, unfolding methods offer improved interpretability and performance. However, their theoretical stability or regularity in solving inverse problems remains subject to certain limitations. To address this, we reevaluate unfolded DNNs and observe that their algorithmically-driven cascading structure exhibits a closer resemblance to iterative regularization. Recognizing this, we propose a modified training approach and configure termination criteria for unfolded DNNs, thereby establishing the unfolding method as an iterative regularization technique. Specifically, our method involves the joint learning of a convex penalty function using an input-convex neural network (ICNN) to quantify distance to a real data manifold. Then, we train a DNN unfolded from the proximal gradient descent algorithm, incorporating this learned penalty. Additionally, we introduce a new termination criterion for the unfolded DNN. Under the assumption that the real data manifold intersects the solutions of the inverse problem with a unique real solution, even when measurements contain perturbations, we provide a theoretical proof of the stable convergence of the unfolded DNN to this solution. Furthermore, we demonstrate with an example of MRI reconstruction that the proposed method outperforms original unfolding methods and traditional regularization methods in terms of reconstruction quality, stability, and convergence speed.","PeriodicalId":50275,"journal":{"name":"Inverse Problems","volume":"14 6","pages":""},"PeriodicalIF":2.0000,"publicationDate":"2024-01-03","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"Inverse Problems","FirstCategoryId":"100","ListUrlMain":"https://doi.org/10.1088/1361-6420/ad1a3c","RegionNum":2,"RegionCategory":"数学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"Q1","JCRName":"MATHEMATICS, APPLIED","Score":null,"Total":0}
引用次数: 0

Abstract

Deep unfolding methods have gained significant popularity in the field of inverse problems as they have driven the design of deep neural networks (DNNs) using iterative algorithms. In contrast to general DNNs, unfolding methods offer improved interpretability and performance. However, their theoretical stability or regularity in solving inverse problems remains subject to certain limitations. To address this, we reevaluate unfolded DNNs and observe that their algorithmically-driven cascading structure exhibits a closer resemblance to iterative regularization. Recognizing this, we propose a modified training approach and configure termination criteria for unfolded DNNs, thereby establishing the unfolding method as an iterative regularization technique. Specifically, our method involves the joint learning of a convex penalty function using an input-convex neural network (ICNN) to quantify distance to a real data manifold. Then, we train a DNN unfolded from the proximal gradient descent algorithm, incorporating this learned penalty. Additionally, we introduce a new termination criterion for the unfolded DNN. Under the assumption that the real data manifold intersects the solutions of the inverse problem with a unique real solution, even when measurements contain perturbations, we provide a theoretical proof of the stable convergence of the unfolded DNN to this solution. Furthermore, we demonstrate with an example of MRI reconstruction that the proposed method outperforms original unfolding methods and traditional regularization methods in terms of reconstruction quality, stability, and convergence speed.
深度展开作为成像逆问题的迭代正则化
深度展开方法在逆问题领域大受欢迎,因为它们推动了使用迭代算法的深度神经网络(DNN)的设计。与一般的 DNNs 相比,展开方法具有更好的可解释性和性能。然而,它们在解决逆问题时的理论稳定性或规律性仍然受到某些限制。为了解决这个问题,我们重新评估了展开 DNN,发现其算法驱动的级联结构与迭代正则化更为相似。认识到这一点后,我们提出了一种改进的训练方法,并为展开 DNN 配置了终止标准,从而将展开方法确立为一种迭代正则化技术。具体来说,我们的方法涉及使用输入-凸神经网络(ICNN)联合学习凸惩罚函数,以量化与真实数据流形的距离。然后,我们通过近似梯度下降算法,结合学习到的惩罚函数,训练一个展开的 DNN。此外,我们还为展开的 DNN 引入了一个新的终止准则。假设真实数据流形与逆问题的解相交,即使测量包含扰动,也有唯一的真实解,我们从理论上证明了展开 DNN 对该解的稳定收敛。此外,我们还以核磁共振成像重建为例证明,所提出的方法在重建质量、稳定性和收敛速度方面都优于原始的展开方法和传统的正则化方法。
本文章由计算机程序翻译,如有差异,请以英文原文为准。
求助全文
约1分钟内获得全文 求助全文
来源期刊
Inverse Problems
Inverse Problems 数学-物理:数学物理
CiteScore
4.40
自引率
14.30%
发文量
115
审稿时长
2.3 months
期刊介绍: An interdisciplinary journal combining mathematical and experimental papers on inverse problems with theoretical, numerical and practical approaches to their solution. As well as applied mathematicians, physical scientists and engineers, the readership includes those working in geophysics, radar, optics, biology, acoustics, communication theory, signal processing and imaging, among others. The emphasis is on publishing original contributions to methods of solving mathematical, physical and applied problems. To be publishable in this journal, papers must meet the highest standards of scientific quality, contain significant and original new science and should present substantial advancement in the field. Due to the broad scope of the journal, we require that authors provide sufficient introductory material to appeal to the wide readership and that articles which are not explicitly applied include a discussion of possible applications.
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
确定
请完成安全验证×
copy
已复制链接
快去分享给好友吧!
我知道了
右上角分享
点击右上角分享
0
联系我们:info@booksci.cn Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。 Copyright © 2023 布克学术 All rights reserved.
京ICP备2023020795号-1
ghs 京公网安备 11010802042870号
Book学术文献互助
Book学术文献互助群
群 号:481959085
Book学术官方微信