On dynamical system modeling of Learned Primal-Dual with a linear operator $\mathcal{K}$: Stability and convergence properties

Jinshu Huang, Yiming Gao, Chunlin Wu
{"title":"On dynamical system modeling of Learned Primal-Dual with a linear operator $\\mathcal{K}$: Stability and convergence properties","authors":"Jinshu Huang, Yiming Gao, Chunlin Wu","doi":"10.1088/1361-6420/ad49ca","DOIUrl":null,"url":null,"abstract":"\n Learned Primal-Dual (LPD) is a deep learning based method for composite optimization problems that is based on unrolling/unfolding the primal-dual hybrid gradient algorithm. While achieving great successes in applications, the mathematical interpretation of LPD as a truncated iterative scheme is not necessarily sufficient to fully understand its properties. In this paper, we study the LPD with a general linear operator. We model the forward propagation of LPD as a system of difference equations and a system of differential equations in discrete- and continuous-time settings (for primal and dual variables/trajectories), which are named discrete-time LPD and continuous-time LPD, respectively. Forward analyses such as stabilities and the convergence of the state variables of the discrete-time LPD to the solution of continuous-time LPD are given. Moreover, we analyze the learning problems with/without regularization terms of both discrete-time and continuous-time LPD from the optimal control viewpoint. We prove convergence results of their optimal solutions with respect to the network state initialization and training data, showing in some sense the topological stability of the learning problems. We also establish convergence from the solution of the discrete-time LPD learning problem to that of the continuous-time LPD learning problem through a piecewise linear extension, under some appropriate assumptions on the space of learnable parameters. This study demonstrates theoretically the robustness of the LPD structure and the associated training process, and can induce some future research and applications.","PeriodicalId":508687,"journal":{"name":"Inverse Problems","volume":" 3","pages":""},"PeriodicalIF":0.0000,"publicationDate":"2024-05-10","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"Inverse Problems","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/10.1088/1361-6420/ad49ca","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
引用次数: 0

Abstract

Learned Primal-Dual (LPD) is a deep learning based method for composite optimization problems that is based on unrolling/unfolding the primal-dual hybrid gradient algorithm. While achieving great successes in applications, the mathematical interpretation of LPD as a truncated iterative scheme is not necessarily sufficient to fully understand its properties. In this paper, we study the LPD with a general linear operator. We model the forward propagation of LPD as a system of difference equations and a system of differential equations in discrete- and continuous-time settings (for primal and dual variables/trajectories), which are named discrete-time LPD and continuous-time LPD, respectively. Forward analyses such as stabilities and the convergence of the state variables of the discrete-time LPD to the solution of continuous-time LPD are given. Moreover, we analyze the learning problems with/without regularization terms of both discrete-time and continuous-time LPD from the optimal control viewpoint. We prove convergence results of their optimal solutions with respect to the network state initialization and training data, showing in some sense the topological stability of the learning problems. We also establish convergence from the solution of the discrete-time LPD learning problem to that of the continuous-time LPD learning problem through a piecewise linear extension, under some appropriate assumptions on the space of learnable parameters. This study demonstrates theoretically the robustness of the LPD structure and the associated training process, and can induce some future research and applications.
关于具有线性算子 $\mathcal{K}$ 的 Learned Primal-Dual 的动力学系统建模:稳定性与收敛性
学习原始双算法(Learned Primal-Dual,LPD)是一种基于深度学习的复合优化问题方法,它以展开/折叠原始双混合梯度算法为基础。虽然在应用中取得了巨大成功,但将 LPD 数学解释为截断迭代方案并不一定足以完全理解其特性。在本文中,我们研究了具有一般线性算子的 LPD。我们将 LPD 的前向传播建模为离散时间和连续时间设置下的差分方程组和微分方程组(对于主变量和对偶变量/轨迹),分别命名为离散时间 LPD 和连续时间 LPD。我们给出了离散时间 LPD 的稳定性和状态变量对连续时间 LPD 解的收敛性等前瞻性分析。此外,我们还从最优控制的角度分析了离散时间和连续时间 LPD 有/无正则化项的学习问题。我们证明了它们的最优解对于网络状态初始化和训练数据的收敛结果,在某种意义上显示了学习问题的拓扑稳定性。我们还通过对可学习参数空间的一些适当假设,建立了离散时间 LPD 学习问题解与连续时间 LPD 学习问题解通过片断线性扩展的收敛性。这项研究从理论上证明了 LPD 结构和相关训练过程的稳健性,并能促进未来的研究和应用。
本文章由计算机程序翻译,如有差异,请以英文原文为准。
求助全文
约1分钟内获得全文 求助全文
来源期刊
自引率
0.00%
发文量
0
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
确定
请完成安全验证×
copy
已复制链接
快去分享给好友吧!
我知道了
右上角分享
点击右上角分享
0
联系我们:info@booksci.cn Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。 Copyright © 2023 布克学术 All rights reserved.
京ICP备2023020795号-1
ghs 京公网安备 11010802042870号
Book学术文献互助
Book学术文献互助群
群 号:481959085
Book学术官方微信