Convergence Acceleration of Markov Chain Monte Carlo-Based Gradient Descent by Deep Unfolding

IF 1.5 4区 物理与天体物理 Q2 PHYSICS, MULTIDISCIPLINARY
Ryo Hagiwara, Satoshi Takabe
{"title":"Convergence Acceleration of Markov Chain Monte Carlo-Based Gradient Descent by Deep Unfolding","authors":"Ryo Hagiwara, Satoshi Takabe","doi":"10.7566/jpsj.93.063801","DOIUrl":null,"url":null,"abstract":"This study proposes a trainable sampling-based solver for combinatorial optimization problems (COPs) using a deep-learning technique called deep unfolding. The proposed solver is based on the Ohzeki method that combines Markov-chain Monte-Carlo (MCMC) and gradient descent, and its step sizes are trained by minimizing a loss function. In the training process, we propose a sampling-based gradient estimation that substitutes auto-differentiation with a variance estimation, thereby circumventing the failure of back propagation due to the non-differentiability of MCMC. The numerical results for a few COPs demonstrated that the proposed solver significantly accelerated the convergence speed compared with the original Ohzeki method.","PeriodicalId":17304,"journal":{"name":"Journal of the Physical Society of Japan","volume":"161 1","pages":""},"PeriodicalIF":1.5000,"publicationDate":"2024-05-02","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"Journal of the Physical Society of Japan","FirstCategoryId":"101","ListUrlMain":"https://doi.org/10.7566/jpsj.93.063801","RegionNum":4,"RegionCategory":"物理与天体物理","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"Q2","JCRName":"PHYSICS, MULTIDISCIPLINARY","Score":null,"Total":0}
引用次数: 0

Abstract

This study proposes a trainable sampling-based solver for combinatorial optimization problems (COPs) using a deep-learning technique called deep unfolding. The proposed solver is based on the Ohzeki method that combines Markov-chain Monte-Carlo (MCMC) and gradient descent, and its step sizes are trained by minimizing a loss function. In the training process, we propose a sampling-based gradient estimation that substitutes auto-differentiation with a variance estimation, thereby circumventing the failure of back propagation due to the non-differentiability of MCMC. The numerical results for a few COPs demonstrated that the proposed solver significantly accelerated the convergence speed compared with the original Ohzeki method.
通过深度展开加速基于马尔可夫链蒙特卡洛的梯度下降收敛
本研究利用一种称为深度展开的深度学习技术,为组合优化问题(COPs)提出了一种基于采样的可训练求解器。所提出的求解器基于 Ohzeki 方法,该方法结合了马尔可夫链蒙特卡洛(MCMC)和梯度下降,其步长通过最小化损失函数进行训练。在训练过程中,我们提出了一种基于采样的梯度估计方法,用方差估计代替自动差分,从而避免了因 MCMC 的不可差分性而导致的反向传播失败。对几个 COP 的数值结果表明,与最初的 Ohzeki 方法相比,所提出的求解器明显加快了收敛速度。
本文章由计算机程序翻译,如有差异,请以英文原文为准。
求助全文
约1分钟内获得全文 求助全文
来源期刊
CiteScore
3.40
自引率
17.60%
发文量
325
审稿时长
3 months
期刊介绍: The papers published in JPSJ should treat fundamental and novel problems of physics scientifically and logically, and contribute to the development in the understanding of physics. The concrete objects are listed below. Subjects Covered JPSJ covers all the fields of physics including (but not restricted to) Elementary particles and fields Nuclear physics Atomic and Molecular Physics Fluid Dynamics Plasma physics Physics of Condensed Matter Metal, Superconductor, Semiconductor, Magnetic Materials, Dielectric Materials Physics of Nanoscale Materials Optics and Quantum Electronics Physics of Complex Systems Mathematical Physics Chemical physics Biophysics Geophysics Astrophysics.
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
确定
请完成安全验证×
copy
已复制链接
快去分享给好友吧!
我知道了
右上角分享
点击右上角分享
0
联系我们:info@booksci.cn Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。 Copyright © 2023 布克学术 All rights reserved.
京ICP备2023020795号-1
ghs 京公网安备 11010802042870号
Book学术文献互助
Book学术文献互助群
群 号:481959085
Book学术官方微信