一种加速的直流规划双近邻梯度算法

IF 1.1 4区 管理学 Q4 OPERATIONS RESEARCH & MANAGEMENT SCIENCE
Gaoxi Li, Ying Yi, Yingquan Huang
{"title":"一种加速的直流规划双近邻梯度算法","authors":"Gaoxi Li, Ying Yi, Yingquan Huang","doi":"10.1142/s0217595923500288","DOIUrl":null,"url":null,"abstract":"The double-proximal gradient algorithm (DPGA) is a new variant of the classical difference-of-convex algorithm (DCA) for solving difference-of-convex (DC) optimization problems. In this paper, we propose an accelerated version of the double-proximal gradient algorithm for DC programming, in which the objective function consists of three convex modules (only one module is smooth). We establish convergence of the sequence generated by our algorithm if the objective function satisfies the Kurdyka–[Formula: see text]ojasiewicz (K[Formula: see text]) property and show that its convergence rate is not weaker than DPGA. Compared with DPGA, the numerical experiments on an image processing model show that the number of iterations of ADPGA is reduced by 43.57% and the running time is reduced by 43.47% on average.","PeriodicalId":55455,"journal":{"name":"Asia-Pacific Journal of Operational Research","volume":null,"pages":null},"PeriodicalIF":1.1000,"publicationDate":"2023-11-04","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":"{\"title\":\"An accelerated double-proximal gradient algorithm for DC programming\",\"authors\":\"Gaoxi Li, Ying Yi, Yingquan Huang\",\"doi\":\"10.1142/s0217595923500288\",\"DOIUrl\":null,\"url\":null,\"abstract\":\"The double-proximal gradient algorithm (DPGA) is a new variant of the classical difference-of-convex algorithm (DCA) for solving difference-of-convex (DC) optimization problems. In this paper, we propose an accelerated version of the double-proximal gradient algorithm for DC programming, in which the objective function consists of three convex modules (only one module is smooth). We establish convergence of the sequence generated by our algorithm if the objective function satisfies the Kurdyka–[Formula: see text]ojasiewicz (K[Formula: see text]) property and show that its convergence rate is not weaker than DPGA. Compared with DPGA, the numerical experiments on an image processing model show that the number of iterations of ADPGA is reduced by 43.57% and the running time is reduced by 43.47% on average.\",\"PeriodicalId\":55455,\"journal\":{\"name\":\"Asia-Pacific Journal of Operational Research\",\"volume\":null,\"pages\":null},\"PeriodicalIF\":1.1000,\"publicationDate\":\"2023-11-04\",\"publicationTypes\":\"Journal Article\",\"fieldsOfStudy\":null,\"isOpenAccess\":false,\"openAccessPdf\":\"\",\"citationCount\":\"0\",\"resultStr\":null,\"platform\":\"Semanticscholar\",\"paperid\":null,\"PeriodicalName\":\"Asia-Pacific Journal of Operational Research\",\"FirstCategoryId\":\"1085\",\"ListUrlMain\":\"https://doi.org/10.1142/s0217595923500288\",\"RegionNum\":4,\"RegionCategory\":\"管理学\",\"ArticlePicture\":[],\"TitleCN\":null,\"AbstractTextCN\":null,\"PMCID\":null,\"EPubDate\":\"\",\"PubModel\":\"\",\"JCR\":\"Q4\",\"JCRName\":\"OPERATIONS RESEARCH & MANAGEMENT SCIENCE\",\"Score\":null,\"Total\":0}","platform":"Semanticscholar","paperid":null,"PeriodicalName":"Asia-Pacific Journal of Operational Research","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/10.1142/s0217595923500288","RegionNum":4,"RegionCategory":"管理学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"Q4","JCRName":"OPERATIONS RESEARCH & MANAGEMENT SCIENCE","Score":null,"Total":0}
引用次数: 0

摘要

双近端梯度算法(DPGA)是对经典凸差分算法(DCA)的改进,用于求解凸差分优化问题。在本文中,我们提出了一种加速版本的双邻点梯度算法用于DC规划,其中目标函数由三个凸模块组成(其中只有一个模块是光滑的)。在目标函数满足Kurdyka -[公式:见文]ojasiewicz (K[公式:见文])性质的条件下,建立了算法生成序列的收敛性,并证明其收敛速度不弱于DPGA。在一个图像处理模型上进行的数值实验表明,与DPGA相比,ADPGA的迭代次数平均减少43.57%,运行时间平均减少43.47%。
本文章由计算机程序翻译,如有差异,请以英文原文为准。
An accelerated double-proximal gradient algorithm for DC programming
The double-proximal gradient algorithm (DPGA) is a new variant of the classical difference-of-convex algorithm (DCA) for solving difference-of-convex (DC) optimization problems. In this paper, we propose an accelerated version of the double-proximal gradient algorithm for DC programming, in which the objective function consists of three convex modules (only one module is smooth). We establish convergence of the sequence generated by our algorithm if the objective function satisfies the Kurdyka–[Formula: see text]ojasiewicz (K[Formula: see text]) property and show that its convergence rate is not weaker than DPGA. Compared with DPGA, the numerical experiments on an image processing model show that the number of iterations of ADPGA is reduced by 43.57% and the running time is reduced by 43.47% on average.
求助全文
通过发布文献求助,成功后即可免费获取论文全文。 去求助
来源期刊
Asia-Pacific Journal of Operational Research
Asia-Pacific Journal of Operational Research 管理科学-运筹学与管理科学
CiteScore
2.00
自引率
14.30%
发文量
44
审稿时长
14.2 months
期刊介绍: The Asia-Pacific Journal of Operational Research (APJOR) provides a forum for practitioners, academics and researchers in Operational Research and related fields, within and beyond the Asia-Pacific region. APJOR will place submissions in one of the following categories: General, Theoretical, OR Practice, Reviewer Survey, OR Education, and Communications (including short articles and letters). Theoretical papers should carry significant methodological developments. Emphasis is on originality, quality and importance, with particular emphasis given to the practical significance of the results. Practical papers, illustrating the application of Operation Research, are of special interest.
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
确定
请完成安全验证×
copy
已复制链接
快去分享给好友吧!
我知道了
右上角分享
点击右上角分享
0
联系我们:info@booksci.cn Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。 Copyright © 2023 布克学术 All rights reserved.
京ICP备2023020795号-1
ghs 京公网安备 11010802042870号
Book学术文献互助
Book学术文献互助群
群 号:481959085
Book学术官方微信