Matthias J Ehrhardt, Erlend S Riis, Torbjørn Ringholm, Carola-Bibiane Schönlieb
{"title":"平滑优化的几何整合方法:离散梯度法的基础","authors":"Matthias J Ehrhardt, Erlend S Riis, Torbjørn Ringholm, Carola-Bibiane Schönlieb","doi":"10.1093/imanum/drae037","DOIUrl":null,"url":null,"abstract":"Discrete gradient methods are geometric integration techniques that can preserve the dissipative structure of gradient flows. Due to the monotonic decay of the function values, they are well suited for general convex and nonconvex optimization problems. Both zero- and first-order algorithms can be derived from the discrete gradient method by selecting different discrete gradients. In this paper, we present a thorough analysis of the discrete gradient method for optimization that provides a solid theoretical foundation. We show that the discrete gradient method is well-posed by proving the existence of iterates for any positive time step, as well as uniqueness in some cases, and propose an efficient method for solving the associated discrete gradient equation. Moreover, we establish an $\\text{O}(1/k)$ convergence rate for convex objectives and prove linear convergence if instead the Polyak–Łojasiewicz inequality is satisfied. The analysis is carried out for three discrete gradients—the Gonzalez discrete gradient, the mean value discrete gradient, and the Itoh–Abe discrete gradient—as well as for a randomised Itoh–Abe method. Our theoretical results are illustrated with a variety of numerical experiments, and we furthermore demonstrate that the methods are robust with respect to stiffness.","PeriodicalId":56295,"journal":{"name":"IMA Journal of Numerical Analysis","volume":null,"pages":null},"PeriodicalIF":2.3000,"publicationDate":"2024-07-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":"{\"title\":\"A geometric integration approach to smooth optimization: foundations of the discrete gradient method\",\"authors\":\"Matthias J Ehrhardt, Erlend S Riis, Torbjørn Ringholm, Carola-Bibiane Schönlieb\",\"doi\":\"10.1093/imanum/drae037\",\"DOIUrl\":null,\"url\":null,\"abstract\":\"Discrete gradient methods are geometric integration techniques that can preserve the dissipative structure of gradient flows. Due to the monotonic decay of the function values, they are well suited for general convex and nonconvex optimization problems. Both zero- and first-order algorithms can be derived from the discrete gradient method by selecting different discrete gradients. In this paper, we present a thorough analysis of the discrete gradient method for optimization that provides a solid theoretical foundation. We show that the discrete gradient method is well-posed by proving the existence of iterates for any positive time step, as well as uniqueness in some cases, and propose an efficient method for solving the associated discrete gradient equation. Moreover, we establish an $\\\\text{O}(1/k)$ convergence rate for convex objectives and prove linear convergence if instead the Polyak–Łojasiewicz inequality is satisfied. The analysis is carried out for three discrete gradients—the Gonzalez discrete gradient, the mean value discrete gradient, and the Itoh–Abe discrete gradient—as well as for a randomised Itoh–Abe method. Our theoretical results are illustrated with a variety of numerical experiments, and we furthermore demonstrate that the methods are robust with respect to stiffness.\",\"PeriodicalId\":56295,\"journal\":{\"name\":\"IMA Journal of Numerical Analysis\",\"volume\":null,\"pages\":null},\"PeriodicalIF\":2.3000,\"publicationDate\":\"2024-07-01\",\"publicationTypes\":\"Journal Article\",\"fieldsOfStudy\":null,\"isOpenAccess\":false,\"openAccessPdf\":\"\",\"citationCount\":\"0\",\"resultStr\":null,\"platform\":\"Semanticscholar\",\"paperid\":null,\"PeriodicalName\":\"IMA Journal of Numerical Analysis\",\"FirstCategoryId\":\"100\",\"ListUrlMain\":\"https://doi.org/10.1093/imanum/drae037\",\"RegionNum\":2,\"RegionCategory\":\"数学\",\"ArticlePicture\":[],\"TitleCN\":null,\"AbstractTextCN\":null,\"PMCID\":null,\"EPubDate\":\"\",\"PubModel\":\"\",\"JCR\":\"Q1\",\"JCRName\":\"MATHEMATICS, APPLIED\",\"Score\":null,\"Total\":0}","platform":"Semanticscholar","paperid":null,"PeriodicalName":"IMA Journal of Numerical Analysis","FirstCategoryId":"100","ListUrlMain":"https://doi.org/10.1093/imanum/drae037","RegionNum":2,"RegionCategory":"数学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"Q1","JCRName":"MATHEMATICS, APPLIED","Score":null,"Total":0}
A geometric integration approach to smooth optimization: foundations of the discrete gradient method
Discrete gradient methods are geometric integration techniques that can preserve the dissipative structure of gradient flows. Due to the monotonic decay of the function values, they are well suited for general convex and nonconvex optimization problems. Both zero- and first-order algorithms can be derived from the discrete gradient method by selecting different discrete gradients. In this paper, we present a thorough analysis of the discrete gradient method for optimization that provides a solid theoretical foundation. We show that the discrete gradient method is well-posed by proving the existence of iterates for any positive time step, as well as uniqueness in some cases, and propose an efficient method for solving the associated discrete gradient equation. Moreover, we establish an $\text{O}(1/k)$ convergence rate for convex objectives and prove linear convergence if instead the Polyak–Łojasiewicz inequality is satisfied. The analysis is carried out for three discrete gradients—the Gonzalez discrete gradient, the mean value discrete gradient, and the Itoh–Abe discrete gradient—as well as for a randomised Itoh–Abe method. Our theoretical results are illustrated with a variety of numerical experiments, and we furthermore demonstrate that the methods are robust with respect to stiffness.
期刊介绍:
The IMA Journal of Numerical Analysis (IMAJNA) publishes original contributions to all fields of numerical analysis; articles will be accepted which treat the theory, development or use of practical algorithms and interactions between these aspects. Occasional survey articles are also published.