非线性欠定最小二乘问题的分割预处理方案

IF 1.8 3区 数学 Q1 MATHEMATICS
Nadja Vater, Alfio Borzì
{"title":"非线性欠定最小二乘问题的分割预处理方案","authors":"Nadja Vater, Alfio Borzì","doi":"10.1002/nla.2558","DOIUrl":null,"url":null,"abstract":"The convergence of preconditioned gradient methods for nonlinear underdetermined least squares problems arising in, for example, supervised learning of overparameterized neural networks is investigated. In this general setting, conditions are given that guarantee the existence of global minimizers that correspond to zero residuals and a proof of the convergence of a gradient method to these global minima is presented. In order to accelerate convergence of the gradient method, different preconditioning strategies are developed and analyzed. In particular, a left randomized preconditioner and a right coarse‐level correction preconditioner are combined and investigated. It is demonstrated that the resulting split preconditioned two‐level gradient method incorporates the advantages of both approaches and performs very efficiently.","PeriodicalId":49731,"journal":{"name":"Numerical Linear Algebra with Applications","volume":null,"pages":null},"PeriodicalIF":1.8000,"publicationDate":"2024-04-18","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":"{\"title\":\"A split preconditioning scheme for nonlinear underdetermined least squares problems\",\"authors\":\"Nadja Vater, Alfio Borzì\",\"doi\":\"10.1002/nla.2558\",\"DOIUrl\":null,\"url\":null,\"abstract\":\"The convergence of preconditioned gradient methods for nonlinear underdetermined least squares problems arising in, for example, supervised learning of overparameterized neural networks is investigated. In this general setting, conditions are given that guarantee the existence of global minimizers that correspond to zero residuals and a proof of the convergence of a gradient method to these global minima is presented. In order to accelerate convergence of the gradient method, different preconditioning strategies are developed and analyzed. In particular, a left randomized preconditioner and a right coarse‐level correction preconditioner are combined and investigated. It is demonstrated that the resulting split preconditioned two‐level gradient method incorporates the advantages of both approaches and performs very efficiently.\",\"PeriodicalId\":49731,\"journal\":{\"name\":\"Numerical Linear Algebra with Applications\",\"volume\":null,\"pages\":null},\"PeriodicalIF\":1.8000,\"publicationDate\":\"2024-04-18\",\"publicationTypes\":\"Journal Article\",\"fieldsOfStudy\":null,\"isOpenAccess\":false,\"openAccessPdf\":\"\",\"citationCount\":\"0\",\"resultStr\":null,\"platform\":\"Semanticscholar\",\"paperid\":null,\"PeriodicalName\":\"Numerical Linear Algebra with Applications\",\"FirstCategoryId\":\"100\",\"ListUrlMain\":\"https://doi.org/10.1002/nla.2558\",\"RegionNum\":3,\"RegionCategory\":\"数学\",\"ArticlePicture\":[],\"TitleCN\":null,\"AbstractTextCN\":null,\"PMCID\":null,\"EPubDate\":\"\",\"PubModel\":\"\",\"JCR\":\"Q1\",\"JCRName\":\"MATHEMATICS\",\"Score\":null,\"Total\":0}","platform":"Semanticscholar","paperid":null,"PeriodicalName":"Numerical Linear Algebra with Applications","FirstCategoryId":"100","ListUrlMain":"https://doi.org/10.1002/nla.2558","RegionNum":3,"RegionCategory":"数学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"Q1","JCRName":"MATHEMATICS","Score":null,"Total":0}
引用次数: 0

摘要

本论文研究了在过参数化神经网络的监督学习等过程中出现的非线性欠定最小二乘问题的预条件梯度方法的收敛性。在这种一般情况下,给出了保证与零残差相对应的全局最小值存在的条件,并给出了梯度方法收敛到这些全局最小值的证明。为了加速梯度法的收敛,我们开发并分析了不同的预处理策略。特别是,结合并研究了左侧随机预处理和右侧粗级校正预处理。结果表明,分离式预处理两级梯度法融合了两种方法的优点,并且非常高效。
本文章由计算机程序翻译,如有差异,请以英文原文为准。
A split preconditioning scheme for nonlinear underdetermined least squares problems
The convergence of preconditioned gradient methods for nonlinear underdetermined least squares problems arising in, for example, supervised learning of overparameterized neural networks is investigated. In this general setting, conditions are given that guarantee the existence of global minimizers that correspond to zero residuals and a proof of the convergence of a gradient method to these global minima is presented. In order to accelerate convergence of the gradient method, different preconditioning strategies are developed and analyzed. In particular, a left randomized preconditioner and a right coarse‐level correction preconditioner are combined and investigated. It is demonstrated that the resulting split preconditioned two‐level gradient method incorporates the advantages of both approaches and performs very efficiently.
求助全文
通过发布文献求助,成功后即可免费获取论文全文。 去求助
来源期刊
CiteScore
3.40
自引率
2.30%
发文量
50
审稿时长
12 months
期刊介绍: Manuscripts submitted to Numerical Linear Algebra with Applications should include large-scale broad-interest applications in which challenging computational results are integral to the approach investigated and analysed. Manuscripts that, in the Editor’s view, do not satisfy these conditions will not be accepted for review. Numerical Linear Algebra with Applications receives submissions in areas that address developing, analysing and applying linear algebra algorithms for solving problems arising in multilinear (tensor) algebra, in statistics, such as Markov Chains, as well as in deterministic and stochastic modelling of large-scale networks, algorithm development, performance analysis or related computational aspects. Topics covered include: Standard and Generalized Conjugate Gradients, Multigrid and Other Iterative Methods; Preconditioning Methods; Direct Solution Methods; Numerical Methods for Eigenproblems; Newton-like Methods for Nonlinear Equations; Parallel and Vectorizable Algorithms in Numerical Linear Algebra; Application of Methods of Numerical Linear Algebra in Science, Engineering and Economics.
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
确定
请完成安全验证×
copy
已复制链接
快去分享给好友吧!
我知道了
右上角分享
点击右上角分享
0
联系我们:info@booksci.cn Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。 Copyright © 2023 布克学术 All rights reserved.
京ICP备2023020795号-1
ghs 京公网安备 11010802042870号
Book学术文献互助
Book学术文献互助群
群 号:481959085
Book学术官方微信