鲁棒随机拟牛顿方法及其在机器学习中的应用

Lin Wu, Tuo Shi, Yongbin Wang, Huiming Chen, W. Lam
{"title":"鲁棒随机拟牛顿方法及其在机器学习中的应用","authors":"Lin Wu, Tuo Shi, Yongbin Wang, Huiming Chen, W. Lam","doi":"10.1109/ICCST53801.2021.00041","DOIUrl":null,"url":null,"abstract":"We study a novel stochastic version of damped and regularized limited memory Broyden-Fletcher-Goldfarb-Shanno (Sd-REG-LBFGS) method for nonconvex optimization problems in this paper. Different from BFGS updating scheme for the strongly convex problems, it is a challenge to preserve the product of the correction pairs positive at each iteration for BFGS in nonconvex case. While utilizing regularization scheme to ensure the robustness of the second-order methods. The proposed method is able to keep the product of correction pairs being away from zero above a specified degree. To make the proposed method robust and computationally efficient, we propose to update curvature information at a spaced interval, in which the average of iterate points is utilized. The numerical experiments have shown that our proposed algorithm has better performance than SdLBFGS or they have almost the same performance. Especially in the problems which utilizes few samples, the proposed method has avoided the ill-conditioned problem and exhibited superior performance.","PeriodicalId":222463,"journal":{"name":"2021 International Conference on Culture-oriented Science & Technology (ICCST)","volume":null,"pages":null},"PeriodicalIF":0.0000,"publicationDate":"2021-11-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":"{\"title\":\"A Robust Stochastic Quasi-Newton Method with the Application in Machine Learning\",\"authors\":\"Lin Wu, Tuo Shi, Yongbin Wang, Huiming Chen, W. Lam\",\"doi\":\"10.1109/ICCST53801.2021.00041\",\"DOIUrl\":null,\"url\":null,\"abstract\":\"We study a novel stochastic version of damped and regularized limited memory Broyden-Fletcher-Goldfarb-Shanno (Sd-REG-LBFGS) method for nonconvex optimization problems in this paper. Different from BFGS updating scheme for the strongly convex problems, it is a challenge to preserve the product of the correction pairs positive at each iteration for BFGS in nonconvex case. While utilizing regularization scheme to ensure the robustness of the second-order methods. The proposed method is able to keep the product of correction pairs being away from zero above a specified degree. To make the proposed method robust and computationally efficient, we propose to update curvature information at a spaced interval, in which the average of iterate points is utilized. The numerical experiments have shown that our proposed algorithm has better performance than SdLBFGS or they have almost the same performance. Especially in the problems which utilizes few samples, the proposed method has avoided the ill-conditioned problem and exhibited superior performance.\",\"PeriodicalId\":222463,\"journal\":{\"name\":\"2021 International Conference on Culture-oriented Science & Technology (ICCST)\",\"volume\":null,\"pages\":null},\"PeriodicalIF\":0.0000,\"publicationDate\":\"2021-11-01\",\"publicationTypes\":\"Journal Article\",\"fieldsOfStudy\":null,\"isOpenAccess\":false,\"openAccessPdf\":\"\",\"citationCount\":\"0\",\"resultStr\":null,\"platform\":\"Semanticscholar\",\"paperid\":null,\"PeriodicalName\":\"2021 International Conference on Culture-oriented Science & Technology (ICCST)\",\"FirstCategoryId\":\"1085\",\"ListUrlMain\":\"https://doi.org/10.1109/ICCST53801.2021.00041\",\"RegionNum\":0,\"RegionCategory\":null,\"ArticlePicture\":[],\"TitleCN\":null,\"AbstractTextCN\":null,\"PMCID\":null,\"EPubDate\":\"\",\"PubModel\":\"\",\"JCR\":\"\",\"JCRName\":\"\",\"Score\":null,\"Total\":0}","platform":"Semanticscholar","paperid":null,"PeriodicalName":"2021 International Conference on Culture-oriented Science & Technology (ICCST)","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/10.1109/ICCST53801.2021.00041","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
引用次数: 0

摘要

本文研究了一种新的随机版本的阻尼正则化有限记忆Broyden-Fletcher-Goldfarb-Shanno (Sd-REG-LBFGS)方法求解非凸优化问题。与强凸问题的BFGS更新方案不同,在非凸情况下,BFGS在每次迭代时保持校正对积为正是一个挑战。同时利用正则化方案保证二阶方法的鲁棒性。该方法能够在一定程度上保持校正对的乘积远离零。为了提高方法的鲁棒性和计算效率,我们提出以间隔更新曲率信息,其中利用迭代点的平均值。数值实验表明,本文提出的算法比SdLBFGS算法性能更好或性能基本相同。特别是在使用少量样本的问题中,该方法避免了病态问题,表现出优异的性能。
本文章由计算机程序翻译,如有差异,请以英文原文为准。
A Robust Stochastic Quasi-Newton Method with the Application in Machine Learning
We study a novel stochastic version of damped and regularized limited memory Broyden-Fletcher-Goldfarb-Shanno (Sd-REG-LBFGS) method for nonconvex optimization problems in this paper. Different from BFGS updating scheme for the strongly convex problems, it is a challenge to preserve the product of the correction pairs positive at each iteration for BFGS in nonconvex case. While utilizing regularization scheme to ensure the robustness of the second-order methods. The proposed method is able to keep the product of correction pairs being away from zero above a specified degree. To make the proposed method robust and computationally efficient, we propose to update curvature information at a spaced interval, in which the average of iterate points is utilized. The numerical experiments have shown that our proposed algorithm has better performance than SdLBFGS or they have almost the same performance. Especially in the problems which utilizes few samples, the proposed method has avoided the ill-conditioned problem and exhibited superior performance.
求助全文
通过发布文献求助,成功后即可免费获取论文全文。 去求助
来源期刊
自引率
0.00%
发文量
0
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
确定
请完成安全验证×
copy
已复制链接
快去分享给好友吧!
我知道了
右上角分享
点击右上角分享
0
联系我们:info@booksci.cn Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。 Copyright © 2023 布克学术 All rights reserved.
京ICP备2023020795号-1
ghs 京公网安备 11010802042870号
Book学术文献互助
Book学术文献互助群
群 号:481959085
Book学术官方微信