Sgd的学习率和动量系数及其交互作用对神经网络性能影响的统计检验

Bingchuan Chen, Ai-Xiang Chen, Xiaolong Chai, Rui Bian
{"title":"Sgd的学习率和动量系数及其交互作用对神经网络性能影响的统计检验","authors":"Bingchuan Chen, Ai-Xiang Chen, Xiaolong Chai, Rui Bian","doi":"10.1109/ICDSBA48748.2019.00065","DOIUrl":null,"url":null,"abstract":"Stochastic Gradient Descent (SGD) is a well-received algorithm for large-scale optimization in neural networks for its low iteration cost. However, due to Gradient variance, it often has difficulty in finding optimal learning rate and thus suffers from slow convergence. Using momentum is proven to be a simple effective way of overcoming the slow convergence problem of SDG as long as momentum is properly set. According to the performance metrics, this paper proposes a novel statistical model for analyzing the performance of neural networks. The model takes into account learning rate and momentum, and the method can be used to evaluate and verify their interaction effects on neural network performance. Our study shows that the interaction effects are significant. When momentum has a value smaller than 0.5, the impact on the training time is not statistically noticeable.","PeriodicalId":382429,"journal":{"name":"2019 3rd International Conference on Data Science and Business Analytics (ICDSBA)","volume":"8 1","pages":"0"},"PeriodicalIF":0.0000,"publicationDate":"2019-10-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"3","resultStr":"{\"title\":\"A Statistical Test of The Effect of Learning Rate and Momentum Coefficient of Sgd and Its Interaction on Neural Network Performance\",\"authors\":\"Bingchuan Chen, Ai-Xiang Chen, Xiaolong Chai, Rui Bian\",\"doi\":\"10.1109/ICDSBA48748.2019.00065\",\"DOIUrl\":null,\"url\":null,\"abstract\":\"Stochastic Gradient Descent (SGD) is a well-received algorithm for large-scale optimization in neural networks for its low iteration cost. However, due to Gradient variance, it often has difficulty in finding optimal learning rate and thus suffers from slow convergence. Using momentum is proven to be a simple effective way of overcoming the slow convergence problem of SDG as long as momentum is properly set. According to the performance metrics, this paper proposes a novel statistical model for analyzing the performance of neural networks. The model takes into account learning rate and momentum, and the method can be used to evaluate and verify their interaction effects on neural network performance. Our study shows that the interaction effects are significant. When momentum has a value smaller than 0.5, the impact on the training time is not statistically noticeable.\",\"PeriodicalId\":382429,\"journal\":{\"name\":\"2019 3rd International Conference on Data Science and Business Analytics (ICDSBA)\",\"volume\":\"8 1\",\"pages\":\"0\"},\"PeriodicalIF\":0.0000,\"publicationDate\":\"2019-10-01\",\"publicationTypes\":\"Journal Article\",\"fieldsOfStudy\":null,\"isOpenAccess\":false,\"openAccessPdf\":\"\",\"citationCount\":\"3\",\"resultStr\":null,\"platform\":\"Semanticscholar\",\"paperid\":null,\"PeriodicalName\":\"2019 3rd International Conference on Data Science and Business Analytics (ICDSBA)\",\"FirstCategoryId\":\"1085\",\"ListUrlMain\":\"https://doi.org/10.1109/ICDSBA48748.2019.00065\",\"RegionNum\":0,\"RegionCategory\":null,\"ArticlePicture\":[],\"TitleCN\":null,\"AbstractTextCN\":null,\"PMCID\":null,\"EPubDate\":\"\",\"PubModel\":\"\",\"JCR\":\"\",\"JCRName\":\"\",\"Score\":null,\"Total\":0}","platform":"Semanticscholar","paperid":null,"PeriodicalName":"2019 3rd International Conference on Data Science and Business Analytics (ICDSBA)","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/10.1109/ICDSBA48748.2019.00065","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
引用次数: 3

摘要

随机梯度下降算法(SGD)因其迭代成本低而被广泛应用于神经网络的大规模优化。但由于梯度方差的存在,往往难以找到最优学习率,收敛速度慢。利用动量被证明是克服可持续发展目标缓慢收敛问题的一种简单有效的方法,只要动量设置得当。根据性能指标,提出了一种新的神经网络性能统计模型。该模型考虑了学习率和动量,该方法可以用来评估和验证它们的交互作用对神经网络性能的影响。我们的研究表明,相互作用的影响是显著的。当动量值小于0.5时,对训练时间的影响在统计上不显著。
本文章由计算机程序翻译,如有差异,请以英文原文为准。
A Statistical Test of The Effect of Learning Rate and Momentum Coefficient of Sgd and Its Interaction on Neural Network Performance
Stochastic Gradient Descent (SGD) is a well-received algorithm for large-scale optimization in neural networks for its low iteration cost. However, due to Gradient variance, it often has difficulty in finding optimal learning rate and thus suffers from slow convergence. Using momentum is proven to be a simple effective way of overcoming the slow convergence problem of SDG as long as momentum is properly set. According to the performance metrics, this paper proposes a novel statistical model for analyzing the performance of neural networks. The model takes into account learning rate and momentum, and the method can be used to evaluate and verify their interaction effects on neural network performance. Our study shows that the interaction effects are significant. When momentum has a value smaller than 0.5, the impact on the training time is not statistically noticeable.
求助全文
通过发布文献求助,成功后即可免费获取论文全文。 去求助
来源期刊
自引率
0.00%
发文量
0
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
确定
请完成安全验证×
copy
已复制链接
快去分享给好友吧!
我知道了
右上角分享
点击右上角分享
0
联系我们:info@booksci.cn Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。 Copyright © 2023 布克学术 All rights reserved.
京ICP备2023020795号-1
ghs 京公网安备 11010802042870号
Book学术文献互助
Book学术文献互助群
群 号:481959085
Book学术官方微信