{"title":"A Statistical Test of The Effect of Learning Rate and Momentum Coefficient of Sgd and Its Interaction on Neural Network Performance","authors":"Bingchuan Chen, Ai-Xiang Chen, Xiaolong Chai, Rui Bian","doi":"10.1109/ICDSBA48748.2019.00065","DOIUrl":null,"url":null,"abstract":"Stochastic Gradient Descent (SGD) is a well-received algorithm for large-scale optimization in neural networks for its low iteration cost. However, due to Gradient variance, it often has difficulty in finding optimal learning rate and thus suffers from slow convergence. Using momentum is proven to be a simple effective way of overcoming the slow convergence problem of SDG as long as momentum is properly set. According to the performance metrics, this paper proposes a novel statistical model for analyzing the performance of neural networks. The model takes into account learning rate and momentum, and the method can be used to evaluate and verify their interaction effects on neural network performance. Our study shows that the interaction effects are significant. When momentum has a value smaller than 0.5, the impact on the training time is not statistically noticeable.","PeriodicalId":382429,"journal":{"name":"2019 3rd International Conference on Data Science and Business Analytics (ICDSBA)","volume":"8 1","pages":"0"},"PeriodicalIF":0.0000,"publicationDate":"2019-10-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"3","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"2019 3rd International Conference on Data Science and Business Analytics (ICDSBA)","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/10.1109/ICDSBA48748.2019.00065","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
引用次数: 3
Abstract
Stochastic Gradient Descent (SGD) is a well-received algorithm for large-scale optimization in neural networks for its low iteration cost. However, due to Gradient variance, it often has difficulty in finding optimal learning rate and thus suffers from slow convergence. Using momentum is proven to be a simple effective way of overcoming the slow convergence problem of SDG as long as momentum is properly set. According to the performance metrics, this paper proposes a novel statistical model for analyzing the performance of neural networks. The model takes into account learning rate and momentum, and the method can be used to evaluate and verify their interaction effects on neural network performance. Our study shows that the interaction effects are significant. When momentum has a value smaller than 0.5, the impact on the training time is not statistically noticeable.