{"title":"Sgd的学习率和动量系数及其交互作用对神经网络性能影响的统计检验","authors":"Bingchuan Chen, Ai-Xiang Chen, Xiaolong Chai, Rui Bian","doi":"10.1109/ICDSBA48748.2019.00065","DOIUrl":null,"url":null,"abstract":"Stochastic Gradient Descent (SGD) is a well-received algorithm for large-scale optimization in neural networks for its low iteration cost. However, due to Gradient variance, it often has difficulty in finding optimal learning rate and thus suffers from slow convergence. Using momentum is proven to be a simple effective way of overcoming the slow convergence problem of SDG as long as momentum is properly set. According to the performance metrics, this paper proposes a novel statistical model for analyzing the performance of neural networks. The model takes into account learning rate and momentum, and the method can be used to evaluate and verify their interaction effects on neural network performance. Our study shows that the interaction effects are significant. When momentum has a value smaller than 0.5, the impact on the training time is not statistically noticeable.","PeriodicalId":382429,"journal":{"name":"2019 3rd International Conference on Data Science and Business Analytics (ICDSBA)","volume":"8 1","pages":"0"},"PeriodicalIF":0.0000,"publicationDate":"2019-10-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"3","resultStr":"{\"title\":\"A Statistical Test of The Effect of Learning Rate and Momentum Coefficient of Sgd and Its Interaction on Neural Network Performance\",\"authors\":\"Bingchuan Chen, Ai-Xiang Chen, Xiaolong Chai, Rui Bian\",\"doi\":\"10.1109/ICDSBA48748.2019.00065\",\"DOIUrl\":null,\"url\":null,\"abstract\":\"Stochastic Gradient Descent (SGD) is a well-received algorithm for large-scale optimization in neural networks for its low iteration cost. However, due to Gradient variance, it often has difficulty in finding optimal learning rate and thus suffers from slow convergence. Using momentum is proven to be a simple effective way of overcoming the slow convergence problem of SDG as long as momentum is properly set. According to the performance metrics, this paper proposes a novel statistical model for analyzing the performance of neural networks. The model takes into account learning rate and momentum, and the method can be used to evaluate and verify their interaction effects on neural network performance. Our study shows that the interaction effects are significant. When momentum has a value smaller than 0.5, the impact on the training time is not statistically noticeable.\",\"PeriodicalId\":382429,\"journal\":{\"name\":\"2019 3rd International Conference on Data Science and Business Analytics (ICDSBA)\",\"volume\":\"8 1\",\"pages\":\"0\"},\"PeriodicalIF\":0.0000,\"publicationDate\":\"2019-10-01\",\"publicationTypes\":\"Journal Article\",\"fieldsOfStudy\":null,\"isOpenAccess\":false,\"openAccessPdf\":\"\",\"citationCount\":\"3\",\"resultStr\":null,\"platform\":\"Semanticscholar\",\"paperid\":null,\"PeriodicalName\":\"2019 3rd International Conference on Data Science and Business Analytics (ICDSBA)\",\"FirstCategoryId\":\"1085\",\"ListUrlMain\":\"https://doi.org/10.1109/ICDSBA48748.2019.00065\",\"RegionNum\":0,\"RegionCategory\":null,\"ArticlePicture\":[],\"TitleCN\":null,\"AbstractTextCN\":null,\"PMCID\":null,\"EPubDate\":\"\",\"PubModel\":\"\",\"JCR\":\"\",\"JCRName\":\"\",\"Score\":null,\"Total\":0}","platform":"Semanticscholar","paperid":null,"PeriodicalName":"2019 3rd International Conference on Data Science and Business Analytics (ICDSBA)","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/10.1109/ICDSBA48748.2019.00065","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
A Statistical Test of The Effect of Learning Rate and Momentum Coefficient of Sgd and Its Interaction on Neural Network Performance
Stochastic Gradient Descent (SGD) is a well-received algorithm for large-scale optimization in neural networks for its low iteration cost. However, due to Gradient variance, it often has difficulty in finding optimal learning rate and thus suffers from slow convergence. Using momentum is proven to be a simple effective way of overcoming the slow convergence problem of SDG as long as momentum is properly set. According to the performance metrics, this paper proposes a novel statistical model for analyzing the performance of neural networks. The model takes into account learning rate and momentum, and the method can be used to evaluate and verify their interaction effects on neural network performance. Our study shows that the interaction effects are significant. When momentum has a value smaller than 0.5, the impact on the training time is not statistically noticeable.