Enhong Liu, Dan Su, Liangming Chen, Long Jin, Xiuchun Xiao, Dongyang Fu
{"title":"应用拉格朗日型1步前移数值微分法优化深度学习中的SGD算法的尝试","authors":"Enhong Liu, Dan Su, Liangming Chen, Long Jin, Xiuchun Xiao, Dongyang Fu","doi":"10.1109/ICIST52614.2021.9440607","DOIUrl":null,"url":null,"abstract":"The form of the original stochastic gradient descent (SGD) algorithm accords with the definition of forward Euler method, which has some inherent defects. Thus, in order to improve the original SGD algorithm, a Lagrange-type 1-step-ahead numerical differentiation method based parameter update algorithm is presented and validated. Instinctively, the new algorithm can fix some inherent flaws of the SGD algorithm. However, a series of experimental results show that the Lagrange-type 1-step-ahead numerical differentiation method cannot be applied to reduce the computational error of SGD. In addition, this method makes the model appear the phenomenon of non-convergence. Finally, on the basis of comparative experiments, the divergence phenomenon is analyzed and explained.","PeriodicalId":371599,"journal":{"name":"2021 11th International Conference on Information Science and Technology (ICIST)","volume":"699 1","pages":"0"},"PeriodicalIF":0.0000,"publicationDate":"2021-05-21","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":"{\"title\":\"An attempt of applying the Lagrange-type 1-step-ahead numerical differentiation method to optimize the SGD algorithm in deep learning\",\"authors\":\"Enhong Liu, Dan Su, Liangming Chen, Long Jin, Xiuchun Xiao, Dongyang Fu\",\"doi\":\"10.1109/ICIST52614.2021.9440607\",\"DOIUrl\":null,\"url\":null,\"abstract\":\"The form of the original stochastic gradient descent (SGD) algorithm accords with the definition of forward Euler method, which has some inherent defects. Thus, in order to improve the original SGD algorithm, a Lagrange-type 1-step-ahead numerical differentiation method based parameter update algorithm is presented and validated. Instinctively, the new algorithm can fix some inherent flaws of the SGD algorithm. However, a series of experimental results show that the Lagrange-type 1-step-ahead numerical differentiation method cannot be applied to reduce the computational error of SGD. In addition, this method makes the model appear the phenomenon of non-convergence. Finally, on the basis of comparative experiments, the divergence phenomenon is analyzed and explained.\",\"PeriodicalId\":371599,\"journal\":{\"name\":\"2021 11th International Conference on Information Science and Technology (ICIST)\",\"volume\":\"699 1\",\"pages\":\"0\"},\"PeriodicalIF\":0.0000,\"publicationDate\":\"2021-05-21\",\"publicationTypes\":\"Journal Article\",\"fieldsOfStudy\":null,\"isOpenAccess\":false,\"openAccessPdf\":\"\",\"citationCount\":\"0\",\"resultStr\":null,\"platform\":\"Semanticscholar\",\"paperid\":null,\"PeriodicalName\":\"2021 11th International Conference on Information Science and Technology (ICIST)\",\"FirstCategoryId\":\"1085\",\"ListUrlMain\":\"https://doi.org/10.1109/ICIST52614.2021.9440607\",\"RegionNum\":0,\"RegionCategory\":null,\"ArticlePicture\":[],\"TitleCN\":null,\"AbstractTextCN\":null,\"PMCID\":null,\"EPubDate\":\"\",\"PubModel\":\"\",\"JCR\":\"\",\"JCRName\":\"\",\"Score\":null,\"Total\":0}","platform":"Semanticscholar","paperid":null,"PeriodicalName":"2021 11th International Conference on Information Science and Technology (ICIST)","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/10.1109/ICIST52614.2021.9440607","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
An attempt of applying the Lagrange-type 1-step-ahead numerical differentiation method to optimize the SGD algorithm in deep learning
The form of the original stochastic gradient descent (SGD) algorithm accords with the definition of forward Euler method, which has some inherent defects. Thus, in order to improve the original SGD algorithm, a Lagrange-type 1-step-ahead numerical differentiation method based parameter update algorithm is presented and validated. Instinctively, the new algorithm can fix some inherent flaws of the SGD algorithm. However, a series of experimental results show that the Lagrange-type 1-step-ahead numerical differentiation method cannot be applied to reduce the computational error of SGD. In addition, this method makes the model appear the phenomenon of non-convergence. Finally, on the basis of comparative experiments, the divergence phenomenon is analyzed and explained.