{"title":"Recurrent Neural Networks with Fractional Order Gradient Method","authors":"Honggang Yang, Rui Fan, Jiejie Chen, Mengfei Xu","doi":"10.1109/icaci55529.2022.9837518","DOIUrl":null,"url":null,"abstract":"In view of the possibility that Recurrent Neural Network(RNN)’s stochastic gradient descent method will converge to the local optimum problem, two fractional stochastic gradient descent methods are proposed in this paper. The methods respectively use the fractional order substitution derivative part defined by Caputo and the fractional order substitution difference form defined by Riemann Liouville to improve the updating method of network parameters. Combining with the gradient descent characteristics, the influence of fractional order on the training results is discussed, and two adaptive order adjustment methods are proposed. Experiments on MNIST and FashionMNIST datasets show that the fractional stochastic gradient optimization algorithm can improve the classification accuracy and training speed of recurrent neural network.","PeriodicalId":412347,"journal":{"name":"2022 14th International Conference on Advanced Computational Intelligence (ICACI)","volume":"25 1","pages":"0"},"PeriodicalIF":0.0000,"publicationDate":"2022-07-15","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"3","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"2022 14th International Conference on Advanced Computational Intelligence (ICACI)","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/10.1109/icaci55529.2022.9837518","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
引用次数: 3
Abstract
In view of the possibility that Recurrent Neural Network(RNN)’s stochastic gradient descent method will converge to the local optimum problem, two fractional stochastic gradient descent methods are proposed in this paper. The methods respectively use the fractional order substitution derivative part defined by Caputo and the fractional order substitution difference form defined by Riemann Liouville to improve the updating method of network parameters. Combining with the gradient descent characteristics, the influence of fractional order on the training results is discussed, and two adaptive order adjustment methods are proposed. Experiments on MNIST and FashionMNIST datasets show that the fractional stochastic gradient optimization algorithm can improve the classification accuracy and training speed of recurrent neural network.