{"title":"基于LSTM和GRU的水稻产量预测","authors":"Yu Qiu","doi":"10.1117/12.2674760","DOIUrl":null,"url":null,"abstract":"Rice yield prediction is a vital problem in the national agriculture and economy. The development of deep learning overcomes the obstacles of traditional machine learning and shows superior performance in solving complicated problems. Especially for natural language processing (NLP) models such as LSTM and GRU, these models outperform the time series data, thus having great potential for complex agricultural spatiotemporal data with high dimensionality and nonlinearity. However, there is little discussion about performance of these two models in rice yield prediction. In this article, we adopted two popular NLP models to build and test 12 different model frameworks based on optimal hyperparameter configurations. And we compared model depth as well as bidirectional setting on the rice yield prediction by observing the performance of MSE losses throughout the training process. The results illustrated that both simple and complex models had outstanding fitting for small-sample training, and the depth and direction of the models did not significantly impact the performance of the experiment. But the complex model notably increases the training cost and decreases the convergence rate, implying that it’s not necessarily suitable for time-series problems with small-sample data. Further, the results could provide insights into a deep learning framework construction and hyperparameter selection for subsequent studies with comparable characteristics.","PeriodicalId":286364,"journal":{"name":"Conference on Computer Graphics, Artificial Intelligence, and Data Processing","volume":"72 4","pages":"0"},"PeriodicalIF":0.0000,"publicationDate":"2023-05-23","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":"{\"title\":\"Rice yield prediction based on LSTM and GRU\",\"authors\":\"Yu Qiu\",\"doi\":\"10.1117/12.2674760\",\"DOIUrl\":null,\"url\":null,\"abstract\":\"Rice yield prediction is a vital problem in the national agriculture and economy. The development of deep learning overcomes the obstacles of traditional machine learning and shows superior performance in solving complicated problems. Especially for natural language processing (NLP) models such as LSTM and GRU, these models outperform the time series data, thus having great potential for complex agricultural spatiotemporal data with high dimensionality and nonlinearity. However, there is little discussion about performance of these two models in rice yield prediction. In this article, we adopted two popular NLP models to build and test 12 different model frameworks based on optimal hyperparameter configurations. And we compared model depth as well as bidirectional setting on the rice yield prediction by observing the performance of MSE losses throughout the training process. The results illustrated that both simple and complex models had outstanding fitting for small-sample training, and the depth and direction of the models did not significantly impact the performance of the experiment. But the complex model notably increases the training cost and decreases the convergence rate, implying that it’s not necessarily suitable for time-series problems with small-sample data. Further, the results could provide insights into a deep learning framework construction and hyperparameter selection for subsequent studies with comparable characteristics.\",\"PeriodicalId\":286364,\"journal\":{\"name\":\"Conference on Computer Graphics, Artificial Intelligence, and Data Processing\",\"volume\":\"72 4\",\"pages\":\"0\"},\"PeriodicalIF\":0.0000,\"publicationDate\":\"2023-05-23\",\"publicationTypes\":\"Journal Article\",\"fieldsOfStudy\":null,\"isOpenAccess\":false,\"openAccessPdf\":\"\",\"citationCount\":\"0\",\"resultStr\":null,\"platform\":\"Semanticscholar\",\"paperid\":null,\"PeriodicalName\":\"Conference on Computer Graphics, Artificial Intelligence, and Data Processing\",\"FirstCategoryId\":\"1085\",\"ListUrlMain\":\"https://doi.org/10.1117/12.2674760\",\"RegionNum\":0,\"RegionCategory\":null,\"ArticlePicture\":[],\"TitleCN\":null,\"AbstractTextCN\":null,\"PMCID\":null,\"EPubDate\":\"\",\"PubModel\":\"\",\"JCR\":\"\",\"JCRName\":\"\",\"Score\":null,\"Total\":0}","platform":"Semanticscholar","paperid":null,"PeriodicalName":"Conference on Computer Graphics, Artificial Intelligence, and Data Processing","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/10.1117/12.2674760","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
Rice yield prediction is a vital problem in the national agriculture and economy. The development of deep learning overcomes the obstacles of traditional machine learning and shows superior performance in solving complicated problems. Especially for natural language processing (NLP) models such as LSTM and GRU, these models outperform the time series data, thus having great potential for complex agricultural spatiotemporal data with high dimensionality and nonlinearity. However, there is little discussion about performance of these two models in rice yield prediction. In this article, we adopted two popular NLP models to build and test 12 different model frameworks based on optimal hyperparameter configurations. And we compared model depth as well as bidirectional setting on the rice yield prediction by observing the performance of MSE losses throughout the training process. The results illustrated that both simple and complex models had outstanding fitting for small-sample training, and the depth and direction of the models did not significantly impact the performance of the experiment. But the complex model notably increases the training cost and decreases the convergence rate, implying that it’s not necessarily suitable for time-series problems with small-sample data. Further, the results could provide insights into a deep learning framework construction and hyperparameter selection for subsequent studies with comparable characteristics.