{"title":"使用机器学习来预测程序的运行时间","authors":"Xinyi Li, Yiyuan Wang, Ying Qian, Liang Dou","doi":"10.1145/3507548.3507588","DOIUrl":null,"url":null,"abstract":"The prediction of program running time can be used to improve scheduling performance of distributed systems. In 2011, Google released a data set documenting the vast amount of information in the Google cluster. However, most of the existing running time prediction models only consider the coarse-grained characteristics of the running environment without considering the influence of the time series data of the running environment on the prediction results. Based on this, this paper innovatively proposes a model to predict the running time of the program, which predicts the future running time through historical information. At the same time, we also propose a new data processing and feature extraction scheme for Google cluster data sets. The results show that our model greatly outperforms the classical model on the Google cluster data set, and the root-mean-square error index of running time under different prediction modes is reduced by more than 60% and 40%, respectively. We hope that the model proposed in this paper can provide new research ideas for cloud computing system design.","PeriodicalId":414908,"journal":{"name":"Proceedings of the 2021 5th International Conference on Computer Science and Artificial Intelligence","volume":"31 1","pages":"0"},"PeriodicalIF":0.0000,"publicationDate":"2021-12-04","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":"{\"title\":\"Use Machine Learning to Predict the Running Time of the Program\",\"authors\":\"Xinyi Li, Yiyuan Wang, Ying Qian, Liang Dou\",\"doi\":\"10.1145/3507548.3507588\",\"DOIUrl\":null,\"url\":null,\"abstract\":\"The prediction of program running time can be used to improve scheduling performance of distributed systems. In 2011, Google released a data set documenting the vast amount of information in the Google cluster. However, most of the existing running time prediction models only consider the coarse-grained characteristics of the running environment without considering the influence of the time series data of the running environment on the prediction results. Based on this, this paper innovatively proposes a model to predict the running time of the program, which predicts the future running time through historical information. At the same time, we also propose a new data processing and feature extraction scheme for Google cluster data sets. The results show that our model greatly outperforms the classical model on the Google cluster data set, and the root-mean-square error index of running time under different prediction modes is reduced by more than 60% and 40%, respectively. We hope that the model proposed in this paper can provide new research ideas for cloud computing system design.\",\"PeriodicalId\":414908,\"journal\":{\"name\":\"Proceedings of the 2021 5th International Conference on Computer Science and Artificial Intelligence\",\"volume\":\"31 1\",\"pages\":\"0\"},\"PeriodicalIF\":0.0000,\"publicationDate\":\"2021-12-04\",\"publicationTypes\":\"Journal Article\",\"fieldsOfStudy\":null,\"isOpenAccess\":false,\"openAccessPdf\":\"\",\"citationCount\":\"0\",\"resultStr\":null,\"platform\":\"Semanticscholar\",\"paperid\":null,\"PeriodicalName\":\"Proceedings of the 2021 5th International Conference on Computer Science and Artificial Intelligence\",\"FirstCategoryId\":\"1085\",\"ListUrlMain\":\"https://doi.org/10.1145/3507548.3507588\",\"RegionNum\":0,\"RegionCategory\":null,\"ArticlePicture\":[],\"TitleCN\":null,\"AbstractTextCN\":null,\"PMCID\":null,\"EPubDate\":\"\",\"PubModel\":\"\",\"JCR\":\"\",\"JCRName\":\"\",\"Score\":null,\"Total\":0}","platform":"Semanticscholar","paperid":null,"PeriodicalName":"Proceedings of the 2021 5th International Conference on Computer Science and Artificial Intelligence","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/10.1145/3507548.3507588","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
Use Machine Learning to Predict the Running Time of the Program
The prediction of program running time can be used to improve scheduling performance of distributed systems. In 2011, Google released a data set documenting the vast amount of information in the Google cluster. However, most of the existing running time prediction models only consider the coarse-grained characteristics of the running environment without considering the influence of the time series data of the running environment on the prediction results. Based on this, this paper innovatively proposes a model to predict the running time of the program, which predicts the future running time through historical information. At the same time, we also propose a new data processing and feature extraction scheme for Google cluster data sets. The results show that our model greatly outperforms the classical model on the Google cluster data set, and the root-mean-square error index of running time under different prediction modes is reduced by more than 60% and 40%, respectively. We hope that the model proposed in this paper can provide new research ideas for cloud computing system design.