N. Gorgolis, I. Hatzilygeroudis, Z. Istenes, Lazlo-Grad Gyenne
{"title":"基于遗传算法的LSTM网络模型超参数优化","authors":"N. Gorgolis, I. Hatzilygeroudis, Z. Istenes, Lazlo-Grad Gyenne","doi":"10.1109/IISA.2019.8900675","DOIUrl":null,"url":null,"abstract":"Next word prediction is an important problem in the domain of NLP, hence in modern artificial intelligence. It draws both scientific and industrial interest, as it consists the core of many processes, like autocorrection, text generation, review prediction etc. Currently, the most efficient and common approach used is classification, using artificial neural networks (ANNs). One of the main drawbacks of ANNs is fine – tuning their hyperparameters, a procedure which is essential to the performance of the model. On the other hand, the approaches usually used for fine – tuning are either computationally unaffordable (e.g. grid search) or of uncertain efficiency (e.g. trial & error). As a response to the above, through the current paper is presented a simple genetic algorithm approach, which is used for the hyperparameter tuning of a common language model and it achieves tuning efficiency without following an exhaustive search.","PeriodicalId":371385,"journal":{"name":"2019 10th International Conference on Information, Intelligence, Systems and Applications (IISA)","volume":"42 1","pages":"0"},"PeriodicalIF":0.0000,"publicationDate":"2019-07-15","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"24","resultStr":"{\"title\":\"Hyperparameter Optimization of LSTM Network Models through Genetic Algorithm\",\"authors\":\"N. Gorgolis, I. Hatzilygeroudis, Z. Istenes, Lazlo-Grad Gyenne\",\"doi\":\"10.1109/IISA.2019.8900675\",\"DOIUrl\":null,\"url\":null,\"abstract\":\"Next word prediction is an important problem in the domain of NLP, hence in modern artificial intelligence. It draws both scientific and industrial interest, as it consists the core of many processes, like autocorrection, text generation, review prediction etc. Currently, the most efficient and common approach used is classification, using artificial neural networks (ANNs). One of the main drawbacks of ANNs is fine – tuning their hyperparameters, a procedure which is essential to the performance of the model. On the other hand, the approaches usually used for fine – tuning are either computationally unaffordable (e.g. grid search) or of uncertain efficiency (e.g. trial & error). As a response to the above, through the current paper is presented a simple genetic algorithm approach, which is used for the hyperparameter tuning of a common language model and it achieves tuning efficiency without following an exhaustive search.\",\"PeriodicalId\":371385,\"journal\":{\"name\":\"2019 10th International Conference on Information, Intelligence, Systems and Applications (IISA)\",\"volume\":\"42 1\",\"pages\":\"0\"},\"PeriodicalIF\":0.0000,\"publicationDate\":\"2019-07-15\",\"publicationTypes\":\"Journal Article\",\"fieldsOfStudy\":null,\"isOpenAccess\":false,\"openAccessPdf\":\"\",\"citationCount\":\"24\",\"resultStr\":null,\"platform\":\"Semanticscholar\",\"paperid\":null,\"PeriodicalName\":\"2019 10th International Conference on Information, Intelligence, Systems and Applications (IISA)\",\"FirstCategoryId\":\"1085\",\"ListUrlMain\":\"https://doi.org/10.1109/IISA.2019.8900675\",\"RegionNum\":0,\"RegionCategory\":null,\"ArticlePicture\":[],\"TitleCN\":null,\"AbstractTextCN\":null,\"PMCID\":null,\"EPubDate\":\"\",\"PubModel\":\"\",\"JCR\":\"\",\"JCRName\":\"\",\"Score\":null,\"Total\":0}","platform":"Semanticscholar","paperid":null,"PeriodicalName":"2019 10th International Conference on Information, Intelligence, Systems and Applications (IISA)","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/10.1109/IISA.2019.8900675","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
Hyperparameter Optimization of LSTM Network Models through Genetic Algorithm
Next word prediction is an important problem in the domain of NLP, hence in modern artificial intelligence. It draws both scientific and industrial interest, as it consists the core of many processes, like autocorrection, text generation, review prediction etc. Currently, the most efficient and common approach used is classification, using artificial neural networks (ANNs). One of the main drawbacks of ANNs is fine – tuning their hyperparameters, a procedure which is essential to the performance of the model. On the other hand, the approaches usually used for fine – tuning are either computationally unaffordable (e.g. grid search) or of uncertain efficiency (e.g. trial & error). As a response to the above, through the current paper is presented a simple genetic algorithm approach, which is used for the hyperparameter tuning of a common language model and it achieves tuning efficiency without following an exhaustive search.