N. Gorgolis, I. Hatzilygeroudis, Z. Istenes, Lazlo-Grad Gyenne
{"title":"Hyperparameter Optimization of LSTM Network Models through Genetic Algorithm","authors":"N. Gorgolis, I. Hatzilygeroudis, Z. Istenes, Lazlo-Grad Gyenne","doi":"10.1109/IISA.2019.8900675","DOIUrl":null,"url":null,"abstract":"Next word prediction is an important problem in the domain of NLP, hence in modern artificial intelligence. It draws both scientific and industrial interest, as it consists the core of many processes, like autocorrection, text generation, review prediction etc. Currently, the most efficient and common approach used is classification, using artificial neural networks (ANNs). One of the main drawbacks of ANNs is fine – tuning their hyperparameters, a procedure which is essential to the performance of the model. On the other hand, the approaches usually used for fine – tuning are either computationally unaffordable (e.g. grid search) or of uncertain efficiency (e.g. trial & error). As a response to the above, through the current paper is presented a simple genetic algorithm approach, which is used for the hyperparameter tuning of a common language model and it achieves tuning efficiency without following an exhaustive search.","PeriodicalId":371385,"journal":{"name":"2019 10th International Conference on Information, Intelligence, Systems and Applications (IISA)","volume":"42 1","pages":"0"},"PeriodicalIF":0.0000,"publicationDate":"2019-07-15","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"24","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"2019 10th International Conference on Information, Intelligence, Systems and Applications (IISA)","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/10.1109/IISA.2019.8900675","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
引用次数: 24
Abstract
Next word prediction is an important problem in the domain of NLP, hence in modern artificial intelligence. It draws both scientific and industrial interest, as it consists the core of many processes, like autocorrection, text generation, review prediction etc. Currently, the most efficient and common approach used is classification, using artificial neural networks (ANNs). One of the main drawbacks of ANNs is fine – tuning their hyperparameters, a procedure which is essential to the performance of the model. On the other hand, the approaches usually used for fine – tuning are either computationally unaffordable (e.g. grid search) or of uncertain efficiency (e.g. trial & error). As a response to the above, through the current paper is presented a simple genetic algorithm approach, which is used for the hyperparameter tuning of a common language model and it achieves tuning efficiency without following an exhaustive search.