The Best Parameter Tuning on RNN Layers for Indonesian Text Classification

Awaliyatul Hikmah, Sumarni Adi, Mulia Sulistiyono
{"title":"The Best Parameter Tuning on RNN Layers for Indonesian Text Classification","authors":"Awaliyatul Hikmah, Sumarni Adi, Mulia Sulistiyono","doi":"10.1109/ISRITI51436.2020.9315425","DOIUrl":null,"url":null,"abstract":"Recurrent Neural Network (RNN) is a deep learning architecture commonly used to process time series and sequence data. Various architectures have been developed to improve the performance of the algorithm in terms of both accuracy and computation time. Besides, the use of appropriate parameter values when building a neural network model also plays an important role in the quality and the outcome of the learning model. In this study, the model trained using RNN-Vanilla, LSTM, and GRU each with 4 different combinations of parameter settings, namely bidirectional mode (True, False), the number of neuron units on each layer (64, 128, 256), the number of RNN layers on the neural network (1, 2, 3), and the batch size when training the model (32, 64, 128). By combining all the parameter values, 162 trials were carried out to perform the task of classifying Indonesian language customer support tickets with four category classes. This study gives the result that the same network architecture but with different parameter combinations results in significant differences in the level of accuracy. The lowest accuracy of all experiments was 32.874% and the highest accuracy resulted was 84.369%. Overall, by calculating the average accuracy of each parameter value, the results obtained are: GRU has the best performance, accuracy tends to increase by activating bidirectional mode, increasing the number of neuron units in the hidden layer, and reducing the batch size. Meanwhile, the addition of the number of RNN layers on the neural network has no impact on increasing the level of accuracy.","PeriodicalId":325920,"journal":{"name":"2020 3rd International Seminar on Research of Information Technology and Intelligent Systems (ISRITI)","volume":null,"pages":null},"PeriodicalIF":0.0000,"publicationDate":"2020-12-10","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"3","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"2020 3rd International Seminar on Research of Information Technology and Intelligent Systems (ISRITI)","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/10.1109/ISRITI51436.2020.9315425","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
引用次数: 3

Abstract

Recurrent Neural Network (RNN) is a deep learning architecture commonly used to process time series and sequence data. Various architectures have been developed to improve the performance of the algorithm in terms of both accuracy and computation time. Besides, the use of appropriate parameter values when building a neural network model also plays an important role in the quality and the outcome of the learning model. In this study, the model trained using RNN-Vanilla, LSTM, and GRU each with 4 different combinations of parameter settings, namely bidirectional mode (True, False), the number of neuron units on each layer (64, 128, 256), the number of RNN layers on the neural network (1, 2, 3), and the batch size when training the model (32, 64, 128). By combining all the parameter values, 162 trials were carried out to perform the task of classifying Indonesian language customer support tickets with four category classes. This study gives the result that the same network architecture but with different parameter combinations results in significant differences in the level of accuracy. The lowest accuracy of all experiments was 32.874% and the highest accuracy resulted was 84.369%. Overall, by calculating the average accuracy of each parameter value, the results obtained are: GRU has the best performance, accuracy tends to increase by activating bidirectional mode, increasing the number of neuron units in the hidden layer, and reducing the batch size. Meanwhile, the addition of the number of RNN layers on the neural network has no impact on increasing the level of accuracy.
印尼语文本分类RNN层的最佳参数调优
递归神经网络(RNN)是一种深度学习架构,通常用于处理时间序列和序列数据。为了提高算法在精度和计算时间方面的性能,已经开发了各种体系结构。此外,在构建神经网络模型时,使用合适的参数值对学习模型的质量和结果也起着重要的作用。在本研究中,使用RNN- vanilla、LSTM和GRU训练的模型各有4种不同的参数设置组合,即双向模式(True, False)、每层神经元单元数(64、128、256)、神经网络上的RNN层数(1、2、3)和训练模型时的批大小(32、64、128)。通过组合所有参数值,进行了162次试验,以执行将印尼语客户支持票分为四类的任务。研究结果表明,在相同的网络结构下,不同的参数组合会导致准确率水平的显著差异。所有实验的最低准确率为32.874%,最高准确率为84.369%。总体而言,通过计算各参数值的平均准确率,得到的结果是:GRU的性能最好,激活双向模式、增加隐藏层神经元单元数、减小批处理大小,准确率有提高的趋势。同时,在神经网络上增加RNN层数对提高准确率水平没有影响。
本文章由计算机程序翻译,如有差异,请以英文原文为准。
求助全文
约1分钟内获得全文 求助全文
来源期刊
自引率
0.00%
发文量
0
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
确定
请完成安全验证×
copy
已复制链接
快去分享给好友吧!
我知道了
右上角分享
点击右上角分享
0
联系我们:info@booksci.cn Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。 Copyright © 2023 布克学术 All rights reserved.
京ICP备2023020795号-1
ghs 京公网安备 11010802042870号
Book学术文献互助
Book学术文献互助群
群 号:481959085
Book学术官方微信