使用 FNet 对酒店客户评论进行情感分析

Q2 Mathematics
Shovan Bhowmik, Rifat Sadik, Wahiduzzaman Akanda, Juboraj Roy Pavel
{"title":"使用 FNet 对酒店客户评论进行情感分析","authors":"Shovan Bhowmik, Rifat Sadik, Wahiduzzaman Akanda, Juboraj Roy Pavel","doi":"10.11591/eei.v13i2.6301","DOIUrl":null,"url":null,"abstract":"Recent research has focused on opinion mining from public sentiments using natural language processing (NLP) and machine learning (ML) techniques. Transformer-based models, such as bidirectional encoder representations from transformers (BERT), excel in extracting semantic information but are resourceintensive. Google’s new research, mixing tokens with fourier transform, also known as FNet, replaced BERT’s attention mechanism with a non-parameterized fourier transform, aiming to reduce training time without compromising performance. This study fine-tuned the FNet model with a publicly available Kaggle hotel review dataset and investigated the performance of this dataset in both FNet and BERT architectures along with conventional machine learning models such as long short-term memory (LSTM) and support vector machine (SVM). Results revealed that FNet significantly reduces the training time by almost 20% and memory utilization by nearly 60% compared to BERT. The highest test accuracy observed in this experiment by FNet was 80.27% which is nearly 97.85% of BERT’s performance with identical parameters.","PeriodicalId":37619,"journal":{"name":"Bulletin of Electrical Engineering and Informatics","volume":"46 3","pages":""},"PeriodicalIF":0.0000,"publicationDate":"2024-04-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":"{\"title\":\"Sentiment analysis with hotel customer reviews using FNet\",\"authors\":\"Shovan Bhowmik, Rifat Sadik, Wahiduzzaman Akanda, Juboraj Roy Pavel\",\"doi\":\"10.11591/eei.v13i2.6301\",\"DOIUrl\":null,\"url\":null,\"abstract\":\"Recent research has focused on opinion mining from public sentiments using natural language processing (NLP) and machine learning (ML) techniques. Transformer-based models, such as bidirectional encoder representations from transformers (BERT), excel in extracting semantic information but are resourceintensive. Google’s new research, mixing tokens with fourier transform, also known as FNet, replaced BERT’s attention mechanism with a non-parameterized fourier transform, aiming to reduce training time without compromising performance. This study fine-tuned the FNet model with a publicly available Kaggle hotel review dataset and investigated the performance of this dataset in both FNet and BERT architectures along with conventional machine learning models such as long short-term memory (LSTM) and support vector machine (SVM). Results revealed that FNet significantly reduces the training time by almost 20% and memory utilization by nearly 60% compared to BERT. The highest test accuracy observed in this experiment by FNet was 80.27% which is nearly 97.85% of BERT’s performance with identical parameters.\",\"PeriodicalId\":37619,\"journal\":{\"name\":\"Bulletin of Electrical Engineering and Informatics\",\"volume\":\"46 3\",\"pages\":\"\"},\"PeriodicalIF\":0.0000,\"publicationDate\":\"2024-04-01\",\"publicationTypes\":\"Journal Article\",\"fieldsOfStudy\":null,\"isOpenAccess\":false,\"openAccessPdf\":\"\",\"citationCount\":\"0\",\"resultStr\":null,\"platform\":\"Semanticscholar\",\"paperid\":null,\"PeriodicalName\":\"Bulletin of Electrical Engineering and Informatics\",\"FirstCategoryId\":\"1085\",\"ListUrlMain\":\"https://doi.org/10.11591/eei.v13i2.6301\",\"RegionNum\":0,\"RegionCategory\":null,\"ArticlePicture\":[],\"TitleCN\":null,\"AbstractTextCN\":null,\"PMCID\":null,\"EPubDate\":\"\",\"PubModel\":\"\",\"JCR\":\"Q2\",\"JCRName\":\"Mathematics\",\"Score\":null,\"Total\":0}","platform":"Semanticscholar","paperid":null,"PeriodicalName":"Bulletin of Electrical Engineering and Informatics","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/10.11591/eei.v13i2.6301","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"Q2","JCRName":"Mathematics","Score":null,"Total":0}
引用次数: 0

摘要

近期的研究重点是利用自然语言处理(NLP)和机器学习(ML)技术从公众情绪中挖掘观点。基于变换器的模型,如来自变换器的双向编码器表示(BERT),在提取语义信息方面表现出色,但却是资源密集型的。谷歌的新研究 "混合标记与傅立叶变换"(又称 FNet)用非参数化的傅立叶变换取代了 BERT 的注意机制,旨在减少训练时间,同时不影响性能。本研究利用公开的 Kaggle 酒店点评数据集对 FNet 模型进行了微调,并研究了该数据集在 FNet 和 BERT 架构下的性能,以及长短期记忆(LSTM)和支持向量机(SVM)等传统机器学习模型的性能。结果显示,与 BERT 相比,FNet 大幅减少了近 20% 的训练时间和近 60% 的内存使用率。在该实验中,FNet 的最高测试准确率为 80.27%,在参数相同的情况下,接近 BERT 性能的 97.85%。
本文章由计算机程序翻译,如有差异,请以英文原文为准。
Sentiment analysis with hotel customer reviews using FNet
Recent research has focused on opinion mining from public sentiments using natural language processing (NLP) and machine learning (ML) techniques. Transformer-based models, such as bidirectional encoder representations from transformers (BERT), excel in extracting semantic information but are resourceintensive. Google’s new research, mixing tokens with fourier transform, also known as FNet, replaced BERT’s attention mechanism with a non-parameterized fourier transform, aiming to reduce training time without compromising performance. This study fine-tuned the FNet model with a publicly available Kaggle hotel review dataset and investigated the performance of this dataset in both FNet and BERT architectures along with conventional machine learning models such as long short-term memory (LSTM) and support vector machine (SVM). Results revealed that FNet significantly reduces the training time by almost 20% and memory utilization by nearly 60% compared to BERT. The highest test accuracy observed in this experiment by FNet was 80.27% which is nearly 97.85% of BERT’s performance with identical parameters.
求助全文
通过发布文献求助,成功后即可免费获取论文全文。 去求助
来源期刊
Bulletin of Electrical Engineering and Informatics
Bulletin of Electrical Engineering and Informatics Computer Science-Computer Science (miscellaneous)
CiteScore
3.60
自引率
0.00%
发文量
0
期刊介绍: Bulletin of Electrical Engineering and Informatics publishes original papers in the field of electrical, computer and informatics engineering which covers, but not limited to, the following scope: Computer Science, Computer Engineering and Informatics[...] Electronics[...] Electrical and Power Engineering[...] Telecommunication and Information Technology[...]Instrumentation and Control Engineering[...]
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
确定
请完成安全验证×
copy
已复制链接
快去分享给好友吧!
我知道了
右上角分享
点击右上角分享
0
联系我们:info@booksci.cn Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。 Copyright © 2023 布克学术 All rights reserved.
京ICP备2023020795号-1
ghs 京公网安备 11010802042870号
Book学术文献互助
Book学术文献互助群
群 号:481959085
Book学术官方微信