Sentiment analysis on COVID tweets using COVID-Twitter-BERT with auxiliary sentence approach

Hung-Yeh Lin, Teng-Sheng Moh
{"title":"Sentiment analysis on COVID tweets using COVID-Twitter-BERT with auxiliary sentence approach","authors":"Hung-Yeh Lin, Teng-Sheng Moh","doi":"10.1145/3409334.3452074","DOIUrl":null,"url":null,"abstract":"Sentiment analysis is a fascinating area as a natural language understanding benchmark to evaluate customers' feedback and needs. Moreover, sentiment analysis can be applied to understand the people's reactions to public events such as the presidential elections and disease pandemics. Recent works in sentiment analysis on COVID-19 present a domain-targeted Bidirectional Encoder Representations from Transformer (BERT) language model, COVID-Twitter BERT (CT-BERT). However, there is little improvement in text classification using a BERT-based language model directly. Therefore, an auxiliary approach using BERT was proposed. This method converts single-sentence classification into pair-sentence classification, which solves the performance issue of BERT in text classification tasks. In this paper, we combine a pre-trained BERT model from COVID-related tweets and the auxiliary-sentence method to achieve better classification performance on COVID tweets sentiment analysis. We show that converting single-sentence classification into pair-sentence classification extends the dataset and obtains higher accuracies and F1 scores. However, we expect a domain-specific language model would perform better than a general language model. In our results, we show that the performance of CT-BERT does not necessarily outperform BERT specifically in understanding sentiments.","PeriodicalId":148741,"journal":{"name":"Proceedings of the 2021 ACM Southeast Conference","volume":"7 1","pages":"0"},"PeriodicalIF":0.0000,"publicationDate":"2021-04-15","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"6","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"Proceedings of the 2021 ACM Southeast Conference","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/10.1145/3409334.3452074","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
引用次数: 6

Abstract

Sentiment analysis is a fascinating area as a natural language understanding benchmark to evaluate customers' feedback and needs. Moreover, sentiment analysis can be applied to understand the people's reactions to public events such as the presidential elections and disease pandemics. Recent works in sentiment analysis on COVID-19 present a domain-targeted Bidirectional Encoder Representations from Transformer (BERT) language model, COVID-Twitter BERT (CT-BERT). However, there is little improvement in text classification using a BERT-based language model directly. Therefore, an auxiliary approach using BERT was proposed. This method converts single-sentence classification into pair-sentence classification, which solves the performance issue of BERT in text classification tasks. In this paper, we combine a pre-trained BERT model from COVID-related tweets and the auxiliary-sentence method to achieve better classification performance on COVID tweets sentiment analysis. We show that converting single-sentence classification into pair-sentence classification extends the dataset and obtains higher accuracies and F1 scores. However, we expect a domain-specific language model would perform better than a general language model. In our results, we show that the performance of CT-BERT does not necessarily outperform BERT specifically in understanding sentiments.
使用COVID- twitter - bert辅助句方法对COVID推文进行情感分析
情感分析是一个很有吸引力的领域,它可以作为自然语言理解的基准来评估客户的反馈和需求。此外,还可以运用情绪分析来了解国民对总统选举、传染病等公共事件的反应。最近关于COVID-19情绪分析的工作提出了一种面向领域的双向编码器表示转换器(BERT)语言模型,即COVID-Twitter BERT (CT-BERT)。然而,直接使用基于bert的语言模型在文本分类方面几乎没有改进。因此,提出了一种基于BERT的辅助方法。该方法将单句分类转化为对句分类,解决了BERT在文本分类任务中的性能问题。在本文中,我们将来自COVID相关推文的预训练BERT模型与辅助句方法相结合,以获得更好的COVID推文情感分析分类性能。我们表明,将单句分类转换为成对分类扩展了数据集,并获得了更高的准确率和F1分数。然而,我们期望特定于领域的语言模型比一般的语言模型表现得更好。在我们的结果中,我们表明CT-BERT的表现不一定优于BERT,特别是在理解情绪方面。
本文章由计算机程序翻译,如有差异,请以英文原文为准。
求助全文
约1分钟内获得全文 求助全文
来源期刊
自引率
0.00%
发文量
0
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
确定
请完成安全验证×
copy
已复制链接
快去分享给好友吧!
我知道了
右上角分享
点击右上角分享
0
联系我们:info@booksci.cn Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。 Copyright © 2023 布克学术 All rights reserved.
京ICP备2023020795号-1
ghs 京公网安备 11010802042870号
Book学术文献互助
Book学术文献互助群
群 号:481959085
Book学术官方微信