A multimodal sentiment recognition method based on attention mechanism

Bo Liu, Jidong Zhang, Yuxiao Xu, Jianqiang Li, Yan Pei, Guanzhi Qu
{"title":"A multimodal sentiment recognition method based on attention mechanism","authors":"Bo Liu, Jidong Zhang, Yuxiao Xu, Jianqiang Li, Yan Pei, Guanzhi Qu","doi":"10.1145/3582099.3582131","DOIUrl":null,"url":null,"abstract":"Effective sentiment analysis on social media data can help to better understand the public's sentiment and opinion tendencies. Combining multimodal content for sentiment classification uses the correlation information of data between modalities, thereby avoiding the situation that a single modality does not fully grasp the overall sentiment. This paper proposes a multimodal sentiment recognition model based on the attention mechanism. Through transfer learning, the latest pre-trained model is used to extract preliminary features of text and image, and the attention mechanism is deployed to achieve further feature extraction of prominent image key regions and text keywords, better mining the internal information of modalities and learning the interaction between modalities. In view of the different contribution of each modal to sentiment classification, a decision-level fusion method is proposed to design fusion rules to integrate the classification results of each modal to obtain the final sentiment recognition result. This model integrates various unimodal features well, and effectively mines the emotional information expressed in Internet social media comments. This method is experimentally tested on the Twitter dataset, and the results show that the classification accuracy of sentiment recognition is significantly improved compared with the single-modal method.","PeriodicalId":222372,"journal":{"name":"Proceedings of the 2022 5th Artificial Intelligence and Cloud Computing Conference","volume":"29 1","pages":"0"},"PeriodicalIF":0.0000,"publicationDate":"2022-12-17","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"Proceedings of the 2022 5th Artificial Intelligence and Cloud Computing Conference","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/10.1145/3582099.3582131","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
引用次数: 0

Abstract

Effective sentiment analysis on social media data can help to better understand the public's sentiment and opinion tendencies. Combining multimodal content for sentiment classification uses the correlation information of data between modalities, thereby avoiding the situation that a single modality does not fully grasp the overall sentiment. This paper proposes a multimodal sentiment recognition model based on the attention mechanism. Through transfer learning, the latest pre-trained model is used to extract preliminary features of text and image, and the attention mechanism is deployed to achieve further feature extraction of prominent image key regions and text keywords, better mining the internal information of modalities and learning the interaction between modalities. In view of the different contribution of each modal to sentiment classification, a decision-level fusion method is proposed to design fusion rules to integrate the classification results of each modal to obtain the final sentiment recognition result. This model integrates various unimodal features well, and effectively mines the emotional information expressed in Internet social media comments. This method is experimentally tested on the Twitter dataset, and the results show that the classification accuracy of sentiment recognition is significantly improved compared with the single-modal method.
一种基于注意机制的多模态情感识别方法
对社交媒体数据进行有效的情绪分析,可以更好地了解公众的情绪和意见倾向。结合多模态内容进行情感分类利用了模态之间数据的相关信息,避免了单一模态不能完全把握整体情感的情况。提出了一种基于注意机制的多模态情感识别模型。通过迁移学习,利用最新的预训练模型提取文本和图像的初步特征,并利用注意机制实现对图像突出关键区域和文本关键词的进一步特征提取,更好地挖掘模态的内部信息,学习模态之间的交互。鉴于各模态对情感分类的贡献不同,提出了一种决策级融合方法,设计融合规则,将各模态的分类结果进行融合,得到最终的情感识别结果。该模型很好地整合了各种单峰特征,有效地挖掘了网络社交媒体评论中表达的情感信息。该方法在Twitter数据集上进行了实验测试,结果表明,与单模态方法相比,情感识别的分类准确率有显著提高。
本文章由计算机程序翻译,如有差异,请以英文原文为准。
求助全文
约1分钟内获得全文 求助全文
来源期刊
自引率
0.00%
发文量
0
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
确定
请完成安全验证×
copy
已复制链接
快去分享给好友吧!
我知道了
右上角分享
点击右上角分享
0
联系我们:info@booksci.cn Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。 Copyright © 2023 布克学术 All rights reserved.
京ICP备2023020795号-1
ghs 京公网安备 11010802042870号
Book学术文献互助
Book学术文献互助群
群 号:604180095
Book学术官方微信