Sentiment Analysis of Persian Instagram Post: a Multimodal Deep Learning Approach

Aria Naseri Karimvand, R. Chegeni, Mohammad Ehsan Basiri, Shahla Nemati
{"title":"Sentiment Analysis of Persian Instagram Post: a Multimodal Deep Learning Approach","authors":"Aria Naseri Karimvand, R. Chegeni, Mohammad Ehsan Basiri, Shahla Nemati","doi":"10.1109/ICWR51868.2021.9443026","DOIUrl":null,"url":null,"abstract":"Instagram is a popular social media that has a wide range of active users from ordinary people to artists and official users. Instagram posts are widely used by users to share text, image or video. Many users use text to describe or complement the images they share. To analyze the sentiment of such posts, both the content of the text and the image should be considered at the same time. This requires modelling of the relationship between the text and image modalities. To address this problem, we propose a multimodal deep learning method. The proposed method utilizes a bi-directional gated recurrent unit (bi-GRU) for processing text comments and a 2-dimensional convolutional neural network (2CNN) for analyzing images. In order to assess the performance of the proposed model, we introduce a new dataset of Instagram posts, MPerInst, containing 512 pairs of images and their corresponding comments written in the Persian language. Implementation results shows that employing both text and image modalities improves polarity detection accuracy and F1-scrore by 23% and 0.24 compared to using only image and text modalities, respectively. Moreover, the proposed model outperforms 11 similar deep fusion models by 11% and 0.1 in terms of accuracy and F1-score. Both the dataset and the codes of our proposed model are publicly available for probable future use.","PeriodicalId":377597,"journal":{"name":"2021 7th International Conference on Web Research (ICWR)","volume":"1 1","pages":"0"},"PeriodicalIF":0.0000,"publicationDate":"2021-05-19","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"7","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"2021 7th International Conference on Web Research (ICWR)","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/10.1109/ICWR51868.2021.9443026","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
引用次数: 7

Abstract

Instagram is a popular social media that has a wide range of active users from ordinary people to artists and official users. Instagram posts are widely used by users to share text, image or video. Many users use text to describe or complement the images they share. To analyze the sentiment of such posts, both the content of the text and the image should be considered at the same time. This requires modelling of the relationship between the text and image modalities. To address this problem, we propose a multimodal deep learning method. The proposed method utilizes a bi-directional gated recurrent unit (bi-GRU) for processing text comments and a 2-dimensional convolutional neural network (2CNN) for analyzing images. In order to assess the performance of the proposed model, we introduce a new dataset of Instagram posts, MPerInst, containing 512 pairs of images and their corresponding comments written in the Persian language. Implementation results shows that employing both text and image modalities improves polarity detection accuracy and F1-scrore by 23% and 0.24 compared to using only image and text modalities, respectively. Moreover, the proposed model outperforms 11 similar deep fusion models by 11% and 0.1 in terms of accuracy and F1-score. Both the dataset and the codes of our proposed model are publicly available for probable future use.
波斯语Instagram帖子的情感分析:一种多模式深度学习方法
Instagram是一个很受欢迎的社交媒体,活跃用户范围很广,从普通人到艺术家和官方用户。Instagram帖子被用户广泛用于分享文字、图片或视频。许多用户使用文字来描述或补充他们分享的图片。要分析这类帖子的情感,需要同时考虑文字和图片的内容。这需要对文本和图像模式之间的关系进行建模。为了解决这个问题,我们提出了一种多模态深度学习方法。该方法使用双向门控循环单元(bi-GRU)处理文本注释,使用二维卷积神经网络(2CNN)分析图像。为了评估所提出模型的性能,我们引入了一个新的Instagram帖子数据集MPerInst,其中包含512对图像及其对应的波斯语评论。实现结果表明,与仅使用图像和文本模式相比,同时使用文本和图像模式可将极性检测精度和f1分数分别提高23%和0.24。此外,该模型在精度和f1评分方面分别比11个类似的深度融合模型高出11%和0.1。我们提出的模型的数据集和代码都是公开的,以便将来可能使用。
本文章由计算机程序翻译,如有差异,请以英文原文为准。
求助全文
约1分钟内获得全文 求助全文
来源期刊
自引率
0.00%
发文量
0
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
确定
请完成安全验证×
copy
已复制链接
快去分享给好友吧!
我知道了
右上角分享
点击右上角分享
0
联系我们:info@booksci.cn Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。 Copyright © 2023 布克学术 All rights reserved.
京ICP备2023020795号-1
ghs 京公网安备 11010802042870号
Book学术文献互助
Book学术文献互助群
群 号:604180095
Book学术官方微信