Aria Naseri Karimvand, R. Chegeni, Mohammad Ehsan Basiri, Shahla Nemati
{"title":"波斯语Instagram帖子的情感分析:一种多模式深度学习方法","authors":"Aria Naseri Karimvand, R. Chegeni, Mohammad Ehsan Basiri, Shahla Nemati","doi":"10.1109/ICWR51868.2021.9443026","DOIUrl":null,"url":null,"abstract":"Instagram is a popular social media that has a wide range of active users from ordinary people to artists and official users. Instagram posts are widely used by users to share text, image or video. Many users use text to describe or complement the images they share. To analyze the sentiment of such posts, both the content of the text and the image should be considered at the same time. This requires modelling of the relationship between the text and image modalities. To address this problem, we propose a multimodal deep learning method. The proposed method utilizes a bi-directional gated recurrent unit (bi-GRU) for processing text comments and a 2-dimensional convolutional neural network (2CNN) for analyzing images. In order to assess the performance of the proposed model, we introduce a new dataset of Instagram posts, MPerInst, containing 512 pairs of images and their corresponding comments written in the Persian language. Implementation results shows that employing both text and image modalities improves polarity detection accuracy and F1-scrore by 23% and 0.24 compared to using only image and text modalities, respectively. Moreover, the proposed model outperforms 11 similar deep fusion models by 11% and 0.1 in terms of accuracy and F1-score. Both the dataset and the codes of our proposed model are publicly available for probable future use.","PeriodicalId":377597,"journal":{"name":"2021 7th International Conference on Web Research (ICWR)","volume":"1 1","pages":"0"},"PeriodicalIF":0.0000,"publicationDate":"2021-05-19","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"7","resultStr":"{\"title\":\"Sentiment Analysis of Persian Instagram Post: a Multimodal Deep Learning Approach\",\"authors\":\"Aria Naseri Karimvand, R. Chegeni, Mohammad Ehsan Basiri, Shahla Nemati\",\"doi\":\"10.1109/ICWR51868.2021.9443026\",\"DOIUrl\":null,\"url\":null,\"abstract\":\"Instagram is a popular social media that has a wide range of active users from ordinary people to artists and official users. Instagram posts are widely used by users to share text, image or video. Many users use text to describe or complement the images they share. To analyze the sentiment of such posts, both the content of the text and the image should be considered at the same time. This requires modelling of the relationship between the text and image modalities. To address this problem, we propose a multimodal deep learning method. The proposed method utilizes a bi-directional gated recurrent unit (bi-GRU) for processing text comments and a 2-dimensional convolutional neural network (2CNN) for analyzing images. In order to assess the performance of the proposed model, we introduce a new dataset of Instagram posts, MPerInst, containing 512 pairs of images and their corresponding comments written in the Persian language. Implementation results shows that employing both text and image modalities improves polarity detection accuracy and F1-scrore by 23% and 0.24 compared to using only image and text modalities, respectively. Moreover, the proposed model outperforms 11 similar deep fusion models by 11% and 0.1 in terms of accuracy and F1-score. Both the dataset and the codes of our proposed model are publicly available for probable future use.\",\"PeriodicalId\":377597,\"journal\":{\"name\":\"2021 7th International Conference on Web Research (ICWR)\",\"volume\":\"1 1\",\"pages\":\"0\"},\"PeriodicalIF\":0.0000,\"publicationDate\":\"2021-05-19\",\"publicationTypes\":\"Journal Article\",\"fieldsOfStudy\":null,\"isOpenAccess\":false,\"openAccessPdf\":\"\",\"citationCount\":\"7\",\"resultStr\":null,\"platform\":\"Semanticscholar\",\"paperid\":null,\"PeriodicalName\":\"2021 7th International Conference on Web Research (ICWR)\",\"FirstCategoryId\":\"1085\",\"ListUrlMain\":\"https://doi.org/10.1109/ICWR51868.2021.9443026\",\"RegionNum\":0,\"RegionCategory\":null,\"ArticlePicture\":[],\"TitleCN\":null,\"AbstractTextCN\":null,\"PMCID\":null,\"EPubDate\":\"\",\"PubModel\":\"\",\"JCR\":\"\",\"JCRName\":\"\",\"Score\":null,\"Total\":0}","platform":"Semanticscholar","paperid":null,"PeriodicalName":"2021 7th International Conference on Web Research (ICWR)","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/10.1109/ICWR51868.2021.9443026","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
Sentiment Analysis of Persian Instagram Post: a Multimodal Deep Learning Approach
Instagram is a popular social media that has a wide range of active users from ordinary people to artists and official users. Instagram posts are widely used by users to share text, image or video. Many users use text to describe or complement the images they share. To analyze the sentiment of such posts, both the content of the text and the image should be considered at the same time. This requires modelling of the relationship between the text and image modalities. To address this problem, we propose a multimodal deep learning method. The proposed method utilizes a bi-directional gated recurrent unit (bi-GRU) for processing text comments and a 2-dimensional convolutional neural network (2CNN) for analyzing images. In order to assess the performance of the proposed model, we introduce a new dataset of Instagram posts, MPerInst, containing 512 pairs of images and their corresponding comments written in the Persian language. Implementation results shows that employing both text and image modalities improves polarity detection accuracy and F1-scrore by 23% and 0.24 compared to using only image and text modalities, respectively. Moreover, the proposed model outperforms 11 similar deep fusion models by 11% and 0.1 in terms of accuracy and F1-score. Both the dataset and the codes of our proposed model are publicly available for probable future use.