{"title":"Deep Learning approach for text, image, and GIF multimodal sentiment analysis","authors":"Amirhossein Shirzad, Hadi Zare, M. Teimouri","doi":"10.1109/ICCKE50421.2020.9303676","DOIUrl":null,"url":null,"abstract":"Recently, social media users are increasingly using visual media like GIFs, videos and images as powerful means of expressing their emotions and sentiments. We created a multimodal sentiment analysis tool for different forms of tweets in Python to compute the sentiment score not only base on the tweet's text, but also to consider GIFs and images to gain better accuracy for the tweet's overall sentiment score. For image sentiment, we use fine-tuned CNN, for texts we use VADER, and for the GIFs we apply both image sentiment and facial expression analysis in every frame of the file. In our work, we demonstrate that by utilizing both textual and image features we can achieve better results compared to other models that only rely on either images or text features. The final sentiment score for the incoming tweets will be calculated by aggregating the output scores obtained from each of our text, image, and GIF modules.","PeriodicalId":402043,"journal":{"name":"2020 10th International Conference on Computer and Knowledge Engineering (ICCKE)","volume":"32 6","pages":"0"},"PeriodicalIF":0.0000,"publicationDate":"2020-10-29","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"2","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"2020 10th International Conference on Computer and Knowledge Engineering (ICCKE)","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/10.1109/ICCKE50421.2020.9303676","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
引用次数: 2
Abstract
Recently, social media users are increasingly using visual media like GIFs, videos and images as powerful means of expressing their emotions and sentiments. We created a multimodal sentiment analysis tool for different forms of tweets in Python to compute the sentiment score not only base on the tweet's text, but also to consider GIFs and images to gain better accuracy for the tweet's overall sentiment score. For image sentiment, we use fine-tuned CNN, for texts we use VADER, and for the GIFs we apply both image sentiment and facial expression analysis in every frame of the file. In our work, we demonstrate that by utilizing both textual and image features we can achieve better results compared to other models that only rely on either images or text features. The final sentiment score for the incoming tweets will be calculated by aggregating the output scores obtained from each of our text, image, and GIF modules.