Visual Sentiment Prediction by Merging Hand-Craft and CNN Features

Wang Fengjiao, Masaki Aono
{"title":"Visual Sentiment Prediction by Merging Hand-Craft and CNN Features","authors":"Wang Fengjiao, Masaki Aono","doi":"10.1109/ICAICTA.2018.8541312","DOIUrl":null,"url":null,"abstract":"Nowadays, more and more people are getting used to social media such as Instagram, Facebook, Twitter, and Flickr to post images and texts to express their sentiment and emotions on almost all events and subjects. In consequence, analyzing sentiment of the huge number of images and texts on social networks has become more indispensable. Most of current research has focused on analyzing sentiment of textual data, while only few research has focused on sentiment analysis of image data. Some of these research has considered handcraft image features, the others has utilized Convolutional Neural Network (CNN) features. However, no research to our knowledge has considered mixing both hand-craft and CNN features. In this paper, we attempt to merge CNN which has shown remarkable achievements in Computer Vision recently, with handcraft features such as Color Histogram (CH) and Bag-of-Visual Words (BoVW) with some local features such as SURF and SIFT to predict sentiment of images. Furthermore, because it is often the case that the large amount of training data may not be easily obtained in the area of visual sentiment, we employ both data augmentation and transfer learning from a pre-trained CNN such as VGG16 trained with ImageNet dataset. With the handshake of hand-craft and End-to-End features from CNN, we attempt to attain the improvement of the performance of the proposed visual sentiment prediction framework. We conducted experiments on an image dataset from Twitter with polarity labels (\"positive\" and \"negative\"). The results of experiments demonstrate that our proposed visual sentimental prediction framework outperforms the current state-of-the-art methods.","PeriodicalId":184882,"journal":{"name":"2018 5th International Conference on Advanced Informatics: Concept Theory and Applications (ICAICTA)","volume":"1 1","pages":"0"},"PeriodicalIF":0.0000,"publicationDate":"2018-08-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"11","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"2018 5th International Conference on Advanced Informatics: Concept Theory and Applications (ICAICTA)","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/10.1109/ICAICTA.2018.8541312","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
引用次数: 11

Abstract

Nowadays, more and more people are getting used to social media such as Instagram, Facebook, Twitter, and Flickr to post images and texts to express their sentiment and emotions on almost all events and subjects. In consequence, analyzing sentiment of the huge number of images and texts on social networks has become more indispensable. Most of current research has focused on analyzing sentiment of textual data, while only few research has focused on sentiment analysis of image data. Some of these research has considered handcraft image features, the others has utilized Convolutional Neural Network (CNN) features. However, no research to our knowledge has considered mixing both hand-craft and CNN features. In this paper, we attempt to merge CNN which has shown remarkable achievements in Computer Vision recently, with handcraft features such as Color Histogram (CH) and Bag-of-Visual Words (BoVW) with some local features such as SURF and SIFT to predict sentiment of images. Furthermore, because it is often the case that the large amount of training data may not be easily obtained in the area of visual sentiment, we employ both data augmentation and transfer learning from a pre-trained CNN such as VGG16 trained with ImageNet dataset. With the handshake of hand-craft and End-to-End features from CNN, we attempt to attain the improvement of the performance of the proposed visual sentiment prediction framework. We conducted experiments on an image dataset from Twitter with polarity labels ("positive" and "negative"). The results of experiments demonstrate that our proposed visual sentimental prediction framework outperforms the current state-of-the-art methods.
结合手工和CNN特征的视觉情感预测
如今,越来越多的人习惯了社交媒体,如Instagram, Facebook, Twitter和Flickr,通过发布图片和文本来表达他们对几乎所有事件和主题的情绪和情感。因此,对社交网络上大量的图片和文字进行情感分析变得更加不可或缺。目前的研究大多集中在文本数据的情感分析上,而对图像数据情感分析的研究很少。其中一些研究考虑了手工图像特征,其他研究则利用了卷积神经网络(CNN)特征。然而,据我们所知,没有研究考虑将手工和CNN特征混合在一起。在本文中,我们尝试将近年来在计算机视觉领域取得显著成就的CNN与手工特征(如颜色直方图(CH)和视觉词袋(BoVW))以及一些局部特征(如SURF和SIFT)合并,以预测图像的情感。此外,由于在视觉情感领域通常不容易获得大量的训练数据,因此我们使用数据增强和从预训练的CNN(如使用ImageNet数据集训练的VGG16)中迁移学习。我们尝试利用hand-craft的握手和CNN的端到端特征来提高所提出的视觉情感预测框架的性能。我们在带有极性标签(“正”和“负”)的Twitter图像数据集上进行了实验。实验结果表明,我们提出的视觉情感预测框架优于目前最先进的方法。
本文章由计算机程序翻译,如有差异,请以英文原文为准。
求助全文
约1分钟内获得全文 求助全文
来源期刊
自引率
0.00%
发文量
0
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
确定
请完成安全验证×
copy
已复制链接
快去分享给好友吧!
我知道了
右上角分享
点击右上角分享
0
联系我们:info@booksci.cn Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。 Copyright © 2023 布克学术 All rights reserved.
京ICP备2023020795号-1
ghs 京公网安备 11010802042870号
Book学术文献互助
Book学术文献互助群
群 号:481959085
Book学术官方微信