基于内容图像检索和多输入卷积神经网络的视觉情感分类混合模型

IF 3.7 2区 计算机科学 Q1 COMPUTER SCIENCE, ARTIFICIAL INTELLIGENCE
Israa K. Salman Al-Tameemi, Mohammad-Reza Feizi-Derakhshi, Zari Farhadi, Amir-Reza Feizi-Derakhshi
{"title":"基于内容图像检索和多输入卷积神经网络的视觉情感分类混合模型","authors":"Israa K. Salman Al-Tameemi,&nbsp;Mohammad-Reza Feizi-Derakhshi,&nbsp;Zari Farhadi,&nbsp;Amir-Reza Feizi-Derakhshi","doi":"10.1155/int/5581601","DOIUrl":null,"url":null,"abstract":"<p>With the exponential growth of multimedia content, visual sentiment classification has emerged as a significant research area. However, it poses unique challenges due to the complexity and subjective nature of the visual information. This can be attributed to the significant presence of semantically ambiguous images within the current benchmark datasets, which enhances the performance of sentiment analysis but ignores the differences between various annotators. Moreover, most current methods concentrate on improving local emotional representations that focus on object extraction procedures rather than utilizing robust features that can effectively indicate the relevance of objects within an image through color information. Motivated by these observations, this paper addresses the need for efficient algorithms for labeling and classifying sentiment from visual images by introducing a novel hybrid model, which combines content-based image retrieval (CBIR) and a multi-input convolutional neural network (CNN). The CBIR model extracts color features from all dataset images, creating a numerical representation for each. It compares a query image to dataset images’ features to find similar features. This process continues until the images are grouped according to color similarity, which allows accurate sentimental categories based on similar features and feelings. Then, a multi-input CNN model is utilized to extract and efficiently incorporate high-level contextual visual information. This model comprises 70 layers, with six branches, each containing 11 layers. It seeks to facilitate the fusion of complementary information by incorporating multiple input categories that differ according to the color features extracted by the CBIR technique. This feature enables the model to understand the target and generate more precise predictions fully. The proposed model demonstrates significant improvements over existing algorithms, as evidenced by evaluations of six benchmark datasets of varying sizes. Also, it outperforms the state of the art in sentiment classification accuracy, getting 87.88%, 84.62%, 84.1%, 83.7%, 80.7%, and 91.2% accuracy for the EmotionROI, ArtPhoto, Twitter I, Twitter II, Abstract, and FI datasets, respectively. Furthermore, the model is evaluated on two newly collected large datasets, which confirm its scalability and robustness in handling large-scale sentiment classification tasks, and thus achieves a significant accuracy of 85.21% and 83.72% with the BGETTY and Twitter datasets, respectively. This paper contributes to the advancement of visual sentiment classification by offering a comprehensive solution for analyzing sentiment from images and laying the foundation for further research.</p>","PeriodicalId":14089,"journal":{"name":"International Journal of Intelligent Systems","volume":"2025 1","pages":""},"PeriodicalIF":3.7000,"publicationDate":"2025-10-16","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://onlinelibrary.wiley.com/doi/epdf/10.1155/int/5581601","citationCount":"0","resultStr":"{\"title\":\"Hybrid Model for Visual Sentiment Classification Using Content-Based Image Retrieval and Multi-Input Convolutional Neural Network\",\"authors\":\"Israa K. Salman Al-Tameemi,&nbsp;Mohammad-Reza Feizi-Derakhshi,&nbsp;Zari Farhadi,&nbsp;Amir-Reza Feizi-Derakhshi\",\"doi\":\"10.1155/int/5581601\",\"DOIUrl\":null,\"url\":null,\"abstract\":\"<p>With the exponential growth of multimedia content, visual sentiment classification has emerged as a significant research area. However, it poses unique challenges due to the complexity and subjective nature of the visual information. This can be attributed to the significant presence of semantically ambiguous images within the current benchmark datasets, which enhances the performance of sentiment analysis but ignores the differences between various annotators. Moreover, most current methods concentrate on improving local emotional representations that focus on object extraction procedures rather than utilizing robust features that can effectively indicate the relevance of objects within an image through color information. Motivated by these observations, this paper addresses the need for efficient algorithms for labeling and classifying sentiment from visual images by introducing a novel hybrid model, which combines content-based image retrieval (CBIR) and a multi-input convolutional neural network (CNN). The CBIR model extracts color features from all dataset images, creating a numerical representation for each. It compares a query image to dataset images’ features to find similar features. This process continues until the images are grouped according to color similarity, which allows accurate sentimental categories based on similar features and feelings. Then, a multi-input CNN model is utilized to extract and efficiently incorporate high-level contextual visual information. This model comprises 70 layers, with six branches, each containing 11 layers. It seeks to facilitate the fusion of complementary information by incorporating multiple input categories that differ according to the color features extracted by the CBIR technique. This feature enables the model to understand the target and generate more precise predictions fully. The proposed model demonstrates significant improvements over existing algorithms, as evidenced by evaluations of six benchmark datasets of varying sizes. Also, it outperforms the state of the art in sentiment classification accuracy, getting 87.88%, 84.62%, 84.1%, 83.7%, 80.7%, and 91.2% accuracy for the EmotionROI, ArtPhoto, Twitter I, Twitter II, Abstract, and FI datasets, respectively. Furthermore, the model is evaluated on two newly collected large datasets, which confirm its scalability and robustness in handling large-scale sentiment classification tasks, and thus achieves a significant accuracy of 85.21% and 83.72% with the BGETTY and Twitter datasets, respectively. This paper contributes to the advancement of visual sentiment classification by offering a comprehensive solution for analyzing sentiment from images and laying the foundation for further research.</p>\",\"PeriodicalId\":14089,\"journal\":{\"name\":\"International Journal of Intelligent Systems\",\"volume\":\"2025 1\",\"pages\":\"\"},\"PeriodicalIF\":3.7000,\"publicationDate\":\"2025-10-16\",\"publicationTypes\":\"Journal Article\",\"fieldsOfStudy\":null,\"isOpenAccess\":false,\"openAccessPdf\":\"https://onlinelibrary.wiley.com/doi/epdf/10.1155/int/5581601\",\"citationCount\":\"0\",\"resultStr\":null,\"platform\":\"Semanticscholar\",\"paperid\":null,\"PeriodicalName\":\"International Journal of Intelligent Systems\",\"FirstCategoryId\":\"94\",\"ListUrlMain\":\"https://onlinelibrary.wiley.com/doi/10.1155/int/5581601\",\"RegionNum\":2,\"RegionCategory\":\"计算机科学\",\"ArticlePicture\":[],\"TitleCN\":null,\"AbstractTextCN\":null,\"PMCID\":null,\"EPubDate\":\"\",\"PubModel\":\"\",\"JCR\":\"Q1\",\"JCRName\":\"COMPUTER SCIENCE, ARTIFICIAL INTELLIGENCE\",\"Score\":null,\"Total\":0}","platform":"Semanticscholar","paperid":null,"PeriodicalName":"International Journal of Intelligent Systems","FirstCategoryId":"94","ListUrlMain":"https://onlinelibrary.wiley.com/doi/10.1155/int/5581601","RegionNum":2,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"Q1","JCRName":"COMPUTER SCIENCE, ARTIFICIAL INTELLIGENCE","Score":null,"Total":0}
引用次数: 0

摘要

随着多媒体内容呈指数级增长,视觉情感分类已成为一个重要的研究领域。然而,由于视觉信息的复杂性和主观性,它提出了独特的挑战。这可以归因于当前基准数据集中存在语义模糊的图像,这增强了情感分析的性能,但忽略了不同注释器之间的差异。此外,目前的大多数方法都集中在改进局部情感表征上,这些表征关注的是对象提取过程,而不是利用鲁棒特征,通过颜色信息有效地指示图像中对象的相关性。基于这些观察结果,本文通过引入一种结合基于内容的图像检索(CBIR)和多输入卷积神经网络(CNN)的新型混合模型,解决了对视觉图像情感标记和分类的高效算法的需求。CBIR模型从所有数据集图像中提取颜色特征,为每个图像创建一个数字表示。它将查询图像与数据集图像的特征进行比较,以找到相似的特征。这个过程一直持续到图像根据颜色相似度分组,这样就可以根据相似的特征和感觉进行准确的情感分类。然后,利用多输入CNN模型提取并高效融合高级上下文视觉信息。该模型包括70层,6个分支,每个分支包含11层。它寻求通过合并根据CBIR技术提取的颜色特征不同的多个输入类别来促进互补信息的融合。这一特性使模型能够充分理解目标并生成更精确的预测。通过对六个不同大小的基准数据集的评估,证明了所提出的模型比现有算法有了显著的改进。此外,它在情感分类准确率方面也优于目前的技术水平,在EmotionROI、ArtPhoto、Twitter I、Twitter II、Abstract和FI数据集上分别获得87.88%、84.62%、84.1%、83.7%、80.7%和91.2%的准确率。此外,在两个新收集的大型数据集上对该模型进行了评估,验证了该模型在处理大规模情感分类任务方面的可扩展性和鲁棒性,在BGETTY和Twitter数据集上分别达到了85.21%和83.72%的显著准确率。本文为图像情感分析提供了一个全面的解决方案,为进一步的研究奠定了基础,有助于视觉情感分类的发展。
本文章由计算机程序翻译,如有差异,请以英文原文为准。

Hybrid Model for Visual Sentiment Classification Using Content-Based Image Retrieval and Multi-Input Convolutional Neural Network

Hybrid Model for Visual Sentiment Classification Using Content-Based Image Retrieval and Multi-Input Convolutional Neural Network

With the exponential growth of multimedia content, visual sentiment classification has emerged as a significant research area. However, it poses unique challenges due to the complexity and subjective nature of the visual information. This can be attributed to the significant presence of semantically ambiguous images within the current benchmark datasets, which enhances the performance of sentiment analysis but ignores the differences between various annotators. Moreover, most current methods concentrate on improving local emotional representations that focus on object extraction procedures rather than utilizing robust features that can effectively indicate the relevance of objects within an image through color information. Motivated by these observations, this paper addresses the need for efficient algorithms for labeling and classifying sentiment from visual images by introducing a novel hybrid model, which combines content-based image retrieval (CBIR) and a multi-input convolutional neural network (CNN). The CBIR model extracts color features from all dataset images, creating a numerical representation for each. It compares a query image to dataset images’ features to find similar features. This process continues until the images are grouped according to color similarity, which allows accurate sentimental categories based on similar features and feelings. Then, a multi-input CNN model is utilized to extract and efficiently incorporate high-level contextual visual information. This model comprises 70 layers, with six branches, each containing 11 layers. It seeks to facilitate the fusion of complementary information by incorporating multiple input categories that differ according to the color features extracted by the CBIR technique. This feature enables the model to understand the target and generate more precise predictions fully. The proposed model demonstrates significant improvements over existing algorithms, as evidenced by evaluations of six benchmark datasets of varying sizes. Also, it outperforms the state of the art in sentiment classification accuracy, getting 87.88%, 84.62%, 84.1%, 83.7%, 80.7%, and 91.2% accuracy for the EmotionROI, ArtPhoto, Twitter I, Twitter II, Abstract, and FI datasets, respectively. Furthermore, the model is evaluated on two newly collected large datasets, which confirm its scalability and robustness in handling large-scale sentiment classification tasks, and thus achieves a significant accuracy of 85.21% and 83.72% with the BGETTY and Twitter datasets, respectively. This paper contributes to the advancement of visual sentiment classification by offering a comprehensive solution for analyzing sentiment from images and laying the foundation for further research.

求助全文
通过发布文献求助,成功后即可免费获取论文全文。 去求助
来源期刊
International Journal of Intelligent Systems
International Journal of Intelligent Systems 工程技术-计算机:人工智能
CiteScore
11.30
自引率
14.30%
发文量
304
审稿时长
9 months
期刊介绍: The International Journal of Intelligent Systems serves as a forum for individuals interested in tapping into the vast theories based on intelligent systems construction. With its peer-reviewed format, the journal explores several fascinating editorials written by today''s experts in the field. Because new developments are being introduced each day, there''s much to be learned — examination, analysis creation, information retrieval, man–computer interactions, and more. The International Journal of Intelligent Systems uses charts and illustrations to demonstrate these ground-breaking issues, and encourages readers to share their thoughts and experiences.
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
确定
请完成安全验证×
copy
已复制链接
快去分享给好友吧!
我知道了
右上角分享
点击右上角分享
0
联系我们:info@booksci.cn Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。 Copyright © 2023 布克学术 All rights reserved.
京ICP备2023020795号-1
ghs 京公网安备 11010802042870号
Book学术文献互助
Book学术文献互助群
群 号:604180095
Book学术官方微信