Improved Multi-attention Neural Networks for Image Emotion Regression and the Initial Introduction of CAPS

Rending Wang, Dongmei Ma
{"title":"Improved Multi-attention Neural Networks for Image Emotion Regression and the Initial Introduction of CAPS","authors":"Rending Wang, Dongmei Ma","doi":"10.54097/92w2rc31","DOIUrl":null,"url":null,"abstract":"Image sentiment analysis is a large class of tasks for classifying or regressing images containing emotional stimuli, and it is believed in psychological research that different groups produce different emotions for the same stimuli. In order to study the influence of cultural background on image sentiment analysis, it is necessary to introduce a dataset of image sentiment stimuli that can represent cultural groups. In this paper, we introduce the Chinese Affective Picture System (CAPS), which represents Chinese culture, and revise and test this dataset. The PDANet model has the best performance among the current image sentiment regression models, but due to the difficulty of extracting cross-channel information from the attention module it uses, image long-distance information correlation and other shortcomings, this paper proposes an image emotion regression multiple attention networks, introduces the SimAM attention mechanism, and improves the loss function to make it more consistent with the psychological theory, and proposes a 10-fold cross-validation for CAPS. The network achieves MSE=0.0188, R2=0.359 on IAPS, and MSE=0.0169, R2=0.463 on NAPS, which is better than PDANet; the best training result of CAPS is MSE=0.0083, R2=0.625, and the paired-sample t-test of the results shows that all the three dimensions are significantly positively correlated, with correlation coefficients r=0.942, 0.895 and 0.943, respectively, showing good internal consistency and excellent application prospect of CAPS.","PeriodicalId":504530,"journal":{"name":"Frontiers in Computing and Intelligent Systems","volume":null,"pages":null},"PeriodicalIF":0.0000,"publicationDate":"2024-05-10","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"Frontiers in Computing and Intelligent Systems","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/10.54097/92w2rc31","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
引用次数: 0

Abstract

Image sentiment analysis is a large class of tasks for classifying or regressing images containing emotional stimuli, and it is believed in psychological research that different groups produce different emotions for the same stimuli. In order to study the influence of cultural background on image sentiment analysis, it is necessary to introduce a dataset of image sentiment stimuli that can represent cultural groups. In this paper, we introduce the Chinese Affective Picture System (CAPS), which represents Chinese culture, and revise and test this dataset. The PDANet model has the best performance among the current image sentiment regression models, but due to the difficulty of extracting cross-channel information from the attention module it uses, image long-distance information correlation and other shortcomings, this paper proposes an image emotion regression multiple attention networks, introduces the SimAM attention mechanism, and improves the loss function to make it more consistent with the psychological theory, and proposes a 10-fold cross-validation for CAPS. The network achieves MSE=0.0188, R2=0.359 on IAPS, and MSE=0.0169, R2=0.463 on NAPS, which is better than PDANet; the best training result of CAPS is MSE=0.0083, R2=0.625, and the paired-sample t-test of the results shows that all the three dimensions are significantly positively correlated, with correlation coefficients r=0.942, 0.895 and 0.943, respectively, showing good internal consistency and excellent application prospect of CAPS.
用于图像情感回归的改进型多注意神经网络和 CAPS 的初步引入
图像情感分析是对包含情感刺激的图像进行分类或回归的一大类任务,心理学研究认为,不同群体对相同的刺激会产生不同的情感。为了研究文化背景对图像情感分析的影响,有必要引入一个能代表文化群体的图像情感刺激数据集。本文引入了代表中国文化的中国情感图像系统(CAPS),并对该数据集进行了修订和测试。PDANet 模型在目前的图像情感回归模型中性能最好,但由于其使用的注意力模块难以提取跨通道信息、图像长距离信息相关等缺点,本文提出了一种图像情感回归多重注意力网络,引入了 SimAM 注意机制,并改进了损失函数,使其更符合心理学理论,并针对 CAPS 提出了 10 倍交叉验证。该网络在IAPS上达到MSE=0.0188,R2=0.359,在NAPS上达到MSE=0.0169,R2=0.463,优于PDANet;CAPS的最佳训练结果为MSE=0.0083,R2=0.625,结果的配对样本t检验表明三个维度均显著正相关,相关系数r分别为0.942、0.895和0.943,显示了CAPS良好的内部一致性和极好的应用前景。
本文章由计算机程序翻译,如有差异,请以英文原文为准。
求助全文
约1分钟内获得全文 求助全文
来源期刊
自引率
0.00%
发文量
0
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
确定
请完成安全验证×
copy
已复制链接
快去分享给好友吧!
我知道了
右上角分享
点击右上角分享
0
联系我们:info@booksci.cn Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。 Copyright © 2023 布克学术 All rights reserved.
京ICP备2023020795号-1
ghs 京公网安备 11010802042870号
Book学术文献互助
Book学术文献互助群
群 号:481959085
Book学术官方微信