跨模态BERT模型在心理社会网络中增强多模态情感分析。

IF 3 3区 心理学 Q1 PSYCHOLOGY, MULTIDISCIPLINARY
Jian Feng
{"title":"跨模态BERT模型在心理社会网络中增强多模态情感分析。","authors":"Jian Feng","doi":"10.1186/s40359-025-03443-z","DOIUrl":null,"url":null,"abstract":"<p><strong>Background: </strong>Human emotions in psychological social networks often involve complex interactions across multiple modalities. Information derived from various channels can synergistically complement one another, leading to a more nuanced depiction of an individual's emotional landscape. Multimodal sentiment analysis emerges as a potent tool to process this diverse array of content, facilitating efficient amalgamation of emotions and quantification of emotional intensity.</p><p><strong>Methods: </strong>This paper proposes a cross-modal BERT model and a cross-modal psychological-emotional fusion (CPEF) model for sentiment analysis, integrating visual, audio, and textual modalities. The model initially processes images and audio through dedicated sub-networks for feature extraction and reduction. These features are then passed through the Masked Multimodal Attention (MMA) module, which amalgamates image and audio features via self-attention, yielding a bimodal attention matrix. Subsequently, textual information is fed into the MMA module, undergoing feature extraction through a pre-trained BERT model. The textual information is then fused with the bimodal attention matrix via the pre-trained BERT model, facilitating emotional fusion across modalities.</p><p><strong>Results: </strong>The experimental results on the CMU-MOSEI dataset showcase the effectiveness of the proposed CPEF model, outperforming comparative models, achieving an impressive accuracy rate of 83.9% and F1 Score of 84.1%, notably improving the quantification of negative, neutral, and positive affective energy.</p><p><strong>Conclusions: </strong>Such advancements contribute to the precise detection of mental health status and the cultivation of a positive and sustainable social network environment.</p>","PeriodicalId":37867,"journal":{"name":"BMC Psychology","volume":"13 1","pages":"1081"},"PeriodicalIF":3.0000,"publicationDate":"2025-09-29","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.ncbi.nlm.nih.gov/pmc/articles/PMC12482056/pdf/","citationCount":"0","resultStr":"{\"title\":\"Cross-modal BERT model for enhanced multimodal sentiment analysis in psychological social networks.\",\"authors\":\"Jian Feng\",\"doi\":\"10.1186/s40359-025-03443-z\",\"DOIUrl\":null,\"url\":null,\"abstract\":\"<p><strong>Background: </strong>Human emotions in psychological social networks often involve complex interactions across multiple modalities. Information derived from various channels can synergistically complement one another, leading to a more nuanced depiction of an individual's emotional landscape. Multimodal sentiment analysis emerges as a potent tool to process this diverse array of content, facilitating efficient amalgamation of emotions and quantification of emotional intensity.</p><p><strong>Methods: </strong>This paper proposes a cross-modal BERT model and a cross-modal psychological-emotional fusion (CPEF) model for sentiment analysis, integrating visual, audio, and textual modalities. The model initially processes images and audio through dedicated sub-networks for feature extraction and reduction. These features are then passed through the Masked Multimodal Attention (MMA) module, which amalgamates image and audio features via self-attention, yielding a bimodal attention matrix. Subsequently, textual information is fed into the MMA module, undergoing feature extraction through a pre-trained BERT model. The textual information is then fused with the bimodal attention matrix via the pre-trained BERT model, facilitating emotional fusion across modalities.</p><p><strong>Results: </strong>The experimental results on the CMU-MOSEI dataset showcase the effectiveness of the proposed CPEF model, outperforming comparative models, achieving an impressive accuracy rate of 83.9% and F1 Score of 84.1%, notably improving the quantification of negative, neutral, and positive affective energy.</p><p><strong>Conclusions: </strong>Such advancements contribute to the precise detection of mental health status and the cultivation of a positive and sustainable social network environment.</p>\",\"PeriodicalId\":37867,\"journal\":{\"name\":\"BMC Psychology\",\"volume\":\"13 1\",\"pages\":\"1081\"},\"PeriodicalIF\":3.0000,\"publicationDate\":\"2025-09-29\",\"publicationTypes\":\"Journal Article\",\"fieldsOfStudy\":null,\"isOpenAccess\":false,\"openAccessPdf\":\"https://www.ncbi.nlm.nih.gov/pmc/articles/PMC12482056/pdf/\",\"citationCount\":\"0\",\"resultStr\":null,\"platform\":\"Semanticscholar\",\"paperid\":null,\"PeriodicalName\":\"BMC Psychology\",\"FirstCategoryId\":\"102\",\"ListUrlMain\":\"https://doi.org/10.1186/s40359-025-03443-z\",\"RegionNum\":3,\"RegionCategory\":\"心理学\",\"ArticlePicture\":[],\"TitleCN\":null,\"AbstractTextCN\":null,\"PMCID\":null,\"EPubDate\":\"\",\"PubModel\":\"\",\"JCR\":\"Q1\",\"JCRName\":\"PSYCHOLOGY, MULTIDISCIPLINARY\",\"Score\":null,\"Total\":0}","platform":"Semanticscholar","paperid":null,"PeriodicalName":"BMC Psychology","FirstCategoryId":"102","ListUrlMain":"https://doi.org/10.1186/s40359-025-03443-z","RegionNum":3,"RegionCategory":"心理学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"Q1","JCRName":"PSYCHOLOGY, MULTIDISCIPLINARY","Score":null,"Total":0}
引用次数: 0

摘要

背景:心理社会网络中的人类情绪通常涉及多种方式的复杂相互作用。来自不同渠道的信息可以相互补充,从而更细致地描述个人的情感景观。多模态情绪分析作为一种强有力的工具出现,可以处理这种多样化的内容,促进情绪的有效融合和情绪强度的量化。方法:结合视觉、音频和文本模式,提出了跨模态BERT模型和跨模态心理情感融合(CPEF)模型。该模型首先通过专用的子网络处理图像和音频,进行特征提取和约简。这些特征然后通过屏蔽多模态注意(MMA)模块,该模块通过自我注意合并图像和音频特征,产生双峰注意矩阵。随后,将文本信息输入MMA模块,通过预训练的BERT模型进行特征提取。然后通过预训练的BERT模型将文本信息与双峰注意矩阵融合,促进跨模态的情感融合。结果:CMU-MOSEI数据集上的实验结果显示了CPEF模型的有效性,优于对比模型,准确率达到83.9%,F1 Score达到84.1%,显著改善了负、中性和正情感能量的量化。结论:这些进步有助于心理健康状况的准确检测和积极、可持续的社会网络环境的培养。
本文章由计算机程序翻译,如有差异,请以英文原文为准。

Cross-modal BERT model for enhanced multimodal sentiment analysis in psychological social networks.

Cross-modal BERT model for enhanced multimodal sentiment analysis in psychological social networks.

Cross-modal BERT model for enhanced multimodal sentiment analysis in psychological social networks.

Cross-modal BERT model for enhanced multimodal sentiment analysis in psychological social networks.

Background: Human emotions in psychological social networks often involve complex interactions across multiple modalities. Information derived from various channels can synergistically complement one another, leading to a more nuanced depiction of an individual's emotional landscape. Multimodal sentiment analysis emerges as a potent tool to process this diverse array of content, facilitating efficient amalgamation of emotions and quantification of emotional intensity.

Methods: This paper proposes a cross-modal BERT model and a cross-modal psychological-emotional fusion (CPEF) model for sentiment analysis, integrating visual, audio, and textual modalities. The model initially processes images and audio through dedicated sub-networks for feature extraction and reduction. These features are then passed through the Masked Multimodal Attention (MMA) module, which amalgamates image and audio features via self-attention, yielding a bimodal attention matrix. Subsequently, textual information is fed into the MMA module, undergoing feature extraction through a pre-trained BERT model. The textual information is then fused with the bimodal attention matrix via the pre-trained BERT model, facilitating emotional fusion across modalities.

Results: The experimental results on the CMU-MOSEI dataset showcase the effectiveness of the proposed CPEF model, outperforming comparative models, achieving an impressive accuracy rate of 83.9% and F1 Score of 84.1%, notably improving the quantification of negative, neutral, and positive affective energy.

Conclusions: Such advancements contribute to the precise detection of mental health status and the cultivation of a positive and sustainable social network environment.

求助全文
通过发布文献求助,成功后即可免费获取论文全文。 去求助
来源期刊
BMC Psychology
BMC Psychology Psychology-Psychology (all)
CiteScore
3.90
自引率
2.80%
发文量
265
审稿时长
24 weeks
期刊介绍: BMC Psychology is an open access, peer-reviewed journal that considers manuscripts on all aspects of psychology, human behavior and the mind, including developmental, clinical, cognitive, experimental, health and social psychology, as well as personality and individual differences. The journal welcomes quantitative and qualitative research methods, including animal studies.
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
确定
请完成安全验证×
copy
已复制链接
快去分享给好友吧!
我知道了
右上角分享
点击右上角分享
0
联系我们:info@booksci.cn Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。 Copyright © 2023 布克学术 All rights reserved.
京ICP备2023020795号-1
ghs 京公网安备 11010802042870号
Book学术文献互助
Book学术文献互助群
群 号:604180095
Book学术官方微信