Xin Nie;Laurence T. Yang;Zhe Li;Xianjun Deng;Fulan Fan;Zecan Yang
{"title":"Interpretable Multimodal Tucker Fusion Model With Information Filtering for Multimodal Sentiment Analysis","authors":"Xin Nie;Laurence T. Yang;Zhe Li;Xianjun Deng;Fulan Fan;Zecan Yang","doi":"10.1109/TCSS.2024.3459929","DOIUrl":null,"url":null,"abstract":"Multimodal sentiment analysis (MSA) integrates multiple sources of sentiment information for processing and has demonstrated superior performance compared to single-modal sentiment analysis, making it widely applicable in domains such as human–computer interaction and public opinion supervision. However, current MSA models heavily rely on black-box deep learning (DL) methods, which lack interpretability. Additionally, effectively integrating multimodal data, reducing noise and redundancy, as well as bridging the semantic gap between heterogeneous data remain challenging issues in multimodal DL. To address these challenges, we propose an interpretable multimodal Tucker fusion model with information filtering (IMTFMIF). We are the first to utilize the multimodal Tucker fusion model for MSA tasks. This approach maps multimodal data into a unified tensor space for fusion, effectively reducing modal heterogeneity and eliminating redundant information while maintaining interpretability. Furthermore, mutual information is employed to filter out task-irrelevant information and explain the association between input and output from an information flow perspective. We propose a novel approach to enhance the comprehension of multimodal data and optimize model performance in MSA tasks. Finally, extensive experiments conducted on three public multimodal datasets demonstrate that our proposed IMTFMIF achieves competitive performance compared to state-of-the-art methods.","PeriodicalId":13044,"journal":{"name":"IEEE Transactions on Computational Social Systems","volume":"12 3","pages":"1351-1364"},"PeriodicalIF":4.5000,"publicationDate":"2024-12-24","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"IEEE Transactions on Computational Social Systems","FirstCategoryId":"94","ListUrlMain":"https://ieeexplore.ieee.org/document/10813576/","RegionNum":2,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"Q1","JCRName":"COMPUTER SCIENCE, CYBERNETICS","Score":null,"Total":0}
引用次数: 0
Abstract
Multimodal sentiment analysis (MSA) integrates multiple sources of sentiment information for processing and has demonstrated superior performance compared to single-modal sentiment analysis, making it widely applicable in domains such as human–computer interaction and public opinion supervision. However, current MSA models heavily rely on black-box deep learning (DL) methods, which lack interpretability. Additionally, effectively integrating multimodal data, reducing noise and redundancy, as well as bridging the semantic gap between heterogeneous data remain challenging issues in multimodal DL. To address these challenges, we propose an interpretable multimodal Tucker fusion model with information filtering (IMTFMIF). We are the first to utilize the multimodal Tucker fusion model for MSA tasks. This approach maps multimodal data into a unified tensor space for fusion, effectively reducing modal heterogeneity and eliminating redundant information while maintaining interpretability. Furthermore, mutual information is employed to filter out task-irrelevant information and explain the association between input and output from an information flow perspective. We propose a novel approach to enhance the comprehension of multimodal data and optimize model performance in MSA tasks. Finally, extensive experiments conducted on three public multimodal datasets demonstrate that our proposed IMTFMIF achieves competitive performance compared to state-of-the-art methods.
期刊介绍:
IEEE Transactions on Computational Social Systems focuses on such topics as modeling, simulation, analysis and understanding of social systems from the quantitative and/or computational perspective. "Systems" include man-man, man-machine and machine-machine organizations and adversarial situations as well as social media structures and their dynamics. More specifically, the proposed transactions publishes articles on modeling the dynamics of social systems, methodologies for incorporating and representing socio-cultural and behavioral aspects in computational modeling, analysis of social system behavior and structure, and paradigms for social systems modeling and simulation. The journal also features articles on social network dynamics, social intelligence and cognition, social systems design and architectures, socio-cultural modeling and representation, and computational behavior modeling, and their applications.