{"title":"基于上下文推理和情绪转移意识的图注意在对话中的情绪识别","authors":"Juan Yang, Puling Wei, Xu Du, Jun Shen","doi":"10.1007/s40747-025-01903-y","DOIUrl":null,"url":null,"abstract":"<p>Emotion recognition in conversations has recently emerged as a hot research topic owing to its increasingly important role in developing intelligent empathy services. Thoroughly exploring the conversational context and accurately capturing emotion-shift information are highly crucial for accurate emotion recognition in conversations. However, existing studies generally failed to fully understand the complex conversational context due to their insufficient capabilities in extracting and integrating emotional cues. Moreover, they mainly focused on the speaker’s emotion inertia while paying less attention to explore multi-perspective emotion-shift patterns. To address these limitations, this study proposes a novel multimodal approach, namely, GAT-CRESA (Graph ATtention based on Contextual Reasoning and Emotion-Shift Awareness). Specifically, the multi-turn global contextual reasoning module iteratively performs contextual perception and cognitive reasoning for efficiently understanding the global conversational context. Then, GAT-CRESA explores emotion-shift information among utterances from both the speaker-dependent and the global context-based perspectives. Next, the emotion-shift awareness graphs are constructed for extracting significant local-level conversational context, where edge relations are determined by the learnt emotion-shift labels. Finally, the outputs of graphs are concatenated for final emotion recognition. The loss of emotion prediction task is combined together with those of two perspective’s emotion-shift learning for guiding the training process. Experimental results show that our GAT-CRESA achieves new state-of-art records with 72.77% ACC and 72.81% wa-F1 on IEMOCAP, and 65.44% ACC and 65.04% wa-F1 on MELD, respectively. The ablation results also indicate the effectiveness and rationality of each component in our approach.</p>","PeriodicalId":10524,"journal":{"name":"Complex & Intelligent Systems","volume":"51 1","pages":""},"PeriodicalIF":5.0000,"publicationDate":"2025-05-12","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":"{\"title\":\"Graph attention based on contextual reasoning and emotion-shift awareness for emotion recognition in conversations\",\"authors\":\"Juan Yang, Puling Wei, Xu Du, Jun Shen\",\"doi\":\"10.1007/s40747-025-01903-y\",\"DOIUrl\":null,\"url\":null,\"abstract\":\"<p>Emotion recognition in conversations has recently emerged as a hot research topic owing to its increasingly important role in developing intelligent empathy services. Thoroughly exploring the conversational context and accurately capturing emotion-shift information are highly crucial for accurate emotion recognition in conversations. However, existing studies generally failed to fully understand the complex conversational context due to their insufficient capabilities in extracting and integrating emotional cues. Moreover, they mainly focused on the speaker’s emotion inertia while paying less attention to explore multi-perspective emotion-shift patterns. To address these limitations, this study proposes a novel multimodal approach, namely, GAT-CRESA (Graph ATtention based on Contextual Reasoning and Emotion-Shift Awareness). Specifically, the multi-turn global contextual reasoning module iteratively performs contextual perception and cognitive reasoning for efficiently understanding the global conversational context. Then, GAT-CRESA explores emotion-shift information among utterances from both the speaker-dependent and the global context-based perspectives. Next, the emotion-shift awareness graphs are constructed for extracting significant local-level conversational context, where edge relations are determined by the learnt emotion-shift labels. Finally, the outputs of graphs are concatenated for final emotion recognition. The loss of emotion prediction task is combined together with those of two perspective’s emotion-shift learning for guiding the training process. Experimental results show that our GAT-CRESA achieves new state-of-art records with 72.77% ACC and 72.81% wa-F1 on IEMOCAP, and 65.44% ACC and 65.04% wa-F1 on MELD, respectively. The ablation results also indicate the effectiveness and rationality of each component in our approach.</p>\",\"PeriodicalId\":10524,\"journal\":{\"name\":\"Complex & Intelligent Systems\",\"volume\":\"51 1\",\"pages\":\"\"},\"PeriodicalIF\":5.0000,\"publicationDate\":\"2025-05-12\",\"publicationTypes\":\"Journal Article\",\"fieldsOfStudy\":null,\"isOpenAccess\":false,\"openAccessPdf\":\"\",\"citationCount\":\"0\",\"resultStr\":null,\"platform\":\"Semanticscholar\",\"paperid\":null,\"PeriodicalName\":\"Complex & Intelligent Systems\",\"FirstCategoryId\":\"94\",\"ListUrlMain\":\"https://doi.org/10.1007/s40747-025-01903-y\",\"RegionNum\":2,\"RegionCategory\":\"计算机科学\",\"ArticlePicture\":[],\"TitleCN\":null,\"AbstractTextCN\":null,\"PMCID\":null,\"EPubDate\":\"\",\"PubModel\":\"\",\"JCR\":\"Q1\",\"JCRName\":\"COMPUTER SCIENCE, ARTIFICIAL INTELLIGENCE\",\"Score\":null,\"Total\":0}","platform":"Semanticscholar","paperid":null,"PeriodicalName":"Complex & Intelligent Systems","FirstCategoryId":"94","ListUrlMain":"https://doi.org/10.1007/s40747-025-01903-y","RegionNum":2,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"Q1","JCRName":"COMPUTER SCIENCE, ARTIFICIAL INTELLIGENCE","Score":null,"Total":0}
Graph attention based on contextual reasoning and emotion-shift awareness for emotion recognition in conversations
Emotion recognition in conversations has recently emerged as a hot research topic owing to its increasingly important role in developing intelligent empathy services. Thoroughly exploring the conversational context and accurately capturing emotion-shift information are highly crucial for accurate emotion recognition in conversations. However, existing studies generally failed to fully understand the complex conversational context due to their insufficient capabilities in extracting and integrating emotional cues. Moreover, they mainly focused on the speaker’s emotion inertia while paying less attention to explore multi-perspective emotion-shift patterns. To address these limitations, this study proposes a novel multimodal approach, namely, GAT-CRESA (Graph ATtention based on Contextual Reasoning and Emotion-Shift Awareness). Specifically, the multi-turn global contextual reasoning module iteratively performs contextual perception and cognitive reasoning for efficiently understanding the global conversational context. Then, GAT-CRESA explores emotion-shift information among utterances from both the speaker-dependent and the global context-based perspectives. Next, the emotion-shift awareness graphs are constructed for extracting significant local-level conversational context, where edge relations are determined by the learnt emotion-shift labels. Finally, the outputs of graphs are concatenated for final emotion recognition. The loss of emotion prediction task is combined together with those of two perspective’s emotion-shift learning for guiding the training process. Experimental results show that our GAT-CRESA achieves new state-of-art records with 72.77% ACC and 72.81% wa-F1 on IEMOCAP, and 65.44% ACC and 65.04% wa-F1 on MELD, respectively. The ablation results also indicate the effectiveness and rationality of each component in our approach.
期刊介绍:
Complex & Intelligent Systems aims to provide a forum for presenting and discussing novel approaches, tools and techniques meant for attaining a cross-fertilization between the broad fields of complex systems, computational simulation, and intelligent analytics and visualization. The transdisciplinary research that the journal focuses on will expand the boundaries of our understanding by investigating the principles and processes that underlie many of the most profound problems facing society today.