Global-View and Speaker-Aware Emotion Cause Extraction in Conversations

IF 4.1 2区 计算机科学 Q1 ACOUSTICS
Jiaming An;Zixiang Ding;Ke Li;Rui Xia
{"title":"Global-View and Speaker-Aware Emotion Cause Extraction in Conversations","authors":"Jiaming An;Zixiang Ding;Ke Li;Rui Xia","doi":"10.1109/TASLP.2023.3319990","DOIUrl":null,"url":null,"abstract":"Emotion cause extraction in conversations, the task of recognizing and extracting the causes behind the emotions in a conversation, is a new and under-explored task. It was previously treated as an utterance-level task, that can only extract cause of one emotion from one utterance at a time and is difficult to model the correlation between different emotions and causes in the conversation. The role of speakers was also not fully utilized in the previous methods. In this article, we introduce a global-view and speaker-aware conversational emotion cause extraction framework. It can fully model the interaction between utterances and emotions in the conversation and simultaneously extract all the causes corresponding to all emotions or one given emotion in a conversation, and can be applied to both real-time and non-real-time task settings. We further propose a Speaker-aware Couple-Decoder Module and a Speaker-Emotion Graph Attention Network, to better model the role of speakers in the conversation. The experimental results prove our approach's advantages in both emotion cause extraction performance and computational efficiency.","PeriodicalId":13332,"journal":{"name":"IEEE/ACM Transactions on Audio, Speech, and Language Processing","volume":"31 ","pages":"3814-3823"},"PeriodicalIF":4.1000,"publicationDate":"2023-10-09","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"IEEE/ACM Transactions on Audio, Speech, and Language Processing","FirstCategoryId":"94","ListUrlMain":"https://ieeexplore.ieee.org/document/10274611/","RegionNum":2,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"Q1","JCRName":"ACOUSTICS","Score":null,"Total":0}
引用次数: 0

Abstract

Emotion cause extraction in conversations, the task of recognizing and extracting the causes behind the emotions in a conversation, is a new and under-explored task. It was previously treated as an utterance-level task, that can only extract cause of one emotion from one utterance at a time and is difficult to model the correlation between different emotions and causes in the conversation. The role of speakers was also not fully utilized in the previous methods. In this article, we introduce a global-view and speaker-aware conversational emotion cause extraction framework. It can fully model the interaction between utterances and emotions in the conversation and simultaneously extract all the causes corresponding to all emotions or one given emotion in a conversation, and can be applied to both real-time and non-real-time task settings. We further propose a Speaker-aware Couple-Decoder Module and a Speaker-Emotion Graph Attention Network, to better model the role of speakers in the conversation. The experimental results prove our approach's advantages in both emotion cause extraction performance and computational efficiency.
全局观与说话人感知的会话情感原因提取
会话中的情绪原因提取,即识别和提取会话中情绪背后的原因,是一项新的、未被充分探索的任务。它以前被视为一个话语层面的任务,一次只能从一个话语中提取一种情绪的原因,并且很难对对话中不同情绪和原因之间的相关性进行建模。发言者的作用在以前的方法中也没有得到充分利用。在本文中,我们介绍了一个全局观和说话人感知的会话情感原因提取框架。它可以完全建模对话中话语和情绪之间的互动,同时提取对话中所有情绪或一种给定情绪对应的所有原因,可以应用于实时和非实时任务设置。我们进一步提出了一个说话人感知耦合解码器模块和说话人情感图注意力网络,以更好地模拟说话人在对话中的角色。实验结果证明了我们的方法在情感原因提取性能和计算效率方面的优势。
本文章由计算机程序翻译,如有差异,请以英文原文为准。
求助全文
约1分钟内获得全文 求助全文
来源期刊
IEEE/ACM Transactions on Audio, Speech, and Language Processing
IEEE/ACM Transactions on Audio, Speech, and Language Processing ACOUSTICS-ENGINEERING, ELECTRICAL & ELECTRONIC
CiteScore
11.30
自引率
11.10%
发文量
217
期刊介绍: The IEEE/ACM Transactions on Audio, Speech, and Language Processing covers audio, speech and language processing and the sciences that support them. In audio processing: transducers, room acoustics, active sound control, human audition, analysis/synthesis/coding of music, and consumer audio. In speech processing: areas such as speech analysis, synthesis, coding, speech and speaker recognition, speech production and perception, and speech enhancement. In language processing: speech and text analysis, understanding, generation, dialog management, translation, summarization, question answering and document indexing and retrieval, as well as general language modeling.
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
确定
请完成安全验证×
copy
已复制链接
快去分享给好友吧!
我知道了
右上角分享
点击右上角分享
0
联系我们:info@booksci.cn Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。 Copyright © 2023 布克学术 All rights reserved.
京ICP备2023020795号-1
ghs 京公网安备 11010802042870号
Book学术文献互助
Book学术文献互助群
群 号:481959085
Book学术官方微信