A Multimodal Myanmar Emotion Dataset for Emotion Recognition.

Khin Pa Pa Aung, Hao-Long Yin, Tian-Fang Ma, Wei-Long Zheng, Bao-Liang Lu
{"title":"A Multimodal Myanmar Emotion Dataset for Emotion Recognition.","authors":"Khin Pa Pa Aung, Hao-Long Yin, Tian-Fang Ma, Wei-Long Zheng, Bao-Liang Lu","doi":"10.1109/EMBC53108.2024.10782660","DOIUrl":null,"url":null,"abstract":"<p><p>Effective emotion recognition is vital for human interaction and has an impact on several fields such as psychology, social sciences, human-computer interaction, and emotional artificial intelligence. This study centers on the innovative contribution of a novel Myanmar emotion dataset to enhance emotion recognition technology in diverse cultural contexts. Our unique dataset is derived from a carefully designed emotion elicitation paradigm, using 15 video clips per session for three emotions (positive, neutral, and negative), with five clips per emotion. We collected electroencephalogram (EEG) signals and eye-tracking data from 20 subjects, and each subject took three sessions spaced over several days. Notably, all video clips used in experiments have been well rated by Myanmar citizens through the Self-Assessment Manikin scale. We validated the proposed dataset's uniqueness using three baseline unimodal classification methods, alongside two traditional multimodal approaches and a deep multimodal approach (DCCA-AM) under subject-dependent and subject-independent settings. Unimodal classification achieved accuracies ranging from 62.57% to 77.05%, while multimodal fusion techniques achieved accuracies ranging from 75.43% to 87.91%. These results underscore the effectiveness of the models, and highlighting the value of our unique dataset for cross-cultural applications.</p>","PeriodicalId":72237,"journal":{"name":"Annual International Conference of the IEEE Engineering in Medicine and Biology Society. IEEE Engineering in Medicine and Biology Society. Annual International Conference","volume":"2024 ","pages":"1-4"},"PeriodicalIF":0.0000,"publicationDate":"2024-07-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"Annual International Conference of the IEEE Engineering in Medicine and Biology Society. IEEE Engineering in Medicine and Biology Society. Annual International Conference","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/10.1109/EMBC53108.2024.10782660","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
引用次数: 0

Abstract

Effective emotion recognition is vital for human interaction and has an impact on several fields such as psychology, social sciences, human-computer interaction, and emotional artificial intelligence. This study centers on the innovative contribution of a novel Myanmar emotion dataset to enhance emotion recognition technology in diverse cultural contexts. Our unique dataset is derived from a carefully designed emotion elicitation paradigm, using 15 video clips per session for three emotions (positive, neutral, and negative), with five clips per emotion. We collected electroencephalogram (EEG) signals and eye-tracking data from 20 subjects, and each subject took three sessions spaced over several days. Notably, all video clips used in experiments have been well rated by Myanmar citizens through the Self-Assessment Manikin scale. We validated the proposed dataset's uniqueness using three baseline unimodal classification methods, alongside two traditional multimodal approaches and a deep multimodal approach (DCCA-AM) under subject-dependent and subject-independent settings. Unimodal classification achieved accuracies ranging from 62.57% to 77.05%, while multimodal fusion techniques achieved accuracies ranging from 75.43% to 87.91%. These results underscore the effectiveness of the models, and highlighting the value of our unique dataset for cross-cultural applications.

用于情绪识别的多模态缅甸情绪数据集。
有效的情感识别对于人类互动至关重要,对心理学、社会科学、人机交互、情感人工智能等多个领域都有影响。本研究的中心是一个新的缅甸情绪数据集的创新贡献,以增强不同文化背景下的情绪识别技术。我们独特的数据集源自精心设计的情绪激发范例,每次会话使用15个视频片段来表达三种情绪(积极、中性和消极),每种情绪有5个片段。我们收集了20名受试者的脑电图(EEG)信号和眼动追踪数据,每个受试者在几天内进行了三个疗程。值得注意的是,实验中使用的所有视频片段都得到了缅甸公民通过自我评估模型量表的良好评价。我们使用三种基线单模态分类方法,以及两种传统多模态方法和一种深度多模态方法(DCCA-AM)在主题依赖和主题独立设置下验证了所提出数据集的唯一性。单模态分类的准确率为62.57% ~ 77.05%,而多模态融合的准确率为75.43% ~ 87.91%。这些结果强调了模型的有效性,并突出了我们独特的数据集在跨文化应用中的价值。
本文章由计算机程序翻译,如有差异,请以英文原文为准。
求助全文
约1分钟内获得全文 求助全文
来源期刊
CiteScore
0.80
自引率
0.00%
发文量
0
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
确定
请完成安全验证×
copy
已复制链接
快去分享给好友吧!
我知道了
右上角分享
点击右上角分享
0
联系我们:info@booksci.cn Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。 Copyright © 2023 布克学术 All rights reserved.
京ICP备2023020795号-1
ghs 京公网安备 11010802042870号
Book学术文献互助
Book学术文献互助群
群 号:604180095
Book学术官方微信