Khin Pa Pa Aung, Hao-Long Yin, Tian-Fang Ma, Wei-Long Zheng, Bao-Liang Lu
{"title":"用于情绪识别的多模态缅甸情绪数据集。","authors":"Khin Pa Pa Aung, Hao-Long Yin, Tian-Fang Ma, Wei-Long Zheng, Bao-Liang Lu","doi":"10.1109/EMBC53108.2024.10782660","DOIUrl":null,"url":null,"abstract":"<p><p>Effective emotion recognition is vital for human interaction and has an impact on several fields such as psychology, social sciences, human-computer interaction, and emotional artificial intelligence. This study centers on the innovative contribution of a novel Myanmar emotion dataset to enhance emotion recognition technology in diverse cultural contexts. Our unique dataset is derived from a carefully designed emotion elicitation paradigm, using 15 video clips per session for three emotions (positive, neutral, and negative), with five clips per emotion. We collected electroencephalogram (EEG) signals and eye-tracking data from 20 subjects, and each subject took three sessions spaced over several days. Notably, all video clips used in experiments have been well rated by Myanmar citizens through the Self-Assessment Manikin scale. We validated the proposed dataset's uniqueness using three baseline unimodal classification methods, alongside two traditional multimodal approaches and a deep multimodal approach (DCCA-AM) under subject-dependent and subject-independent settings. Unimodal classification achieved accuracies ranging from 62.57% to 77.05%, while multimodal fusion techniques achieved accuracies ranging from 75.43% to 87.91%. These results underscore the effectiveness of the models, and highlighting the value of our unique dataset for cross-cultural applications.</p>","PeriodicalId":72237,"journal":{"name":"Annual International Conference of the IEEE Engineering in Medicine and Biology Society. IEEE Engineering in Medicine and Biology Society. Annual International Conference","volume":"2024 ","pages":"1-4"},"PeriodicalIF":0.0000,"publicationDate":"2024-07-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":"{\"title\":\"A Multimodal Myanmar Emotion Dataset for Emotion Recognition.\",\"authors\":\"Khin Pa Pa Aung, Hao-Long Yin, Tian-Fang Ma, Wei-Long Zheng, Bao-Liang Lu\",\"doi\":\"10.1109/EMBC53108.2024.10782660\",\"DOIUrl\":null,\"url\":null,\"abstract\":\"<p><p>Effective emotion recognition is vital for human interaction and has an impact on several fields such as psychology, social sciences, human-computer interaction, and emotional artificial intelligence. This study centers on the innovative contribution of a novel Myanmar emotion dataset to enhance emotion recognition technology in diverse cultural contexts. Our unique dataset is derived from a carefully designed emotion elicitation paradigm, using 15 video clips per session for three emotions (positive, neutral, and negative), with five clips per emotion. We collected electroencephalogram (EEG) signals and eye-tracking data from 20 subjects, and each subject took three sessions spaced over several days. Notably, all video clips used in experiments have been well rated by Myanmar citizens through the Self-Assessment Manikin scale. We validated the proposed dataset's uniqueness using three baseline unimodal classification methods, alongside two traditional multimodal approaches and a deep multimodal approach (DCCA-AM) under subject-dependent and subject-independent settings. Unimodal classification achieved accuracies ranging from 62.57% to 77.05%, while multimodal fusion techniques achieved accuracies ranging from 75.43% to 87.91%. These results underscore the effectiveness of the models, and highlighting the value of our unique dataset for cross-cultural applications.</p>\",\"PeriodicalId\":72237,\"journal\":{\"name\":\"Annual International Conference of the IEEE Engineering in Medicine and Biology Society. IEEE Engineering in Medicine and Biology Society. Annual International Conference\",\"volume\":\"2024 \",\"pages\":\"1-4\"},\"PeriodicalIF\":0.0000,\"publicationDate\":\"2024-07-01\",\"publicationTypes\":\"Journal Article\",\"fieldsOfStudy\":null,\"isOpenAccess\":false,\"openAccessPdf\":\"\",\"citationCount\":\"0\",\"resultStr\":null,\"platform\":\"Semanticscholar\",\"paperid\":null,\"PeriodicalName\":\"Annual International Conference of the IEEE Engineering in Medicine and Biology Society. IEEE Engineering in Medicine and Biology Society. Annual International Conference\",\"FirstCategoryId\":\"1085\",\"ListUrlMain\":\"https://doi.org/10.1109/EMBC53108.2024.10782660\",\"RegionNum\":0,\"RegionCategory\":null,\"ArticlePicture\":[],\"TitleCN\":null,\"AbstractTextCN\":null,\"PMCID\":null,\"EPubDate\":\"\",\"PubModel\":\"\",\"JCR\":\"\",\"JCRName\":\"\",\"Score\":null,\"Total\":0}","platform":"Semanticscholar","paperid":null,"PeriodicalName":"Annual International Conference of the IEEE Engineering in Medicine and Biology Society. IEEE Engineering in Medicine and Biology Society. Annual International Conference","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/10.1109/EMBC53108.2024.10782660","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
A Multimodal Myanmar Emotion Dataset for Emotion Recognition.
Effective emotion recognition is vital for human interaction and has an impact on several fields such as psychology, social sciences, human-computer interaction, and emotional artificial intelligence. This study centers on the innovative contribution of a novel Myanmar emotion dataset to enhance emotion recognition technology in diverse cultural contexts. Our unique dataset is derived from a carefully designed emotion elicitation paradigm, using 15 video clips per session for three emotions (positive, neutral, and negative), with five clips per emotion. We collected electroencephalogram (EEG) signals and eye-tracking data from 20 subjects, and each subject took three sessions spaced over several days. Notably, all video clips used in experiments have been well rated by Myanmar citizens through the Self-Assessment Manikin scale. We validated the proposed dataset's uniqueness using three baseline unimodal classification methods, alongside two traditional multimodal approaches and a deep multimodal approach (DCCA-AM) under subject-dependent and subject-independent settings. Unimodal classification achieved accuracies ranging from 62.57% to 77.05%, while multimodal fusion techniques achieved accuracies ranging from 75.43% to 87.91%. These results underscore the effectiveness of the models, and highlighting the value of our unique dataset for cross-cultural applications.