声门波形对言语情绪的贡献:两两比较研究

Zhongzhe Xiao, Ying Chen, Zhi Tao
{"title":"声门波形对言语情绪的贡献:两两比较研究","authors":"Zhongzhe Xiao, Ying Chen, Zhi Tao","doi":"10.1109/PIC.2018.8706134","DOIUrl":null,"url":null,"abstract":"In this work, we investigated the contribution of the glottal waveform in human vocal emotion expressing. Seven emotional states including moderate and intense versions of three emotional families as anger, joy, and sadness, plus a neutral state are considered, with speech samples in Mandarin Chinese. The glottal waveform extracted from speech samples of different emotion states are first analyzed in both time domain and frequency domain to discover their differences. Comparative emotion classifications are then taken out based on features extracted from two sources: original whole speech signal, or only glottal wave signal. Two sets of experiments are performed, as the generation of a performance-driven hierarchical classifier architecture, and pairwise classification on individual emotional states. The low difference between accuracies obtained from the two sources proved that a majority of emotional cues in speech could be conveyed through glottal waveform. The best distinguishable emotional pair by glottal waveform is intense anger against moderate sadness, with the accuracy up to 92.45%. It is also concluded in this work that glottal waveform represent better valence cues than arousal cues of emotion.","PeriodicalId":236106,"journal":{"name":"2018 IEEE International Conference on Progress in Informatics and Computing (PIC)","volume":"47 1","pages":"0"},"PeriodicalIF":0.0000,"publicationDate":"2018-08-30","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"1","resultStr":"{\"title\":\"Contribution of Glottal Waveform in Speech Emotion: A Comparative Pairwise Investigation\",\"authors\":\"Zhongzhe Xiao, Ying Chen, Zhi Tao\",\"doi\":\"10.1109/PIC.2018.8706134\",\"DOIUrl\":null,\"url\":null,\"abstract\":\"In this work, we investigated the contribution of the glottal waveform in human vocal emotion expressing. Seven emotional states including moderate and intense versions of three emotional families as anger, joy, and sadness, plus a neutral state are considered, with speech samples in Mandarin Chinese. The glottal waveform extracted from speech samples of different emotion states are first analyzed in both time domain and frequency domain to discover their differences. Comparative emotion classifications are then taken out based on features extracted from two sources: original whole speech signal, or only glottal wave signal. Two sets of experiments are performed, as the generation of a performance-driven hierarchical classifier architecture, and pairwise classification on individual emotional states. The low difference between accuracies obtained from the two sources proved that a majority of emotional cues in speech could be conveyed through glottal waveform. The best distinguishable emotional pair by glottal waveform is intense anger against moderate sadness, with the accuracy up to 92.45%. It is also concluded in this work that glottal waveform represent better valence cues than arousal cues of emotion.\",\"PeriodicalId\":236106,\"journal\":{\"name\":\"2018 IEEE International Conference on Progress in Informatics and Computing (PIC)\",\"volume\":\"47 1\",\"pages\":\"0\"},\"PeriodicalIF\":0.0000,\"publicationDate\":\"2018-08-30\",\"publicationTypes\":\"Journal Article\",\"fieldsOfStudy\":null,\"isOpenAccess\":false,\"openAccessPdf\":\"\",\"citationCount\":\"1\",\"resultStr\":null,\"platform\":\"Semanticscholar\",\"paperid\":null,\"PeriodicalName\":\"2018 IEEE International Conference on Progress in Informatics and Computing (PIC)\",\"FirstCategoryId\":\"1085\",\"ListUrlMain\":\"https://doi.org/10.1109/PIC.2018.8706134\",\"RegionNum\":0,\"RegionCategory\":null,\"ArticlePicture\":[],\"TitleCN\":null,\"AbstractTextCN\":null,\"PMCID\":null,\"EPubDate\":\"\",\"PubModel\":\"\",\"JCR\":\"\",\"JCRName\":\"\",\"Score\":null,\"Total\":0}","platform":"Semanticscholar","paperid":null,"PeriodicalName":"2018 IEEE International Conference on Progress in Informatics and Computing (PIC)","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/10.1109/PIC.2018.8706134","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
引用次数: 1

摘要

在这项工作中,我们研究了声门波形在人类声音情感表达中的贡献。考虑了七种情绪状态,包括愤怒、喜悦和悲伤三种情绪家庭的中度和强烈版本,以及中性状态,并使用普通话语音样本。首先对不同情绪状态下的语音样本提取的声门波形进行时域和频域分析,发现它们之间的差异。然后根据原始的完整语音信号和仅声门波信号提取的特征进行比较情绪分类。两组实验分别进行了性能驱动的分层分类器体系结构的生成和个体情绪状态的两两分类。两种声门波形的准确率差异较小,证明语音中的大部分情感线索都可以通过声门波形传递。声门波形区分强烈愤怒和中度悲伤情绪对效果最好,准确率高达92.45%。本研究还得出声门波形比情绪唤醒波形更能代表效价信号的结论。
本文章由计算机程序翻译,如有差异,请以英文原文为准。
Contribution of Glottal Waveform in Speech Emotion: A Comparative Pairwise Investigation
In this work, we investigated the contribution of the glottal waveform in human vocal emotion expressing. Seven emotional states including moderate and intense versions of three emotional families as anger, joy, and sadness, plus a neutral state are considered, with speech samples in Mandarin Chinese. The glottal waveform extracted from speech samples of different emotion states are first analyzed in both time domain and frequency domain to discover their differences. Comparative emotion classifications are then taken out based on features extracted from two sources: original whole speech signal, or only glottal wave signal. Two sets of experiments are performed, as the generation of a performance-driven hierarchical classifier architecture, and pairwise classification on individual emotional states. The low difference between accuracies obtained from the two sources proved that a majority of emotional cues in speech could be conveyed through glottal waveform. The best distinguishable emotional pair by glottal waveform is intense anger against moderate sadness, with the accuracy up to 92.45%. It is also concluded in this work that glottal waveform represent better valence cues than arousal cues of emotion.
求助全文
通过发布文献求助,成功后即可免费获取论文全文。 去求助
来源期刊
自引率
0.00%
发文量
0
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
确定
请完成安全验证×
copy
已复制链接
快去分享给好友吧!
我知道了
右上角分享
点击右上角分享
0
联系我们:info@booksci.cn Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。 Copyright © 2023 布克学术 All rights reserved.
京ICP备2023020795号-1
ghs 京公网安备 11010802042870号
Book学术文献互助
Book学术文献互助群
群 号:604180095
Book学术官方微信