Speaker-independent negative emotion recognition

M. Kotti, F. Paternò, Constantine Kotropoulos
{"title":"Speaker-independent negative emotion recognition","authors":"M. Kotti, F. Paternò, Constantine Kotropoulos","doi":"10.1109/CIP.2010.5604091","DOIUrl":null,"url":null,"abstract":"This work aims to provide a method able to distinguish between negative and non-negative emotions in vocal interaction. A large pool of 1418 features is extracted for that purpose. Several of those features are tested in emotion recognition for the first time. Next, feature selection is applied separately to male and female utterances. In particular, a bidirectional Best First search with backtracking is applied. The first contribution is the demonstration that a significant number of features, first tested here, are retained after feature selection. The selected features are then fed as input to support vector machines with various kernel functions as well as to the K nearest neighbors classifier. The second contribution is in the speaker-independent experiments conducted in order to cope with the limited number of speakers present in the commonly used emotion speech corpora. Speaker-independent systems are known to be more robust and present a better generalization ability than the speaker-dependent ones. Experimental results are reported for the Berlin emotional speech database. The best performing classifier is found to be the support vector machine with the Gaussian radial basis function kernel. Correctly classified utterances are 86.73%±3.95% for male subjects and 91.73%±4.18% for female subjects. The last contribution is in the statistical analysis of the performance of the support vector machine classifier against the K nearest neighbors classifier as well as the statistical analysis of the various support vector machine kernels impact","PeriodicalId":171474,"journal":{"name":"2010 2nd International Workshop on Cognitive Information Processing","volume":"29 1","pages":"0"},"PeriodicalIF":0.0000,"publicationDate":"2010-06-14","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"10","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"2010 2nd International Workshop on Cognitive Information Processing","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/10.1109/CIP.2010.5604091","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
引用次数: 10

Abstract

This work aims to provide a method able to distinguish between negative and non-negative emotions in vocal interaction. A large pool of 1418 features is extracted for that purpose. Several of those features are tested in emotion recognition for the first time. Next, feature selection is applied separately to male and female utterances. In particular, a bidirectional Best First search with backtracking is applied. The first contribution is the demonstration that a significant number of features, first tested here, are retained after feature selection. The selected features are then fed as input to support vector machines with various kernel functions as well as to the K nearest neighbors classifier. The second contribution is in the speaker-independent experiments conducted in order to cope with the limited number of speakers present in the commonly used emotion speech corpora. Speaker-independent systems are known to be more robust and present a better generalization ability than the speaker-dependent ones. Experimental results are reported for the Berlin emotional speech database. The best performing classifier is found to be the support vector machine with the Gaussian radial basis function kernel. Correctly classified utterances are 86.73%±3.95% for male subjects and 91.73%±4.18% for female subjects. The last contribution is in the statistical analysis of the performance of the support vector machine classifier against the K nearest neighbors classifier as well as the statistical analysis of the various support vector machine kernels impact
不依赖说话人的负面情绪识别
本研究旨在提供一种能够区分声音互动中消极和非消极情绪的方法。为此目的提取了1418个特征的大池。其中一些特征首次在情绪识别中进行了测试。接下来,将特征选择分别应用于男性和女性话语。特别地,应用了带回溯的双向最佳优先搜索。第一个贡献是证明了在特征选择之后保留了大量的特征(在这里首次测试)。然后将选择的特征作为输入馈送给具有各种核函数的支持向量机以及K近邻分类器。第二个贡献是在独立于说话人的实验中进行的,以应对常用的情感语音语料库中有限的说话人数量。与演讲者相关的系统相比,演讲者无关的系统具有更强的鲁棒性和更好的泛化能力。报道了柏林情感语音数据库的实验结果。结果表明,具有高斯径向基函数核的支持向量机是性能最好的分类器。男性被试的正确分类率为86.73%±3.95%,女性被试的正确分类率为91.73%±4.18%。最后的贡献是统计分析了支持向量机分类器对K近邻分类器的性能,以及统计分析了各种支持向量机核的影响
本文章由计算机程序翻译,如有差异,请以英文原文为准。
求助全文
约1分钟内获得全文 求助全文
来源期刊
自引率
0.00%
发文量
0
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
确定
请完成安全验证×
copy
已复制链接
快去分享给好友吧!
我知道了
右上角分享
点击右上角分享
0
联系我们:info@booksci.cn Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。 Copyright © 2023 布克学术 All rights reserved.
京ICP备2023020795号-1
ghs 京公网安备 11010802042870号
Book学术文献互助
Book学术文献互助群
群 号:481959085
Book学术官方微信