C. Shahnaz, S. Sultana, S. Fattah, R. H. M. Rafi, I. Ahmmed, Weiping Zhu, M. Ahmad
{"title":"Emotion recognition based on EMD-Wavelet analysis of speech signals","authors":"C. Shahnaz, S. Sultana, S. Fattah, R. H. M. Rafi, I. Ahmmed, Weiping Zhu, M. Ahmad","doi":"10.1109/ICDSP.2015.7251881","DOIUrl":null,"url":null,"abstract":"In this paper, a speech emotion recognition method is proposed based on wavelet analysis on decomposed speech data obtained via empirical mode decomposition (EMD). Instead of analyzing the given speech signal directly, first the intrinsic mode functions (IMFs) are extracted by using the EMD and then the discrete wavelet transform (DWT) is performed only on the selected dominant IMFs. Both approximate and detail DWT coefficients of the dominant IMF are taken into consideration. It is found that some higher order statistics of these EMD-DWT coefficients corresponding to different emotions exhibit distinguishing characteristics and these statistical parameters are chosen as the desired features. For the purpose of classification, K nearest neighbor (KNN) classifier is employed along with the hierarchical clustering. Extensive simulations are carried out on widely used EMO-DB speech emotion database containing four class emotions, namely angry, happy, sad and neutral. Simulation results show that the proposed EMD-Wavelet based feature can provide quite satisfactory recognition performance with reduced feature dimension.","PeriodicalId":216293,"journal":{"name":"2015 IEEE International Conference on Digital Signal Processing (DSP)","volume":"22 1","pages":"0"},"PeriodicalIF":0.0000,"publicationDate":"2015-07-21","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"7","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"2015 IEEE International Conference on Digital Signal Processing (DSP)","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/10.1109/ICDSP.2015.7251881","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
引用次数: 7
Abstract
In this paper, a speech emotion recognition method is proposed based on wavelet analysis on decomposed speech data obtained via empirical mode decomposition (EMD). Instead of analyzing the given speech signal directly, first the intrinsic mode functions (IMFs) are extracted by using the EMD and then the discrete wavelet transform (DWT) is performed only on the selected dominant IMFs. Both approximate and detail DWT coefficients of the dominant IMF are taken into consideration. It is found that some higher order statistics of these EMD-DWT coefficients corresponding to different emotions exhibit distinguishing characteristics and these statistical parameters are chosen as the desired features. For the purpose of classification, K nearest neighbor (KNN) classifier is employed along with the hierarchical clustering. Extensive simulations are carried out on widely used EMO-DB speech emotion database containing four class emotions, namely angry, happy, sad and neutral. Simulation results show that the proposed EMD-Wavelet based feature can provide quite satisfactory recognition performance with reduced feature dimension.