{"title":"A mobile emotion recognition system based on speech signals and facial images","authors":"Yu-Hao Wu, Shu-Jing Lin, Don-Lin Yang","doi":"10.1109/ICSEC.2013.6694781","DOIUrl":null,"url":null,"abstract":"Smartphones are used daily for personal and business communications, and they have become a primary medium to capture human emotions. By recognizing the emotions of speakers during a conversation, one can deliver or understand messages better, make successful negotiations, and provide more personal services. Therefore, we developed an emotion recognition system on a mobile platform based on speech signals and facial images. This research has two phases, a training phase and a testing phase. In the first phase, speech signals and facial images are processed through data preprocessing, feature extraction, and SVM classifier construction steps. In the second phase, the participants generated video recordings as test data. These data were transformed for feature extraction and classified into four emotion classes by using the generated classifiers. Feature selection methods were exploited to choose useful features. We proposed an adjustable weighted segmentation method to determine the final results of emotion recognition. Various experiments were performed using real world simulations to evaluate the proposed system. The result showed an average accuracy rate of 87 percent with the highest accuracy rate at 91 percent. Facial images were also used to improve emotion recognition especially during periods of silence in conversations.","PeriodicalId":191620,"journal":{"name":"2013 International Computer Science and Engineering Conference (ICSEC)","volume":"313 1","pages":"0"},"PeriodicalIF":0.0000,"publicationDate":"2013-09-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"8","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"2013 International Computer Science and Engineering Conference (ICSEC)","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/10.1109/ICSEC.2013.6694781","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
引用次数: 8
Abstract
Smartphones are used daily for personal and business communications, and they have become a primary medium to capture human emotions. By recognizing the emotions of speakers during a conversation, one can deliver or understand messages better, make successful negotiations, and provide more personal services. Therefore, we developed an emotion recognition system on a mobile platform based on speech signals and facial images. This research has two phases, a training phase and a testing phase. In the first phase, speech signals and facial images are processed through data preprocessing, feature extraction, and SVM classifier construction steps. In the second phase, the participants generated video recordings as test data. These data were transformed for feature extraction and classified into four emotion classes by using the generated classifiers. Feature selection methods were exploited to choose useful features. We proposed an adjustable weighted segmentation method to determine the final results of emotion recognition. Various experiments were performed using real world simulations to evaluate the proposed system. The result showed an average accuracy rate of 87 percent with the highest accuracy rate at 91 percent. Facial images were also used to improve emotion recognition especially during periods of silence in conversations.