K. Priya, S. M. Mansoor Roomi, P. Shanmugavadivu, M. Sethuraman, P. Kalaivani
{"title":"基于音频和情感线索的面试表现自动评估系统","authors":"K. Priya, S. M. Mansoor Roomi, P. Shanmugavadivu, M. Sethuraman, P. Kalaivani","doi":"10.1109/ICACCS.2019.8728458","DOIUrl":null,"url":null,"abstract":"Automatic analysis and performance evaluation of interviewees is a largely unexplored and challenging problem. The proposed work provides a computational structure to enumerate the performance of the interviewee in the context of communication to give their performance feedback from the analysis of multimodal signals such as facial images and voice. A video captured during the interview is split into audio and visual frames. From the visual frames, the face is detected and their emotions are analyzed with the help of Histogram of Oriented Gradients. The facial expressions are classified by using Support Vector Machine. The classified facial expressions are happy, fear, sad, neutral, surprise, disgust and angry. From the audio cues,, the Mel Frequency Cepstral Coefficient features are extracted and this is categorized as fluent, non-fluent-pause and non-fluent-stammer. Both emotion and fluency of the candidate are fused to find performance score. This automated analysis provides the ratings for interview behavior such as poor, medium, high performance.","PeriodicalId":249139,"journal":{"name":"2019 5th International Conference on Advanced Computing & Communication Systems (ICACCS)","volume":"15 1","pages":"0"},"PeriodicalIF":0.0000,"publicationDate":"2019-03-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"1","resultStr":"{\"title\":\"An Automated System for the Assesment of Interview Performance through Audio & Emotion Cues\",\"authors\":\"K. Priya, S. M. Mansoor Roomi, P. Shanmugavadivu, M. Sethuraman, P. Kalaivani\",\"doi\":\"10.1109/ICACCS.2019.8728458\",\"DOIUrl\":null,\"url\":null,\"abstract\":\"Automatic analysis and performance evaluation of interviewees is a largely unexplored and challenging problem. The proposed work provides a computational structure to enumerate the performance of the interviewee in the context of communication to give their performance feedback from the analysis of multimodal signals such as facial images and voice. A video captured during the interview is split into audio and visual frames. From the visual frames, the face is detected and their emotions are analyzed with the help of Histogram of Oriented Gradients. The facial expressions are classified by using Support Vector Machine. The classified facial expressions are happy, fear, sad, neutral, surprise, disgust and angry. From the audio cues,, the Mel Frequency Cepstral Coefficient features are extracted and this is categorized as fluent, non-fluent-pause and non-fluent-stammer. Both emotion and fluency of the candidate are fused to find performance score. This automated analysis provides the ratings for interview behavior such as poor, medium, high performance.\",\"PeriodicalId\":249139,\"journal\":{\"name\":\"2019 5th International Conference on Advanced Computing & Communication Systems (ICACCS)\",\"volume\":\"15 1\",\"pages\":\"0\"},\"PeriodicalIF\":0.0000,\"publicationDate\":\"2019-03-01\",\"publicationTypes\":\"Journal Article\",\"fieldsOfStudy\":null,\"isOpenAccess\":false,\"openAccessPdf\":\"\",\"citationCount\":\"1\",\"resultStr\":null,\"platform\":\"Semanticscholar\",\"paperid\":null,\"PeriodicalName\":\"2019 5th International Conference on Advanced Computing & Communication Systems (ICACCS)\",\"FirstCategoryId\":\"1085\",\"ListUrlMain\":\"https://doi.org/10.1109/ICACCS.2019.8728458\",\"RegionNum\":0,\"RegionCategory\":null,\"ArticlePicture\":[],\"TitleCN\":null,\"AbstractTextCN\":null,\"PMCID\":null,\"EPubDate\":\"\",\"PubModel\":\"\",\"JCR\":\"\",\"JCRName\":\"\",\"Score\":null,\"Total\":0}","platform":"Semanticscholar","paperid":null,"PeriodicalName":"2019 5th International Conference on Advanced Computing & Communication Systems (ICACCS)","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/10.1109/ICACCS.2019.8728458","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
An Automated System for the Assesment of Interview Performance through Audio & Emotion Cues
Automatic analysis and performance evaluation of interviewees is a largely unexplored and challenging problem. The proposed work provides a computational structure to enumerate the performance of the interviewee in the context of communication to give their performance feedback from the analysis of multimodal signals such as facial images and voice. A video captured during the interview is split into audio and visual frames. From the visual frames, the face is detected and their emotions are analyzed with the help of Histogram of Oriented Gradients. The facial expressions are classified by using Support Vector Machine. The classified facial expressions are happy, fear, sad, neutral, surprise, disgust and angry. From the audio cues,, the Mel Frequency Cepstral Coefficient features are extracted and this is categorized as fluent, non-fluent-pause and non-fluent-stammer. Both emotion and fluency of the candidate are fused to find performance score. This automated analysis provides the ratings for interview behavior such as poor, medium, high performance.