{"title":"基于局部二值模式和马尔可夫模型的面部表情和语音情感检测的比较分析:计算机视觉与面部识别","authors":"Kennedy Chengeta","doi":"10.1145/3271553.3271574","DOIUrl":null,"url":null,"abstract":"Emotion detection has been achieved widely in facial and voice recognition separately with considerable success. The 6 emotional categories coming out of the classification include anger, fear, disgust, happiness and surprise. These can be infered from one's facial expressions both in the form of micro and macro expressions. In facial expressions the emotions are derived by feature extracting the facial expressions in different facial poses and classifying the expression feature vectors derived. Similarly automatic classification of a person's speech's affective state has also been used in signal processing to give insights into the nature of emotions. Speech being a critical tool for communication has been used to derive the emotional state of a human being. Different approaches have been successfully used to derive emotional states either in the form of facial expression recognition or speech emotional recognition being used. Less work has looked at fusing the two approaches to see if this improves emotional recognition accuracy. The study analyses the strengths of both and also limitations of either. The study reveals that emotional derivation based on facial expression recognition and acoustic information complement each other and a fusion of the two leads to better performance and results compared to the audio or acoustic recognition alone.","PeriodicalId":414782,"journal":{"name":"Proceedings of the 2nd International Conference on Vision, Image and Signal Processing","volume":"23 1","pages":"0"},"PeriodicalIF":0.0000,"publicationDate":"2018-08-27","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"2","resultStr":"{\"title\":\"Comparative Analysis of Emotion Detection from Facial Expressions and Voice Using Local Binary Patterns and Markov Models: Computer Vision and Facial Recognition\",\"authors\":\"Kennedy Chengeta\",\"doi\":\"10.1145/3271553.3271574\",\"DOIUrl\":null,\"url\":null,\"abstract\":\"Emotion detection has been achieved widely in facial and voice recognition separately with considerable success. The 6 emotional categories coming out of the classification include anger, fear, disgust, happiness and surprise. These can be infered from one's facial expressions both in the form of micro and macro expressions. In facial expressions the emotions are derived by feature extracting the facial expressions in different facial poses and classifying the expression feature vectors derived. Similarly automatic classification of a person's speech's affective state has also been used in signal processing to give insights into the nature of emotions. Speech being a critical tool for communication has been used to derive the emotional state of a human being. Different approaches have been successfully used to derive emotional states either in the form of facial expression recognition or speech emotional recognition being used. Less work has looked at fusing the two approaches to see if this improves emotional recognition accuracy. The study analyses the strengths of both and also limitations of either. The study reveals that emotional derivation based on facial expression recognition and acoustic information complement each other and a fusion of the two leads to better performance and results compared to the audio or acoustic recognition alone.\",\"PeriodicalId\":414782,\"journal\":{\"name\":\"Proceedings of the 2nd International Conference on Vision, Image and Signal Processing\",\"volume\":\"23 1\",\"pages\":\"0\"},\"PeriodicalIF\":0.0000,\"publicationDate\":\"2018-08-27\",\"publicationTypes\":\"Journal Article\",\"fieldsOfStudy\":null,\"isOpenAccess\":false,\"openAccessPdf\":\"\",\"citationCount\":\"2\",\"resultStr\":null,\"platform\":\"Semanticscholar\",\"paperid\":null,\"PeriodicalName\":\"Proceedings of the 2nd International Conference on Vision, Image and Signal Processing\",\"FirstCategoryId\":\"1085\",\"ListUrlMain\":\"https://doi.org/10.1145/3271553.3271574\",\"RegionNum\":0,\"RegionCategory\":null,\"ArticlePicture\":[],\"TitleCN\":null,\"AbstractTextCN\":null,\"PMCID\":null,\"EPubDate\":\"\",\"PubModel\":\"\",\"JCR\":\"\",\"JCRName\":\"\",\"Score\":null,\"Total\":0}","platform":"Semanticscholar","paperid":null,"PeriodicalName":"Proceedings of the 2nd International Conference on Vision, Image and Signal Processing","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/10.1145/3271553.3271574","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
Comparative Analysis of Emotion Detection from Facial Expressions and Voice Using Local Binary Patterns and Markov Models: Computer Vision and Facial Recognition
Emotion detection has been achieved widely in facial and voice recognition separately with considerable success. The 6 emotional categories coming out of the classification include anger, fear, disgust, happiness and surprise. These can be infered from one's facial expressions both in the form of micro and macro expressions. In facial expressions the emotions are derived by feature extracting the facial expressions in different facial poses and classifying the expression feature vectors derived. Similarly automatic classification of a person's speech's affective state has also been used in signal processing to give insights into the nature of emotions. Speech being a critical tool for communication has been used to derive the emotional state of a human being. Different approaches have been successfully used to derive emotional states either in the form of facial expression recognition or speech emotional recognition being used. Less work has looked at fusing the two approaches to see if this improves emotional recognition accuracy. The study analyses the strengths of both and also limitations of either. The study reveals that emotional derivation based on facial expression recognition and acoustic information complement each other and a fusion of the two leads to better performance and results compared to the audio or acoustic recognition alone.