{"title":"基于空间语境与面部表情时间动态解耦的情绪识别","authors":"R. Alazrai, K. M. Yousef, M. Daoud","doi":"10.1109/ISNCC.2019.8909141","DOIUrl":null,"url":null,"abstract":"This paper presents an emotion recognition approach based on decoupling the spatial context from the temporal dynamics of facial expressions in video sequences. In particular, each emotional state is represented as a set of temporal phases, where each temporal phase exhibits different temporal dynamics such as the expressing speed and the variable length of each phase. In this work, we have developed an algorithm for automatically detecting the temporal phases of human facial expressions by employing the concept of mutual information to define a similarity measure among different video frames. Moreover, we have developed a two-layer framework for emotional state recognition. The first layer utilizes the spatial context to classify the frames in an input video into emotional-specific temporal phases using a support vector machine classifier. In the second layer, dynamic time warping is used to classify the sequence of labels associated with the video frames, which is generated in the first layer, into a specific emotional state. In order to validate the performance of the proposed approach, we have conducted extensive computer simulations and the results show an average classification accuracy of 93.53% using the extended Cohn-Kanade facial-expression database.","PeriodicalId":187178,"journal":{"name":"2019 International Symposium on Networks, Computers and Communications (ISNCC)","volume":"70 1","pages":"0"},"PeriodicalIF":0.0000,"publicationDate":"2019-06-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"2","resultStr":"{\"title\":\"Emotion Recognition Based on Decoupling the Spatial Context from the Temporal Dynamics of Facial Expressions\",\"authors\":\"R. Alazrai, K. M. Yousef, M. Daoud\",\"doi\":\"10.1109/ISNCC.2019.8909141\",\"DOIUrl\":null,\"url\":null,\"abstract\":\"This paper presents an emotion recognition approach based on decoupling the spatial context from the temporal dynamics of facial expressions in video sequences. In particular, each emotional state is represented as a set of temporal phases, where each temporal phase exhibits different temporal dynamics such as the expressing speed and the variable length of each phase. In this work, we have developed an algorithm for automatically detecting the temporal phases of human facial expressions by employing the concept of mutual information to define a similarity measure among different video frames. Moreover, we have developed a two-layer framework for emotional state recognition. The first layer utilizes the spatial context to classify the frames in an input video into emotional-specific temporal phases using a support vector machine classifier. In the second layer, dynamic time warping is used to classify the sequence of labels associated with the video frames, which is generated in the first layer, into a specific emotional state. In order to validate the performance of the proposed approach, we have conducted extensive computer simulations and the results show an average classification accuracy of 93.53% using the extended Cohn-Kanade facial-expression database.\",\"PeriodicalId\":187178,\"journal\":{\"name\":\"2019 International Symposium on Networks, Computers and Communications (ISNCC)\",\"volume\":\"70 1\",\"pages\":\"0\"},\"PeriodicalIF\":0.0000,\"publicationDate\":\"2019-06-01\",\"publicationTypes\":\"Journal Article\",\"fieldsOfStudy\":null,\"isOpenAccess\":false,\"openAccessPdf\":\"\",\"citationCount\":\"2\",\"resultStr\":null,\"platform\":\"Semanticscholar\",\"paperid\":null,\"PeriodicalName\":\"2019 International Symposium on Networks, Computers and Communications (ISNCC)\",\"FirstCategoryId\":\"1085\",\"ListUrlMain\":\"https://doi.org/10.1109/ISNCC.2019.8909141\",\"RegionNum\":0,\"RegionCategory\":null,\"ArticlePicture\":[],\"TitleCN\":null,\"AbstractTextCN\":null,\"PMCID\":null,\"EPubDate\":\"\",\"PubModel\":\"\",\"JCR\":\"\",\"JCRName\":\"\",\"Score\":null,\"Total\":0}","platform":"Semanticscholar","paperid":null,"PeriodicalName":"2019 International Symposium on Networks, Computers and Communications (ISNCC)","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/10.1109/ISNCC.2019.8909141","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
Emotion Recognition Based on Decoupling the Spatial Context from the Temporal Dynamics of Facial Expressions
This paper presents an emotion recognition approach based on decoupling the spatial context from the temporal dynamics of facial expressions in video sequences. In particular, each emotional state is represented as a set of temporal phases, where each temporal phase exhibits different temporal dynamics such as the expressing speed and the variable length of each phase. In this work, we have developed an algorithm for automatically detecting the temporal phases of human facial expressions by employing the concept of mutual information to define a similarity measure among different video frames. Moreover, we have developed a two-layer framework for emotional state recognition. The first layer utilizes the spatial context to classify the frames in an input video into emotional-specific temporal phases using a support vector machine classifier. In the second layer, dynamic time warping is used to classify the sequence of labels associated with the video frames, which is generated in the first layer, into a specific emotional state. In order to validate the performance of the proposed approach, we have conducted extensive computer simulations and the results show an average classification accuracy of 93.53% using the extended Cohn-Kanade facial-expression database.