{"title":"Facial expression recognition by learning spatiotemporal features with multi-layer independent subspace analysis","authors":"ChenHan Lin, Fei Long, Yongjie Zhan","doi":"10.1109/CISP-BMEI.2017.8301920","DOIUrl":null,"url":null,"abstract":"We propose to learn spatiotemporal features for video-based facial expression recognition with multi-layer independent subspace analysis (ISA) algorithm. On the first layer, a set of ISA filters are learned from small 3D patches of the video data, and then more abstract and powerful features on the second layer are learned from the feature responses of the first layer. Two public facial expression databases, extended Cohn-Kanade and MMI are used to evaluate our method. Experimental results show that the features learned by multi-layer architecture achieve better recognition performance than that of single-layer model. Furthermore, our method outperforms popular hand-crafted features, and the overall accuracy of our method is comparable to some related feature learning based methods.","PeriodicalId":6474,"journal":{"name":"2017 10th International Congress on Image and Signal Processing, BioMedical Engineering and Informatics (CISP-BMEI)","volume":"45 1","pages":"1-6"},"PeriodicalIF":0.0000,"publicationDate":"2017-10-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"5","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"2017 10th International Congress on Image and Signal Processing, BioMedical Engineering and Informatics (CISP-BMEI)","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/10.1109/CISP-BMEI.2017.8301920","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
引用次数: 5
Abstract
We propose to learn spatiotemporal features for video-based facial expression recognition with multi-layer independent subspace analysis (ISA) algorithm. On the first layer, a set of ISA filters are learned from small 3D patches of the video data, and then more abstract and powerful features on the second layer are learned from the feature responses of the first layer. Two public facial expression databases, extended Cohn-Kanade and MMI are used to evaluate our method. Experimental results show that the features learned by multi-layer architecture achieve better recognition performance than that of single-layer model. Furthermore, our method outperforms popular hand-crafted features, and the overall accuracy of our method is comparable to some related feature learning based methods.