{"title":"Relation-aware Network for Facial Expression Recognition","authors":"Xin Ma, Yingdong Ma","doi":"10.1109/FG57933.2023.10042525","DOIUrl":null,"url":null,"abstract":"Facial expression recognition (FER) is a challenging computer vision task due to problems including intra-class variation, occlusion, head-pose variation, etc. The convolutional neural networks (CNNs) have been widely adopted to implement facial expression classification. While convolutional operation captures local information effectively, CNN-models ignore relations between pixels and channels. In this work, we present a Relation-aware Network (RANet) for facial expression classification. RANet is composed of two relational attention modules to construct relationships of spatial positions and channels. Global relationships help RANet focusing on discriminative facial regions to alleviate the above problems. The separable convolution has been applied to compute spatial attention efficiently. Experimental results demonstrate that our proposed method achieves 89.57% and 65.09% accuracy rate on the RAF-DB dataset and the AffectNet-7 dataset, respectively.","PeriodicalId":318766,"journal":{"name":"2023 IEEE 17th International Conference on Automatic Face and Gesture Recognition (FG)","volume":"19 1","pages":"0"},"PeriodicalIF":0.0000,"publicationDate":"2023-01-05","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"1","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"2023 IEEE 17th International Conference on Automatic Face and Gesture Recognition (FG)","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/10.1109/FG57933.2023.10042525","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
引用次数: 1
Abstract
Facial expression recognition (FER) is a challenging computer vision task due to problems including intra-class variation, occlusion, head-pose variation, etc. The convolutional neural networks (CNNs) have been widely adopted to implement facial expression classification. While convolutional operation captures local information effectively, CNN-models ignore relations between pixels and channels. In this work, we present a Relation-aware Network (RANet) for facial expression classification. RANet is composed of two relational attention modules to construct relationships of spatial positions and channels. Global relationships help RANet focusing on discriminative facial regions to alleviate the above problems. The separable convolution has been applied to compute spatial attention efficiently. Experimental results demonstrate that our proposed method achieves 89.57% and 65.09% accuracy rate on the RAF-DB dataset and the AffectNet-7 dataset, respectively.