Jinglin Zhang , Qiangchang Wang , Jing Li , Yilong Yin
{"title":"A noise-robust and generalizable framework for facial expression recognition","authors":"Jinglin Zhang , Qiangchang Wang , Jing Li , Yilong Yin","doi":"10.1016/j.ins.2025.122457","DOIUrl":null,"url":null,"abstract":"<div><div>Facial Expression Recognition (FER) shows promising applicability in various real-world contexts, including criminal investigations and digital entertainment. Existing cross-domain FER methods primarily focus on spatial domain features sensitive to noise. However, these methods may propagate noise from the source domain to unseen target domains, degrading recognition performance. To address this, we propose a Noise-Robust and Generalizable framework for FER (NR-GFER), mainly comprising Residual Adapter (RA), Fourier Prompt (FP) modules, and a cross-stage unified fusion mechanism. Specifically, the RA module flexibly transfers the generalization ability of a visual-language large model to FER. Leveraging the residual mechanism improves the discriminative ability of spatial domain features. However, the domain gap may lead FER models to capture source domain-specific noise, which adversely affects performance on target domains. To mitigate this, the FP module extracts frequency domain features via the Fourier transform, integrates them with prompts, and reconstructs them back to the spatial domain through the inverse Fourier transform, thus reducing the negative impact of noise from the source domain. Finally, the cross-stage unified fusion mechanism that bridges intra-module and inter-module semantic priorities, simplifying hyperparameter optimization. Comprehensive evaluations across seven in-the-wild FER datasets confirm that our NR-GFER achieves state-of-the-art performance.</div></div>","PeriodicalId":51063,"journal":{"name":"Information Sciences","volume":"719 ","pages":"Article 122457"},"PeriodicalIF":6.8000,"publicationDate":"2025-06-26","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"Information Sciences","FirstCategoryId":"94","ListUrlMain":"https://www.sciencedirect.com/science/article/pii/S0020025525005894","RegionNum":1,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"0","JCRName":"COMPUTER SCIENCE, INFORMATION SYSTEMS","Score":null,"Total":0}
引用次数: 0
Abstract
Facial Expression Recognition (FER) shows promising applicability in various real-world contexts, including criminal investigations and digital entertainment. Existing cross-domain FER methods primarily focus on spatial domain features sensitive to noise. However, these methods may propagate noise from the source domain to unseen target domains, degrading recognition performance. To address this, we propose a Noise-Robust and Generalizable framework for FER (NR-GFER), mainly comprising Residual Adapter (RA), Fourier Prompt (FP) modules, and a cross-stage unified fusion mechanism. Specifically, the RA module flexibly transfers the generalization ability of a visual-language large model to FER. Leveraging the residual mechanism improves the discriminative ability of spatial domain features. However, the domain gap may lead FER models to capture source domain-specific noise, which adversely affects performance on target domains. To mitigate this, the FP module extracts frequency domain features via the Fourier transform, integrates them with prompts, and reconstructs them back to the spatial domain through the inverse Fourier transform, thus reducing the negative impact of noise from the source domain. Finally, the cross-stage unified fusion mechanism that bridges intra-module and inter-module semantic priorities, simplifying hyperparameter optimization. Comprehensive evaluations across seven in-the-wild FER datasets confirm that our NR-GFER achieves state-of-the-art performance.
期刊介绍:
Informatics and Computer Science Intelligent Systems Applications is an esteemed international journal that focuses on publishing original and creative research findings in the field of information sciences. We also feature a limited number of timely tutorial and surveying contributions.
Our journal aims to cater to a diverse audience, including researchers, developers, managers, strategic planners, graduate students, and anyone interested in staying up-to-date with cutting-edge research in information science, knowledge engineering, and intelligent systems. While readers are expected to share a common interest in information science, they come from varying backgrounds such as engineering, mathematics, statistics, physics, computer science, cell biology, molecular biology, management science, cognitive science, neurobiology, behavioral sciences, and biochemistry.