Qingdu Li, Keting Fu, Jian Liu, Yishan Li, Qinze Ren, Kang Xu, Junxiu Fu, Na Liu, Ye Yuan
{"title":"Optimizing Class Imbalance in Facial Expression Recognition Using Dynamic Intra-Class Clustering.","authors":"Qingdu Li, Keting Fu, Jian Liu, Yishan Li, Qinze Ren, Kang Xu, Junxiu Fu, Na Liu, Ye Yuan","doi":"10.3390/biomimetics10050296","DOIUrl":null,"url":null,"abstract":"<p><p>While deep neural networks demonstrate robust performance in visual tasks, the long-tail distribution of real-world data leads to significant recognition accuracy degradation in critical scenarios such as medical human-robot affective interaction, particularly the misidentification of low-frequency negative emotions (e.g., fear and disgust) that may trigger psychological resistance in patients. Here, we propose a method based on dynamic intra-class clustering (DICC) to optimize the class imbalance problem in facial expression recognition tasks. The DICC method dynamically adjusts the distribution of majority classes by clustering them into subclasses and generating pseudo-labels, which helps the model learn more discriminative features and improve classification accuracy. By comparing with existing methods, we demonstrate that the DICC method can help the model achieve superior performance across various facial expression datasets. In this study, we conducted an in-depth evaluation of the DICC method against baseline methods using the FER2013, MMAFEDB, and Emotion-Domestic datasets, achieving improvements in classification accuracy of 1.73%, 1.97%, and 5.48%, respectively. This indicates that the DICC method can effectively enhance classification precision, especially in the recognition of minority class samples. This approach provides a novel perspective for addressing the class imbalance challenge in facial expression recognition and offers a reference for future research and applications in related fields.</p>","PeriodicalId":8907,"journal":{"name":"Biomimetics","volume":"10 5","pages":""},"PeriodicalIF":3.4000,"publicationDate":"2025-05-08","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.ncbi.nlm.nih.gov/pmc/articles/PMC12109554/pdf/","citationCount":"0","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"Biomimetics","FirstCategoryId":"5","ListUrlMain":"https://doi.org/10.3390/biomimetics10050296","RegionNum":3,"RegionCategory":"医学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"Q1","JCRName":"ENGINEERING, MULTIDISCIPLINARY","Score":null,"Total":0}
引用次数: 0
Abstract
While deep neural networks demonstrate robust performance in visual tasks, the long-tail distribution of real-world data leads to significant recognition accuracy degradation in critical scenarios such as medical human-robot affective interaction, particularly the misidentification of low-frequency negative emotions (e.g., fear and disgust) that may trigger psychological resistance in patients. Here, we propose a method based on dynamic intra-class clustering (DICC) to optimize the class imbalance problem in facial expression recognition tasks. The DICC method dynamically adjusts the distribution of majority classes by clustering them into subclasses and generating pseudo-labels, which helps the model learn more discriminative features and improve classification accuracy. By comparing with existing methods, we demonstrate that the DICC method can help the model achieve superior performance across various facial expression datasets. In this study, we conducted an in-depth evaluation of the DICC method against baseline methods using the FER2013, MMAFEDB, and Emotion-Domestic datasets, achieving improvements in classification accuracy of 1.73%, 1.97%, and 5.48%, respectively. This indicates that the DICC method can effectively enhance classification precision, especially in the recognition of minority class samples. This approach provides a novel perspective for addressing the class imbalance challenge in facial expression recognition and offers a reference for future research and applications in related fields.