Xinyang Zhou, Zhijie Wen, Yuandi Zhao, Jun Shi, Shihui Ying
{"title":"基于多层次协同学习的长尾图像分类中噪声标签的消除","authors":"Xinyang Zhou, Zhijie Wen, Yuandi Zhao, Jun Shi, Shihui Ying","doi":"10.1007/s10489-025-06809-3","DOIUrl":null,"url":null,"abstract":"<div><p>Label noise and class imbalance are two types of data bias that have attracted widespread attention in the past, but few methods can address both of them simultaneously. Recently, some works have begun to explore handling the two biases concurrently. In this article, we combine feature-level sample selection with logit-level knowledge distillation and logit adjustment to form a more complete collaborative training framework using two neural networks, which is termed <b>D</b>ynamic <b>N</b>oise and <b>I</b>mbalance <b>W</b>eighted <b>D</b>istillation (DNIWD). Firstly, we construct two types of sample sets, which are dynamic high-confidence set and basic confidence set. Based on the former, we estimate the centroids for each class in the latent space and select clean and easy examples for the peer network based on the uncertainty. Secondly, based on the latter, we perform knowledge distillation between the existing two networks to facilitate the learning of all classes, letting the network adaptively adjust the weight of distillation loss based on its own outputs. Meanwhile, we add an auxiliary classifier to each network and apply an improved balanced loss to train it, in order to boost the generalization performance of tail classes in more severe cases of class imbalance and provide balanced predictions for constructing confidence sample sets. Compared to state-of-the-art methods, <b>DNIWD</b> achieves significant improvement on synthetic and real-world datasets.</p></div>","PeriodicalId":8041,"journal":{"name":"Applied Intelligence","volume":"55 14","pages":""},"PeriodicalIF":3.5000,"publicationDate":"2025-09-10","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":"{\"title\":\"Mitigating noisy labels in long-tailed image classification via multi-level collaborative learning\",\"authors\":\"Xinyang Zhou, Zhijie Wen, Yuandi Zhao, Jun Shi, Shihui Ying\",\"doi\":\"10.1007/s10489-025-06809-3\",\"DOIUrl\":null,\"url\":null,\"abstract\":\"<div><p>Label noise and class imbalance are two types of data bias that have attracted widespread attention in the past, but few methods can address both of them simultaneously. Recently, some works have begun to explore handling the two biases concurrently. In this article, we combine feature-level sample selection with logit-level knowledge distillation and logit adjustment to form a more complete collaborative training framework using two neural networks, which is termed <b>D</b>ynamic <b>N</b>oise and <b>I</b>mbalance <b>W</b>eighted <b>D</b>istillation (DNIWD). Firstly, we construct two types of sample sets, which are dynamic high-confidence set and basic confidence set. Based on the former, we estimate the centroids for each class in the latent space and select clean and easy examples for the peer network based on the uncertainty. Secondly, based on the latter, we perform knowledge distillation between the existing two networks to facilitate the learning of all classes, letting the network adaptively adjust the weight of distillation loss based on its own outputs. Meanwhile, we add an auxiliary classifier to each network and apply an improved balanced loss to train it, in order to boost the generalization performance of tail classes in more severe cases of class imbalance and provide balanced predictions for constructing confidence sample sets. Compared to state-of-the-art methods, <b>DNIWD</b> achieves significant improvement on synthetic and real-world datasets.</p></div>\",\"PeriodicalId\":8041,\"journal\":{\"name\":\"Applied Intelligence\",\"volume\":\"55 14\",\"pages\":\"\"},\"PeriodicalIF\":3.5000,\"publicationDate\":\"2025-09-10\",\"publicationTypes\":\"Journal Article\",\"fieldsOfStudy\":null,\"isOpenAccess\":false,\"openAccessPdf\":\"\",\"citationCount\":\"0\",\"resultStr\":null,\"platform\":\"Semanticscholar\",\"paperid\":null,\"PeriodicalName\":\"Applied Intelligence\",\"FirstCategoryId\":\"94\",\"ListUrlMain\":\"https://link.springer.com/article/10.1007/s10489-025-06809-3\",\"RegionNum\":2,\"RegionCategory\":\"计算机科学\",\"ArticlePicture\":[],\"TitleCN\":null,\"AbstractTextCN\":null,\"PMCID\":null,\"EPubDate\":\"\",\"PubModel\":\"\",\"JCR\":\"Q2\",\"JCRName\":\"COMPUTER SCIENCE, ARTIFICIAL INTELLIGENCE\",\"Score\":null,\"Total\":0}","platform":"Semanticscholar","paperid":null,"PeriodicalName":"Applied Intelligence","FirstCategoryId":"94","ListUrlMain":"https://link.springer.com/article/10.1007/s10489-025-06809-3","RegionNum":2,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"Q2","JCRName":"COMPUTER SCIENCE, ARTIFICIAL INTELLIGENCE","Score":null,"Total":0}
Mitigating noisy labels in long-tailed image classification via multi-level collaborative learning
Label noise and class imbalance are two types of data bias that have attracted widespread attention in the past, but few methods can address both of them simultaneously. Recently, some works have begun to explore handling the two biases concurrently. In this article, we combine feature-level sample selection with logit-level knowledge distillation and logit adjustment to form a more complete collaborative training framework using two neural networks, which is termed Dynamic Noise and Imbalance Weighted Distillation (DNIWD). Firstly, we construct two types of sample sets, which are dynamic high-confidence set and basic confidence set. Based on the former, we estimate the centroids for each class in the latent space and select clean and easy examples for the peer network based on the uncertainty. Secondly, based on the latter, we perform knowledge distillation between the existing two networks to facilitate the learning of all classes, letting the network adaptively adjust the weight of distillation loss based on its own outputs. Meanwhile, we add an auxiliary classifier to each network and apply an improved balanced loss to train it, in order to boost the generalization performance of tail classes in more severe cases of class imbalance and provide balanced predictions for constructing confidence sample sets. Compared to state-of-the-art methods, DNIWD achieves significant improvement on synthetic and real-world datasets.
期刊介绍:
With a focus on research in artificial intelligence and neural networks, this journal addresses issues involving solutions of real-life manufacturing, defense, management, government and industrial problems which are too complex to be solved through conventional approaches and require the simulation of intelligent thought processes, heuristics, applications of knowledge, and distributed and parallel processing. The integration of these multiple approaches in solving complex problems is of particular importance.
The journal presents new and original research and technological developments, addressing real and complex issues applicable to difficult problems. It provides a medium for exchanging scientific research and technological achievements accomplished by the international community.