Mitigating noisy labels in long-tailed image classification via multi-level collaborative learning

IF 3.5 2区 计算机科学 Q2 COMPUTER SCIENCE, ARTIFICIAL INTELLIGENCE
Xinyang Zhou, Zhijie Wen, Yuandi Zhao, Jun Shi, Shihui Ying
{"title":"Mitigating noisy labels in long-tailed image classification via multi-level collaborative learning","authors":"Xinyang Zhou,&nbsp;Zhijie Wen,&nbsp;Yuandi Zhao,&nbsp;Jun Shi,&nbsp;Shihui Ying","doi":"10.1007/s10489-025-06809-3","DOIUrl":null,"url":null,"abstract":"<div><p>Label noise and class imbalance are two types of data bias that have attracted widespread attention in the past, but few methods can address both of them simultaneously. Recently, some works have begun to explore handling the two biases concurrently. In this article, we combine feature-level sample selection with logit-level knowledge distillation and logit adjustment to form a more complete collaborative training framework using two neural networks, which is termed <b>D</b>ynamic <b>N</b>oise and <b>I</b>mbalance <b>W</b>eighted <b>D</b>istillation (DNIWD). Firstly, we construct two types of sample sets, which are dynamic high-confidence set and basic confidence set. Based on the former, we estimate the centroids for each class in the latent space and select clean and easy examples for the peer network based on the uncertainty. Secondly, based on the latter, we perform knowledge distillation between the existing two networks to facilitate the learning of all classes, letting the network adaptively adjust the weight of distillation loss based on its own outputs. Meanwhile, we add an auxiliary classifier to each network and apply an improved balanced loss to train it, in order to boost the generalization performance of tail classes in more severe cases of class imbalance and provide balanced predictions for constructing confidence sample sets. Compared to state-of-the-art methods, <b>DNIWD</b> achieves significant improvement on synthetic and real-world datasets.</p></div>","PeriodicalId":8041,"journal":{"name":"Applied Intelligence","volume":"55 14","pages":""},"PeriodicalIF":3.5000,"publicationDate":"2025-09-10","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"Applied Intelligence","FirstCategoryId":"94","ListUrlMain":"https://link.springer.com/article/10.1007/s10489-025-06809-3","RegionNum":2,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"Q2","JCRName":"COMPUTER SCIENCE, ARTIFICIAL INTELLIGENCE","Score":null,"Total":0}
引用次数: 0

Abstract

Label noise and class imbalance are two types of data bias that have attracted widespread attention in the past, but few methods can address both of them simultaneously. Recently, some works have begun to explore handling the two biases concurrently. In this article, we combine feature-level sample selection with logit-level knowledge distillation and logit adjustment to form a more complete collaborative training framework using two neural networks, which is termed Dynamic Noise and Imbalance Weighted Distillation (DNIWD). Firstly, we construct two types of sample sets, which are dynamic high-confidence set and basic confidence set. Based on the former, we estimate the centroids for each class in the latent space and select clean and easy examples for the peer network based on the uncertainty. Secondly, based on the latter, we perform knowledge distillation between the existing two networks to facilitate the learning of all classes, letting the network adaptively adjust the weight of distillation loss based on its own outputs. Meanwhile, we add an auxiliary classifier to each network and apply an improved balanced loss to train it, in order to boost the generalization performance of tail classes in more severe cases of class imbalance and provide balanced predictions for constructing confidence sample sets. Compared to state-of-the-art methods, DNIWD achieves significant improvement on synthetic and real-world datasets.

基于多层次协同学习的长尾图像分类中噪声标签的消除
标签噪声和类不平衡是过去引起广泛关注的两种数据偏差,但很少有方法可以同时解决这两种偏差。最近,一些研究开始探讨如何同时处理这两种偏见。在本文中,我们将特征级样本选择与logit级知识蒸馏和logit调整相结合,形成了一个使用两个神经网络的更完整的协同训练框架,称为动态噪声和不平衡加权蒸馏(DNIWD)。首先,构造了两类样本集,即动态高置信度集和基本置信度集。在前者的基础上,我们估计潜在空间中每个类的质心,并根据不确定性为对等网络选择干净简单的样例。其次,在后者的基础上,我们在现有的两个网络之间进行知识蒸馏,以方便所有类的学习,让网络根据自己的输出自适应调整蒸馏损失的权重。同时,我们在每个网络中添加一个辅助分类器,并使用改进的平衡损失对其进行训练,以提高尾类在更严重的类失衡情况下的泛化性能,并为构建置信样本集提供平衡预测。与最先进的方法相比,DNIWD在合成和真实数据集上取得了显着改进。
本文章由计算机程序翻译,如有差异,请以英文原文为准。
求助全文
约1分钟内获得全文 求助全文
来源期刊
Applied Intelligence
Applied Intelligence 工程技术-计算机:人工智能
CiteScore
6.60
自引率
20.80%
发文量
1361
审稿时长
5.9 months
期刊介绍: With a focus on research in artificial intelligence and neural networks, this journal addresses issues involving solutions of real-life manufacturing, defense, management, government and industrial problems which are too complex to be solved through conventional approaches and require the simulation of intelligent thought processes, heuristics, applications of knowledge, and distributed and parallel processing. The integration of these multiple approaches in solving complex problems is of particular importance. The journal presents new and original research and technological developments, addressing real and complex issues applicable to difficult problems. It provides a medium for exchanging scientific research and technological achievements accomplished by the international community.
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
确定
请完成安全验证×
copy
已复制链接
快去分享给好友吧!
我知道了
右上角分享
点击右上角分享
0
联系我们:info@booksci.cn Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。 Copyright © 2023 布克学术 All rights reserved.
京ICP备2023020795号-1
ghs 京公网安备 11010802042870号
Book学术文献互助
Book学术文献互助群
群 号:604180095
Book学术官方微信