Robust label propagation based on prior-guided cross domain data augmentation for few-shot unsupervised domain adaptation

IF 7.6 1区 计算机科学 Q1 COMPUTER SCIENCE, ARTIFICIAL INTELLIGENCE
Peng Zhao , Jiakun Shi , Ping Ye , Huiting Liu , Xia Ji
{"title":"Robust label propagation based on prior-guided cross domain data augmentation for few-shot unsupervised domain adaptation","authors":"Peng Zhao ,&nbsp;Jiakun Shi ,&nbsp;Ping Ye ,&nbsp;Huiting Liu ,&nbsp;Xia Ji","doi":"10.1016/j.knosys.2025.114432","DOIUrl":null,"url":null,"abstract":"<div><div>Few-shot unsupervised domain adaptation (FS-UDA) aims to leverage knowledge from an imbalanced, labeled source domain and apply it to an unlabeled target domain. The primary difficulties of FS-UDA stem from the disparity in data distributions across source and target domains, coupled with uneven class representation in the source data. Label propagation (LP) is commonly used in domain adaptation scenarios. However, in FS-UDA tasks, LP disproportionately favors the normal classes because the source domain suffers from imbalanced class distribution, which results in insufficient feature representation and a large domain gap for the few-shot classes. To tackle these problems, we introduce a new robust LP approach that leverages prior-guided cross-domain data augmentation for FS-UDA. Unlike conventional approaches that solely utilize source domain visual data for few-shot class augmentation, our proposed method employs contrastive language image pretraining-derived semantic priors to supervise visual feature extractor training and optimize few-shot prototypes. It enhances domain-invariant feature learning while mitigating cross-domain distribution mismatches. We introduce the visual information from the target domain to perform data augmentation via style transfer, obtaining more diverse class-specific information. Subsequently, we capture intradomain and interdomain relationships more accurately by constructing intradomain and interdomain graphs independently for all samples (original and augmented) from both domains, which facilitates more effective LP and makes LP robust to few-shot classes. Furthermore, we introduce an adaptive graph regularization loss to dynamically adjust class weights, enhance intraclass compactness within domains, and reduce intraclass distribution discrepancies between different domains. Comprehensive experiments validate that the proposed method achieves superior performance compared to existing state-of-the-art methods across various FS-UDA tasks. The proposed method achieves 77.3 % and 61.7 % average accuracies for few-shot classes on the Office-31 and Office-Home datasets, respectively.</div></div>","PeriodicalId":49939,"journal":{"name":"Knowledge-Based Systems","volume":"329 ","pages":"Article 114432"},"PeriodicalIF":7.6000,"publicationDate":"2025-09-09","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"Knowledge-Based Systems","FirstCategoryId":"94","ListUrlMain":"https://www.sciencedirect.com/science/article/pii/S0950705125014716","RegionNum":1,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"Q1","JCRName":"COMPUTER SCIENCE, ARTIFICIAL INTELLIGENCE","Score":null,"Total":0}
引用次数: 0

Abstract

Few-shot unsupervised domain adaptation (FS-UDA) aims to leverage knowledge from an imbalanced, labeled source domain and apply it to an unlabeled target domain. The primary difficulties of FS-UDA stem from the disparity in data distributions across source and target domains, coupled with uneven class representation in the source data. Label propagation (LP) is commonly used in domain adaptation scenarios. However, in FS-UDA tasks, LP disproportionately favors the normal classes because the source domain suffers from imbalanced class distribution, which results in insufficient feature representation and a large domain gap for the few-shot classes. To tackle these problems, we introduce a new robust LP approach that leverages prior-guided cross-domain data augmentation for FS-UDA. Unlike conventional approaches that solely utilize source domain visual data for few-shot class augmentation, our proposed method employs contrastive language image pretraining-derived semantic priors to supervise visual feature extractor training and optimize few-shot prototypes. It enhances domain-invariant feature learning while mitigating cross-domain distribution mismatches. We introduce the visual information from the target domain to perform data augmentation via style transfer, obtaining more diverse class-specific information. Subsequently, we capture intradomain and interdomain relationships more accurately by constructing intradomain and interdomain graphs independently for all samples (original and augmented) from both domains, which facilitates more effective LP and makes LP robust to few-shot classes. Furthermore, we introduce an adaptive graph regularization loss to dynamically adjust class weights, enhance intraclass compactness within domains, and reduce intraclass distribution discrepancies between different domains. Comprehensive experiments validate that the proposed method achieves superior performance compared to existing state-of-the-art methods across various FS-UDA tasks. The proposed method achieves 77.3 % and 61.7 % average accuracies for few-shot classes on the Office-31 and Office-Home datasets, respectively.
基于先验引导跨域数据增强的鲁棒标签传播,用于少镜头无监督域自适应
少射无监督域自适应(FS-UDA)旨在利用来自不平衡的标记源域的知识并将其应用于未标记的目标域。FS-UDA的主要困难源于源域和目标域之间数据分布的差异,以及源数据中不均匀的类表示。标签传播(Label propagation, LP)是一种常用的域适应场景。然而,在FS-UDA任务中,由于源域的类分布不平衡,LP不成比例地倾向于正常类,这导致特征表示不足,并且对于少数射击类存在较大的域间隙。为了解决这些问题,我们引入了一种新的鲁棒LP方法,该方法利用FS-UDA的先验引导跨域数据增强。与传统方法仅利用源域视觉数据进行少镜头类增强不同,我们提出的方法采用对比语言图像预训练衍生的语义先验来监督视觉特征提取器的训练并优化少镜头原型。它增强了域不变特征学习,同时减轻了跨域分布不匹配。我们引入目标域的视觉信息,通过风格迁移进行数据增强,获得更多样化的类特定信息。随后,我们通过对来自两个域的所有样本(原始和增强)独立构建域内和域间图来更准确地捕获域内和域间关系,从而促进了更有效的LP,并使LP对少数射击类具有鲁棒性。此外,我们引入自适应图正则化损失来动态调整类权值,增强域内的类内紧密性,减少不同域间的类内分布差异。综合实验证明,与现有的最先进的方法相比,所提出的方法在各种FS-UDA任务中具有优越的性能。该方法在Office-31和Office-Home数据集上的平均准确率分别达到77.3%和61.7%。
本文章由计算机程序翻译,如有差异,请以英文原文为准。
求助全文
约1分钟内获得全文 求助全文
来源期刊
Knowledge-Based Systems
Knowledge-Based Systems 工程技术-计算机:人工智能
CiteScore
14.80
自引率
12.50%
发文量
1245
审稿时长
7.8 months
期刊介绍: Knowledge-Based Systems, an international and interdisciplinary journal in artificial intelligence, publishes original, innovative, and creative research results in the field. It focuses on knowledge-based and other artificial intelligence techniques-based systems. The journal aims to support human prediction and decision-making through data science and computation techniques, provide a balanced coverage of theory and practical study, and encourage the development and implementation of knowledge-based intelligence models, methods, systems, and software tools. Applications in business, government, education, engineering, and healthcare are emphasized.
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
确定
请完成安全验证×
copy
已复制链接
快去分享给好友吧!
我知道了
右上角分享
点击右上角分享
0
联系我们:info@booksci.cn Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。 Copyright © 2023 布克学术 All rights reserved.
京ICP备2023020795号-1
ghs 京公网安备 11010802042870号
Book学术文献互助
Book学术文献互助群
群 号:604180095
Book学术官方微信