Adaptive integrated weight unsupervised multi-source domain adaptation without source data

IF 5 2区 计算机科学 Q1 COMPUTER SCIENCE, ARTIFICIAL INTELLIGENCE
Zhirui Wang, Liu Yang, Yahong Han
{"title":"Adaptive integrated weight unsupervised multi-source domain adaptation without source data","authors":"Zhirui Wang, Liu Yang, Yahong Han","doi":"10.1007/s40747-025-01871-3","DOIUrl":null,"url":null,"abstract":"<p>Unsupervised multi-source domain adaptation methods transfer knowledge learned from multiple labeled source domains to an unlabeled target domain. Existing methods assume that all source domain data can be accessed directly. However, such an assumption is unrealistic and causes data privacy concerns, especially when the source domain labels include personal information. In such a setting, it is prohibited to minimize domain gaps by pairwise calculation of the data from the source and target domains. Therefore, this work addresses the source-free unsupervised multi-source domain adaptation problem, where only the source models are available during the adaptation. We propose trust center sample-based source-free domain adaptation (TSDA) method to solve this problem. The key idea is to leverage the pre-trained models from the source domain and progressively train the target model in a self-learning manner. Because target samples with low entropy measured from the pre-trained source model achieve high accuracy, the trust center samples are selected first using the entropy function. Then pseudo labels are assigned for target samples based on a self-supervised pseudo-labeling strategy. For multiple source domains, corresponding target models are learned based on the assigned pseudo labels. Finally, multiple target models are integrated to predict the label for unlabeled target data. Extensive experiment results on some benchmark datasets and generated adversarial samples demonstrate that our approach outperforms existing UMDA methods, even though some methods can always access source data.</p>","PeriodicalId":10524,"journal":{"name":"Complex & Intelligent Systems","volume":"138 1","pages":""},"PeriodicalIF":5.0000,"publicationDate":"2025-04-22","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"Complex & Intelligent Systems","FirstCategoryId":"94","ListUrlMain":"https://doi.org/10.1007/s40747-025-01871-3","RegionNum":2,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"Q1","JCRName":"COMPUTER SCIENCE, ARTIFICIAL INTELLIGENCE","Score":null,"Total":0}
引用次数: 0

Abstract

Unsupervised multi-source domain adaptation methods transfer knowledge learned from multiple labeled source domains to an unlabeled target domain. Existing methods assume that all source domain data can be accessed directly. However, such an assumption is unrealistic and causes data privacy concerns, especially when the source domain labels include personal information. In such a setting, it is prohibited to minimize domain gaps by pairwise calculation of the data from the source and target domains. Therefore, this work addresses the source-free unsupervised multi-source domain adaptation problem, where only the source models are available during the adaptation. We propose trust center sample-based source-free domain adaptation (TSDA) method to solve this problem. The key idea is to leverage the pre-trained models from the source domain and progressively train the target model in a self-learning manner. Because target samples with low entropy measured from the pre-trained source model achieve high accuracy, the trust center samples are selected first using the entropy function. Then pseudo labels are assigned for target samples based on a self-supervised pseudo-labeling strategy. For multiple source domains, corresponding target models are learned based on the assigned pseudo labels. Finally, multiple target models are integrated to predict the label for unlabeled target data. Extensive experiment results on some benchmark datasets and generated adversarial samples demonstrate that our approach outperforms existing UMDA methods, even though some methods can always access source data.

无源数据的自适应积分权无监督多源域自适应
无监督多源领域自适应方法将从多个标记的源领域学习到的知识转移到未标记的目标领域。现有方法假定可以直接访问所有源域数据。然而,这样的假设是不现实的,并且会引起数据隐私问题,特别是当源域标签包含个人信息时。在这种情况下,禁止通过对源域和目标域的数据进行两两计算来最小化域间隙。因此,本文解决了无源无监督多源域自适应问题,即在自适应过程中只有源模型可用。为此,提出了基于信任中心样本的无源域自适应(TSDA)方法。关键思想是利用来自源域的预训练模型,并以自学习的方式逐步训练目标模型。由于从预训练源模型中测量的低熵目标样本具有较高的准确率,因此首先使用熵函数选择信任中心样本。然后基于自监督伪标记策略为目标样本分配伪标签。对于多个源域,根据指定的伪标签学习相应的目标模型。最后,结合多个目标模型对未标记的目标数据进行标签预测。在一些基准数据集和生成的对抗性样本上的大量实验结果表明,尽管一些方法总是可以访问源数据,但我们的方法优于现有的UMDA方法。
本文章由计算机程序翻译,如有差异,请以英文原文为准。
求助全文
约1分钟内获得全文 求助全文
来源期刊
Complex & Intelligent Systems
Complex & Intelligent Systems COMPUTER SCIENCE, ARTIFICIAL INTELLIGENCE-
CiteScore
9.60
自引率
10.30%
发文量
297
期刊介绍: Complex & Intelligent Systems aims to provide a forum for presenting and discussing novel approaches, tools and techniques meant for attaining a cross-fertilization between the broad fields of complex systems, computational simulation, and intelligent analytics and visualization. The transdisciplinary research that the journal focuses on will expand the boundaries of our understanding by investigating the principles and processes that underlie many of the most profound problems facing society today.
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
确定
请完成安全验证×
copy
已复制链接
快去分享给好友吧!
我知道了
右上角分享
点击右上角分享
0
联系我们:info@booksci.cn Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。 Copyright © 2023 布克学术 All rights reserved.
京ICP备2023020795号-1
ghs 京公网安备 11010802042870号
Book学术文献互助
Book学术文献互助群
群 号:481959085
Book学术官方微信