Ying Jin;Zhangjie Cao;Ximei Wang;Jianmin Wang;Mingsheng Long
{"title":"One Fits Many: Class Confusion Loss for Versatile Domain Adaptation","authors":"Ying Jin;Zhangjie Cao;Ximei Wang;Jianmin Wang;Mingsheng Long","doi":"10.1109/TPAMI.2024.3392565","DOIUrl":null,"url":null,"abstract":"In the open world, various label sets and domain configurations give rise to a variety of Domain Adaptation (DA) setups, including closed-set, partial-set, open-set, and universal DA, as well as multi-source and multi-target DA. It is notable that existing DA methods are generally designed only for a specific setup, and may under-perform in setups they are not tailored to. This paper shifts the common paradigm of DA to Versatile Domain Adaptation (VDA), where one method can handle several different DA setups without any modification. Towards this goal, we first delve into a general inductive bias: class confusion, and then uncover that reducing such pairwise class confusion leads to significant transfer gains. With this insight, we propose one general class confusion loss (CC-Loss) to learn many setups. We estimate class confusion based only on classifier predictions and minimize the class confusion to enable accurate target predictions. Further, we improve the loss by enforcing the consistency of confusion matrices under different data augmentations to encourage its invariance to distribution perturbations. Experiments on 2D vision and 3D vision benchmarks show that the CC-Loss performs competitively in different mainstream DA setups.","PeriodicalId":94034,"journal":{"name":"IEEE transactions on pattern analysis and machine intelligence","volume":"46 11","pages":"7251-7266"},"PeriodicalIF":18.6000,"publicationDate":"2024-04-23","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"IEEE transactions on pattern analysis and machine intelligence","FirstCategoryId":"1085","ListUrlMain":"https://ieeexplore.ieee.org/document/10506994/","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
引用次数: 0
Abstract
In the open world, various label sets and domain configurations give rise to a variety of Domain Adaptation (DA) setups, including closed-set, partial-set, open-set, and universal DA, as well as multi-source and multi-target DA. It is notable that existing DA methods are generally designed only for a specific setup, and may under-perform in setups they are not tailored to. This paper shifts the common paradigm of DA to Versatile Domain Adaptation (VDA), where one method can handle several different DA setups without any modification. Towards this goal, we first delve into a general inductive bias: class confusion, and then uncover that reducing such pairwise class confusion leads to significant transfer gains. With this insight, we propose one general class confusion loss (CC-Loss) to learn many setups. We estimate class confusion based only on classifier predictions and minimize the class confusion to enable accurate target predictions. Further, we improve the loss by enforcing the consistency of confusion matrices under different data augmentations to encourage its invariance to distribution perturbations. Experiments on 2D vision and 3D vision benchmarks show that the CC-Loss performs competitively in different mainstream DA setups.
在开放世界中,各种标签集和域配置产生了各种域适应(DA)设置,包括封闭集、部分集、开放集和通用 DA,以及多源和多目标 DA。值得注意的是,现有的 DA 方法通常只针对特定的设置而设计,在不适合的设置中可能表现不佳。本文将常见的检测范式转变为多功能领域适应(VDA),即一种方法无需任何修改即可处理多种不同的检测设置。为了实现这一目标,我们首先深入研究了一种普遍的归纳偏差:类别混淆,然后发现减少这种成对类别混淆会带来显著的转移收益。基于这一认识,我们提出了一种通用的类混淆损失(CC-Loss)来学习多种设置。我们仅根据分类器的预测来估计类混淆,并将类混淆最小化,从而实现准确的目标预测。此外,我们还通过强化不同数据增强下混淆矩阵的一致性来改进损失,从而提高其对分布扰动的不变性。对二维视觉和三维视觉基准的实验表明,CC-Loss 在不同的主流 DA 设置中表现出很强的竞争力。