Zhengshan Wang , Long Chen , Juan He , Linyao Yang , Fei-Yue Wang
{"title":"Exploring Latent Transferability of feature components","authors":"Zhengshan Wang , Long Chen , Juan He , Linyao Yang , Fei-Yue Wang","doi":"10.1016/j.patcog.2024.111184","DOIUrl":null,"url":null,"abstract":"<div><div>Feature disentanglement techniques have been widely employed to extract transferable (domain-invariant) features from non-transferable (domain-specific) features in Unsupervised Domain Adaptation (UDA). However, due to the complex interplay among high-dimensional features, the separated “non-transferable” features may still be partially informative. Suppressing or disregarding them, as commonly employed in previous methods, can overlook the inherent transferability. In this work, we introduce two concepts: Partially Transferable Class Features and Partially Transferable Domain Features (PTCF and PTDF), and propose a succinct feature disentanglement technique. Different with prior works, we do not seek to thoroughly peel off the non-transferable features, as it is challenging practically. Instead, we take the two-stage strategy consisting of rough feature disentanglement and dynamic adjustment. We name our model as ELT because it can systematically Explore Latent Transferability of feature components. ELT can automatically evaluate the transferability of internal feature components, dynamically giving more attention to features with high transferability and less to features with low transferability, effectively solving the problem of negative transfer. Extensive experimental results have proved its efficiency. The code and supplementary file will be available at <span><span>https://github.com/njtjmc/ELT</span><svg><path></path></svg></span>.</div></div>","PeriodicalId":49713,"journal":{"name":"Pattern Recognition","volume":"160 ","pages":"Article 111184"},"PeriodicalIF":7.5000,"publicationDate":"2024-11-20","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"Pattern Recognition","FirstCategoryId":"94","ListUrlMain":"https://www.sciencedirect.com/science/article/pii/S003132032400935X","RegionNum":1,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"Q1","JCRName":"COMPUTER SCIENCE, ARTIFICIAL INTELLIGENCE","Score":null,"Total":0}
引用次数: 0
Abstract
Feature disentanglement techniques have been widely employed to extract transferable (domain-invariant) features from non-transferable (domain-specific) features in Unsupervised Domain Adaptation (UDA). However, due to the complex interplay among high-dimensional features, the separated “non-transferable” features may still be partially informative. Suppressing or disregarding them, as commonly employed in previous methods, can overlook the inherent transferability. In this work, we introduce two concepts: Partially Transferable Class Features and Partially Transferable Domain Features (PTCF and PTDF), and propose a succinct feature disentanglement technique. Different with prior works, we do not seek to thoroughly peel off the non-transferable features, as it is challenging practically. Instead, we take the two-stage strategy consisting of rough feature disentanglement and dynamic adjustment. We name our model as ELT because it can systematically Explore Latent Transferability of feature components. ELT can automatically evaluate the transferability of internal feature components, dynamically giving more attention to features with high transferability and less to features with low transferability, effectively solving the problem of negative transfer. Extensive experimental results have proved its efficiency. The code and supplementary file will be available at https://github.com/njtjmc/ELT.
期刊介绍:
The field of Pattern Recognition is both mature and rapidly evolving, playing a crucial role in various related fields such as computer vision, image processing, text analysis, and neural networks. It closely intersects with machine learning and is being applied in emerging areas like biometrics, bioinformatics, multimedia data analysis, and data science. The journal Pattern Recognition, established half a century ago during the early days of computer science, has since grown significantly in scope and influence.