Zelin Yang , Lin Xu , Shiyang Yan , Haixia Bi , Fan Li
{"title":"IDEAL: Independent domain embedding augmentation learning","authors":"Zelin Yang , Lin Xu , Shiyang Yan , Haixia Bi , Fan Li","doi":"10.1016/j.patcog.2025.112024","DOIUrl":null,"url":null,"abstract":"<div><div>Deep metric learning is fundamental to open-set pattern recognition and has become a focal point of research in recent years. Significant efforts have been devoted to designing sampling, mining, and weighting strategies within algorithmic-level deep metric learning (DML) loss objectives. However, less attention has been paid to input-level but essential data transformations. In this paper, we develop a novel mechanism, independent domain embedding augmentation learning (IDEAL) method. It can simultaneously learn multiple independent embedding spaces for multiple domains generated by predefined data transformations. Our IDEAL is orthogonal to existing DML techniques and can be seamlessly combined with one DML approach for enhanced performance. Empirical results on visual retrieval tasks demonstrate the superiority of the proposed method. For instance, IDEAL significantly improves the performance of both Multi-Similarity (MS) Loss and Hypergraph-Induced Semantic Tuplet (HIST) loss. Specifically, it boosts the Recall<span><math><mrow><mi>@</mi><mn>1</mn></mrow></math></span> from 84.5% <span><math><mo>→</mo></math></span> 87.1% for MS Loss on Cars-196 and from 65.8% <span><math><mo>→</mo></math></span> 69.5% on CUB-200. Similarly, for HIST loss, IDEAL improves the performance on Cars-196 from 87.4% <span><math><mo>→</mo></math></span> 90.3%, on CUB-200 from 69.7% to 72.3%. It significantly outperforms methods using basic network architectures (e.g., ResNet-50, BN-Inception), such as XBM and Intra-Batch. The source code of our proposed method is available at <span><span>https://github.com/emdata-ailab/Ideal-learning</span><svg><path></path></svg></span>.</div></div>","PeriodicalId":49713,"journal":{"name":"Pattern Recognition","volume":"170 ","pages":"Article 112024"},"PeriodicalIF":7.5000,"publicationDate":"2025-07-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"Pattern Recognition","FirstCategoryId":"94","ListUrlMain":"https://www.sciencedirect.com/science/article/pii/S0031320325006843","RegionNum":1,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"Q1","JCRName":"COMPUTER SCIENCE, ARTIFICIAL INTELLIGENCE","Score":null,"Total":0}
引用次数: 0
Abstract
Deep metric learning is fundamental to open-set pattern recognition and has become a focal point of research in recent years. Significant efforts have been devoted to designing sampling, mining, and weighting strategies within algorithmic-level deep metric learning (DML) loss objectives. However, less attention has been paid to input-level but essential data transformations. In this paper, we develop a novel mechanism, independent domain embedding augmentation learning (IDEAL) method. It can simultaneously learn multiple independent embedding spaces for multiple domains generated by predefined data transformations. Our IDEAL is orthogonal to existing DML techniques and can be seamlessly combined with one DML approach for enhanced performance. Empirical results on visual retrieval tasks demonstrate the superiority of the proposed method. For instance, IDEAL significantly improves the performance of both Multi-Similarity (MS) Loss and Hypergraph-Induced Semantic Tuplet (HIST) loss. Specifically, it boosts the Recall from 84.5% 87.1% for MS Loss on Cars-196 and from 65.8% 69.5% on CUB-200. Similarly, for HIST loss, IDEAL improves the performance on Cars-196 from 87.4% 90.3%, on CUB-200 from 69.7% to 72.3%. It significantly outperforms methods using basic network architectures (e.g., ResNet-50, BN-Inception), such as XBM and Intra-Batch. The source code of our proposed method is available at https://github.com/emdata-ailab/Ideal-learning.
期刊介绍:
The field of Pattern Recognition is both mature and rapidly evolving, playing a crucial role in various related fields such as computer vision, image processing, text analysis, and neural networks. It closely intersects with machine learning and is being applied in emerging areas like biometrics, bioinformatics, multimedia data analysis, and data science. The journal Pattern Recognition, established half a century ago during the early days of computer science, has since grown significantly in scope and influence.