{"title":"基于高频的多谱关注域泛化","authors":"Surong Ying, Xinghao Song, Hongpeng Wang","doi":"10.1007/s10462-025-11217-7","DOIUrl":null,"url":null,"abstract":"<div><p>Deep learning models have made great progress in many vision tasks, but they suffer from domain shift problem when exposed to out-of-distribution scenarios. Domain generalization (DG) is proposed to learn a model from several observable source domains that can generalize well to unknown target domains. Although recent advances in DG works have achieved promising performance, there is a high demand for computational resource, especially those that employ meta-learning or ensemble learning strategies. However, some pioneering works propose to replace convolutional neural network (CNN) as the backbone architecture with multi-layer perceptron (MLP)-like models that can not only learn long-range spatial dependencies but also reduce network parameters using Fourier transform-based techniques. Inspired by this, in this paper, we propose a high-frequency-based multi-spectral attention (HMCA) to facilitate a lightweight MLP-like model to learn global domain-invariant features by focusing on high-frequency components sufficiently. Moreover, we adopt a data augmentation strategy based on Fourier transform to simulate domain shift, thus enabling the model to pay more attention on robust features. Extensive experiments on benchmark datasets demonstrate that our method is superior to the existing CNN-based and MLP-based DG methods.</p></div>","PeriodicalId":8449,"journal":{"name":"Artificial Intelligence Review","volume":"58 8","pages":""},"PeriodicalIF":10.7000,"publicationDate":"2025-05-14","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://link.springer.com/content/pdf/10.1007/s10462-025-11217-7.pdf","citationCount":"0","resultStr":"{\"title\":\"High-frequency-based multi-spectral attention for domain generalization\",\"authors\":\"Surong Ying, Xinghao Song, Hongpeng Wang\",\"doi\":\"10.1007/s10462-025-11217-7\",\"DOIUrl\":null,\"url\":null,\"abstract\":\"<div><p>Deep learning models have made great progress in many vision tasks, but they suffer from domain shift problem when exposed to out-of-distribution scenarios. Domain generalization (DG) is proposed to learn a model from several observable source domains that can generalize well to unknown target domains. Although recent advances in DG works have achieved promising performance, there is a high demand for computational resource, especially those that employ meta-learning or ensemble learning strategies. However, some pioneering works propose to replace convolutional neural network (CNN) as the backbone architecture with multi-layer perceptron (MLP)-like models that can not only learn long-range spatial dependencies but also reduce network parameters using Fourier transform-based techniques. Inspired by this, in this paper, we propose a high-frequency-based multi-spectral attention (HMCA) to facilitate a lightweight MLP-like model to learn global domain-invariant features by focusing on high-frequency components sufficiently. Moreover, we adopt a data augmentation strategy based on Fourier transform to simulate domain shift, thus enabling the model to pay more attention on robust features. Extensive experiments on benchmark datasets demonstrate that our method is superior to the existing CNN-based and MLP-based DG methods.</p></div>\",\"PeriodicalId\":8449,\"journal\":{\"name\":\"Artificial Intelligence Review\",\"volume\":\"58 8\",\"pages\":\"\"},\"PeriodicalIF\":10.7000,\"publicationDate\":\"2025-05-14\",\"publicationTypes\":\"Journal Article\",\"fieldsOfStudy\":null,\"isOpenAccess\":false,\"openAccessPdf\":\"https://link.springer.com/content/pdf/10.1007/s10462-025-11217-7.pdf\",\"citationCount\":\"0\",\"resultStr\":null,\"platform\":\"Semanticscholar\",\"paperid\":null,\"PeriodicalName\":\"Artificial Intelligence Review\",\"FirstCategoryId\":\"94\",\"ListUrlMain\":\"https://link.springer.com/article/10.1007/s10462-025-11217-7\",\"RegionNum\":2,\"RegionCategory\":\"计算机科学\",\"ArticlePicture\":[],\"TitleCN\":null,\"AbstractTextCN\":null,\"PMCID\":null,\"EPubDate\":\"\",\"PubModel\":\"\",\"JCR\":\"Q1\",\"JCRName\":\"COMPUTER SCIENCE, ARTIFICIAL INTELLIGENCE\",\"Score\":null,\"Total\":0}","platform":"Semanticscholar","paperid":null,"PeriodicalName":"Artificial Intelligence Review","FirstCategoryId":"94","ListUrlMain":"https://link.springer.com/article/10.1007/s10462-025-11217-7","RegionNum":2,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"Q1","JCRName":"COMPUTER SCIENCE, ARTIFICIAL INTELLIGENCE","Score":null,"Total":0}
High-frequency-based multi-spectral attention for domain generalization
Deep learning models have made great progress in many vision tasks, but they suffer from domain shift problem when exposed to out-of-distribution scenarios. Domain generalization (DG) is proposed to learn a model from several observable source domains that can generalize well to unknown target domains. Although recent advances in DG works have achieved promising performance, there is a high demand for computational resource, especially those that employ meta-learning or ensemble learning strategies. However, some pioneering works propose to replace convolutional neural network (CNN) as the backbone architecture with multi-layer perceptron (MLP)-like models that can not only learn long-range spatial dependencies but also reduce network parameters using Fourier transform-based techniques. Inspired by this, in this paper, we propose a high-frequency-based multi-spectral attention (HMCA) to facilitate a lightweight MLP-like model to learn global domain-invariant features by focusing on high-frequency components sufficiently. Moreover, we adopt a data augmentation strategy based on Fourier transform to simulate domain shift, thus enabling the model to pay more attention on robust features. Extensive experiments on benchmark datasets demonstrate that our method is superior to the existing CNN-based and MLP-based DG methods.
期刊介绍:
Artificial Intelligence Review, a fully open access journal, publishes cutting-edge research in artificial intelligence and cognitive science. It features critical evaluations of applications, techniques, and algorithms, providing a platform for both researchers and application developers. The journal includes refereed survey and tutorial articles, along with reviews and commentary on significant developments in the field.