High-frequency-based multi-spectral attention for domain generalization

IF 10.7 2区 计算机科学 Q1 COMPUTER SCIENCE, ARTIFICIAL INTELLIGENCE
Surong Ying, Xinghao Song, Hongpeng Wang
{"title":"High-frequency-based multi-spectral attention for domain generalization","authors":"Surong Ying,&nbsp;Xinghao Song,&nbsp;Hongpeng Wang","doi":"10.1007/s10462-025-11217-7","DOIUrl":null,"url":null,"abstract":"<div><p>Deep learning models have made great progress in many vision tasks, but they suffer from domain shift problem when exposed to out-of-distribution scenarios. Domain generalization (DG) is proposed to learn a model from several observable source domains that can generalize well to unknown target domains. Although recent advances in DG works have achieved promising performance, there is a high demand for computational resource, especially those that employ meta-learning or ensemble learning strategies. However, some pioneering works propose to replace convolutional neural network (CNN) as the backbone architecture with multi-layer perceptron (MLP)-like models that can not only learn long-range spatial dependencies but also reduce network parameters using Fourier transform-based techniques. Inspired by this, in this paper, we propose a high-frequency-based multi-spectral attention (HMCA) to facilitate a lightweight MLP-like model to learn global domain-invariant features by focusing on high-frequency components sufficiently. Moreover, we adopt a data augmentation strategy based on Fourier transform to simulate domain shift, thus enabling the model to pay more attention on robust features. Extensive experiments on benchmark datasets demonstrate that our method is superior to the existing CNN-based and MLP-based DG methods.</p></div>","PeriodicalId":8449,"journal":{"name":"Artificial Intelligence Review","volume":"58 8","pages":""},"PeriodicalIF":10.7000,"publicationDate":"2025-05-14","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://link.springer.com/content/pdf/10.1007/s10462-025-11217-7.pdf","citationCount":"0","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"Artificial Intelligence Review","FirstCategoryId":"94","ListUrlMain":"https://link.springer.com/article/10.1007/s10462-025-11217-7","RegionNum":2,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"Q1","JCRName":"COMPUTER SCIENCE, ARTIFICIAL INTELLIGENCE","Score":null,"Total":0}
引用次数: 0

Abstract

Deep learning models have made great progress in many vision tasks, but they suffer from domain shift problem when exposed to out-of-distribution scenarios. Domain generalization (DG) is proposed to learn a model from several observable source domains that can generalize well to unknown target domains. Although recent advances in DG works have achieved promising performance, there is a high demand for computational resource, especially those that employ meta-learning or ensemble learning strategies. However, some pioneering works propose to replace convolutional neural network (CNN) as the backbone architecture with multi-layer perceptron (MLP)-like models that can not only learn long-range spatial dependencies but also reduce network parameters using Fourier transform-based techniques. Inspired by this, in this paper, we propose a high-frequency-based multi-spectral attention (HMCA) to facilitate a lightweight MLP-like model to learn global domain-invariant features by focusing on high-frequency components sufficiently. Moreover, we adopt a data augmentation strategy based on Fourier transform to simulate domain shift, thus enabling the model to pay more attention on robust features. Extensive experiments on benchmark datasets demonstrate that our method is superior to the existing CNN-based and MLP-based DG methods.

基于高频的多谱关注域泛化
深度学习模型在许多视觉任务中都取得了很大的进步,但在非分布场景中存在领域转移问题。领域泛化(DG)是一种从多个可观察的源域学习模型的方法,可以很好地泛化到未知的目标域。尽管近年来在分布式学习方面的研究取得了很好的进展,但对计算资源的需求很大,特别是那些采用元学习或集成学习策略的研究。然而,一些开创性的工作提出用多层感知器(MLP)类模型取代卷积神经网络(CNN)作为主干架构,该模型不仅可以学习远程空间依赖关系,还可以使用基于傅里叶变换的技术减少网络参数。受此启发,本文提出了一种基于高频的多频谱注意(HMCA),通过充分关注高频成分,促进轻量级mlp模型学习全局域不变特征。此外,我们采用基于傅里叶变换的数据增强策略来模拟域移位,从而使模型更加关注鲁棒性特征。在基准数据集上的大量实验表明,我们的方法优于现有的基于cnn和基于mlp的DG方法。
本文章由计算机程序翻译,如有差异,请以英文原文为准。
求助全文
约1分钟内获得全文 求助全文
来源期刊
Artificial Intelligence Review
Artificial Intelligence Review 工程技术-计算机:人工智能
CiteScore
22.00
自引率
3.30%
发文量
194
审稿时长
5.3 months
期刊介绍: Artificial Intelligence Review, a fully open access journal, publishes cutting-edge research in artificial intelligence and cognitive science. It features critical evaluations of applications, techniques, and algorithms, providing a platform for both researchers and application developers. The journal includes refereed survey and tutorial articles, along with reviews and commentary on significant developments in the field.
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
确定
请完成安全验证×
copy
已复制链接
快去分享给好友吧!
我知道了
右上角分享
点击右上角分享
0
联系我们:info@booksci.cn Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。 Copyright © 2023 布克学术 All rights reserved.
京ICP备2023020795号-1
ghs 京公网安备 11010802042870号
Book学术文献互助
Book学术文献互助群
群 号:604180095
Book学术官方微信