基于Kolmogorov-Arnold适配器的多尺度视觉混合专家显著目标检测

IF 5.5 2区 计算机科学 Q1 COMPUTER SCIENCE, ARTIFICIAL INTELLIGENCE
Chaojun Cen , Fei Li , Ping Hu , Zhenbo Li
{"title":"基于Kolmogorov-Arnold适配器的多尺度视觉混合专家显著目标检测","authors":"Chaojun Cen ,&nbsp;Fei Li ,&nbsp;Ping Hu ,&nbsp;Zhenbo Li","doi":"10.1016/j.neucom.2025.130349","DOIUrl":null,"url":null,"abstract":"<div><div>Diverse domains and object variations make salient object detection (SOD) a challenging task in computer vision. Many previous studies have adopted multi-scale neural networks with attention mechanisms. Although they are popular, the design of their networks lacks sufficient flexibility, which hinders their generalization across objects of different scales and domains. To address the above issue, we propose a novel mixture-of-experts salient object detection (MoESOD) approach. We design a multi-scale mixture-of-experts (MMoE) module, essentially large neural networks, to improve the model’s expressive power and generalization ability without significantly increasing computational cost. By leveraging expert competition and collaboration strategies, the MMoE module effectively integrates contributions from different experts. The MMoE module not only captures multi-scale features but also effectively fuses semantic information across scales through the expert gating mechanism. Additionally, the novel Kolmogorov–Arnold adapter (KAA) is designed to enhance the model’s flexibility, allowing it to adapt easily to SOD tasks across different domains. Comprehensive experiments show that MoESOD consistently achieves higher performance than, or at least comparable performance to, state-of-the-art methods on <em>17</em> different SOD benchmarks and <em>1</em> downstream tasks. To the best of our knowledge, this is the first study to explore Kolmogorov–Arnold network within the SOD community.</div></div>","PeriodicalId":19268,"journal":{"name":"Neurocomputing","volume":"644 ","pages":"Article 130349"},"PeriodicalIF":5.5000,"publicationDate":"2025-05-12","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":"{\"title\":\"A multi-scale vision mixture-of-experts for salient object detection with Kolmogorov–Arnold adapter\",\"authors\":\"Chaojun Cen ,&nbsp;Fei Li ,&nbsp;Ping Hu ,&nbsp;Zhenbo Li\",\"doi\":\"10.1016/j.neucom.2025.130349\",\"DOIUrl\":null,\"url\":null,\"abstract\":\"<div><div>Diverse domains and object variations make salient object detection (SOD) a challenging task in computer vision. Many previous studies have adopted multi-scale neural networks with attention mechanisms. Although they are popular, the design of their networks lacks sufficient flexibility, which hinders their generalization across objects of different scales and domains. To address the above issue, we propose a novel mixture-of-experts salient object detection (MoESOD) approach. We design a multi-scale mixture-of-experts (MMoE) module, essentially large neural networks, to improve the model’s expressive power and generalization ability without significantly increasing computational cost. By leveraging expert competition and collaboration strategies, the MMoE module effectively integrates contributions from different experts. The MMoE module not only captures multi-scale features but also effectively fuses semantic information across scales through the expert gating mechanism. Additionally, the novel Kolmogorov–Arnold adapter (KAA) is designed to enhance the model’s flexibility, allowing it to adapt easily to SOD tasks across different domains. Comprehensive experiments show that MoESOD consistently achieves higher performance than, or at least comparable performance to, state-of-the-art methods on <em>17</em> different SOD benchmarks and <em>1</em> downstream tasks. To the best of our knowledge, this is the first study to explore Kolmogorov–Arnold network within the SOD community.</div></div>\",\"PeriodicalId\":19268,\"journal\":{\"name\":\"Neurocomputing\",\"volume\":\"644 \",\"pages\":\"Article 130349\"},\"PeriodicalIF\":5.5000,\"publicationDate\":\"2025-05-12\",\"publicationTypes\":\"Journal Article\",\"fieldsOfStudy\":null,\"isOpenAccess\":false,\"openAccessPdf\":\"\",\"citationCount\":\"0\",\"resultStr\":null,\"platform\":\"Semanticscholar\",\"paperid\":null,\"PeriodicalName\":\"Neurocomputing\",\"FirstCategoryId\":\"94\",\"ListUrlMain\":\"https://www.sciencedirect.com/science/article/pii/S0925231225010215\",\"RegionNum\":2,\"RegionCategory\":\"计算机科学\",\"ArticlePicture\":[],\"TitleCN\":null,\"AbstractTextCN\":null,\"PMCID\":null,\"EPubDate\":\"\",\"PubModel\":\"\",\"JCR\":\"Q1\",\"JCRName\":\"COMPUTER SCIENCE, ARTIFICIAL INTELLIGENCE\",\"Score\":null,\"Total\":0}","platform":"Semanticscholar","paperid":null,"PeriodicalName":"Neurocomputing","FirstCategoryId":"94","ListUrlMain":"https://www.sciencedirect.com/science/article/pii/S0925231225010215","RegionNum":2,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"Q1","JCRName":"COMPUTER SCIENCE, ARTIFICIAL INTELLIGENCE","Score":null,"Total":0}
引用次数: 0

摘要

不同的领域和目标变化使得显著目标检测(SOD)成为计算机视觉中一个具有挑战性的任务。以往的许多研究都采用了具有注意机制的多尺度神经网络。尽管它们很受欢迎,但它们的网络设计缺乏足够的灵活性,这阻碍了它们在不同尺度和领域对象上的泛化。为了解决上述问题,我们提出了一种新的混合专家显著目标检测(MoESOD)方法。为了在不显著增加计算成本的情况下提高模型的表达能力和泛化能力,我们设计了一个多尺度专家混合(MMoE)模块,本质上是大型神经网络。通过利用专家竞争和协作策略,MMoE模块有效地整合了不同专家的贡献。MMoE模块不仅捕获了多尺度特征,而且通过专家门控机制有效地融合了跨尺度的语义信息。此外,新型Kolmogorov-Arnold适配器(KAA)旨在增强模型的灵活性,使其能够轻松适应不同领域的SOD任务。综合实验表明,在17个不同的SOD基准测试和1个下游任务中,MoESOD的性能始终高于或至少与最先进的方法相当。据我们所知,这是第一个在SOD群落中探索Kolmogorov-Arnold网络的研究。
本文章由计算机程序翻译,如有差异,请以英文原文为准。
A multi-scale vision mixture-of-experts for salient object detection with Kolmogorov–Arnold adapter
Diverse domains and object variations make salient object detection (SOD) a challenging task in computer vision. Many previous studies have adopted multi-scale neural networks with attention mechanisms. Although they are popular, the design of their networks lacks sufficient flexibility, which hinders their generalization across objects of different scales and domains. To address the above issue, we propose a novel mixture-of-experts salient object detection (MoESOD) approach. We design a multi-scale mixture-of-experts (MMoE) module, essentially large neural networks, to improve the model’s expressive power and generalization ability without significantly increasing computational cost. By leveraging expert competition and collaboration strategies, the MMoE module effectively integrates contributions from different experts. The MMoE module not only captures multi-scale features but also effectively fuses semantic information across scales through the expert gating mechanism. Additionally, the novel Kolmogorov–Arnold adapter (KAA) is designed to enhance the model’s flexibility, allowing it to adapt easily to SOD tasks across different domains. Comprehensive experiments show that MoESOD consistently achieves higher performance than, or at least comparable performance to, state-of-the-art methods on 17 different SOD benchmarks and 1 downstream tasks. To the best of our knowledge, this is the first study to explore Kolmogorov–Arnold network within the SOD community.
求助全文
通过发布文献求助,成功后即可免费获取论文全文。 去求助
来源期刊
Neurocomputing
Neurocomputing 工程技术-计算机:人工智能
CiteScore
13.10
自引率
10.00%
发文量
1382
审稿时长
70 days
期刊介绍: Neurocomputing publishes articles describing recent fundamental contributions in the field of neurocomputing. Neurocomputing theory, practice and applications are the essential topics being covered.
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
确定
请完成安全验证×
copy
已复制链接
快去分享给好友吧!
我知道了
右上角分享
点击右上角分享
0
联系我们:info@booksci.cn Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。 Copyright © 2023 布克学术 All rights reserved.
京ICP备2023020795号-1
ghs 京公网安备 11010802042870号
Book学术文献互助
Book学术文献互助群
群 号:481959085
Book学术官方微信