{"title":"基于Kolmogorov-Arnold适配器的多尺度视觉混合专家显著目标检测","authors":"Chaojun Cen , Fei Li , Ping Hu , Zhenbo Li","doi":"10.1016/j.neucom.2025.130349","DOIUrl":null,"url":null,"abstract":"<div><div>Diverse domains and object variations make salient object detection (SOD) a challenging task in computer vision. Many previous studies have adopted multi-scale neural networks with attention mechanisms. Although they are popular, the design of their networks lacks sufficient flexibility, which hinders their generalization across objects of different scales and domains. To address the above issue, we propose a novel mixture-of-experts salient object detection (MoESOD) approach. We design a multi-scale mixture-of-experts (MMoE) module, essentially large neural networks, to improve the model’s expressive power and generalization ability without significantly increasing computational cost. By leveraging expert competition and collaboration strategies, the MMoE module effectively integrates contributions from different experts. The MMoE module not only captures multi-scale features but also effectively fuses semantic information across scales through the expert gating mechanism. Additionally, the novel Kolmogorov–Arnold adapter (KAA) is designed to enhance the model’s flexibility, allowing it to adapt easily to SOD tasks across different domains. Comprehensive experiments show that MoESOD consistently achieves higher performance than, or at least comparable performance to, state-of-the-art methods on <em>17</em> different SOD benchmarks and <em>1</em> downstream tasks. To the best of our knowledge, this is the first study to explore Kolmogorov–Arnold network within the SOD community.</div></div>","PeriodicalId":19268,"journal":{"name":"Neurocomputing","volume":"644 ","pages":"Article 130349"},"PeriodicalIF":5.5000,"publicationDate":"2025-05-12","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":"{\"title\":\"A multi-scale vision mixture-of-experts for salient object detection with Kolmogorov–Arnold adapter\",\"authors\":\"Chaojun Cen , Fei Li , Ping Hu , Zhenbo Li\",\"doi\":\"10.1016/j.neucom.2025.130349\",\"DOIUrl\":null,\"url\":null,\"abstract\":\"<div><div>Diverse domains and object variations make salient object detection (SOD) a challenging task in computer vision. Many previous studies have adopted multi-scale neural networks with attention mechanisms. Although they are popular, the design of their networks lacks sufficient flexibility, which hinders their generalization across objects of different scales and domains. To address the above issue, we propose a novel mixture-of-experts salient object detection (MoESOD) approach. We design a multi-scale mixture-of-experts (MMoE) module, essentially large neural networks, to improve the model’s expressive power and generalization ability without significantly increasing computational cost. By leveraging expert competition and collaboration strategies, the MMoE module effectively integrates contributions from different experts. The MMoE module not only captures multi-scale features but also effectively fuses semantic information across scales through the expert gating mechanism. Additionally, the novel Kolmogorov–Arnold adapter (KAA) is designed to enhance the model’s flexibility, allowing it to adapt easily to SOD tasks across different domains. Comprehensive experiments show that MoESOD consistently achieves higher performance than, or at least comparable performance to, state-of-the-art methods on <em>17</em> different SOD benchmarks and <em>1</em> downstream tasks. To the best of our knowledge, this is the first study to explore Kolmogorov–Arnold network within the SOD community.</div></div>\",\"PeriodicalId\":19268,\"journal\":{\"name\":\"Neurocomputing\",\"volume\":\"644 \",\"pages\":\"Article 130349\"},\"PeriodicalIF\":5.5000,\"publicationDate\":\"2025-05-12\",\"publicationTypes\":\"Journal Article\",\"fieldsOfStudy\":null,\"isOpenAccess\":false,\"openAccessPdf\":\"\",\"citationCount\":\"0\",\"resultStr\":null,\"platform\":\"Semanticscholar\",\"paperid\":null,\"PeriodicalName\":\"Neurocomputing\",\"FirstCategoryId\":\"94\",\"ListUrlMain\":\"https://www.sciencedirect.com/science/article/pii/S0925231225010215\",\"RegionNum\":2,\"RegionCategory\":\"计算机科学\",\"ArticlePicture\":[],\"TitleCN\":null,\"AbstractTextCN\":null,\"PMCID\":null,\"EPubDate\":\"\",\"PubModel\":\"\",\"JCR\":\"Q1\",\"JCRName\":\"COMPUTER SCIENCE, ARTIFICIAL INTELLIGENCE\",\"Score\":null,\"Total\":0}","platform":"Semanticscholar","paperid":null,"PeriodicalName":"Neurocomputing","FirstCategoryId":"94","ListUrlMain":"https://www.sciencedirect.com/science/article/pii/S0925231225010215","RegionNum":2,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"Q1","JCRName":"COMPUTER SCIENCE, ARTIFICIAL INTELLIGENCE","Score":null,"Total":0}
A multi-scale vision mixture-of-experts for salient object detection with Kolmogorov–Arnold adapter
Diverse domains and object variations make salient object detection (SOD) a challenging task in computer vision. Many previous studies have adopted multi-scale neural networks with attention mechanisms. Although they are popular, the design of their networks lacks sufficient flexibility, which hinders their generalization across objects of different scales and domains. To address the above issue, we propose a novel mixture-of-experts salient object detection (MoESOD) approach. We design a multi-scale mixture-of-experts (MMoE) module, essentially large neural networks, to improve the model’s expressive power and generalization ability without significantly increasing computational cost. By leveraging expert competition and collaboration strategies, the MMoE module effectively integrates contributions from different experts. The MMoE module not only captures multi-scale features but also effectively fuses semantic information across scales through the expert gating mechanism. Additionally, the novel Kolmogorov–Arnold adapter (KAA) is designed to enhance the model’s flexibility, allowing it to adapt easily to SOD tasks across different domains. Comprehensive experiments show that MoESOD consistently achieves higher performance than, or at least comparable performance to, state-of-the-art methods on 17 different SOD benchmarks and 1 downstream tasks. To the best of our knowledge, this is the first study to explore Kolmogorov–Arnold network within the SOD community.
期刊介绍:
Neurocomputing publishes articles describing recent fundamental contributions in the field of neurocomputing. Neurocomputing theory, practice and applications are the essential topics being covered.