{"title":"SFMamba: a novel spatial-frequency collaborative learning for multimodal medical image fusion with mamba","authors":"Zhaijuan Ding, Zhaisheng Ding, Yunzhe Men, Yanyu Liu, Shengyang Luan, Shufang Tian","doi":"10.1007/s10489-026-07252-8","DOIUrl":null,"url":null,"abstract":"<div><p>The objective of medical image fusion is to discern and amalgamate complementary features extracted from multimodal medical images, producing fused representations that are more interpretable and informative. This enhancement facilitates higher diagnostic accuracy and efficiency. While Transformer-based fusion methods excel at capturing long-range dependencies and enable parallel computation, their computational cost rises sharply with increasing input dimensions, often reaching quadratic complexity. To address this challenge, we propose a novel spatial-frequency domain fusion network, termed as SFMamba, which exploits a frequency transformation in conjunction with the Mamba model to fully leverage both spatial and frequency information. An efficient Mamba branch incorporates a spatial-frequency state-space (SFSS) model, reducing computational burden to linear or near-linear complexity. The selective-band feature extraction (SBFE) branch is constructed using a discrete wavelet pyramid, designed to capture multi-scale frequency components consistently across source images. To dynamically and effectively fuse multi-modal complementary information, we introduce a multi-domain feature fusion (MDFF) module that elevates fusion performance. Training is conducted with a multi-teacher learning strategy (MTLS) that integrates pre-trained convolutional neural networks and transformer-based fusion methods to generate multiple pseudo-labels, guiding the network to inherit fused knowledge from these priors. Extensive experiments demonstrate that SFMamba achieves state-of-the-art performance in both subjective and objective evaluations. The code for SFMamba is available at https://github.com/DZSYUNNAN/SFMamba.</p></div>","PeriodicalId":8041,"journal":{"name":"Applied Intelligence","volume":"56 7","pages":""},"PeriodicalIF":3.5000,"publicationDate":"2026-05-07","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"Applied Intelligence","FirstCategoryId":"94","ListUrlMain":"https://link.springer.com/article/10.1007/s10489-026-07252-8","RegionNum":2,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"Q2","JCRName":"COMPUTER SCIENCE, ARTIFICIAL INTELLIGENCE","Score":null,"Total":0}
引用次数: 0
Abstract
The objective of medical image fusion is to discern and amalgamate complementary features extracted from multimodal medical images, producing fused representations that are more interpretable and informative. This enhancement facilitates higher diagnostic accuracy and efficiency. While Transformer-based fusion methods excel at capturing long-range dependencies and enable parallel computation, their computational cost rises sharply with increasing input dimensions, often reaching quadratic complexity. To address this challenge, we propose a novel spatial-frequency domain fusion network, termed as SFMamba, which exploits a frequency transformation in conjunction with the Mamba model to fully leverage both spatial and frequency information. An efficient Mamba branch incorporates a spatial-frequency state-space (SFSS) model, reducing computational burden to linear or near-linear complexity. The selective-band feature extraction (SBFE) branch is constructed using a discrete wavelet pyramid, designed to capture multi-scale frequency components consistently across source images. To dynamically and effectively fuse multi-modal complementary information, we introduce a multi-domain feature fusion (MDFF) module that elevates fusion performance. Training is conducted with a multi-teacher learning strategy (MTLS) that integrates pre-trained convolutional neural networks and transformer-based fusion methods to generate multiple pseudo-labels, guiding the network to inherit fused knowledge from these priors. Extensive experiments demonstrate that SFMamba achieves state-of-the-art performance in both subjective and objective evaluations. The code for SFMamba is available at https://github.com/DZSYUNNAN/SFMamba.
期刊介绍:
With a focus on research in artificial intelligence and neural networks, this journal addresses issues involving solutions of real-life manufacturing, defense, management, government and industrial problems which are too complex to be solved through conventional approaches and require the simulation of intelligent thought processes, heuristics, applications of knowledge, and distributed and parallel processing. The integration of these multiple approaches in solving complex problems is of particular importance.
The journal presents new and original research and technological developments, addressing real and complex issues applicable to difficult problems. It provides a medium for exchanging scientific research and technological achievements accomplished by the international community.