Shiqiang Liu , Weisheng Li , Dan He , Guofen Wang , Yuping Huang
{"title":"SSEFusion: Salient semantic enhancement for multimodal medical image fusion with Mamba and dynamic spiking neural networks","authors":"Shiqiang Liu , Weisheng Li , Dan He , Guofen Wang , Yuping Huang","doi":"10.1016/j.inffus.2025.103031","DOIUrl":null,"url":null,"abstract":"<div><div>Multimodal medical image fusion technology enhances medical representations and plays a vital role in clinical diagnosis. However, fusing medical images remains a challenge due to the stochastic nature of lesions and the complex structures of organs. Although many fusion methods have been proposed recently, most struggle to effectively establish global context dependency while preserving salient semantic features, leading to the loss of crucial medical information. Therefore, we propose a novel salient semantic enhancement fusion (SSEFusion) framework, whose key components include a dual-branch encoder that combines Mamba and spiking neural network (SNN) models (Mamba-SNN encoder), feature interaction attention (FIA) blocks, and a decoder equipped with detail enhancement (DE) blocks. In the encoder, the Mamba-based branch introduces visual state space (VSS) blocks to efficiently establish global dependencies and extract global features for the effective identification of the lesion area. Meanwhile, the SNN-based branch dynamically extracts fine-grained salient features to enhance the retention of medical semantic information in complex structures. Global features and fine-grained salient features semantically interact to achieve feature complementarity through the FIA blocks. Benefiting from the DE block, SSEFusion generates fused images with prominent edge textures. Furthermore, we propose a salient semantic loss based on leaky-integrate-and-fire (LIF) neurons to enhance the guidance in extracting key features. Extensive fusion experiments show that SSEFusion outperforms state-of-the-art fusion methods in terms of salient medical semantic information retention. The code is available at <span><span>https://github.com/Shiqiang-Liu/SSEFusion</span><svg><path></path></svg></span>.</div></div>","PeriodicalId":50367,"journal":{"name":"Information Fusion","volume":"119 ","pages":"Article 103031"},"PeriodicalIF":14.7000,"publicationDate":"2025-02-19","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"Information Fusion","FirstCategoryId":"94","ListUrlMain":"https://www.sciencedirect.com/science/article/pii/S1566253525001046","RegionNum":1,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"Q1","JCRName":"COMPUTER SCIENCE, ARTIFICIAL INTELLIGENCE","Score":null,"Total":0}
引用次数: 0
Abstract
Multimodal medical image fusion technology enhances medical representations and plays a vital role in clinical diagnosis. However, fusing medical images remains a challenge due to the stochastic nature of lesions and the complex structures of organs. Although many fusion methods have been proposed recently, most struggle to effectively establish global context dependency while preserving salient semantic features, leading to the loss of crucial medical information. Therefore, we propose a novel salient semantic enhancement fusion (SSEFusion) framework, whose key components include a dual-branch encoder that combines Mamba and spiking neural network (SNN) models (Mamba-SNN encoder), feature interaction attention (FIA) blocks, and a decoder equipped with detail enhancement (DE) blocks. In the encoder, the Mamba-based branch introduces visual state space (VSS) blocks to efficiently establish global dependencies and extract global features for the effective identification of the lesion area. Meanwhile, the SNN-based branch dynamically extracts fine-grained salient features to enhance the retention of medical semantic information in complex structures. Global features and fine-grained salient features semantically interact to achieve feature complementarity through the FIA blocks. Benefiting from the DE block, SSEFusion generates fused images with prominent edge textures. Furthermore, we propose a salient semantic loss based on leaky-integrate-and-fire (LIF) neurons to enhance the guidance in extracting key features. Extensive fusion experiments show that SSEFusion outperforms state-of-the-art fusion methods in terms of salient medical semantic information retention. The code is available at https://github.com/Shiqiang-Liu/SSEFusion.
期刊介绍:
Information Fusion serves as a central platform for showcasing advancements in multi-sensor, multi-source, multi-process information fusion, fostering collaboration among diverse disciplines driving its progress. It is the leading outlet for sharing research and development in this field, focusing on architectures, algorithms, and applications. Papers dealing with fundamental theoretical analyses as well as those demonstrating their application to real-world problems will be welcome.