{"title":"SAFFusion:一种用于多模态医学图像融合的显著性感知频率融合网络。","authors":"Renhe Liu, Yu Liu, Han Wang, Junxian Li, Kai Hu","doi":"10.1364/BOE.555458","DOIUrl":null,"url":null,"abstract":"<p><p>Medical image fusion integrates complementary information from multimodal medical images to provide comprehensive references for clinical decision-making, such as the diagnosis of Alzheimer's disease and the detection and segmentation of brain tumors. Although traditional and deep learning-based fusion methods have been extensively studied, they often fail to devise targeted strategies that fully utilize distinct regional or feature-specific information. This paper proposes SAFFusion, a saliency-aware frequency fusion network that integrates intensity and texture cues from multimodal medical images. We first introduce Mamba-UNet, a multiscale encoder-decoder architecture enhanced by the Mamba design, to improve global modeling in feature extraction and image reconstruction. By employing the contourlet transform in Mamba-UNet, we replace conventional pooling with multiscale representations and decompose spatial features into high- and low-frequency subbands. A dual-branch frequency feature fusion module then fuses cross-modality information according to the distinct characteristics of these frequency subbands. Furthermore, we apply latent low-rank representation (LatLRR) to assess image saliency and implement adaptive loss constraints to preserve information in salient and non-salient regions. Quantitative results on CT/MRI, SPECT/MRI, and PET/MRI fusion tasks show that SAFFusion outperforms state-of-the-art methods. Qualitative evaluations confirm that SAFFusion effectively merges prominent intensity features and rich textures from multiple sources.</p>","PeriodicalId":8969,"journal":{"name":"Biomedical optics express","volume":"16 6","pages":"2459-2481"},"PeriodicalIF":3.2000,"publicationDate":"2025-05-27","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.ncbi.nlm.nih.gov/pmc/articles/PMC12265500/pdf/","citationCount":"0","resultStr":"{\"title\":\"SAFFusion: a saliency-aware frequency fusion network for multimodal medical image fusion.\",\"authors\":\"Renhe Liu, Yu Liu, Han Wang, Junxian Li, Kai Hu\",\"doi\":\"10.1364/BOE.555458\",\"DOIUrl\":null,\"url\":null,\"abstract\":\"<p><p>Medical image fusion integrates complementary information from multimodal medical images to provide comprehensive references for clinical decision-making, such as the diagnosis of Alzheimer's disease and the detection and segmentation of brain tumors. Although traditional and deep learning-based fusion methods have been extensively studied, they often fail to devise targeted strategies that fully utilize distinct regional or feature-specific information. This paper proposes SAFFusion, a saliency-aware frequency fusion network that integrates intensity and texture cues from multimodal medical images. We first introduce Mamba-UNet, a multiscale encoder-decoder architecture enhanced by the Mamba design, to improve global modeling in feature extraction and image reconstruction. By employing the contourlet transform in Mamba-UNet, we replace conventional pooling with multiscale representations and decompose spatial features into high- and low-frequency subbands. A dual-branch frequency feature fusion module then fuses cross-modality information according to the distinct characteristics of these frequency subbands. Furthermore, we apply latent low-rank representation (LatLRR) to assess image saliency and implement adaptive loss constraints to preserve information in salient and non-salient regions. Quantitative results on CT/MRI, SPECT/MRI, and PET/MRI fusion tasks show that SAFFusion outperforms state-of-the-art methods. Qualitative evaluations confirm that SAFFusion effectively merges prominent intensity features and rich textures from multiple sources.</p>\",\"PeriodicalId\":8969,\"journal\":{\"name\":\"Biomedical optics express\",\"volume\":\"16 6\",\"pages\":\"2459-2481\"},\"PeriodicalIF\":3.2000,\"publicationDate\":\"2025-05-27\",\"publicationTypes\":\"Journal Article\",\"fieldsOfStudy\":null,\"isOpenAccess\":false,\"openAccessPdf\":\"https://www.ncbi.nlm.nih.gov/pmc/articles/PMC12265500/pdf/\",\"citationCount\":\"0\",\"resultStr\":null,\"platform\":\"Semanticscholar\",\"paperid\":null,\"PeriodicalName\":\"Biomedical optics express\",\"FirstCategoryId\":\"3\",\"ListUrlMain\":\"https://doi.org/10.1364/BOE.555458\",\"RegionNum\":2,\"RegionCategory\":\"医学\",\"ArticlePicture\":[],\"TitleCN\":null,\"AbstractTextCN\":null,\"PMCID\":null,\"EPubDate\":\"2025/6/1 0:00:00\",\"PubModel\":\"eCollection\",\"JCR\":\"Q2\",\"JCRName\":\"BIOCHEMICAL RESEARCH METHODS\",\"Score\":null,\"Total\":0}","platform":"Semanticscholar","paperid":null,"PeriodicalName":"Biomedical optics express","FirstCategoryId":"3","ListUrlMain":"https://doi.org/10.1364/BOE.555458","RegionNum":2,"RegionCategory":"医学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"2025/6/1 0:00:00","PubModel":"eCollection","JCR":"Q2","JCRName":"BIOCHEMICAL RESEARCH METHODS","Score":null,"Total":0}
SAFFusion: a saliency-aware frequency fusion network for multimodal medical image fusion.
Medical image fusion integrates complementary information from multimodal medical images to provide comprehensive references for clinical decision-making, such as the diagnosis of Alzheimer's disease and the detection and segmentation of brain tumors. Although traditional and deep learning-based fusion methods have been extensively studied, they often fail to devise targeted strategies that fully utilize distinct regional or feature-specific information. This paper proposes SAFFusion, a saliency-aware frequency fusion network that integrates intensity and texture cues from multimodal medical images. We first introduce Mamba-UNet, a multiscale encoder-decoder architecture enhanced by the Mamba design, to improve global modeling in feature extraction and image reconstruction. By employing the contourlet transform in Mamba-UNet, we replace conventional pooling with multiscale representations and decompose spatial features into high- and low-frequency subbands. A dual-branch frequency feature fusion module then fuses cross-modality information according to the distinct characteristics of these frequency subbands. Furthermore, we apply latent low-rank representation (LatLRR) to assess image saliency and implement adaptive loss constraints to preserve information in salient and non-salient regions. Quantitative results on CT/MRI, SPECT/MRI, and PET/MRI fusion tasks show that SAFFusion outperforms state-of-the-art methods. Qualitative evaluations confirm that SAFFusion effectively merges prominent intensity features and rich textures from multiple sources.
期刊介绍:
The journal''s scope encompasses fundamental research, technology development, biomedical studies and clinical applications. BOEx focuses on the leading edge topics in the field, including:
Tissue optics and spectroscopy
Novel microscopies
Optical coherence tomography
Diffuse and fluorescence tomography
Photoacoustic and multimodal imaging
Molecular imaging and therapies
Nanophotonic biosensing
Optical biophysics/photobiology
Microfluidic optical devices
Vision research.