SMAFusion: Multimodal medical image fusion based on spatial registration and local-global multi-scale feature adaptive fusion

IF 6.5 2区 计算机科学 Q1 COMPUTER SCIENCE, ARTIFICIAL INTELLIGENCE
Wei Guo, Lifang Wang, Jianchao Zeng, Qiang Han, KaiXin Jin, Xiwen Wang
{"title":"SMAFusion: Multimodal medical image fusion based on spatial registration and local-global multi-scale feature adaptive fusion","authors":"Wei Guo,&nbsp;Lifang Wang,&nbsp;Jianchao Zeng,&nbsp;Qiang Han,&nbsp;KaiXin Jin,&nbsp;Xiwen Wang","doi":"10.1016/j.neucom.2025.131039","DOIUrl":null,"url":null,"abstract":"<div><div>Aiming at the current image fusion methods relying entirely on paired images after registration, the fusion efficiency is low, and the existing image registration methods overlook the impact of image distribution differences on registration outcomes, as well as the poor capability of the image fusion methods for local-global multi-scale feature extraction from source images during image fusion, we propose a multimodal medical image fusion method based on spatial registration and local-global multi-scale feature adaptive fusion (SMAFusion). In SMAFusion, the spatial registration module is utilized to minimize the impact of image distribution differences on the registration results. Meanwhile, a local-global multi-scale feature encoder is proposed to fully extract local and global information at different scales. And an adaptive fusion strategy is employed to fuse the multi-scale features of different modal images. The experimental results show that this method has an average improvement of 24 %, 121.04 %, 19.57 %, 40.36 %, and 97.62 % compared to the eight comparative methods in five evaluation indicators: mutual information (MI), difference correlation sum (SCD), visual fidelity (VIFF), gradient fusion performance (QAB/F) and structural similarity (SSIM), respectively. The fused image preserves more details and texture information, and exhibits high consistency with the original image in terms of overall structure while enhancing the fusion ability for unregistered images. The significant improvement in MI and SCD indicates that the fused image better integrates complementary information from multiple modalities, helping doctors obtain more diagnostic clues from a single image and reducing the risk of missed or incorrect diagnoses. The improvements in VIFF and SSIM suggest that the fused image performs better in structure preservation and visual clarity, making lesion areas and anatomical structures clearer, thereby enhancing clinical readability and diagnostic efficiency. The enhancement in QAB/F further demonstrates the ability to preserve image edges and texture details, which is directly beneficial for observing fine structures.</div></div>","PeriodicalId":19268,"journal":{"name":"Neurocomputing","volume":"652 ","pages":"Article 131039"},"PeriodicalIF":6.5000,"publicationDate":"2025-07-22","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"Neurocomputing","FirstCategoryId":"94","ListUrlMain":"https://www.sciencedirect.com/science/article/pii/S0925231225017114","RegionNum":2,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"Q1","JCRName":"COMPUTER SCIENCE, ARTIFICIAL INTELLIGENCE","Score":null,"Total":0}
引用次数: 0

Abstract

Aiming at the current image fusion methods relying entirely on paired images after registration, the fusion efficiency is low, and the existing image registration methods overlook the impact of image distribution differences on registration outcomes, as well as the poor capability of the image fusion methods for local-global multi-scale feature extraction from source images during image fusion, we propose a multimodal medical image fusion method based on spatial registration and local-global multi-scale feature adaptive fusion (SMAFusion). In SMAFusion, the spatial registration module is utilized to minimize the impact of image distribution differences on the registration results. Meanwhile, a local-global multi-scale feature encoder is proposed to fully extract local and global information at different scales. And an adaptive fusion strategy is employed to fuse the multi-scale features of different modal images. The experimental results show that this method has an average improvement of 24 %, 121.04 %, 19.57 %, 40.36 %, and 97.62 % compared to the eight comparative methods in five evaluation indicators: mutual information (MI), difference correlation sum (SCD), visual fidelity (VIFF), gradient fusion performance (QAB/F) and structural similarity (SSIM), respectively. The fused image preserves more details and texture information, and exhibits high consistency with the original image in terms of overall structure while enhancing the fusion ability for unregistered images. The significant improvement in MI and SCD indicates that the fused image better integrates complementary information from multiple modalities, helping doctors obtain more diagnostic clues from a single image and reducing the risk of missed or incorrect diagnoses. The improvements in VIFF and SSIM suggest that the fused image performs better in structure preservation and visual clarity, making lesion areas and anatomical structures clearer, thereby enhancing clinical readability and diagnostic efficiency. The enhancement in QAB/F further demonstrates the ability to preserve image edges and texture details, which is directly beneficial for observing fine structures.
SMAFusion:基于空间配准和局部-全局多尺度特征自适应融合的多模态医学图像融合
针对目前的图像融合方法完全依赖配准后的成对图像,融合效率低,并且现有的图像配准方法忽略了图像分布差异对配准结果的影响,以及图像融合方法在图像融合过程中对源图像局部-全局多尺度特征提取能力差的问题。提出一种基于空间配准和局部-全局多尺度特征自适应融合(SMAFusion)的多模态医学图像融合方法。在SMAFusion中,利用空间配准模块最小化图像分布差异对配准结果的影响。同时,提出了一种局部-全局多尺度特征编码器,以充分提取不同尺度下的局部和全局信息。采用自适应融合策略融合不同模态图像的多尺度特征。实验结果表明,该方法在互信息(MI)、差异相关和(SCD)、视觉逼真度(VIFF)、梯度融合性能(QAB/F)和结构相似度(SSIM) 5个评价指标上,与8种比较方法相比,平均提高了24 %、121.04 %、19.57 %、40.36 %和97.62 %。融合后的图像保留了更多的细节和纹理信息,在整体结构上与原图像保持了较高的一致性,同时增强了对未配准图像的融合能力。MI和SCD的显著改善表明融合图像更好地整合了多模式的互补信息,帮助医生从单一图像中获得更多的诊断线索,降低了漏诊或误诊的风险。VIFF和SSIM的改进表明融合后的图像在结构保存和视觉清晰度方面有更好的表现,病变区域和解剖结构更加清晰,从而提高了临床的可读性和诊断效率。QAB/F的增强进一步证明了保留图像边缘和纹理细节的能力,这直接有利于观察精细结构。
本文章由计算机程序翻译,如有差异,请以英文原文为准。
求助全文
约1分钟内获得全文 求助全文
来源期刊
Neurocomputing
Neurocomputing 工程技术-计算机:人工智能
CiteScore
13.10
自引率
10.00%
发文量
1382
审稿时长
70 days
期刊介绍: Neurocomputing publishes articles describing recent fundamental contributions in the field of neurocomputing. Neurocomputing theory, practice and applications are the essential topics being covered.
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
确定
请完成安全验证×
copy
已复制链接
快去分享给好友吧!
我知道了
右上角分享
点击右上角分享
0
联系我们:info@booksci.cn Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。 Copyright © 2023 布克学术 All rights reserved.
京ICP备2023020795号-1
ghs 京公网安备 11010802042870号
Book学术文献互助
Book学术文献互助群
群 号:604180095
Book学术官方微信