A multi-scale pyramid residual weight network for medical image fusion.

IF 2.9 2区 医学 Q2 RADIOLOGY, NUCLEAR MEDICINE & MEDICAL IMAGING
Quantitative Imaging in Medicine and Surgery Pub Date : 2025-03-03 Epub Date: 2025-02-26 DOI:10.21037/qims-24-851
Yiwei Liu, Shaoze Zhang, Yao Tang, Xihai Zhao, Zuo-Xiang He
{"title":"A multi-scale pyramid residual weight network for medical image fusion.","authors":"Yiwei Liu, Shaoze Zhang, Yao Tang, Xihai Zhao, Zuo-Xiang He","doi":"10.21037/qims-24-851","DOIUrl":null,"url":null,"abstract":"<p><strong>Background: </strong>Due to the inherent limitations of imaging sensors, acquiring medical images that simultaneously provide functional metabolic information and detailed structural organization remains a significant challenge. Multi-modal image fusion has emerged as a critical technology for clinical diagnosis and surgical navigation, as it enables the integration of complementary information from different imaging modalities. However, existing deep learning (DL)-based fusion methods often face difficulties in effectively combining high-frequency detail information with low-frequency contextual information, which frequently leads to the degradation of high-frequency details. Therefore, there is a pressing need for a method that addresses these challenges, preserving both high- and low-frequency information while maintaining clear structural contours. In response to this issue, a novel convolutional neural network (CNN), named the multi-scale pyramid residual weight network (LYWNet), is proposed. The objective of this approach is to improve the fusion process by effectively integrating high- and low-frequency information, thereby enhancing the quality and accuracy of multimodal image fusion. This method aims to overcome the limitations of current fusion techniques and ensure the preservation of both functional and structural details, ultimately contributing to more precise clinical diagnoses and better surgical navigation outcomes.</p><p><strong>Methods: </strong>We propose a novel CNN, LYWNet, designed to address these challenges. LYWNet is composed of three modules: (I) data preprocessing module: utilizes three convolutional layers to extract both deep and shallow features from the input images. (II) Feature extraction module: incorporates three identical multi-scale pyramid residual weight (LYW) blocks in series, each featuring three interactive branches to preserve high-frequency detail information effectively. (III) Image reconstruction module: utilizes a fusion algorithm based on feature distillation to ensure the effective integration of functional and anatomical information. The proposed image fusion algorithm enhances the interaction of contextual cues and retains the metabolic details from functional images while preserving texture details from anatomical images.</p><p><strong>Results: </strong>The proposed LYWNet demonstrated its ability to retain high-frequency details during feature extraction, effectively combining them with low-frequency contextual information. The fusion results exhibited reduced differences between the fused image and the original images. The structural similarity (SSIM) and peak signal-to-noise ratio (PSNR) were 0.5592±0.0536 and 17.3594±1.0211, respectively, for single-photon emission computed tomography-magnetic resonance imaging (SPECT-MRI), 0.5195±0.0730 and 14.5324±1.7365 for PET-MRI; 0.5376±0.0442 and 13.9202±0.7265 for magnetic resonance imaging-computed tomography.</p><p><strong>Conclusions: </strong>LYWNet excels at integrating high-frequency detail information and low-frequency contextual information, addressing the deficiencies of existing DL-based image fusion methods. This approach provides superior fused images that retain the functional metabolic information and anatomical texture, making it a valuable tool for clinical diagnosis and surgical navigation.</p>","PeriodicalId":54267,"journal":{"name":"Quantitative Imaging in Medicine and Surgery","volume":"15 3","pages":"1793-1821"},"PeriodicalIF":2.9000,"publicationDate":"2025-03-03","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.ncbi.nlm.nih.gov/pmc/articles/PMC11948399/pdf/","citationCount":"0","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"Quantitative Imaging in Medicine and Surgery","FirstCategoryId":"3","ListUrlMain":"https://doi.org/10.21037/qims-24-851","RegionNum":2,"RegionCategory":"医学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"2025/2/26 0:00:00","PubModel":"Epub","JCR":"Q2","JCRName":"RADIOLOGY, NUCLEAR MEDICINE & MEDICAL IMAGING","Score":null,"Total":0}
引用次数: 0

Abstract

Background: Due to the inherent limitations of imaging sensors, acquiring medical images that simultaneously provide functional metabolic information and detailed structural organization remains a significant challenge. Multi-modal image fusion has emerged as a critical technology for clinical diagnosis and surgical navigation, as it enables the integration of complementary information from different imaging modalities. However, existing deep learning (DL)-based fusion methods often face difficulties in effectively combining high-frequency detail information with low-frequency contextual information, which frequently leads to the degradation of high-frequency details. Therefore, there is a pressing need for a method that addresses these challenges, preserving both high- and low-frequency information while maintaining clear structural contours. In response to this issue, a novel convolutional neural network (CNN), named the multi-scale pyramid residual weight network (LYWNet), is proposed. The objective of this approach is to improve the fusion process by effectively integrating high- and low-frequency information, thereby enhancing the quality and accuracy of multimodal image fusion. This method aims to overcome the limitations of current fusion techniques and ensure the preservation of both functional and structural details, ultimately contributing to more precise clinical diagnoses and better surgical navigation outcomes.

Methods: We propose a novel CNN, LYWNet, designed to address these challenges. LYWNet is composed of three modules: (I) data preprocessing module: utilizes three convolutional layers to extract both deep and shallow features from the input images. (II) Feature extraction module: incorporates three identical multi-scale pyramid residual weight (LYW) blocks in series, each featuring three interactive branches to preserve high-frequency detail information effectively. (III) Image reconstruction module: utilizes a fusion algorithm based on feature distillation to ensure the effective integration of functional and anatomical information. The proposed image fusion algorithm enhances the interaction of contextual cues and retains the metabolic details from functional images while preserving texture details from anatomical images.

Results: The proposed LYWNet demonstrated its ability to retain high-frequency details during feature extraction, effectively combining them with low-frequency contextual information. The fusion results exhibited reduced differences between the fused image and the original images. The structural similarity (SSIM) and peak signal-to-noise ratio (PSNR) were 0.5592±0.0536 and 17.3594±1.0211, respectively, for single-photon emission computed tomography-magnetic resonance imaging (SPECT-MRI), 0.5195±0.0730 and 14.5324±1.7365 for PET-MRI; 0.5376±0.0442 and 13.9202±0.7265 for magnetic resonance imaging-computed tomography.

Conclusions: LYWNet excels at integrating high-frequency detail information and low-frequency contextual information, addressing the deficiencies of existing DL-based image fusion methods. This approach provides superior fused images that retain the functional metabolic information and anatomical texture, making it a valuable tool for clinical diagnosis and surgical navigation.

一种用于医学图像融合的多尺度金字塔残差权网络。
背景:由于成像传感器的固有局限性,获取同时提供功能代谢信息和详细结构组织的医学图像仍然是一个重大挑战。多模态图像融合已经成为临床诊断和手术导航的关键技术,因为它可以整合来自不同成像模式的互补信息。然而,现有的基于深度学习(DL)的融合方法往往难以有效地将高频细节信息与低频上下文信息相结合,从而导致高频细节信息的退化。因此,迫切需要一种方法来解决这些挑战,在保持清晰结构轮廓的同时保留高频和低频信息。针对这一问题,提出了一种新的卷积神经网络(CNN),称为多尺度金字塔残差权网络(LYWNet)。该方法的目的是通过有效地融合高低频信息来改进融合过程,从而提高多模态图像融合的质量和精度。该方法旨在克服当前融合技术的局限性,确保功能和结构细节的保留,最终有助于更精确的临床诊断和更好的手术导航结果。方法:我们提出了一种新颖的CNN, LYWNet,旨在解决这些挑战。LYWNet由三个模块组成:(1)数据预处理模块:利用三个卷积层从输入图像中提取深层和浅层特征。(二)特征提取模块:将三个相同的多尺度金字塔残差权重(LYW)块串联在一起,每个块具有三个交互分支,有效保存高频细节信息。(三)图像重建模块:采用基于特征蒸馏的融合算法,保证功能信息和解剖信息的有效融合。提出的图像融合算法增强了上下文线索的相互作用,保留了功能图像中的代谢细节,同时保留了解剖图像中的纹理细节。结果:所提出的LYWNet能够在特征提取过程中保留高频细节,并有效地将其与低频上下文信息相结合。融合结果显示融合后的图像与原始图像之间的差异减小。单光子发射计算机断层-磁共振成像(SPECT-MRI)的结构相似性(SSIM)和峰值信噪比(PSNR)分别为0.5592±0.0536和17.3594±1.0211,PET-MRI的结构相似性(SSIM)和峰值信噪比(PSNR)分别为0.5195±0.0730和14.5324±1.7365;磁共振成像-计算机断层扫描为0.5376±0.0442和13.9202±0.7265。结论:LYWNet在融合高频细节信息和低频上下文信息方面表现出色,解决了现有基于dl的图像融合方法的不足。该方法提供了优越的融合图像,保留了功能代谢信息和解剖纹理,使其成为临床诊断和手术导航的宝贵工具。
本文章由计算机程序翻译,如有差异,请以英文原文为准。
求助全文
约1分钟内获得全文 求助全文
来源期刊
Quantitative Imaging in Medicine and Surgery
Quantitative Imaging in Medicine and Surgery Medicine-Radiology, Nuclear Medicine and Imaging
CiteScore
4.20
自引率
17.90%
发文量
252
期刊介绍: Information not localized
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
确定
请完成安全验证×
copy
已复制链接
快去分享给好友吧!
我知道了
右上角分享
点击右上角分享
0
联系我们:info@booksci.cn Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。 Copyright © 2023 布克学术 All rights reserved.
京ICP备2023020795号-1
ghs 京公网安备 11010802042870号
Book学术文献互助
Book学术文献互助群
群 号:481959085
Book学术官方微信