A multi-scale pyramid residual weight network for medical image fusion.

IF 2.9 2区 医学 Q2 RADIOLOGY, NUCLEAR MEDICINE & MEDICAL IMAGING
Quantitative Imaging in Medicine and Surgery Pub Date : 2025-03-03 Epub Date: 2025-02-26 DOI:10.21037/qims-24-851
Yiwei Liu, Shaoze Zhang, Yao Tang, Xihai Zhao, Zuo-Xiang He
{"title":"A multi-scale pyramid residual weight network for medical image fusion.","authors":"Yiwei Liu, Shaoze Zhang, Yao Tang, Xihai Zhao, Zuo-Xiang He","doi":"10.21037/qims-24-851","DOIUrl":null,"url":null,"abstract":"<p><strong>Background: </strong>Due to the inherent limitations of imaging sensors, acquiring medical images that simultaneously provide functional metabolic information and detailed structural organization remains a significant challenge. Multi-modal image fusion has emerged as a critical technology for clinical diagnosis and surgical navigation, as it enables the integration of complementary information from different imaging modalities. However, existing deep learning (DL)-based fusion methods often face difficulties in effectively combining high-frequency detail information with low-frequency contextual information, which frequently leads to the degradation of high-frequency details. Therefore, there is a pressing need for a method that addresses these challenges, preserving both high- and low-frequency information while maintaining clear structural contours. In response to this issue, a novel convolutional neural network (CNN), named the multi-scale pyramid residual weight network (LYWNet), is proposed. The objective of this approach is to improve the fusion process by effectively integrating high- and low-frequency information, thereby enhancing the quality and accuracy of multimodal image fusion. This method aims to overcome the limitations of current fusion techniques and ensure the preservation of both functional and structural details, ultimately contributing to more precise clinical diagnoses and better surgical navigation outcomes.</p><p><strong>Methods: </strong>We propose a novel CNN, LYWNet, designed to address these challenges. LYWNet is composed of three modules: (I) data preprocessing module: utilizes three convolutional layers to extract both deep and shallow features from the input images. (II) Feature extraction module: incorporates three identical multi-scale pyramid residual weight (LYW) blocks in series, each featuring three interactive branches to preserve high-frequency detail information effectively. (III) Image reconstruction module: utilizes a fusion algorithm based on feature distillation to ensure the effective integration of functional and anatomical information. The proposed image fusion algorithm enhances the interaction of contextual cues and retains the metabolic details from functional images while preserving texture details from anatomical images.</p><p><strong>Results: </strong>The proposed LYWNet demonstrated its ability to retain high-frequency details during feature extraction, effectively combining them with low-frequency contextual information. The fusion results exhibited reduced differences between the fused image and the original images. The structural similarity (SSIM) and peak signal-to-noise ratio (PSNR) were 0.5592±0.0536 and 17.3594±1.0211, respectively, for single-photon emission computed tomography-magnetic resonance imaging (SPECT-MRI), 0.5195±0.0730 and 14.5324±1.7365 for PET-MRI; 0.5376±0.0442 and 13.9202±0.7265 for magnetic resonance imaging-computed tomography.</p><p><strong>Conclusions: </strong>LYWNet excels at integrating high-frequency detail information and low-frequency contextual information, addressing the deficiencies of existing DL-based image fusion methods. This approach provides superior fused images that retain the functional metabolic information and anatomical texture, making it a valuable tool for clinical diagnosis and surgical navigation.</p>","PeriodicalId":54267,"journal":{"name":"Quantitative Imaging in Medicine and Surgery","volume":"15 3","pages":"1793-1821"},"PeriodicalIF":2.9000,"publicationDate":"2025-03-03","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.ncbi.nlm.nih.gov/pmc/articles/PMC11948399/pdf/","citationCount":"0","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"Quantitative Imaging in Medicine and Surgery","FirstCategoryId":"3","ListUrlMain":"https://doi.org/10.21037/qims-24-851","RegionNum":2,"RegionCategory":"医学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"2025/2/26 0:00:00","PubModel":"Epub","JCR":"Q2","JCRName":"RADIOLOGY, NUCLEAR MEDICINE & MEDICAL IMAGING","Score":null,"Total":0}
引用次数: 0

Abstract

Background: Due to the inherent limitations of imaging sensors, acquiring medical images that simultaneously provide functional metabolic information and detailed structural organization remains a significant challenge. Multi-modal image fusion has emerged as a critical technology for clinical diagnosis and surgical navigation, as it enables the integration of complementary information from different imaging modalities. However, existing deep learning (DL)-based fusion methods often face difficulties in effectively combining high-frequency detail information with low-frequency contextual information, which frequently leads to the degradation of high-frequency details. Therefore, there is a pressing need for a method that addresses these challenges, preserving both high- and low-frequency information while maintaining clear structural contours. In response to this issue, a novel convolutional neural network (CNN), named the multi-scale pyramid residual weight network (LYWNet), is proposed. The objective of this approach is to improve the fusion process by effectively integrating high- and low-frequency information, thereby enhancing the quality and accuracy of multimodal image fusion. This method aims to overcome the limitations of current fusion techniques and ensure the preservation of both functional and structural details, ultimately contributing to more precise clinical diagnoses and better surgical navigation outcomes.

Methods: We propose a novel CNN, LYWNet, designed to address these challenges. LYWNet is composed of three modules: (I) data preprocessing module: utilizes three convolutional layers to extract both deep and shallow features from the input images. (II) Feature extraction module: incorporates three identical multi-scale pyramid residual weight (LYW) blocks in series, each featuring three interactive branches to preserve high-frequency detail information effectively. (III) Image reconstruction module: utilizes a fusion algorithm based on feature distillation to ensure the effective integration of functional and anatomical information. The proposed image fusion algorithm enhances the interaction of contextual cues and retains the metabolic details from functional images while preserving texture details from anatomical images.

Results: The proposed LYWNet demonstrated its ability to retain high-frequency details during feature extraction, effectively combining them with low-frequency contextual information. The fusion results exhibited reduced differences between the fused image and the original images. The structural similarity (SSIM) and peak signal-to-noise ratio (PSNR) were 0.5592±0.0536 and 17.3594±1.0211, respectively, for single-photon emission computed tomography-magnetic resonance imaging (SPECT-MRI), 0.5195±0.0730 and 14.5324±1.7365 for PET-MRI; 0.5376±0.0442 and 13.9202±0.7265 for magnetic resonance imaging-computed tomography.

Conclusions: LYWNet excels at integrating high-frequency detail information and low-frequency contextual information, addressing the deficiencies of existing DL-based image fusion methods. This approach provides superior fused images that retain the functional metabolic information and anatomical texture, making it a valuable tool for clinical diagnosis and surgical navigation.

求助全文
约1分钟内获得全文 求助全文
来源期刊
Quantitative Imaging in Medicine and Surgery
Quantitative Imaging in Medicine and Surgery Medicine-Radiology, Nuclear Medicine and Imaging
CiteScore
4.20
自引率
17.90%
发文量
252
期刊介绍: Information not localized
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
确定
请完成安全验证×
copy
已复制链接
快去分享给好友吧!
我知道了
右上角分享
点击右上角分享
0
联系我们:info@booksci.cn Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。 Copyright © 2023 布克学术 All rights reserved.
京ICP备2023020795号-1
ghs 京公网安备 11010802042870号
Book学术文献互助
Book学术文献互助群
群 号:481959085
Book学术官方微信