{"title":"M3IF-NSST-MTV: Modified Total variation-based multi-modal medical image fusion using Laplacian energy and morphology in the NSST domain","authors":"Dev Kumar Chaudhary , Prabhishek Singh , Achyut Shankar , Manoj Diwakar","doi":"10.1016/j.imavis.2025.105581","DOIUrl":null,"url":null,"abstract":"<div><div>This paper presents a new multi-modal medical image fusion (M3IF) technique that fuses the medical images obtained from different medical imaging modalities, such as Computed Tomography (CT), Magnetic Resonance Imaging (MRI), Single Photon Emission Computed Tomography (SPECT) or Positron Emission Tomography (PET), into a single image. This single image is enhanced and contains all the important information of the source images. This paper proposes a hybrid M3IF technique, i.e., M3IF-NSST-MTV, where input medical images are decomposed using Non-Subsampled Shearlet Transform (NSST). It decomposes the image into low frequency coefficients (LFCs), and high frequency coefficients (HFCs). The LFCs are fused using Laplacian energy, and HFCs are fused using morphology. The fused image obtained after applying inverse-NSST is directed to the modified Total Variation (TV), that refines the NSST output. This modified TV output is again fused with NSST output using Feature Similarity Index Measure (FSIM) with Correlation Coefficient (CC)-based threshold value. This modified TV refinement process is iterative process. The results of M3IF-NSST-MTV are evaluated at the pre-set number of iterations = 200. The final fusion results of M3IF-NSST-MTV are compared with some of the prevalent non-traditional methods and based on visual quality and quantitative metric-based analysis; it is found that the M3IF-NSST-MTV delivers better results than all the compared methods.</div></div>","PeriodicalId":50374,"journal":{"name":"Image and Vision Computing","volume":"159 ","pages":"Article 105581"},"PeriodicalIF":4.2000,"publicationDate":"2025-05-11","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"Image and Vision Computing","FirstCategoryId":"94","ListUrlMain":"https://www.sciencedirect.com/science/article/pii/S0262885625001696","RegionNum":3,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"Q2","JCRName":"COMPUTER SCIENCE, ARTIFICIAL INTELLIGENCE","Score":null,"Total":0}
引用次数: 0
Abstract
This paper presents a new multi-modal medical image fusion (M3IF) technique that fuses the medical images obtained from different medical imaging modalities, such as Computed Tomography (CT), Magnetic Resonance Imaging (MRI), Single Photon Emission Computed Tomography (SPECT) or Positron Emission Tomography (PET), into a single image. This single image is enhanced and contains all the important information of the source images. This paper proposes a hybrid M3IF technique, i.e., M3IF-NSST-MTV, where input medical images are decomposed using Non-Subsampled Shearlet Transform (NSST). It decomposes the image into low frequency coefficients (LFCs), and high frequency coefficients (HFCs). The LFCs are fused using Laplacian energy, and HFCs are fused using morphology. The fused image obtained after applying inverse-NSST is directed to the modified Total Variation (TV), that refines the NSST output. This modified TV output is again fused with NSST output using Feature Similarity Index Measure (FSIM) with Correlation Coefficient (CC)-based threshold value. This modified TV refinement process is iterative process. The results of M3IF-NSST-MTV are evaluated at the pre-set number of iterations = 200. The final fusion results of M3IF-NSST-MTV are compared with some of the prevalent non-traditional methods and based on visual quality and quantitative metric-based analysis; it is found that the M3IF-NSST-MTV delivers better results than all the compared methods.
期刊介绍:
Image and Vision Computing has as a primary aim the provision of an effective medium of interchange for the results of high quality theoretical and applied research fundamental to all aspects of image interpretation and computer vision. The journal publishes work that proposes new image interpretation and computer vision methodology or addresses the application of such methods to real world scenes. It seeks to strengthen a deeper understanding in the discipline by encouraging the quantitative comparison and performance evaluation of the proposed methodology. The coverage includes: image interpretation, scene modelling, object recognition and tracking, shape analysis, monitoring and surveillance, active vision and robotic systems, SLAM, biologically-inspired computer vision, motion analysis, stereo vision, document image understanding, character and handwritten text recognition, face and gesture recognition, biometrics, vision-based human-computer interaction, human activity and behavior understanding, data fusion from multiple sensor inputs, image databases.