{"title":"Image fusion: A deep Y shaped–residual convolution auto-encoder with MS-SSIM loss function","authors":"M. Gayathri Devi , I.S. Akila","doi":"10.1016/j.jrras.2024.101089","DOIUrl":null,"url":null,"abstract":"<div><p>Image fusion and deep learning are actively investigating fields of research. Their application domains include machine vision, clinical imaging, remote sensing, and other areas, all of which are used to obtain comprehensive information about a specific image. Image fusion is a process that integrates multiple imaging modalities to create a single image, for the sake of providing comprehensive information. Extensive literature shows that various methodologies, requirements, and network types are utilized for diverse modality fusion. This paper addresses the previously described issue by utilizing a unique Y-shaped Residual Convolution Autoencoder Neural Network to combine images from various modalities using the same network specifications and thereby eliminating the need for manual fusion. The combined convolved features are recreated in the decoder part using a symmetric nested residual approach with the encoder. By employing MS-SSIM as the loss function, the network is capable of generating images that are perceptually and pixel-wise indistinguishable from the target images. The fusion results are compared with five other current approaches, and the Y-shaped convolutional autoencoder result demonstrates superior achievement in both quantitative and qualitative aspects.</p></div>","PeriodicalId":16920,"journal":{"name":"Journal of Radiation Research and Applied Sciences","volume":"17 4","pages":"Article 101089"},"PeriodicalIF":1.7000,"publicationDate":"2024-09-06","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.sciencedirect.com/science/article/pii/S1687850724002735/pdfft?md5=899ee8d47f5932bf61dd8889093ce5f7&pid=1-s2.0-S1687850724002735-main.pdf","citationCount":"0","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"Journal of Radiation Research and Applied Sciences","FirstCategoryId":"103","ListUrlMain":"https://www.sciencedirect.com/science/article/pii/S1687850724002735","RegionNum":4,"RegionCategory":"综合性期刊","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"Q2","JCRName":"MULTIDISCIPLINARY SCIENCES","Score":null,"Total":0}
引用次数: 0
Abstract
Image fusion and deep learning are actively investigating fields of research. Their application domains include machine vision, clinical imaging, remote sensing, and other areas, all of which are used to obtain comprehensive information about a specific image. Image fusion is a process that integrates multiple imaging modalities to create a single image, for the sake of providing comprehensive information. Extensive literature shows that various methodologies, requirements, and network types are utilized for diverse modality fusion. This paper addresses the previously described issue by utilizing a unique Y-shaped Residual Convolution Autoencoder Neural Network to combine images from various modalities using the same network specifications and thereby eliminating the need for manual fusion. The combined convolved features are recreated in the decoder part using a symmetric nested residual approach with the encoder. By employing MS-SSIM as the loss function, the network is capable of generating images that are perceptually and pixel-wise indistinguishable from the target images. The fusion results are compared with five other current approaches, and the Y-shaped convolutional autoencoder result demonstrates superior achievement in both quantitative and qualitative aspects.
期刊介绍:
Journal of Radiation Research and Applied Sciences provides a high quality medium for the publication of substantial, original and scientific and technological papers on the development and applications of nuclear, radiation and isotopes in biology, medicine, drugs, biochemistry, microbiology, agriculture, entomology, food technology, chemistry, physics, solid states, engineering, environmental and applied sciences.