基于深度学习和图像增强的红外与可见光图像融合管道

IF 6.5 2区 计算机科学 Q1 COMPUTER SCIENCE, ARTIFICIAL INTELLIGENCE
Jin Qi , Deboch Eyob Abera , Mola Natnael Fanose, Lingfeng Wang, Jian Cheng
{"title":"基于深度学习和图像增强的红外与可见光图像融合管道","authors":"Jin Qi ,&nbsp;Deboch Eyob Abera ,&nbsp;Mola Natnael Fanose,&nbsp;Lingfeng Wang,&nbsp;Jian Cheng","doi":"10.1016/j.neucom.2024.127353","DOIUrl":null,"url":null,"abstract":"<div><p>It is difficult to use supervised machine-learning methods for infrared (IR) and visible (VIS) image fusion (IVF) because of the shortage of ground-truth target fusion images, and image quality and contrast control are rarely considered in existing IVF methods. In this study, we proposed a simple IVF pipeline that converts the IVF problem into a supervised binary classification problem (sharp vs. blur) and uses image enhancement techniques to improve the image quality in three locations in the pipeline. We took a biological vision consistent assumption that the sharp region contains more useful information than the blurred region. A deep binary classifier based on a convolutional neural network (CNN) was designed to compare the sharpness of the infrared region and visible regions. The output score map of the deep classifier was treated as a weight map in the weighted average fusion rule. The proposed deep binary classifier was trained using natural visible images from the MS COCO dataset, rather than images from the IVF domain (called “cross domain training” here). Specifically, our proposed pipeline contains four stages: (1) enhancing the IR and VIS input images by linear transformation and the High-Dynamic-Range Compression (HDRC) method, respectively; (2) inputting the enhanced IR and VIS images to the trained CNN classifier to obtain the weight map; and (3) using a weight map to obtain the weighted average of the enhanced IR and VIS images; and (4) using single scale Retinex (SSR) to enhance the fused image to obtain the final enhanced fusion image. Extensive experimental results on public IVF datasets demonstrate the superior performance of our proposed approach over other state-of-the-art methods in terms of both subjective visual quality and 11 objective metrics. It was demonstrated that the complementary information between the infrared and visible images can be efficiently extracted using our proposed binary classifier, and the fused image quality is significantly improved. The source code is available at <span>https://github.com/eyob12/Deep_infrared_and_visible_image_fusion</span><svg><path></path></svg>.</p></div>","PeriodicalId":19268,"journal":{"name":"Neurocomputing","volume":"578 ","pages":"Article 127353"},"PeriodicalIF":6.5000,"publicationDate":"2024-02-10","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":"{\"title\":\"A deep learning and image enhancement based pipeline for infrared and visible image fusion\",\"authors\":\"Jin Qi ,&nbsp;Deboch Eyob Abera ,&nbsp;Mola Natnael Fanose,&nbsp;Lingfeng Wang,&nbsp;Jian Cheng\",\"doi\":\"10.1016/j.neucom.2024.127353\",\"DOIUrl\":null,\"url\":null,\"abstract\":\"<div><p>It is difficult to use supervised machine-learning methods for infrared (IR) and visible (VIS) image fusion (IVF) because of the shortage of ground-truth target fusion images, and image quality and contrast control are rarely considered in existing IVF methods. In this study, we proposed a simple IVF pipeline that converts the IVF problem into a supervised binary classification problem (sharp vs. blur) and uses image enhancement techniques to improve the image quality in three locations in the pipeline. We took a biological vision consistent assumption that the sharp region contains more useful information than the blurred region. A deep binary classifier based on a convolutional neural network (CNN) was designed to compare the sharpness of the infrared region and visible regions. The output score map of the deep classifier was treated as a weight map in the weighted average fusion rule. The proposed deep binary classifier was trained using natural visible images from the MS COCO dataset, rather than images from the IVF domain (called “cross domain training” here). Specifically, our proposed pipeline contains four stages: (1) enhancing the IR and VIS input images by linear transformation and the High-Dynamic-Range Compression (HDRC) method, respectively; (2) inputting the enhanced IR and VIS images to the trained CNN classifier to obtain the weight map; and (3) using a weight map to obtain the weighted average of the enhanced IR and VIS images; and (4) using single scale Retinex (SSR) to enhance the fused image to obtain the final enhanced fusion image. Extensive experimental results on public IVF datasets demonstrate the superior performance of our proposed approach over other state-of-the-art methods in terms of both subjective visual quality and 11 objective metrics. It was demonstrated that the complementary information between the infrared and visible images can be efficiently extracted using our proposed binary classifier, and the fused image quality is significantly improved. The source code is available at <span>https://github.com/eyob12/Deep_infrared_and_visible_image_fusion</span><svg><path></path></svg>.</p></div>\",\"PeriodicalId\":19268,\"journal\":{\"name\":\"Neurocomputing\",\"volume\":\"578 \",\"pages\":\"Article 127353\"},\"PeriodicalIF\":6.5000,\"publicationDate\":\"2024-02-10\",\"publicationTypes\":\"Journal Article\",\"fieldsOfStudy\":null,\"isOpenAccess\":false,\"openAccessPdf\":\"\",\"citationCount\":\"0\",\"resultStr\":null,\"platform\":\"Semanticscholar\",\"paperid\":null,\"PeriodicalName\":\"Neurocomputing\",\"FirstCategoryId\":\"94\",\"ListUrlMain\":\"https://www.sciencedirect.com/science/article/pii/S0925231224001243\",\"RegionNum\":2,\"RegionCategory\":\"计算机科学\",\"ArticlePicture\":[],\"TitleCN\":null,\"AbstractTextCN\":null,\"PMCID\":null,\"EPubDate\":\"\",\"PubModel\":\"\",\"JCR\":\"Q1\",\"JCRName\":\"COMPUTER SCIENCE, ARTIFICIAL INTELLIGENCE\",\"Score\":null,\"Total\":0}","platform":"Semanticscholar","paperid":null,"PeriodicalName":"Neurocomputing","FirstCategoryId":"94","ListUrlMain":"https://www.sciencedirect.com/science/article/pii/S0925231224001243","RegionNum":2,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"Q1","JCRName":"COMPUTER SCIENCE, ARTIFICIAL INTELLIGENCE","Score":null,"Total":0}
引用次数: 0

摘要

由于缺乏地面实况目标融合图像,因此很难将有监督的机器学习方法用于红外(IR)和可见光(VIS)图像融合(IVF),而且现有的 IVF 方法很少考虑图像质量和对比度控制。在这项研究中,我们提出了一个简单的 IVF 管道,它将 IVF 问题转换为一个有监督的二元分类问题(清晰与模糊),并在管道的三个位置使用图像增强技术来提高图像质量。我们采用了与生物视觉一致的假设,即清晰区域比模糊区域包含更多有用信息。我们设计了一种基于卷积神经网络(CNN)的深度二元分类器,用于比较红外区域和可见光区域的清晰度。深度分类器的输出分数图被视为加权平均融合规则中的权重图。所提出的深度二元分类器是使用 MS COCO 数据集中的自然可见光图像而非 IVF 领域的图像进行训练的(此处称为 "跨领域训练")。具体来说,我们提出的管道包括四个阶段:(1)通过线性变换和高动态范围压缩(HDRC)方法分别增强红外和可见光输入图像;(2)将增强后的红外和可见光图像输入训练好的 CNN 分类器,以获得权重图;(3)使用权重图获得增强后的红外和可见光图像的加权平均值;(4)使用单比例 Retinex(SSR)增强融合图像,以获得最终的增强融合图像。在公共 IVF 数据集上进行的大量实验结果表明,我们提出的方法在主观视觉质量和 11 项客观指标方面都优于其他最先进的方法。实验结果表明,使用我们提出的二元分类器可以有效提取红外图像和可见光图像之间的互补信息,并显著提高融合图像的质量。源代码见 https://github.com/eyob12/Deep_infrared_and_visible_image_fusion。
本文章由计算机程序翻译,如有差异,请以英文原文为准。
A deep learning and image enhancement based pipeline for infrared and visible image fusion

It is difficult to use supervised machine-learning methods for infrared (IR) and visible (VIS) image fusion (IVF) because of the shortage of ground-truth target fusion images, and image quality and contrast control are rarely considered in existing IVF methods. In this study, we proposed a simple IVF pipeline that converts the IVF problem into a supervised binary classification problem (sharp vs. blur) and uses image enhancement techniques to improve the image quality in three locations in the pipeline. We took a biological vision consistent assumption that the sharp region contains more useful information than the blurred region. A deep binary classifier based on a convolutional neural network (CNN) was designed to compare the sharpness of the infrared region and visible regions. The output score map of the deep classifier was treated as a weight map in the weighted average fusion rule. The proposed deep binary classifier was trained using natural visible images from the MS COCO dataset, rather than images from the IVF domain (called “cross domain training” here). Specifically, our proposed pipeline contains four stages: (1) enhancing the IR and VIS input images by linear transformation and the High-Dynamic-Range Compression (HDRC) method, respectively; (2) inputting the enhanced IR and VIS images to the trained CNN classifier to obtain the weight map; and (3) using a weight map to obtain the weighted average of the enhanced IR and VIS images; and (4) using single scale Retinex (SSR) to enhance the fused image to obtain the final enhanced fusion image. Extensive experimental results on public IVF datasets demonstrate the superior performance of our proposed approach over other state-of-the-art methods in terms of both subjective visual quality and 11 objective metrics. It was demonstrated that the complementary information between the infrared and visible images can be efficiently extracted using our proposed binary classifier, and the fused image quality is significantly improved. The source code is available at https://github.com/eyob12/Deep_infrared_and_visible_image_fusion.

求助全文
通过发布文献求助,成功后即可免费获取论文全文。 去求助
来源期刊
Neurocomputing
Neurocomputing 工程技术-计算机:人工智能
CiteScore
13.10
自引率
10.00%
发文量
1382
审稿时长
70 days
期刊介绍: Neurocomputing publishes articles describing recent fundamental contributions in the field of neurocomputing. Neurocomputing theory, practice and applications are the essential topics being covered.
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
确定
请完成安全验证×
copy
已复制链接
快去分享给好友吧!
我知道了
右上角分享
点击右上角分享
0
联系我们:info@booksci.cn Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。 Copyright © 2023 布克学术 All rights reserved.
京ICP备2023020795号-1
ghs 京公网安备 11010802042870号
Book学术文献互助
Book学术文献互助群
群 号:604180095
Book学术官方微信