Unpaired high-quality image-guided infrared and visible image fusion via generative adversarial network

IF 1.3 4区 计算机科学 Q3 COMPUTER SCIENCE, SOFTWARE ENGINEERING
Hang Li, Zheng Guan, Xue Wang, Qiuhan Shao
{"title":"Unpaired high-quality image-guided infrared and visible image fusion via generative adversarial network","authors":"Hang Li,&nbsp;Zheng Guan,&nbsp;Xue Wang,&nbsp;Qiuhan Shao","doi":"10.1016/j.cagd.2024.102325","DOIUrl":null,"url":null,"abstract":"<div><p>Current infrared and visible image fusion (IVIF) methods lack ground truth and require prior knowledge to guide the feature fusion process. However, in the fusion process, these features have not been placed in an equal and well-defined position, which causes the degradation of image quality. To address this challenge, this study develops a new end-to-end model, termed unpaired high-quality image-guided generative adversarial network (UHG-GAN). Specifically, we introduce the high-quality image as the reference standard of the fused image and employ a global discriminator and a local discriminator to identify the distribution difference between the high-quality image and the fused image. Through adversarial learning, the generator can generate images that are more consistent with high-quality expression. In addition, we also designed the laplacian pyramid augmentation (LPA) module in the generator, which integrates multi-scale features of source images across domains so that the generator can more fully extract the structure and texture information. Extensive experiments demonstrate that our method can effectively preserve the target information in the infrared image and the scene information in the visible image and significantly improve the image quality.</p></div>","PeriodicalId":55226,"journal":{"name":"Computer Aided Geometric Design","volume":"111 ","pages":"Article 102325"},"PeriodicalIF":1.3000,"publicationDate":"2024-04-25","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"Computer Aided Geometric Design","FirstCategoryId":"94","ListUrlMain":"https://www.sciencedirect.com/science/article/pii/S0167839624000591","RegionNum":4,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"Q3","JCRName":"COMPUTER SCIENCE, SOFTWARE ENGINEERING","Score":null,"Total":0}
引用次数: 0

Abstract

Current infrared and visible image fusion (IVIF) methods lack ground truth and require prior knowledge to guide the feature fusion process. However, in the fusion process, these features have not been placed in an equal and well-defined position, which causes the degradation of image quality. To address this challenge, this study develops a new end-to-end model, termed unpaired high-quality image-guided generative adversarial network (UHG-GAN). Specifically, we introduce the high-quality image as the reference standard of the fused image and employ a global discriminator and a local discriminator to identify the distribution difference between the high-quality image and the fused image. Through adversarial learning, the generator can generate images that are more consistent with high-quality expression. In addition, we also designed the laplacian pyramid augmentation (LPA) module in the generator, which integrates multi-scale features of source images across domains so that the generator can more fully extract the structure and texture information. Extensive experiments demonstrate that our method can effectively preserve the target information in the infrared image and the scene information in the visible image and significantly improve the image quality.

通过生成式对抗网络实现非配对高质量图像引导的红外和可见光图像融合
目前的红外与可见光图像融合(IVIF)方法缺乏基本事实,需要先验知识来指导特征融合过程。然而,在融合过程中,这些特征并没有被放置在相等和明确的位置上,从而导致图像质量下降。为了应对这一挑战,本研究开发了一种新的端到端模型,称为无配对高质量图像引导生成对抗网络(UHG-GAN)。具体来说,我们引入高质量图像作为融合图像的参考标准,并采用全局判别器和局部判别器来识别高质量图像和融合图像之间的分布差异。通过对抗学习,生成器可以生成更符合高质量表达的图像。此外,我们还在生成器中设计了拉普拉斯金字塔增强(LPA)模块,它可以跨域整合源图像的多尺度特征,从而使生成器可以更充分地提取结构和纹理信息。大量实验证明,我们的方法能有效保留红外图像中的目标信息和可见光图像中的场景信息,并显著提高图像质量。
本文章由计算机程序翻译,如有差异,请以英文原文为准。
求助全文
约1分钟内获得全文 求助全文
来源期刊
Computer Aided Geometric Design
Computer Aided Geometric Design 工程技术-计算机:软件工程
CiteScore
3.50
自引率
13.30%
发文量
57
审稿时长
60 days
期刊介绍: The journal Computer Aided Geometric Design is for researchers, scholars, and software developers dealing with mathematical and computational methods for the description of geometric objects as they arise in areas ranging from CAD/CAM to robotics and scientific visualization. The journal publishes original research papers, survey papers and with quick editorial decisions short communications of at most 3 pages. The primary objects of interest are curves, surfaces, and volumes such as splines (NURBS), meshes, subdivision surfaces as well as algorithms to generate, analyze, and manipulate them. This journal will report on new developments in CAGD and its applications, including but not restricted to the following: -Mathematical and Geometric Foundations- Curve, Surface, and Volume generation- CAGD applications in Numerical Analysis, Computational Geometry, Computer Graphics, or Computer Vision- Industrial, medical, and scientific applications. The aim is to collect and disseminate information on computer aided design in one journal. To provide the user community with methods and algorithms for representing curves and surfaces. To illustrate computer aided geometric design by means of interesting applications. To combine curve and surface methods with computer graphics. To explain scientific phenomena by means of computer graphics. To concentrate on the interaction between theory and application. To expose unsolved problems of the practice. To develop new methods in computer aided geometry.
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
确定
请完成安全验证×
copy
已复制链接
快去分享给好友吧!
我知道了
右上角分享
点击右上角分享
0
联系我们:info@booksci.cn Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。 Copyright © 2023 布克学术 All rights reserved.
京ICP备2023020795号-1
ghs 京公网安备 11010802042870号
Book学术文献互助
Book学术文献互助群
群 号:481959085
Book学术官方微信