红外与可见光图像融合的图表示学习

IF 6.4 2区 计算机科学 Q1 AUTOMATION & CONTROL SYSTEMS
Jing Li;Lu Bai;Bin Yang;Chang Li;Lingfei Ma
{"title":"红外与可见光图像融合的图表示学习","authors":"Jing Li;Lu Bai;Bin Yang;Chang Li;Lingfei Ma","doi":"10.1109/TASE.2025.3557234","DOIUrl":null,"url":null,"abstract":"Infrared and visible image fusion aims to extract complementary features to synthesize a single fused image. In our method, we covert the regular image format into the graph space and conduct graph convolutional networks (GCNs) to extract NLss for the reliable infrared and visible image fusion. More specifically, GCNs are first performed on each intra-modal set to aggregate the features and propagate the inherent information, thereby extracting independent intra-modal NLss. Then, such intra-modal non-local self-similarity (NLss) features of infrared and visible images are concatenated to explore cross-domain NLss inter-modally and reconstruct the fused images. Extensive experiments show the superior performance of our method with the qualitative and quantitative analysis on the TNO, RoadScene and M3FD datasets, respectively, outperforming many state-of-the-art (SOTA) methods for the robust and effective infrared and visible image fusion. Note to Practitioners—This paper was motivated by the problem that the most existing methods that employ convolutional neural networks (CNNs) and transformer-based frameworks mainly extract local features and long-range dependence. However, they often cause overlooking the image’s NLss or information redundancy, resulting in poor infrared and visible image fusion. To address these problems, graph-based data representations can construct relationships among spatially repeatable details or textures with far-space distances, which are more suitable for handling irregular objects. Therefore, it is significant to covert the regular image format into the graph space and conduct graph convolutional networks (GCNs) to extract NLss for the reliable infrared and visible image fusion. In this paper, we develop an infrared and visible image fusion method based on graph representation learning strategy.","PeriodicalId":51060,"journal":{"name":"IEEE Transactions on Automation Science and Engineering","volume":"22 ","pages":"13801-13813"},"PeriodicalIF":6.4000,"publicationDate":"2025-04-07","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":"{\"title\":\"Graph Representation Learning for Infrared and Visible Image Fusion\",\"authors\":\"Jing Li;Lu Bai;Bin Yang;Chang Li;Lingfei Ma\",\"doi\":\"10.1109/TASE.2025.3557234\",\"DOIUrl\":null,\"url\":null,\"abstract\":\"Infrared and visible image fusion aims to extract complementary features to synthesize a single fused image. In our method, we covert the regular image format into the graph space and conduct graph convolutional networks (GCNs) to extract NLss for the reliable infrared and visible image fusion. More specifically, GCNs are first performed on each intra-modal set to aggregate the features and propagate the inherent information, thereby extracting independent intra-modal NLss. Then, such intra-modal non-local self-similarity (NLss) features of infrared and visible images are concatenated to explore cross-domain NLss inter-modally and reconstruct the fused images. Extensive experiments show the superior performance of our method with the qualitative and quantitative analysis on the TNO, RoadScene and M3FD datasets, respectively, outperforming many state-of-the-art (SOTA) methods for the robust and effective infrared and visible image fusion. Note to Practitioners—This paper was motivated by the problem that the most existing methods that employ convolutional neural networks (CNNs) and transformer-based frameworks mainly extract local features and long-range dependence. However, they often cause overlooking the image’s NLss or information redundancy, resulting in poor infrared and visible image fusion. To address these problems, graph-based data representations can construct relationships among spatially repeatable details or textures with far-space distances, which are more suitable for handling irregular objects. Therefore, it is significant to covert the regular image format into the graph space and conduct graph convolutional networks (GCNs) to extract NLss for the reliable infrared and visible image fusion. In this paper, we develop an infrared and visible image fusion method based on graph representation learning strategy.\",\"PeriodicalId\":51060,\"journal\":{\"name\":\"IEEE Transactions on Automation Science and Engineering\",\"volume\":\"22 \",\"pages\":\"13801-13813\"},\"PeriodicalIF\":6.4000,\"publicationDate\":\"2025-04-07\",\"publicationTypes\":\"Journal Article\",\"fieldsOfStudy\":null,\"isOpenAccess\":false,\"openAccessPdf\":\"\",\"citationCount\":\"0\",\"resultStr\":null,\"platform\":\"Semanticscholar\",\"paperid\":null,\"PeriodicalName\":\"IEEE Transactions on Automation Science and Engineering\",\"FirstCategoryId\":\"94\",\"ListUrlMain\":\"https://ieeexplore.ieee.org/document/10955329/\",\"RegionNum\":2,\"RegionCategory\":\"计算机科学\",\"ArticlePicture\":[],\"TitleCN\":null,\"AbstractTextCN\":null,\"PMCID\":null,\"EPubDate\":\"\",\"PubModel\":\"\",\"JCR\":\"Q1\",\"JCRName\":\"AUTOMATION & CONTROL SYSTEMS\",\"Score\":null,\"Total\":0}","platform":"Semanticscholar","paperid":null,"PeriodicalName":"IEEE Transactions on Automation Science and Engineering","FirstCategoryId":"94","ListUrlMain":"https://ieeexplore.ieee.org/document/10955329/","RegionNum":2,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"Q1","JCRName":"AUTOMATION & CONTROL SYSTEMS","Score":null,"Total":0}
引用次数: 0

摘要

红外图像与可见光图像融合的目的是提取互补特征,合成一幅融合图像。在我们的方法中,我们将规则图像格式转换到图空间中,并通过图卷积网络(GCNs)提取NLss,以实现可靠的红外和可见光图像融合。更具体地说,首先对每个模态内集执行GCNs,聚合特征并传播固有信息,从而提取独立的模态内NLss。然后,将红外和可见光图像的模态内非局部自相似(NLss)特征进行连接,进行模态间的跨域非局部自相似探索,重建融合后的图像。大量的实验表明,我们的方法在TNO、RoadScene和M3FD数据集的定性和定量分析上表现优异,在鲁棒和有效的红外和可见光图像融合方面优于许多最先进的(SOTA)方法。从业人员注意事项-本文的动机是基于现有的大多数使用卷积神经网络(cnn)和基于变压器的框架的方法主要提取局部特征和远程依赖关系的问题。然而,它们往往会导致忽略图像的nlless或信息冗余,导致红外和可见光图像融合效果不佳。为了解决这些问题,基于图的数据表示可以构建空间上可重复的细节或具有远空间距离的纹理之间的关系,更适合处理不规则物体。因此,将规则图像格式转换到图空间中,并利用图卷积网络(GCNs)提取NLss,以实现可靠的红外和可见光图像融合具有重要意义。本文提出了一种基于图表示学习策略的红外图像与可见光图像融合方法。
本文章由计算机程序翻译,如有差异,请以英文原文为准。
Graph Representation Learning for Infrared and Visible Image Fusion
Infrared and visible image fusion aims to extract complementary features to synthesize a single fused image. In our method, we covert the regular image format into the graph space and conduct graph convolutional networks (GCNs) to extract NLss for the reliable infrared and visible image fusion. More specifically, GCNs are first performed on each intra-modal set to aggregate the features and propagate the inherent information, thereby extracting independent intra-modal NLss. Then, such intra-modal non-local self-similarity (NLss) features of infrared and visible images are concatenated to explore cross-domain NLss inter-modally and reconstruct the fused images. Extensive experiments show the superior performance of our method with the qualitative and quantitative analysis on the TNO, RoadScene and M3FD datasets, respectively, outperforming many state-of-the-art (SOTA) methods for the robust and effective infrared and visible image fusion. Note to Practitioners—This paper was motivated by the problem that the most existing methods that employ convolutional neural networks (CNNs) and transformer-based frameworks mainly extract local features and long-range dependence. However, they often cause overlooking the image’s NLss or information redundancy, resulting in poor infrared and visible image fusion. To address these problems, graph-based data representations can construct relationships among spatially repeatable details or textures with far-space distances, which are more suitable for handling irregular objects. Therefore, it is significant to covert the regular image format into the graph space and conduct graph convolutional networks (GCNs) to extract NLss for the reliable infrared and visible image fusion. In this paper, we develop an infrared and visible image fusion method based on graph representation learning strategy.
求助全文
通过发布文献求助,成功后即可免费获取论文全文。 去求助
来源期刊
IEEE Transactions on Automation Science and Engineering
IEEE Transactions on Automation Science and Engineering 工程技术-自动化与控制系统
CiteScore
12.50
自引率
14.30%
发文量
404
审稿时长
3.0 months
期刊介绍: The IEEE Transactions on Automation Science and Engineering (T-ASE) publishes fundamental papers on Automation, emphasizing scientific results that advance efficiency, quality, productivity, and reliability. T-ASE encourages interdisciplinary approaches from computer science, control systems, electrical engineering, mathematics, mechanical engineering, operations research, and other fields. T-ASE welcomes results relevant to industries such as agriculture, biotechnology, healthcare, home automation, maintenance, manufacturing, pharmaceuticals, retail, security, service, supply chains, and transportation. T-ASE addresses a research community willing to integrate knowledge across disciplines and industries. For this purpose, each paper includes a Note to Practitioners that summarizes how its results can be applied or how they might be extended to apply in practice.
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
确定
请完成安全验证×
copy
已复制链接
快去分享给好友吧!
我知道了
右上角分享
点击右上角分享
0
联系我们:info@booksci.cn Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。 Copyright © 2023 布克学术 All rights reserved.
京ICP备2023020795号-1
ghs 京公网安备 11010802042870号
Book学术文献互助
Book学术文献互助群
群 号:604180095
Book学术官方微信