{"title":"TCIGFusion:一种用于红外和可见光图像融合的两阶段相关特征交互引导网络","authors":"Jiawei Liu, Guiling Sun, Bowen Zheng, Liang Dong","doi":"10.1016/j.optlaseng.2025.109265","DOIUrl":null,"url":null,"abstract":"<div><div>Infrared and visible image fusion is aimed at generating images with prominent targets and texture details, providing support for downstream applications such as object detection. However, most existing deep learning-based fusion methods involve single-stage training and manually designed fusion rules, which cannot effectively extract and fuse features. Therefore, in this paper, we propose a two-stage correlated feature interactive guided network termed TCIGFusion. In the first stage, a Unet-like dual-branch Transformer module and dynamic large kernel convolution block (DLKB) are used to extract global features from the two source images, while the convolution blocks extract local features from the source images. In the second phase, we designed a cross attention guide module (CAGM) to interactively fuse the heterogeneously related features of the two modalities, avoiding the complexity associated with manually designing fusion rules. Furthermore, to optimize the efficacy of the fusion network, we employ a combination of image reconstruction, decomposition, and gradient loss functions for unsupervised training of the model. The superiority of our TCIGFusion is evidenced by extensive experimentation conducted on multiple public datasets. These experiments demonstrate that our method outperforms other state-of-the-art deep learning approaches, as evaluated through both subjective and objective metrics.</div></div>","PeriodicalId":49719,"journal":{"name":"Optics and Lasers in Engineering","volume":"195 ","pages":"Article 109265"},"PeriodicalIF":3.7000,"publicationDate":"2025-08-21","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":"{\"title\":\"TCIGFusion: A two-stage correlated feature interactive guided network for infrared and visible image fusion\",\"authors\":\"Jiawei Liu, Guiling Sun, Bowen Zheng, Liang Dong\",\"doi\":\"10.1016/j.optlaseng.2025.109265\",\"DOIUrl\":null,\"url\":null,\"abstract\":\"<div><div>Infrared and visible image fusion is aimed at generating images with prominent targets and texture details, providing support for downstream applications such as object detection. However, most existing deep learning-based fusion methods involve single-stage training and manually designed fusion rules, which cannot effectively extract and fuse features. Therefore, in this paper, we propose a two-stage correlated feature interactive guided network termed TCIGFusion. In the first stage, a Unet-like dual-branch Transformer module and dynamic large kernel convolution block (DLKB) are used to extract global features from the two source images, while the convolution blocks extract local features from the source images. In the second phase, we designed a cross attention guide module (CAGM) to interactively fuse the heterogeneously related features of the two modalities, avoiding the complexity associated with manually designing fusion rules. Furthermore, to optimize the efficacy of the fusion network, we employ a combination of image reconstruction, decomposition, and gradient loss functions for unsupervised training of the model. The superiority of our TCIGFusion is evidenced by extensive experimentation conducted on multiple public datasets. These experiments demonstrate that our method outperforms other state-of-the-art deep learning approaches, as evaluated through both subjective and objective metrics.</div></div>\",\"PeriodicalId\":49719,\"journal\":{\"name\":\"Optics and Lasers in Engineering\",\"volume\":\"195 \",\"pages\":\"Article 109265\"},\"PeriodicalIF\":3.7000,\"publicationDate\":\"2025-08-21\",\"publicationTypes\":\"Journal Article\",\"fieldsOfStudy\":null,\"isOpenAccess\":false,\"openAccessPdf\":\"\",\"citationCount\":\"0\",\"resultStr\":null,\"platform\":\"Semanticscholar\",\"paperid\":null,\"PeriodicalName\":\"Optics and Lasers in Engineering\",\"FirstCategoryId\":\"5\",\"ListUrlMain\":\"https://www.sciencedirect.com/science/article/pii/S0143816625004506\",\"RegionNum\":2,\"RegionCategory\":\"工程技术\",\"ArticlePicture\":[],\"TitleCN\":null,\"AbstractTextCN\":null,\"PMCID\":null,\"EPubDate\":\"\",\"PubModel\":\"\",\"JCR\":\"Q2\",\"JCRName\":\"OPTICS\",\"Score\":null,\"Total\":0}","platform":"Semanticscholar","paperid":null,"PeriodicalName":"Optics and Lasers in Engineering","FirstCategoryId":"5","ListUrlMain":"https://www.sciencedirect.com/science/article/pii/S0143816625004506","RegionNum":2,"RegionCategory":"工程技术","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"Q2","JCRName":"OPTICS","Score":null,"Total":0}
TCIGFusion: A two-stage correlated feature interactive guided network for infrared and visible image fusion
Infrared and visible image fusion is aimed at generating images with prominent targets and texture details, providing support for downstream applications such as object detection. However, most existing deep learning-based fusion methods involve single-stage training and manually designed fusion rules, which cannot effectively extract and fuse features. Therefore, in this paper, we propose a two-stage correlated feature interactive guided network termed TCIGFusion. In the first stage, a Unet-like dual-branch Transformer module and dynamic large kernel convolution block (DLKB) are used to extract global features from the two source images, while the convolution blocks extract local features from the source images. In the second phase, we designed a cross attention guide module (CAGM) to interactively fuse the heterogeneously related features of the two modalities, avoiding the complexity associated with manually designing fusion rules. Furthermore, to optimize the efficacy of the fusion network, we employ a combination of image reconstruction, decomposition, and gradient loss functions for unsupervised training of the model. The superiority of our TCIGFusion is evidenced by extensive experimentation conducted on multiple public datasets. These experiments demonstrate that our method outperforms other state-of-the-art deep learning approaches, as evaluated through both subjective and objective metrics.
期刊介绍:
Optics and Lasers in Engineering aims at providing an international forum for the interchange of information on the development of optical techniques and laser technology in engineering. Emphasis is placed on contributions targeted at the practical use of methods and devices, the development and enhancement of solutions and new theoretical concepts for experimental methods.
Optics and Lasers in Engineering reflects the main areas in which optical methods are being used and developed for an engineering environment. Manuscripts should offer clear evidence of novelty and significance. Papers focusing on parameter optimization or computational issues are not suitable. Similarly, papers focussed on an application rather than the optical method fall outside the journal''s scope. The scope of the journal is defined to include the following:
-Optical Metrology-
Optical Methods for 3D visualization and virtual engineering-
Optical Techniques for Microsystems-
Imaging, Microscopy and Adaptive Optics-
Computational Imaging-
Laser methods in manufacturing-
Integrated optical and photonic sensors-
Optics and Photonics in Life Science-
Hyperspectral and spectroscopic methods-
Infrared and Terahertz techniques