{"title":"One-stream one-stage pure transformer for infrared and visible image fusion","authors":"Qianhong Zhang , Qiao Liu , Di Yuan , Yunpeng Liu","doi":"10.1016/j.infrared.2025.106067","DOIUrl":null,"url":null,"abstract":"<div><div>Existing infrared and visible image fusion methods usually use two same backbones to extract deep features of the source images, and then manually design a fusion strategy to fuse them. However, this framework overlooks the importance of feature interaction during the feature extraction stage and does not fully capture the complementary information between the source images. In addition, it requires the design of complex fusion strategies, which limits its robustness and generalization to different scenarios. To this end, we propose a one-stream one-stage pure Transformer-based fusion framework, which simplifies feature extraction and fusion into a unified one-stream pipeline. Specifically, the proposed method consists of a one-stream fusion network and a decomposition network. The fusion network uses several Swin Transformer blocks to extract and fuse two modality features simultaneously. Thanks to the sliding-window-based multi-head attention, the fusion network can acquire local features and global dependencies and seamlessly model their contextual relationships. Due to the lack of effective supervision signals, the fusion network struggles to fully transfer important information from the source images, which can easily lead to the generation of artifacts. To eliminate these artifacts and simultaneously force the fused image to contain richer information, we design a simple decomposition network that decomposes the fusion result into the source images with consistency constraints. Extensive comparative and ablation experiments on four image fusion benchmarks demonstrate that our method achieves favorable results. In addition, the results on downstream tasks, including object detection and semantic segmentation, further show the effectiveness of the proposed method.</div></div>","PeriodicalId":13549,"journal":{"name":"Infrared Physics & Technology","volume":"151 ","pages":"Article 106067"},"PeriodicalIF":3.4000,"publicationDate":"2025-08-18","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"Infrared Physics & Technology","FirstCategoryId":"101","ListUrlMain":"https://www.sciencedirect.com/science/article/pii/S1350449525003603","RegionNum":3,"RegionCategory":"物理与天体物理","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"Q2","JCRName":"INSTRUMENTS & INSTRUMENTATION","Score":null,"Total":0}
引用次数: 0
Abstract
Existing infrared and visible image fusion methods usually use two same backbones to extract deep features of the source images, and then manually design a fusion strategy to fuse them. However, this framework overlooks the importance of feature interaction during the feature extraction stage and does not fully capture the complementary information between the source images. In addition, it requires the design of complex fusion strategies, which limits its robustness and generalization to different scenarios. To this end, we propose a one-stream one-stage pure Transformer-based fusion framework, which simplifies feature extraction and fusion into a unified one-stream pipeline. Specifically, the proposed method consists of a one-stream fusion network and a decomposition network. The fusion network uses several Swin Transformer blocks to extract and fuse two modality features simultaneously. Thanks to the sliding-window-based multi-head attention, the fusion network can acquire local features and global dependencies and seamlessly model their contextual relationships. Due to the lack of effective supervision signals, the fusion network struggles to fully transfer important information from the source images, which can easily lead to the generation of artifacts. To eliminate these artifacts and simultaneously force the fused image to contain richer information, we design a simple decomposition network that decomposes the fusion result into the source images with consistency constraints. Extensive comparative and ablation experiments on four image fusion benchmarks demonstrate that our method achieves favorable results. In addition, the results on downstream tasks, including object detection and semantic segmentation, further show the effectiveness of the proposed method.
期刊介绍:
The Journal covers the entire field of infrared physics and technology: theory, experiment, application, devices and instrumentation. Infrared'' is defined as covering the near, mid and far infrared (terahertz) regions from 0.75um (750nm) to 1mm (300GHz.) Submissions in the 300GHz to 100GHz region may be accepted at the editors discretion if their content is relevant to shorter wavelengths. Submissions must be primarily concerned with and directly relevant to this spectral region.
Its core topics can be summarized as the generation, propagation and detection, of infrared radiation; the associated optics, materials and devices; and its use in all fields of science, industry, engineering and medicine.
Infrared techniques occur in many different fields, notably spectroscopy and interferometry; material characterization and processing; atmospheric physics, astronomy and space research. Scientific aspects include lasers, quantum optics, quantum electronics, image processing and semiconductor physics. Some important applications are medical diagnostics and treatment, industrial inspection and environmental monitoring.