{"title":"Enhanced visible light detail and infrared thermal radiation for dual-mode imaging system via multi-information interaction","authors":"Xiaosong Liu, Huaibin Qiu, Zhuolin Ou, Jiazhen Dou, Jianglei Di, Yuwen Qin","doi":"10.1016/j.jvcir.2025.104583","DOIUrl":null,"url":null,"abstract":"<div><div>In the field of dual-mode optical imaging, image fusion techniques offer significant advantages such as improved spatial resolution and the suppression of redundant information, particularly for visible and infrared image. However, existing fusion methods often overlook the interaction between multiple feature information during the extraction and fusion stages, resulting in the inability to extract both visible light detail and infrared thermal radiation effectively. To address this challenge, we construct a dual-mode imaging system and propose an image fusion method that incorporates Convolution-Swin-Transformer Blocks (CSTBs). The block combines Convolution and Shifted Window Transformer to improve the interaction and extraction between local and global information within the images. On the other hand, our proposed method strengthens the comprehensive interaction and fusion between shallow pixel-level information and deeper semantic representation by fusing local and global feature information at various layers. Furthermore, we introduce a multi-component loss function that balances the complementary features extracted from the source images, with a particular focus on enhancing edge texture, structure, and brightness information. Experimental results demonstrate that our method achieves superior performance in simultaneously enhancing both texture details and thermal radiation. This is evidenced by results on two publicly available datasets, as well as the Target_GDUT dataset captured using our dual-mode optical imaging system.</div></div>","PeriodicalId":54755,"journal":{"name":"Journal of Visual Communication and Image Representation","volume":"112 ","pages":"Article 104583"},"PeriodicalIF":3.1000,"publicationDate":"2025-09-13","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"Journal of Visual Communication and Image Representation","FirstCategoryId":"94","ListUrlMain":"https://www.sciencedirect.com/science/article/pii/S104732032500197X","RegionNum":4,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"Q2","JCRName":"COMPUTER SCIENCE, INFORMATION SYSTEMS","Score":null,"Total":0}
引用次数: 0
Abstract
In the field of dual-mode optical imaging, image fusion techniques offer significant advantages such as improved spatial resolution and the suppression of redundant information, particularly for visible and infrared image. However, existing fusion methods often overlook the interaction between multiple feature information during the extraction and fusion stages, resulting in the inability to extract both visible light detail and infrared thermal radiation effectively. To address this challenge, we construct a dual-mode imaging system and propose an image fusion method that incorporates Convolution-Swin-Transformer Blocks (CSTBs). The block combines Convolution and Shifted Window Transformer to improve the interaction and extraction between local and global information within the images. On the other hand, our proposed method strengthens the comprehensive interaction and fusion between shallow pixel-level information and deeper semantic representation by fusing local and global feature information at various layers. Furthermore, we introduce a multi-component loss function that balances the complementary features extracted from the source images, with a particular focus on enhancing edge texture, structure, and brightness information. Experimental results demonstrate that our method achieves superior performance in simultaneously enhancing both texture details and thermal radiation. This is evidenced by results on two publicly available datasets, as well as the Target_GDUT dataset captured using our dual-mode optical imaging system.
期刊介绍:
The Journal of Visual Communication and Image Representation publishes papers on state-of-the-art visual communication and image representation, with emphasis on novel technologies and theoretical work in this multidisciplinary area of pure and applied research. The field of visual communication and image representation is considered in its broadest sense and covers both digital and analog aspects as well as processing and communication in biological visual systems.