Mingxin Yu , Zhenyang Liang , Ning Li , Mingwei Lin
{"title":"DGTF: A framework utilizing textual semantics for detail-aware and global fusion with a multi-degradation scenarios dataset","authors":"Mingxin Yu , Zhenyang Liang , Ning Li , Mingwei Lin","doi":"10.1016/j.optlastec.2025.113319","DOIUrl":null,"url":null,"abstract":"<div><div>Infrared and visible light fusion aims to integrate information from both modalities to generate high-quality fused images. However, existing methods do not perform well in multiple degraded scenarios. In these scenarios, the source images face quality degradation and information loss. Fusion combined with degradation processing often damages the detailed target information in the images. To address this limitation, this paper proposes a novel Degradation-Text Fusion framework, named DGTF, which leverages cascaded degradation text, object text, and target masks for detail-aware degradation regulation. The framework adjusts the details based on the object text specified in the given input, ensuring that global degradation processing does not compromise the quality of detail fusion. This approach overcomes the limitations of previous methods constrained by global degradation processing. To train and evaluate DGTF, we constructed a new infrared and visible light dataset, Multi-degraded scene text target infrared and visible datasets (MTS), which encompasses seven extreme scenarios, including rain, snow, fog, low light, exposure, infrared noise, and low contrast. Extensive experimental results demonstrate that our method significantly outperforms existing techniques in fusion performance, even without text guidance. Furthermore, tests conducted on the MTS dataset reveal that the detail-regulated fusion results achieved by DGTF far surpass traditional degradation-based fusion methods, effectively enhancing the performance of advanced vision tasks. These findings validate the effectiveness of the proposed detail regulation framework. Our code is available at <span><span>https://github.com/linshenj/DGTF</span><svg><path></path></svg></span>.</div></div>","PeriodicalId":19511,"journal":{"name":"Optics and Laser Technology","volume":"191 ","pages":"Article 113319"},"PeriodicalIF":4.6000,"publicationDate":"2025-06-12","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"Optics and Laser Technology","FirstCategoryId":"101","ListUrlMain":"https://www.sciencedirect.com/science/article/pii/S0030399225009107","RegionNum":2,"RegionCategory":"物理与天体物理","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"Q1","JCRName":"OPTICS","Score":null,"Total":0}
引用次数: 0
Abstract
Infrared and visible light fusion aims to integrate information from both modalities to generate high-quality fused images. However, existing methods do not perform well in multiple degraded scenarios. In these scenarios, the source images face quality degradation and information loss. Fusion combined with degradation processing often damages the detailed target information in the images. To address this limitation, this paper proposes a novel Degradation-Text Fusion framework, named DGTF, which leverages cascaded degradation text, object text, and target masks for detail-aware degradation regulation. The framework adjusts the details based on the object text specified in the given input, ensuring that global degradation processing does not compromise the quality of detail fusion. This approach overcomes the limitations of previous methods constrained by global degradation processing. To train and evaluate DGTF, we constructed a new infrared and visible light dataset, Multi-degraded scene text target infrared and visible datasets (MTS), which encompasses seven extreme scenarios, including rain, snow, fog, low light, exposure, infrared noise, and low contrast. Extensive experimental results demonstrate that our method significantly outperforms existing techniques in fusion performance, even without text guidance. Furthermore, tests conducted on the MTS dataset reveal that the detail-regulated fusion results achieved by DGTF far surpass traditional degradation-based fusion methods, effectively enhancing the performance of advanced vision tasks. These findings validate the effectiveness of the proposed detail regulation framework. Our code is available at https://github.com/linshenj/DGTF.
期刊介绍:
Optics & Laser Technology aims to provide a vehicle for the publication of a broad range of high quality research and review papers in those fields of scientific and engineering research appertaining to the development and application of the technology of optics and lasers. Papers describing original work in these areas are submitted to rigorous refereeing prior to acceptance for publication.
The scope of Optics & Laser Technology encompasses, but is not restricted to, the following areas:
•development in all types of lasers
•developments in optoelectronic devices and photonics
•developments in new photonics and optical concepts
•developments in conventional optics, optical instruments and components
•techniques of optical metrology, including interferometry and optical fibre sensors
•LIDAR and other non-contact optical measurement techniques, including optical methods in heat and fluid flow
•applications of lasers to materials processing, optical NDT display (including holography) and optical communication
•research and development in the field of laser safety including studies of hazards resulting from the applications of lasers (laser safety, hazards of laser fume)
•developments in optical computing and optical information processing
•developments in new optical materials
•developments in new optical characterization methods and techniques
•developments in quantum optics
•developments in light assisted micro and nanofabrication methods and techniques
•developments in nanophotonics and biophotonics
•developments in imaging processing and systems