{"title":"基于模态特征增强的低光环境下可见光与红外图像融合定位","authors":"Shan Su;Li Yan;Yuquan Zhou;Pinzhuo Wang;Changjun Chen","doi":"10.1109/JSEN.2025.3576989","DOIUrl":null,"url":null,"abstract":"In low-light environments, the imaging quality of single-modal red green blue (RGB) sensors severely deteriorates, leading to blurred textures and loss of effective information, which affects the accuracy of downstream tasks. Our goal is to provide robust feature extraction in low-light environments by fusing visible and infrared images. Infrared and visible fusion is an important and effective image enhancement technique, aiming to generate high-quality fused images with prominent targets and rich textures in challenging environments. Hence, we propose an unsupervised enhanced infrared and visible fusion method for low-light environments. The method first designs a module for single-modal image feature enhancement based on infrared and visible images for low-light conditions, initially improving the quality of the original infrared and visible images. Subsequently, a cross-modal feature enhancement module based on edge/texture information extraction and guidance is proposed to enhance the edge structures and texture details in the fused features. Specifically, the network utilizes the ResNet as the backbone for single-modal image feature extraction and enhancement, employing channel and spatial attention mechanisms to enhance single-modal image features. An infrared/visible image edge/texture information guidance module is added, leveraging the complementary edge/texture features provided by the two different modalities to guide the learning of the other modality image, thereby achieving the goal of cross-modal image enhancement in low-light environments. In the fusion stage, the DenseNet is employed as the unsupervised fusion network framework. Based on image information measurements, a patch-based loss function with regional weighting is designed, enabling the fused image to dynamically learn advantageous features from different regions of different modal images, achieving complementary feature enhancement of infrared and visible images. Through qualitative and quantitative analyses of three datasets, compared with eight other state of the art (SOTA) methods, the proposed method demonstrates balanced performance, achieving high-quality fusion of multimodal features in low-light environments.","PeriodicalId":447,"journal":{"name":"IEEE Sensors Journal","volume":"25 15","pages":"28476-28492"},"PeriodicalIF":4.3000,"publicationDate":"2025-06-11","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":"{\"title\":\"Visible and Infrared Image Fusion Based on Modality Feature Enhancement for Localization in Low-Light Environments\",\"authors\":\"Shan Su;Li Yan;Yuquan Zhou;Pinzhuo Wang;Changjun Chen\",\"doi\":\"10.1109/JSEN.2025.3576989\",\"DOIUrl\":null,\"url\":null,\"abstract\":\"In low-light environments, the imaging quality of single-modal red green blue (RGB) sensors severely deteriorates, leading to blurred textures and loss of effective information, which affects the accuracy of downstream tasks. Our goal is to provide robust feature extraction in low-light environments by fusing visible and infrared images. Infrared and visible fusion is an important and effective image enhancement technique, aiming to generate high-quality fused images with prominent targets and rich textures in challenging environments. Hence, we propose an unsupervised enhanced infrared and visible fusion method for low-light environments. The method first designs a module for single-modal image feature enhancement based on infrared and visible images for low-light conditions, initially improving the quality of the original infrared and visible images. Subsequently, a cross-modal feature enhancement module based on edge/texture information extraction and guidance is proposed to enhance the edge structures and texture details in the fused features. Specifically, the network utilizes the ResNet as the backbone for single-modal image feature extraction and enhancement, employing channel and spatial attention mechanisms to enhance single-modal image features. An infrared/visible image edge/texture information guidance module is added, leveraging the complementary edge/texture features provided by the two different modalities to guide the learning of the other modality image, thereby achieving the goal of cross-modal image enhancement in low-light environments. In the fusion stage, the DenseNet is employed as the unsupervised fusion network framework. Based on image information measurements, a patch-based loss function with regional weighting is designed, enabling the fused image to dynamically learn advantageous features from different regions of different modal images, achieving complementary feature enhancement of infrared and visible images. Through qualitative and quantitative analyses of three datasets, compared with eight other state of the art (SOTA) methods, the proposed method demonstrates balanced performance, achieving high-quality fusion of multimodal features in low-light environments.\",\"PeriodicalId\":447,\"journal\":{\"name\":\"IEEE Sensors Journal\",\"volume\":\"25 15\",\"pages\":\"28476-28492\"},\"PeriodicalIF\":4.3000,\"publicationDate\":\"2025-06-11\",\"publicationTypes\":\"Journal Article\",\"fieldsOfStudy\":null,\"isOpenAccess\":false,\"openAccessPdf\":\"\",\"citationCount\":\"0\",\"resultStr\":null,\"platform\":\"Semanticscholar\",\"paperid\":null,\"PeriodicalName\":\"IEEE Sensors Journal\",\"FirstCategoryId\":\"103\",\"ListUrlMain\":\"https://ieeexplore.ieee.org/document/11031096/\",\"RegionNum\":2,\"RegionCategory\":\"综合性期刊\",\"ArticlePicture\":[],\"TitleCN\":null,\"AbstractTextCN\":null,\"PMCID\":null,\"EPubDate\":\"\",\"PubModel\":\"\",\"JCR\":\"Q1\",\"JCRName\":\"ENGINEERING, ELECTRICAL & ELECTRONIC\",\"Score\":null,\"Total\":0}","platform":"Semanticscholar","paperid":null,"PeriodicalName":"IEEE Sensors Journal","FirstCategoryId":"103","ListUrlMain":"https://ieeexplore.ieee.org/document/11031096/","RegionNum":2,"RegionCategory":"综合性期刊","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"Q1","JCRName":"ENGINEERING, ELECTRICAL & ELECTRONIC","Score":null,"Total":0}
Visible and Infrared Image Fusion Based on Modality Feature Enhancement for Localization in Low-Light Environments
In low-light environments, the imaging quality of single-modal red green blue (RGB) sensors severely deteriorates, leading to blurred textures and loss of effective information, which affects the accuracy of downstream tasks. Our goal is to provide robust feature extraction in low-light environments by fusing visible and infrared images. Infrared and visible fusion is an important and effective image enhancement technique, aiming to generate high-quality fused images with prominent targets and rich textures in challenging environments. Hence, we propose an unsupervised enhanced infrared and visible fusion method for low-light environments. The method first designs a module for single-modal image feature enhancement based on infrared and visible images for low-light conditions, initially improving the quality of the original infrared and visible images. Subsequently, a cross-modal feature enhancement module based on edge/texture information extraction and guidance is proposed to enhance the edge structures and texture details in the fused features. Specifically, the network utilizes the ResNet as the backbone for single-modal image feature extraction and enhancement, employing channel and spatial attention mechanisms to enhance single-modal image features. An infrared/visible image edge/texture information guidance module is added, leveraging the complementary edge/texture features provided by the two different modalities to guide the learning of the other modality image, thereby achieving the goal of cross-modal image enhancement in low-light environments. In the fusion stage, the DenseNet is employed as the unsupervised fusion network framework. Based on image information measurements, a patch-based loss function with regional weighting is designed, enabling the fused image to dynamically learn advantageous features from different regions of different modal images, achieving complementary feature enhancement of infrared and visible images. Through qualitative and quantitative analyses of three datasets, compared with eight other state of the art (SOTA) methods, the proposed method demonstrates balanced performance, achieving high-quality fusion of multimodal features in low-light environments.
期刊介绍:
The fields of interest of the IEEE Sensors Journal are the theory, design , fabrication, manufacturing and applications of devices for sensing and transducing physical, chemical and biological phenomena, with emphasis on the electronics and physics aspect of sensors and integrated sensors-actuators. IEEE Sensors Journal deals with the following:
-Sensor Phenomenology, Modelling, and Evaluation
-Sensor Materials, Processing, and Fabrication
-Chemical and Gas Sensors
-Microfluidics and Biosensors
-Optical Sensors
-Physical Sensors: Temperature, Mechanical, Magnetic, and others
-Acoustic and Ultrasonic Sensors
-Sensor Packaging
-Sensor Networks
-Sensor Applications
-Sensor Systems: Signals, Processing, and Interfaces
-Actuators and Sensor Power Systems
-Sensor Signal Processing for high precision and stability (amplification, filtering, linearization, modulation/demodulation) and under harsh conditions (EMC, radiation, humidity, temperature); energy consumption/harvesting
-Sensor Data Processing (soft computing with sensor data, e.g., pattern recognition, machine learning, evolutionary computation; sensor data fusion, processing of wave e.g., electromagnetic and acoustic; and non-wave, e.g., chemical, gravity, particle, thermal, radiative and non-radiative sensor data, detection, estimation and classification based on sensor data)
-Sensors in Industrial Practice