{"title":"RCFusionNet:一种雷达引导的有效多模态感知的两阶段融合方法","authors":"Weifan Xu;Xiangwei Meng;Yong Hu;Hai Deng","doi":"10.1109/JSEN.2025.3592236","DOIUrl":null,"url":null,"abstract":"Multisensor fusion perception plays a pivotal role in advancing autonomous driving. While most existing LiDAR–camera (LC) fusion methods achieve high detection accuracy, LiDAR is inherently vulnerable to extreme weather conditions and is costlier compared to microwave radar. In this work, we introduce a robust radar–camera (RC) fusion model, RCFusionNet, aimed at delivering enhanced and robust detection performances under adverse weather conditions. Specifically, we introduce a camera feature encoder (CFE) to enhance the capability of visual representations, which incorporate multiscale bird’s-eye view (BEV) features to further enhance the detection performance. Additionally, a radar feature encoder (RFE) is introduced to fully exploit radar point cloud data by selectively extracting informative radar features. Furthermore, we adopt a radar-guided two-stage fusion strategy (RGTSFS), applied during both the BEV feature generation and the RC feature fusion phases. Extensive evaluations on the nuScenes benchmark demonstrate the effectiveness of our RCFusionNet.","PeriodicalId":447,"journal":{"name":"IEEE Sensors Journal","volume":"25 17","pages":"33890-33900"},"PeriodicalIF":4.3000,"publicationDate":"2025-07-30","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":"{\"title\":\"RCFusionNet: A Radar-Guided Two-Stage Fusion Approach to Effective Multimodal Perception\",\"authors\":\"Weifan Xu;Xiangwei Meng;Yong Hu;Hai Deng\",\"doi\":\"10.1109/JSEN.2025.3592236\",\"DOIUrl\":null,\"url\":null,\"abstract\":\"Multisensor fusion perception plays a pivotal role in advancing autonomous driving. While most existing LiDAR–camera (LC) fusion methods achieve high detection accuracy, LiDAR is inherently vulnerable to extreme weather conditions and is costlier compared to microwave radar. In this work, we introduce a robust radar–camera (RC) fusion model, RCFusionNet, aimed at delivering enhanced and robust detection performances under adverse weather conditions. Specifically, we introduce a camera feature encoder (CFE) to enhance the capability of visual representations, which incorporate multiscale bird’s-eye view (BEV) features to further enhance the detection performance. Additionally, a radar feature encoder (RFE) is introduced to fully exploit radar point cloud data by selectively extracting informative radar features. Furthermore, we adopt a radar-guided two-stage fusion strategy (RGTSFS), applied during both the BEV feature generation and the RC feature fusion phases. Extensive evaluations on the nuScenes benchmark demonstrate the effectiveness of our RCFusionNet.\",\"PeriodicalId\":447,\"journal\":{\"name\":\"IEEE Sensors Journal\",\"volume\":\"25 17\",\"pages\":\"33890-33900\"},\"PeriodicalIF\":4.3000,\"publicationDate\":\"2025-07-30\",\"publicationTypes\":\"Journal Article\",\"fieldsOfStudy\":null,\"isOpenAccess\":false,\"openAccessPdf\":\"\",\"citationCount\":\"0\",\"resultStr\":null,\"platform\":\"Semanticscholar\",\"paperid\":null,\"PeriodicalName\":\"IEEE Sensors Journal\",\"FirstCategoryId\":\"103\",\"ListUrlMain\":\"https://ieeexplore.ieee.org/document/11104960/\",\"RegionNum\":2,\"RegionCategory\":\"综合性期刊\",\"ArticlePicture\":[],\"TitleCN\":null,\"AbstractTextCN\":null,\"PMCID\":null,\"EPubDate\":\"\",\"PubModel\":\"\",\"JCR\":\"Q1\",\"JCRName\":\"ENGINEERING, ELECTRICAL & ELECTRONIC\",\"Score\":null,\"Total\":0}","platform":"Semanticscholar","paperid":null,"PeriodicalName":"IEEE Sensors Journal","FirstCategoryId":"103","ListUrlMain":"https://ieeexplore.ieee.org/document/11104960/","RegionNum":2,"RegionCategory":"综合性期刊","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"Q1","JCRName":"ENGINEERING, ELECTRICAL & ELECTRONIC","Score":null,"Total":0}
RCFusionNet: A Radar-Guided Two-Stage Fusion Approach to Effective Multimodal Perception
Multisensor fusion perception plays a pivotal role in advancing autonomous driving. While most existing LiDAR–camera (LC) fusion methods achieve high detection accuracy, LiDAR is inherently vulnerable to extreme weather conditions and is costlier compared to microwave radar. In this work, we introduce a robust radar–camera (RC) fusion model, RCFusionNet, aimed at delivering enhanced and robust detection performances under adverse weather conditions. Specifically, we introduce a camera feature encoder (CFE) to enhance the capability of visual representations, which incorporate multiscale bird’s-eye view (BEV) features to further enhance the detection performance. Additionally, a radar feature encoder (RFE) is introduced to fully exploit radar point cloud data by selectively extracting informative radar features. Furthermore, we adopt a radar-guided two-stage fusion strategy (RGTSFS), applied during both the BEV feature generation and the RC feature fusion phases. Extensive evaluations on the nuScenes benchmark demonstrate the effectiveness of our RCFusionNet.
期刊介绍:
The fields of interest of the IEEE Sensors Journal are the theory, design , fabrication, manufacturing and applications of devices for sensing and transducing physical, chemical and biological phenomena, with emphasis on the electronics and physics aspect of sensors and integrated sensors-actuators. IEEE Sensors Journal deals with the following:
-Sensor Phenomenology, Modelling, and Evaluation
-Sensor Materials, Processing, and Fabrication
-Chemical and Gas Sensors
-Microfluidics and Biosensors
-Optical Sensors
-Physical Sensors: Temperature, Mechanical, Magnetic, and others
-Acoustic and Ultrasonic Sensors
-Sensor Packaging
-Sensor Networks
-Sensor Applications
-Sensor Systems: Signals, Processing, and Interfaces
-Actuators and Sensor Power Systems
-Sensor Signal Processing for high precision and stability (amplification, filtering, linearization, modulation/demodulation) and under harsh conditions (EMC, radiation, humidity, temperature); energy consumption/harvesting
-Sensor Data Processing (soft computing with sensor data, e.g., pattern recognition, machine learning, evolutionary computation; sensor data fusion, processing of wave e.g., electromagnetic and acoustic; and non-wave, e.g., chemical, gravity, particle, thermal, radiative and non-radiative sensor data, detection, estimation and classification based on sensor data)
-Sensors in Industrial Practice