{"title":"自动清扫车辆多模态融合感知技术研究","authors":"Yang Zhang;Bo Yang;Wukun Lei;Xiaofei Pei","doi":"10.1109/JSEN.2025.3578375","DOIUrl":null,"url":null,"abstract":"This article proposes a multimodal fusion framework to address the challenges of detecting and tracking specialized vehicles and dynamic targets in complex industrial park environments. The framework integrates LiDAR, a monocular Camera, and an inertial navigation system (INS) to achieve precise obstacle perception and stable tracking through dynamic region of interest (ROI) cropping, optimized point cloud clustering, target detection, and multimodal perception fusion. First, a path-aware dynamic ROI cropping method and a multiregion density-aware seed point cloud ground segmentation approach are introduced to improve adaptability and point cloud processing efficiency. Second, a two-stage refinement strategy method is proposed to enhance target clustering accuracy. Furthermore, by combining the 2-D detection network, a multimodal perception fusion module, and a multiobject tracking (MOT) strategy, the framework significantly improves fusion efficiency and matching accuracy. Field tests demonstrate that the framework achieves excellent performance, with static object localization deviations below 0.8 m and reliable state estimation for dynamic targets. On a custom dataset, the monocular Camera achieves 91.67% accuracy for specialized vehicles, while the fusion framework exhibits strong adaptability and reliability in complex scenarios.","PeriodicalId":447,"journal":{"name":"IEEE Sensors Journal","volume":"25 14","pages":"27743-27753"},"PeriodicalIF":4.3000,"publicationDate":"2025-06-16","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":"{\"title\":\"Research on Multimodal Fusion Perception Technology for Autonomous Sweeping Vehicle\",\"authors\":\"Yang Zhang;Bo Yang;Wukun Lei;Xiaofei Pei\",\"doi\":\"10.1109/JSEN.2025.3578375\",\"DOIUrl\":null,\"url\":null,\"abstract\":\"This article proposes a multimodal fusion framework to address the challenges of detecting and tracking specialized vehicles and dynamic targets in complex industrial park environments. The framework integrates LiDAR, a monocular Camera, and an inertial navigation system (INS) to achieve precise obstacle perception and stable tracking through dynamic region of interest (ROI) cropping, optimized point cloud clustering, target detection, and multimodal perception fusion. First, a path-aware dynamic ROI cropping method and a multiregion density-aware seed point cloud ground segmentation approach are introduced to improve adaptability and point cloud processing efficiency. Second, a two-stage refinement strategy method is proposed to enhance target clustering accuracy. Furthermore, by combining the 2-D detection network, a multimodal perception fusion module, and a multiobject tracking (MOT) strategy, the framework significantly improves fusion efficiency and matching accuracy. Field tests demonstrate that the framework achieves excellent performance, with static object localization deviations below 0.8 m and reliable state estimation for dynamic targets. On a custom dataset, the monocular Camera achieves 91.67% accuracy for specialized vehicles, while the fusion framework exhibits strong adaptability and reliability in complex scenarios.\",\"PeriodicalId\":447,\"journal\":{\"name\":\"IEEE Sensors Journal\",\"volume\":\"25 14\",\"pages\":\"27743-27753\"},\"PeriodicalIF\":4.3000,\"publicationDate\":\"2025-06-16\",\"publicationTypes\":\"Journal Article\",\"fieldsOfStudy\":null,\"isOpenAccess\":false,\"openAccessPdf\":\"\",\"citationCount\":\"0\",\"resultStr\":null,\"platform\":\"Semanticscholar\",\"paperid\":null,\"PeriodicalName\":\"IEEE Sensors Journal\",\"FirstCategoryId\":\"103\",\"ListUrlMain\":\"https://ieeexplore.ieee.org/document/11037394/\",\"RegionNum\":2,\"RegionCategory\":\"综合性期刊\",\"ArticlePicture\":[],\"TitleCN\":null,\"AbstractTextCN\":null,\"PMCID\":null,\"EPubDate\":\"\",\"PubModel\":\"\",\"JCR\":\"Q1\",\"JCRName\":\"ENGINEERING, ELECTRICAL & ELECTRONIC\",\"Score\":null,\"Total\":0}","platform":"Semanticscholar","paperid":null,"PeriodicalName":"IEEE Sensors Journal","FirstCategoryId":"103","ListUrlMain":"https://ieeexplore.ieee.org/document/11037394/","RegionNum":2,"RegionCategory":"综合性期刊","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"Q1","JCRName":"ENGINEERING, ELECTRICAL & ELECTRONIC","Score":null,"Total":0}
Research on Multimodal Fusion Perception Technology for Autonomous Sweeping Vehicle
This article proposes a multimodal fusion framework to address the challenges of detecting and tracking specialized vehicles and dynamic targets in complex industrial park environments. The framework integrates LiDAR, a monocular Camera, and an inertial navigation system (INS) to achieve precise obstacle perception and stable tracking through dynamic region of interest (ROI) cropping, optimized point cloud clustering, target detection, and multimodal perception fusion. First, a path-aware dynamic ROI cropping method and a multiregion density-aware seed point cloud ground segmentation approach are introduced to improve adaptability and point cloud processing efficiency. Second, a two-stage refinement strategy method is proposed to enhance target clustering accuracy. Furthermore, by combining the 2-D detection network, a multimodal perception fusion module, and a multiobject tracking (MOT) strategy, the framework significantly improves fusion efficiency and matching accuracy. Field tests demonstrate that the framework achieves excellent performance, with static object localization deviations below 0.8 m and reliable state estimation for dynamic targets. On a custom dataset, the monocular Camera achieves 91.67% accuracy for specialized vehicles, while the fusion framework exhibits strong adaptability and reliability in complex scenarios.
期刊介绍:
The fields of interest of the IEEE Sensors Journal are the theory, design , fabrication, manufacturing and applications of devices for sensing and transducing physical, chemical and biological phenomena, with emphasis on the electronics and physics aspect of sensors and integrated sensors-actuators. IEEE Sensors Journal deals with the following:
-Sensor Phenomenology, Modelling, and Evaluation
-Sensor Materials, Processing, and Fabrication
-Chemical and Gas Sensors
-Microfluidics and Biosensors
-Optical Sensors
-Physical Sensors: Temperature, Mechanical, Magnetic, and others
-Acoustic and Ultrasonic Sensors
-Sensor Packaging
-Sensor Networks
-Sensor Applications
-Sensor Systems: Signals, Processing, and Interfaces
-Actuators and Sensor Power Systems
-Sensor Signal Processing for high precision and stability (amplification, filtering, linearization, modulation/demodulation) and under harsh conditions (EMC, radiation, humidity, temperature); energy consumption/harvesting
-Sensor Data Processing (soft computing with sensor data, e.g., pattern recognition, machine learning, evolutionary computation; sensor data fusion, processing of wave e.g., electromagnetic and acoustic; and non-wave, e.g., chemical, gravity, particle, thermal, radiative and non-radiative sensor data, detection, estimation and classification based on sensor data)
-Sensors in Industrial Practice