Hao Du, Lu Ren, Yuanda Wang, Xiang Cao, Changyin Sun
{"title":"Advancements in perception system with multi-sensor fusion for embodied agents","authors":"Hao Du, Lu Ren, Yuanda Wang, Xiang Cao, Changyin Sun","doi":"10.1016/j.inffus.2024.102859","DOIUrl":null,"url":null,"abstract":"The multi-sensor data fusion perception technology, as a pivotal technique for achieving complex environmental perception and decision-making, has been garnering extensive attention from researchers. To date, there has been a lack of comprehensive review articles discussing the research progress of multi-sensor fusion perception systems for embodied agents, particularly in terms of analyzing the agent’s perception of itself and the surrounding scene. To address this gap and encourage further research, this study defines key terminology and analyzes datasets from the past two decades, focusing on advancements in multi-sensor fusion SLAM and multi-sensor scene perception. These key designs can aid researchers in gaining a better understanding of the field and initiating research in the domain of multi-sensor fusion perception for embodied agents. In this survey, we begin with a brief introduction to common sensor types and their characteristics. We then delve into the multi-sensor fusion perception datasets tailored for the domains of autonomous driving, drones, unmanned ground vehicles, and unmanned surface vehicles. Following this, we discuss the classification and fundamental principles of existing multi-sensor data fusion SLAM algorithms, and present the experimental outcomes of various classical fusion frameworks. Subsequently, we comprehensively review the technologies of multi-sensor data fusion scene perception, including object detection, semantic segmentation, instance segmentation, and panoramic understanding. Finally, we summarize our findings and discuss potential future developments in multi-sensor fusion perception technology.","PeriodicalId":50367,"journal":{"name":"Information Fusion","volume":"47 1","pages":""},"PeriodicalIF":14.7000,"publicationDate":"2024-12-11","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"Information Fusion","FirstCategoryId":"94","ListUrlMain":"https://doi.org/10.1016/j.inffus.2024.102859","RegionNum":1,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"Q1","JCRName":"COMPUTER SCIENCE, ARTIFICIAL INTELLIGENCE","Score":null,"Total":0}
引用次数: 0
摘要
多传感器数据融合感知技术作为实现复杂环境感知和决策的关键技术,一直受到研究人员的广泛关注。迄今为止,还缺乏全面的综述文章来讨论面向具身代理的多传感器融合感知系统的研究进展,尤其是在分析代理对自身和周围场景的感知方面。为了填补这一空白并鼓励进一步研究,本研究定义了关键术语,并分析了过去二十年的数据集,重点关注多传感器融合 SLAM 和多传感器场景感知方面的进展。这些关键设计可帮助研究人员更好地了解该领域,并启动面向具身代理的多传感器融合感知领域的研究。在这份调查报告中,我们首先简要介绍了常见的传感器类型及其特点。然后,我们深入探讨了为自动驾驶、无人机、无人地面车辆和无人水面车辆等领域量身定制的多传感器融合感知数据集。随后,我们讨论了现有多传感器数据融合 SLAM 算法的分类和基本原理,并介绍了各种经典融合框架的实验结果。随后,我们全面回顾了多传感器数据融合场景感知技术,包括物体检测、语义分割、实例分割和全景理解。最后,我们总结了我们的研究成果,并讨论了多传感器融合感知技术的未来发展潜力。
Advancements in perception system with multi-sensor fusion for embodied agents
The multi-sensor data fusion perception technology, as a pivotal technique for achieving complex environmental perception and decision-making, has been garnering extensive attention from researchers. To date, there has been a lack of comprehensive review articles discussing the research progress of multi-sensor fusion perception systems for embodied agents, particularly in terms of analyzing the agent’s perception of itself and the surrounding scene. To address this gap and encourage further research, this study defines key terminology and analyzes datasets from the past two decades, focusing on advancements in multi-sensor fusion SLAM and multi-sensor scene perception. These key designs can aid researchers in gaining a better understanding of the field and initiating research in the domain of multi-sensor fusion perception for embodied agents. In this survey, we begin with a brief introduction to common sensor types and their characteristics. We then delve into the multi-sensor fusion perception datasets tailored for the domains of autonomous driving, drones, unmanned ground vehicles, and unmanned surface vehicles. Following this, we discuss the classification and fundamental principles of existing multi-sensor data fusion SLAM algorithms, and present the experimental outcomes of various classical fusion frameworks. Subsequently, we comprehensively review the technologies of multi-sensor data fusion scene perception, including object detection, semantic segmentation, instance segmentation, and panoramic understanding. Finally, we summarize our findings and discuss potential future developments in multi-sensor fusion perception technology.
期刊介绍:
Information Fusion serves as a central platform for showcasing advancements in multi-sensor, multi-source, multi-process information fusion, fostering collaboration among diverse disciplines driving its progress. It is the leading outlet for sharing research and development in this field, focusing on architectures, algorithms, and applications. Papers dealing with fundamental theoretical analyses as well as those demonstrating their application to real-world problems will be welcome.