Multi-Uncertainty Captured Multi-Robot Lidar Odometry and Mapping Framework for Large-Scale Environments

Guang-ming Xiong, Junyi Ma, Huilong Yu, Jingyi Xu, Jiahui Xu
{"title":"Multi-Uncertainty Captured Multi-Robot Lidar Odometry and Mapping Framework for Large-Scale Environments","authors":"Guang-ming Xiong, Junyi Ma, Huilong Yu, Jingyi Xu, Jiahui Xu","doi":"10.1142/s2301385023410030","DOIUrl":null,"url":null,"abstract":"Multi-robot simultaneous localization and mapping (MR-SLAM) is of great importance for enhancing the efficiency of large-scale environment exploration. Despite remarkable advances in schemes for cooperation, there is a critical lack of approaches to handle multiple uncertainties inherent to MR-SLAM in large-scale environments. This paper proposes a multi-uncertainty captured multi-robot lidar odometry and mapping (MUC-LOAM) framework, to quantify and utilize the uncertainties of feature points and robot mutual poses in large-scale environments. A proposed hybrid weighting strategy for pose update is integrated into MUC-LOAM to handle feature uncertainty from distance changing and dynamic objects. A devised Bayesian Neural Network (BNN) is proposed to capture mutual pose uncertainty. Then the covariance propagation of quaternions to Euler angles conversion is leveraged to filter out unreliable mutual poses. Another covariance propagation through coordinate transformations in nonlinear optimization improves the accuracy of map merging. The feasibility and enhanced robustness of the proposed framework for large-scale exploration are validated on both public datasets and real-world experiments.","PeriodicalId":164619,"journal":{"name":"Unmanned Syst.","volume":null,"pages":null},"PeriodicalIF":0.0000,"publicationDate":"2022-07-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"2","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"Unmanned Syst.","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/10.1142/s2301385023410030","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
引用次数: 2

Abstract

Multi-robot simultaneous localization and mapping (MR-SLAM) is of great importance for enhancing the efficiency of large-scale environment exploration. Despite remarkable advances in schemes for cooperation, there is a critical lack of approaches to handle multiple uncertainties inherent to MR-SLAM in large-scale environments. This paper proposes a multi-uncertainty captured multi-robot lidar odometry and mapping (MUC-LOAM) framework, to quantify and utilize the uncertainties of feature points and robot mutual poses in large-scale environments. A proposed hybrid weighting strategy for pose update is integrated into MUC-LOAM to handle feature uncertainty from distance changing and dynamic objects. A devised Bayesian Neural Network (BNN) is proposed to capture mutual pose uncertainty. Then the covariance propagation of quaternions to Euler angles conversion is leveraged to filter out unreliable mutual poses. Another covariance propagation through coordinate transformations in nonlinear optimization improves the accuracy of map merging. The feasibility and enhanced robustness of the proposed framework for large-scale exploration are validated on both public datasets and real-world experiments.
面向大尺度环境的多不确定多机器人激光雷达测程与制图框架
多机器人同步定位与制图(MR-SLAM)对于提高大规模环境勘探的效率具有重要意义。尽管在合作方案方面取得了显著进展,但严重缺乏处理大规模环境中MR-SLAM固有的多重不确定性的方法。为了量化和利用大尺度环境下特征点和机器人互位姿的不确定性,提出了一种多不确定性捕获多机器人激光雷达测程与制图(mu - loam)框架。将姿态更新的混合加权策略集成到mu - loam中,以处理来自距离变化和动态目标的特征不确定性。提出了一种设计的贝叶斯神经网络(BNN)来捕获互位不确定性。然后利用四元数的协方差传播到欧拉角转换,过滤掉不可靠的互位姿。非线性优化中的另一种协方差传播方法通过坐标变换提高了地图合并的精度。在公共数据集和实际实验中验证了该框架在大规模勘探中的可行性和增强的鲁棒性。
本文章由计算机程序翻译,如有差异,请以英文原文为准。
求助全文
约1分钟内获得全文 求助全文
来源期刊
自引率
0.00%
发文量
0
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
确定
请完成安全验证×
copy
已复制链接
快去分享给好友吧!
我知道了
右上角分享
点击右上角分享
0
联系我们:info@booksci.cn Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。 Copyright © 2023 布克学术 All rights reserved.
京ICP备2023020795号-1
ghs 京公网安备 11010802042870号
Book学术文献互助
Book学术文献互助群
群 号:481959085
Book学术官方微信