LVIMOT:精确和鲁棒的激光雷达-视觉-惯性定位和多目标跟踪在动态环境中通过紧密耦合集成

IF 12.2 1区 地球科学 Q1 GEOGRAPHY, PHYSICAL
Shaoquan Feng, Xingxing Li, Zhuohao Yan, Yuxuan Zhou, Chunxi Xia
{"title":"LVIMOT:精确和鲁棒的激光雷达-视觉-惯性定位和多目标跟踪在动态环境中通过紧密耦合集成","authors":"Shaoquan Feng,&nbsp;Xingxing Li,&nbsp;Zhuohao Yan,&nbsp;Yuxuan Zhou,&nbsp;Chunxi Xia","doi":"10.1016/j.isprsjprs.2025.10.003","DOIUrl":null,"url":null,"abstract":"<div><div>Most existing LiDAR-visual-inertial (LVI) localization systems typically assume a static environment and often neglect the valuable dynamic object information, which limits localization accuracy and scene understanding in dynamic environments. Meanwhile, precise tracking of surrounding objects is essential for applications such as autonomous driving, augmented reality (AR), and virtual reality (VR). To this end, we propose LVIMOT, a tightly coupled LVI localization and multi-object tracking (MOT) system based on factor graph optimization, capable of jointly estimating the trajectories of the ego-vehicle and surrounding objects. In the proposed method, tracked objects are represented by 2D/3D bounding boxes and continuously associated by combining LiDAR and visual detections with multimodal feature cues. Building upon this, a binary hypothesis testing method for motion status classification is employed to identify object motion status by fusing motion and appearance information. Subsequently, the measurements associated with dynamic objects are fully exploited to construct object-related factors, which are jointly optimized with static features and IMU pre-integration factors within a unified factor graph to refine the trajectories of both the ego-vehicle and tracked objects. Extensive experiments on highly dynamic sequences from the KITTI and nuScenes datasets demonstrate that LVIMOT achieves state-of-the-art performance, with an HOTA of 80.37 for MOT, and the lowest mean translational and rotational absolute trajectory errors (ATE) for self-localization on KITTI (0.36 m, 0.02 rad) and nuScenes (0.13 m, 0.02 rad). The results confirm the effectiveness of integrating object-aware motion constraints and multimodal fusion in improving localization robustness and precision. The source code will be available at <span><span>https://github.com/shqfeng/LVIMOT.git</span><svg><path></path></svg></span>.</div></div>","PeriodicalId":50269,"journal":{"name":"ISPRS Journal of Photogrammetry and Remote Sensing","volume":"230 ","pages":"Pages 675-692"},"PeriodicalIF":12.2000,"publicationDate":"2025-10-10","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":"{\"title\":\"LVIMOT: Accurate and robust LiDAR-visual-inertial localization and multi-object tracking in dynamic environments via tightly coupled integration\",\"authors\":\"Shaoquan Feng,&nbsp;Xingxing Li,&nbsp;Zhuohao Yan,&nbsp;Yuxuan Zhou,&nbsp;Chunxi Xia\",\"doi\":\"10.1016/j.isprsjprs.2025.10.003\",\"DOIUrl\":null,\"url\":null,\"abstract\":\"<div><div>Most existing LiDAR-visual-inertial (LVI) localization systems typically assume a static environment and often neglect the valuable dynamic object information, which limits localization accuracy and scene understanding in dynamic environments. Meanwhile, precise tracking of surrounding objects is essential for applications such as autonomous driving, augmented reality (AR), and virtual reality (VR). To this end, we propose LVIMOT, a tightly coupled LVI localization and multi-object tracking (MOT) system based on factor graph optimization, capable of jointly estimating the trajectories of the ego-vehicle and surrounding objects. In the proposed method, tracked objects are represented by 2D/3D bounding boxes and continuously associated by combining LiDAR and visual detections with multimodal feature cues. Building upon this, a binary hypothesis testing method for motion status classification is employed to identify object motion status by fusing motion and appearance information. Subsequently, the measurements associated with dynamic objects are fully exploited to construct object-related factors, which are jointly optimized with static features and IMU pre-integration factors within a unified factor graph to refine the trajectories of both the ego-vehicle and tracked objects. Extensive experiments on highly dynamic sequences from the KITTI and nuScenes datasets demonstrate that LVIMOT achieves state-of-the-art performance, with an HOTA of 80.37 for MOT, and the lowest mean translational and rotational absolute trajectory errors (ATE) for self-localization on KITTI (0.36 m, 0.02 rad) and nuScenes (0.13 m, 0.02 rad). The results confirm the effectiveness of integrating object-aware motion constraints and multimodal fusion in improving localization robustness and precision. The source code will be available at <span><span>https://github.com/shqfeng/LVIMOT.git</span><svg><path></path></svg></span>.</div></div>\",\"PeriodicalId\":50269,\"journal\":{\"name\":\"ISPRS Journal of Photogrammetry and Remote Sensing\",\"volume\":\"230 \",\"pages\":\"Pages 675-692\"},\"PeriodicalIF\":12.2000,\"publicationDate\":\"2025-10-10\",\"publicationTypes\":\"Journal Article\",\"fieldsOfStudy\":null,\"isOpenAccess\":false,\"openAccessPdf\":\"\",\"citationCount\":\"0\",\"resultStr\":null,\"platform\":\"Semanticscholar\",\"paperid\":null,\"PeriodicalName\":\"ISPRS Journal of Photogrammetry and Remote Sensing\",\"FirstCategoryId\":\"5\",\"ListUrlMain\":\"https://www.sciencedirect.com/science/article/pii/S0924271625003910\",\"RegionNum\":1,\"RegionCategory\":\"地球科学\",\"ArticlePicture\":[],\"TitleCN\":null,\"AbstractTextCN\":null,\"PMCID\":null,\"EPubDate\":\"\",\"PubModel\":\"\",\"JCR\":\"Q1\",\"JCRName\":\"GEOGRAPHY, PHYSICAL\",\"Score\":null,\"Total\":0}","platform":"Semanticscholar","paperid":null,"PeriodicalName":"ISPRS Journal of Photogrammetry and Remote Sensing","FirstCategoryId":"5","ListUrlMain":"https://www.sciencedirect.com/science/article/pii/S0924271625003910","RegionNum":1,"RegionCategory":"地球科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"Q1","JCRName":"GEOGRAPHY, PHYSICAL","Score":null,"Total":0}
引用次数: 0

摘要

现有的激光雷达-视觉惯性(LVI)定位系统通常假设一个静态环境,往往忽略了有价值的动态目标信息,这限制了动态环境下的定位精度和对场景的理解。同时,精确跟踪周围物体对于自动驾驶、增强现实(AR)和虚拟现实(VR)等应用至关重要。为此,我们提出了一种基于因子图优化的紧密耦合LVI定位与多目标跟踪(MOT)系统LVIMOT,该系统能够联合估计自我车辆和周围物体的轨迹。在该方法中,被跟踪对象由2D/3D边界框表示,并通过结合激光雷达和视觉检测与多模态特征线索进行连续关联。在此基础上,采用运动状态分类的二值假设检验方法,通过融合运动和外观信息来识别物体的运动状态。随后,充分利用与动态目标相关的测量数据构建目标相关因子,并在统一的因子图中与静态特征和IMU预积分因子进行联合优化,以细化自我-车辆和被跟踪目标的轨迹。在KITTI和nuScenes的高动态序列数据集上进行的大量实验表明,LVIMOT达到了最先进的性能,MOT的HOTA为80.37,在KITTI (0.36 m, 0.02 rad)和nuScenes (0.13 m, 0.02 rad)上的自定位平均平移和旋转绝对轨迹误差(ATE)最低。结果证实了将目标感知运动约束与多模态融合相结合在提高定位鲁棒性和精度方面的有效性。源代码可从https://github.com/shqfeng/LVIMOT.git获得。
本文章由计算机程序翻译,如有差异,请以英文原文为准。
LVIMOT: Accurate and robust LiDAR-visual-inertial localization and multi-object tracking in dynamic environments via tightly coupled integration
Most existing LiDAR-visual-inertial (LVI) localization systems typically assume a static environment and often neglect the valuable dynamic object information, which limits localization accuracy and scene understanding in dynamic environments. Meanwhile, precise tracking of surrounding objects is essential for applications such as autonomous driving, augmented reality (AR), and virtual reality (VR). To this end, we propose LVIMOT, a tightly coupled LVI localization and multi-object tracking (MOT) system based on factor graph optimization, capable of jointly estimating the trajectories of the ego-vehicle and surrounding objects. In the proposed method, tracked objects are represented by 2D/3D bounding boxes and continuously associated by combining LiDAR and visual detections with multimodal feature cues. Building upon this, a binary hypothesis testing method for motion status classification is employed to identify object motion status by fusing motion and appearance information. Subsequently, the measurements associated with dynamic objects are fully exploited to construct object-related factors, which are jointly optimized with static features and IMU pre-integration factors within a unified factor graph to refine the trajectories of both the ego-vehicle and tracked objects. Extensive experiments on highly dynamic sequences from the KITTI and nuScenes datasets demonstrate that LVIMOT achieves state-of-the-art performance, with an HOTA of 80.37 for MOT, and the lowest mean translational and rotational absolute trajectory errors (ATE) for self-localization on KITTI (0.36 m, 0.02 rad) and nuScenes (0.13 m, 0.02 rad). The results confirm the effectiveness of integrating object-aware motion constraints and multimodal fusion in improving localization robustness and precision. The source code will be available at https://github.com/shqfeng/LVIMOT.git.
求助全文
通过发布文献求助,成功后即可免费获取论文全文。 去求助
来源期刊
ISPRS Journal of Photogrammetry and Remote Sensing
ISPRS Journal of Photogrammetry and Remote Sensing 工程技术-成像科学与照相技术
CiteScore
21.00
自引率
6.30%
发文量
273
审稿时长
40 days
期刊介绍: The ISPRS Journal of Photogrammetry and Remote Sensing (P&RS) serves as the official journal of the International Society for Photogrammetry and Remote Sensing (ISPRS). It acts as a platform for scientists and professionals worldwide who are involved in various disciplines that utilize photogrammetry, remote sensing, spatial information systems, computer vision, and related fields. The journal aims to facilitate communication and dissemination of advancements in these disciplines, while also acting as a comprehensive source of reference and archive. P&RS endeavors to publish high-quality, peer-reviewed research papers that are preferably original and have not been published before. These papers can cover scientific/research, technological development, or application/practical aspects. Additionally, the journal welcomes papers that are based on presentations from ISPRS meetings, as long as they are considered significant contributions to the aforementioned fields. In particular, P&RS encourages the submission of papers that are of broad scientific interest, showcase innovative applications (especially in emerging fields), have an interdisciplinary focus, discuss topics that have received limited attention in P&RS or related journals, or explore new directions in scientific or professional realms. It is preferred that theoretical papers include practical applications, while papers focusing on systems and applications should include a theoretical background.
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
确定
请完成安全验证×
copy
已复制链接
快去分享给好友吧!
我知道了
右上角分享
点击右上角分享
0
联系我们:info@booksci.cn Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。 Copyright © 2023 布克学术 All rights reserved.
京ICP备2023020795号-1
ghs 京公网安备 11010802042870号
Book学术文献互助
Book学术文献互助群
群 号:604180095
Book学术官方微信