实现空中协作立体:无人飞行器的实时跨相机特征关联和相对姿态估计

IF 4.3 2区 综合性期刊 Q1 ENGINEERING, ELECTRICAL & ELECTRONIC
Zhaoying Wang;Wei Dong
{"title":"实现空中协作立体:无人飞行器的实时跨相机特征关联和相对姿态估计","authors":"Zhaoying Wang;Wei Dong","doi":"10.1109/JSEN.2025.3532699","DOIUrl":null,"url":null,"abstract":"The collaborative unmanned aerial vehicles (UAVs) can construct a wide and variable baseline stereo camera, providing potential benefits in flexible and large-scale depth perception. Compared to the fixed-baseline stereo camera, the collaborative stereo system in dynamic flight face the additional challenge of continuously varying baselines, which necessitates real-time cross-camera stereo feature association and real-time relative pose estimation on resource-constrained onboard computers. To tackle these challenges, we propose a real-time dual-channel feature association with a guidance-prediction framework. This framework utilizes a graph neural network (GNN) to periodically guide cross-camera feature associations between two UAVs, while the guided features are continuously predicted to enable real-time feature association. Additionally, we propose a real-time relative multistate-constrained Kalman filter (Rel-MSCKF) algorithm, which efficiently integrates covisual overlapping features with the UAVs’ visual-inertial odometry (VIO), enabling accurate and fast relative pose estimation. Extensive real-world experiments are performed using the widely adopted but resource-constrained NVIDIA NX onboard computer. The results demonstrate that the dual-channel algorithm achieves cross-camera feature association for 30 Hz image streams, with an average runtime of 14 ms, significantly outperforming the conventional cross-camera feature matching algorithms, which typically require over 70 ms. For the relative pose estimation of two cameras, the proposed Rel-MSCKF algorithm achieves pose estimation within 16 ms, outperforming the current pose-graph optimization (PGO) manner with 330 ms. Additionally, we examine the convergence behavior of Rel-MSCKF under various spatial configurations. The system’s robustness is further evaluated under the challenges of asynchronous image acquisition, communication interruptions, and field-of-view (FOV) occlusions. Online video: <uri>https://youtu.be/avxMuOf5Qcw</uri>","PeriodicalId":447,"journal":{"name":"IEEE Sensors Journal","volume":"25 6","pages":"9861-9875"},"PeriodicalIF":4.3000,"publicationDate":"2025-01-28","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":"{\"title\":\"Toward Aerial Collaborative Stereo: Real-Time Cross-Camera Feature Association and Relative Pose Estimation for UAVs\",\"authors\":\"Zhaoying Wang;Wei Dong\",\"doi\":\"10.1109/JSEN.2025.3532699\",\"DOIUrl\":null,\"url\":null,\"abstract\":\"The collaborative unmanned aerial vehicles (UAVs) can construct a wide and variable baseline stereo camera, providing potential benefits in flexible and large-scale depth perception. Compared to the fixed-baseline stereo camera, the collaborative stereo system in dynamic flight face the additional challenge of continuously varying baselines, which necessitates real-time cross-camera stereo feature association and real-time relative pose estimation on resource-constrained onboard computers. To tackle these challenges, we propose a real-time dual-channel feature association with a guidance-prediction framework. This framework utilizes a graph neural network (GNN) to periodically guide cross-camera feature associations between two UAVs, while the guided features are continuously predicted to enable real-time feature association. Additionally, we propose a real-time relative multistate-constrained Kalman filter (Rel-MSCKF) algorithm, which efficiently integrates covisual overlapping features with the UAVs’ visual-inertial odometry (VIO), enabling accurate and fast relative pose estimation. Extensive real-world experiments are performed using the widely adopted but resource-constrained NVIDIA NX onboard computer. The results demonstrate that the dual-channel algorithm achieves cross-camera feature association for 30 Hz image streams, with an average runtime of 14 ms, significantly outperforming the conventional cross-camera feature matching algorithms, which typically require over 70 ms. For the relative pose estimation of two cameras, the proposed Rel-MSCKF algorithm achieves pose estimation within 16 ms, outperforming the current pose-graph optimization (PGO) manner with 330 ms. Additionally, we examine the convergence behavior of Rel-MSCKF under various spatial configurations. The system’s robustness is further evaluated under the challenges of asynchronous image acquisition, communication interruptions, and field-of-view (FOV) occlusions. Online video: <uri>https://youtu.be/avxMuOf5Qcw</uri>\",\"PeriodicalId\":447,\"journal\":{\"name\":\"IEEE Sensors Journal\",\"volume\":\"25 6\",\"pages\":\"9861-9875\"},\"PeriodicalIF\":4.3000,\"publicationDate\":\"2025-01-28\",\"publicationTypes\":\"Journal Article\",\"fieldsOfStudy\":null,\"isOpenAccess\":false,\"openAccessPdf\":\"\",\"citationCount\":\"0\",\"resultStr\":null,\"platform\":\"Semanticscholar\",\"paperid\":null,\"PeriodicalName\":\"IEEE Sensors Journal\",\"FirstCategoryId\":\"103\",\"ListUrlMain\":\"https://ieeexplore.ieee.org/document/10856812/\",\"RegionNum\":2,\"RegionCategory\":\"综合性期刊\",\"ArticlePicture\":[],\"TitleCN\":null,\"AbstractTextCN\":null,\"PMCID\":null,\"EPubDate\":\"\",\"PubModel\":\"\",\"JCR\":\"Q1\",\"JCRName\":\"ENGINEERING, ELECTRICAL & ELECTRONIC\",\"Score\":null,\"Total\":0}","platform":"Semanticscholar","paperid":null,"PeriodicalName":"IEEE Sensors Journal","FirstCategoryId":"103","ListUrlMain":"https://ieeexplore.ieee.org/document/10856812/","RegionNum":2,"RegionCategory":"综合性期刊","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"Q1","JCRName":"ENGINEERING, ELECTRICAL & ELECTRONIC","Score":null,"Total":0}
引用次数: 0

摘要

协同无人机可以构建宽、可变基线立体摄像机,为灵活、大规模的深度感知提供了潜在的优势。与固定基线立体相机相比,动态飞行协同立体系统面临着持续变化基线的额外挑战,这就需要在资源受限的机载计算机上实现实时跨相机立体特征关联和实时相对姿态估计。为了解决这些挑战,我们提出了一种实时双通道特征与指导预测框架的关联。该框架利用图神经网络(GNN)周期性地引导两架无人机之间的跨摄像头特征关联,同时连续预测引导特征以实现实时特征关联。此外,我们提出了一种实时相对多状态约束卡尔曼滤波(Rel-MSCKF)算法,该算法有效地将共视重叠特征与无人机的视觉惯性测距(VIO)相结合,实现了准确、快速的相对姿态估计。使用广泛采用但资源受限的NVIDIA NX板载计算机进行了广泛的实际实验。结果表明,双通道算法实现了30 Hz图像流的跨摄像头特征关联,平均运行时间为14 ms,显著优于传统的跨摄像头特征匹配算法,后者通常需要70 ms以上的时间。对于两台相机的相对姿态估计,本文提出的Rel-MSCKF算法在16 ms内实现了姿态估计,超过了目前的姿态图优化(PGO)方法330 ms。此外,我们还研究了Rel-MSCKF在不同空间构型下的收敛行为。在异步图像采集、通信中断和视场(FOV)遮挡的挑战下,进一步评估了系统的鲁棒性。在线视频:https://youtu.be/avxMuOf5Qcw
本文章由计算机程序翻译,如有差异,请以英文原文为准。
Toward Aerial Collaborative Stereo: Real-Time Cross-Camera Feature Association and Relative Pose Estimation for UAVs
The collaborative unmanned aerial vehicles (UAVs) can construct a wide and variable baseline stereo camera, providing potential benefits in flexible and large-scale depth perception. Compared to the fixed-baseline stereo camera, the collaborative stereo system in dynamic flight face the additional challenge of continuously varying baselines, which necessitates real-time cross-camera stereo feature association and real-time relative pose estimation on resource-constrained onboard computers. To tackle these challenges, we propose a real-time dual-channel feature association with a guidance-prediction framework. This framework utilizes a graph neural network (GNN) to periodically guide cross-camera feature associations between two UAVs, while the guided features are continuously predicted to enable real-time feature association. Additionally, we propose a real-time relative multistate-constrained Kalman filter (Rel-MSCKF) algorithm, which efficiently integrates covisual overlapping features with the UAVs’ visual-inertial odometry (VIO), enabling accurate and fast relative pose estimation. Extensive real-world experiments are performed using the widely adopted but resource-constrained NVIDIA NX onboard computer. The results demonstrate that the dual-channel algorithm achieves cross-camera feature association for 30 Hz image streams, with an average runtime of 14 ms, significantly outperforming the conventional cross-camera feature matching algorithms, which typically require over 70 ms. For the relative pose estimation of two cameras, the proposed Rel-MSCKF algorithm achieves pose estimation within 16 ms, outperforming the current pose-graph optimization (PGO) manner with 330 ms. Additionally, we examine the convergence behavior of Rel-MSCKF under various spatial configurations. The system’s robustness is further evaluated under the challenges of asynchronous image acquisition, communication interruptions, and field-of-view (FOV) occlusions. Online video: https://youtu.be/avxMuOf5Qcw
求助全文
通过发布文献求助,成功后即可免费获取论文全文。 去求助
来源期刊
IEEE Sensors Journal
IEEE Sensors Journal 工程技术-工程:电子与电气
CiteScore
7.70
自引率
14.00%
发文量
2058
审稿时长
5.2 months
期刊介绍: The fields of interest of the IEEE Sensors Journal are the theory, design , fabrication, manufacturing and applications of devices for sensing and transducing physical, chemical and biological phenomena, with emphasis on the electronics and physics aspect of sensors and integrated sensors-actuators. IEEE Sensors Journal deals with the following: -Sensor Phenomenology, Modelling, and Evaluation -Sensor Materials, Processing, and Fabrication -Chemical and Gas Sensors -Microfluidics and Biosensors -Optical Sensors -Physical Sensors: Temperature, Mechanical, Magnetic, and others -Acoustic and Ultrasonic Sensors -Sensor Packaging -Sensor Networks -Sensor Applications -Sensor Systems: Signals, Processing, and Interfaces -Actuators and Sensor Power Systems -Sensor Signal Processing for high precision and stability (amplification, filtering, linearization, modulation/demodulation) and under harsh conditions (EMC, radiation, humidity, temperature); energy consumption/harvesting -Sensor Data Processing (soft computing with sensor data, e.g., pattern recognition, machine learning, evolutionary computation; sensor data fusion, processing of wave e.g., electromagnetic and acoustic; and non-wave, e.g., chemical, gravity, particle, thermal, radiative and non-radiative sensor data, detection, estimation and classification based on sensor data) -Sensors in Industrial Practice
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
确定
请完成安全验证×
copy
已复制链接
快去分享给好友吧!
我知道了
右上角分享
点击右上角分享
0
联系我们:info@booksci.cn Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。 Copyright © 2023 布克学术 All rights reserved.
京ICP备2023020795号-1
ghs 京公网安备 11010802042870号
Book学术文献互助
Book学术文献互助群
群 号:604180095
Book学术官方微信