{"title":"实现空中协作立体:无人飞行器的实时跨相机特征关联和相对姿态估计","authors":"Zhaoying Wang;Wei Dong","doi":"10.1109/JSEN.2025.3532699","DOIUrl":null,"url":null,"abstract":"The collaborative unmanned aerial vehicles (UAVs) can construct a wide and variable baseline stereo camera, providing potential benefits in flexible and large-scale depth perception. Compared to the fixed-baseline stereo camera, the collaborative stereo system in dynamic flight face the additional challenge of continuously varying baselines, which necessitates real-time cross-camera stereo feature association and real-time relative pose estimation on resource-constrained onboard computers. To tackle these challenges, we propose a real-time dual-channel feature association with a guidance-prediction framework. This framework utilizes a graph neural network (GNN) to periodically guide cross-camera feature associations between two UAVs, while the guided features are continuously predicted to enable real-time feature association. Additionally, we propose a real-time relative multistate-constrained Kalman filter (Rel-MSCKF) algorithm, which efficiently integrates covisual overlapping features with the UAVs’ visual-inertial odometry (VIO), enabling accurate and fast relative pose estimation. Extensive real-world experiments are performed using the widely adopted but resource-constrained NVIDIA NX onboard computer. The results demonstrate that the dual-channel algorithm achieves cross-camera feature association for 30 Hz image streams, with an average runtime of 14 ms, significantly outperforming the conventional cross-camera feature matching algorithms, which typically require over 70 ms. For the relative pose estimation of two cameras, the proposed Rel-MSCKF algorithm achieves pose estimation within 16 ms, outperforming the current pose-graph optimization (PGO) manner with 330 ms. Additionally, we examine the convergence behavior of Rel-MSCKF under various spatial configurations. The system’s robustness is further evaluated under the challenges of asynchronous image acquisition, communication interruptions, and field-of-view (FOV) occlusions. Online video: <uri>https://youtu.be/avxMuOf5Qcw</uri>","PeriodicalId":447,"journal":{"name":"IEEE Sensors Journal","volume":"25 6","pages":"9861-9875"},"PeriodicalIF":4.3000,"publicationDate":"2025-01-28","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":"{\"title\":\"Toward Aerial Collaborative Stereo: Real-Time Cross-Camera Feature Association and Relative Pose Estimation for UAVs\",\"authors\":\"Zhaoying Wang;Wei Dong\",\"doi\":\"10.1109/JSEN.2025.3532699\",\"DOIUrl\":null,\"url\":null,\"abstract\":\"The collaborative unmanned aerial vehicles (UAVs) can construct a wide and variable baseline stereo camera, providing potential benefits in flexible and large-scale depth perception. Compared to the fixed-baseline stereo camera, the collaborative stereo system in dynamic flight face the additional challenge of continuously varying baselines, which necessitates real-time cross-camera stereo feature association and real-time relative pose estimation on resource-constrained onboard computers. To tackle these challenges, we propose a real-time dual-channel feature association with a guidance-prediction framework. This framework utilizes a graph neural network (GNN) to periodically guide cross-camera feature associations between two UAVs, while the guided features are continuously predicted to enable real-time feature association. Additionally, we propose a real-time relative multistate-constrained Kalman filter (Rel-MSCKF) algorithm, which efficiently integrates covisual overlapping features with the UAVs’ visual-inertial odometry (VIO), enabling accurate and fast relative pose estimation. Extensive real-world experiments are performed using the widely adopted but resource-constrained NVIDIA NX onboard computer. The results demonstrate that the dual-channel algorithm achieves cross-camera feature association for 30 Hz image streams, with an average runtime of 14 ms, significantly outperforming the conventional cross-camera feature matching algorithms, which typically require over 70 ms. For the relative pose estimation of two cameras, the proposed Rel-MSCKF algorithm achieves pose estimation within 16 ms, outperforming the current pose-graph optimization (PGO) manner with 330 ms. Additionally, we examine the convergence behavior of Rel-MSCKF under various spatial configurations. The system’s robustness is further evaluated under the challenges of asynchronous image acquisition, communication interruptions, and field-of-view (FOV) occlusions. Online video: <uri>https://youtu.be/avxMuOf5Qcw</uri>\",\"PeriodicalId\":447,\"journal\":{\"name\":\"IEEE Sensors Journal\",\"volume\":\"25 6\",\"pages\":\"9861-9875\"},\"PeriodicalIF\":4.3000,\"publicationDate\":\"2025-01-28\",\"publicationTypes\":\"Journal Article\",\"fieldsOfStudy\":null,\"isOpenAccess\":false,\"openAccessPdf\":\"\",\"citationCount\":\"0\",\"resultStr\":null,\"platform\":\"Semanticscholar\",\"paperid\":null,\"PeriodicalName\":\"IEEE Sensors Journal\",\"FirstCategoryId\":\"103\",\"ListUrlMain\":\"https://ieeexplore.ieee.org/document/10856812/\",\"RegionNum\":2,\"RegionCategory\":\"综合性期刊\",\"ArticlePicture\":[],\"TitleCN\":null,\"AbstractTextCN\":null,\"PMCID\":null,\"EPubDate\":\"\",\"PubModel\":\"\",\"JCR\":\"Q1\",\"JCRName\":\"ENGINEERING, ELECTRICAL & ELECTRONIC\",\"Score\":null,\"Total\":0}","platform":"Semanticscholar","paperid":null,"PeriodicalName":"IEEE Sensors Journal","FirstCategoryId":"103","ListUrlMain":"https://ieeexplore.ieee.org/document/10856812/","RegionNum":2,"RegionCategory":"综合性期刊","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"Q1","JCRName":"ENGINEERING, ELECTRICAL & ELECTRONIC","Score":null,"Total":0}
Toward Aerial Collaborative Stereo: Real-Time Cross-Camera Feature Association and Relative Pose Estimation for UAVs
The collaborative unmanned aerial vehicles (UAVs) can construct a wide and variable baseline stereo camera, providing potential benefits in flexible and large-scale depth perception. Compared to the fixed-baseline stereo camera, the collaborative stereo system in dynamic flight face the additional challenge of continuously varying baselines, which necessitates real-time cross-camera stereo feature association and real-time relative pose estimation on resource-constrained onboard computers. To tackle these challenges, we propose a real-time dual-channel feature association with a guidance-prediction framework. This framework utilizes a graph neural network (GNN) to periodically guide cross-camera feature associations between two UAVs, while the guided features are continuously predicted to enable real-time feature association. Additionally, we propose a real-time relative multistate-constrained Kalman filter (Rel-MSCKF) algorithm, which efficiently integrates covisual overlapping features with the UAVs’ visual-inertial odometry (VIO), enabling accurate and fast relative pose estimation. Extensive real-world experiments are performed using the widely adopted but resource-constrained NVIDIA NX onboard computer. The results demonstrate that the dual-channel algorithm achieves cross-camera feature association for 30 Hz image streams, with an average runtime of 14 ms, significantly outperforming the conventional cross-camera feature matching algorithms, which typically require over 70 ms. For the relative pose estimation of two cameras, the proposed Rel-MSCKF algorithm achieves pose estimation within 16 ms, outperforming the current pose-graph optimization (PGO) manner with 330 ms. Additionally, we examine the convergence behavior of Rel-MSCKF under various spatial configurations. The system’s robustness is further evaluated under the challenges of asynchronous image acquisition, communication interruptions, and field-of-view (FOV) occlusions. Online video: https://youtu.be/avxMuOf5Qcw
期刊介绍:
The fields of interest of the IEEE Sensors Journal are the theory, design , fabrication, manufacturing and applications of devices for sensing and transducing physical, chemical and biological phenomena, with emphasis on the electronics and physics aspect of sensors and integrated sensors-actuators. IEEE Sensors Journal deals with the following:
-Sensor Phenomenology, Modelling, and Evaluation
-Sensor Materials, Processing, and Fabrication
-Chemical and Gas Sensors
-Microfluidics and Biosensors
-Optical Sensors
-Physical Sensors: Temperature, Mechanical, Magnetic, and others
-Acoustic and Ultrasonic Sensors
-Sensor Packaging
-Sensor Networks
-Sensor Applications
-Sensor Systems: Signals, Processing, and Interfaces
-Actuators and Sensor Power Systems
-Sensor Signal Processing for high precision and stability (amplification, filtering, linearization, modulation/demodulation) and under harsh conditions (EMC, radiation, humidity, temperature); energy consumption/harvesting
-Sensor Data Processing (soft computing with sensor data, e.g., pattern recognition, machine learning, evolutionary computation; sensor data fusion, processing of wave e.g., electromagnetic and acoustic; and non-wave, e.g., chemical, gravity, particle, thermal, radiative and non-radiative sensor data, detection, estimation and classification based on sensor data)
-Sensors in Industrial Practice