Jun Lai, Suyang Liu, Xiaojia Xiang, Chaoran Li, Dengqing Tang, Han Zhou
{"title":"Performance Analysis of Visual–Inertial–Range Cooperative Localization for Unmanned Autonomous Vehicle Swarm","authors":"Jun Lai, Suyang Liu, Xiaojia Xiang, Chaoran Li, Dengqing Tang, Han Zhou","doi":"10.3390/drones7110651","DOIUrl":null,"url":null,"abstract":"The swarm of small UAVs is an emerging technology that will enable abundant cooperative tasks. To tackle the positioning problem for the UAV swarm, cooperative localization (CL) has been intensively studied since it uses relative measurement to improve the positioning availability and accuracy for the swarm in GPS-denied environments. Besides relying on inter-UAV range measurement, traditional CL algorithms need to place anchors as location references, which limits their applicability. To implement an infrastructure-less swarm navigation system, a consumer-grade camera together with an inertial device can provide rich environment information, which can be recognized as a kind of local location reference. This paper aims to analyze the fundamental performance of visual–inertial–range CL, which is also a popular metric for UAV planning and sensing optimizing, especially for resource-limited environments. Specifically, a closed-form Fisher information matrix (FIM) of visual–inertial–range CL is constructed in Rn×SO(n) manifold. By introducing an equivalent FIM and utilizing of the sparsity of the FIM, the performance of pose estimation can be efficiently calculated. A series of numerical simulations validate its effectiveness for analyzing the CL performance.","PeriodicalId":36448,"journal":{"name":"Drones","volume":"1 1","pages":"0"},"PeriodicalIF":4.4000,"publicationDate":"2023-10-26","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"1","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"Drones","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/10.3390/drones7110651","RegionNum":2,"RegionCategory":"地球科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"Q1","JCRName":"REMOTE SENSING","Score":null,"Total":0}
引用次数: 1
Abstract
The swarm of small UAVs is an emerging technology that will enable abundant cooperative tasks. To tackle the positioning problem for the UAV swarm, cooperative localization (CL) has been intensively studied since it uses relative measurement to improve the positioning availability and accuracy for the swarm in GPS-denied environments. Besides relying on inter-UAV range measurement, traditional CL algorithms need to place anchors as location references, which limits their applicability. To implement an infrastructure-less swarm navigation system, a consumer-grade camera together with an inertial device can provide rich environment information, which can be recognized as a kind of local location reference. This paper aims to analyze the fundamental performance of visual–inertial–range CL, which is also a popular metric for UAV planning and sensing optimizing, especially for resource-limited environments. Specifically, a closed-form Fisher information matrix (FIM) of visual–inertial–range CL is constructed in Rn×SO(n) manifold. By introducing an equivalent FIM and utilizing of the sparsity of the FIM, the performance of pose estimation can be efficiently calculated. A series of numerical simulations validate its effectiveness for analyzing the CL performance.