{"title":"Learning-based monocular visual-inertial odometry with \n \n \n \n S\n \n E\n 2\n \n \n (\n 3\n )\n \n \n \n $S{E}_{2}(3)$\n -EKF","authors":"Chi Guo, Jianlang Hu, Yarong Luo","doi":"10.1002/rob.22349","DOIUrl":null,"url":null,"abstract":"<p>Learning-based visual odometry (VO) becomes popular as it achieves a remarkable performance without manually crafted image processing and burdensome calibration. Meanwhile, the inertial navigation can provide a localization solution to assist VO when the VO produces poor state estimation under challenging visual conditions. Therefore, the combination of learning-based technique and classical state estimation method can further improve the performance of pose estimation. In this paper, we propose a learning-based visual-inertial odometry (VIO) algorithm, which consists of an end-to-end VO network and an <span></span><math>\n <semantics>\n <mrow>\n \n <mrow>\n <mi>S</mi>\n \n <msub>\n <mi>E</mi>\n \n <mn>2</mn>\n </msub>\n \n <mrow>\n <mo>(</mo>\n \n <mn>3</mn>\n \n <mo>)</mo>\n </mrow>\n </mrow>\n </mrow>\n <annotation> $S{E}_{2}(3)$</annotation>\n </semantics></math>-Extended Kalman Filter (EKF). The VO network mainly combines a convolutional neural network with a recurrent neural network, taking advantage of two consecutive monocular images to produce relative pose estimation with associated uncertainties. The <span></span><math>\n <semantics>\n <mrow>\n \n <mrow>\n <mi>S</mi>\n \n <msub>\n <mi>E</mi>\n \n <mn>2</mn>\n </msub>\n \n <mrow>\n <mo>(</mo>\n \n <mn>3</mn>\n \n <mo>)</mo>\n </mrow>\n </mrow>\n </mrow>\n <annotation> $S{E}_{2}(3)$</annotation>\n </semantics></math>-EKF, which is proved to overcome the inconsistency issues of VIO, propagates inertial measurement unit kinematics-based states, and fuses relative measurements and uncertainties from the VO network in its update step. The extensive experimental results on the KITTI data set and the EuRoC data set demonstrate the superior performance of the proposed method compared to other related methods.</p>","PeriodicalId":192,"journal":{"name":"Journal of Field Robotics","volume":"41 6","pages":"1780-1796"},"PeriodicalIF":4.2000,"publicationDate":"2024-04-24","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"Journal of Field Robotics","FirstCategoryId":"94","ListUrlMain":"https://onlinelibrary.wiley.com/doi/10.1002/rob.22349","RegionNum":2,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"Q2","JCRName":"ROBOTICS","Score":null,"Total":0}
引用次数: 0
Abstract
Learning-based visual odometry (VO) becomes popular as it achieves a remarkable performance without manually crafted image processing and burdensome calibration. Meanwhile, the inertial navigation can provide a localization solution to assist VO when the VO produces poor state estimation under challenging visual conditions. Therefore, the combination of learning-based technique and classical state estimation method can further improve the performance of pose estimation. In this paper, we propose a learning-based visual-inertial odometry (VIO) algorithm, which consists of an end-to-end VO network and an -Extended Kalman Filter (EKF). The VO network mainly combines a convolutional neural network with a recurrent neural network, taking advantage of two consecutive monocular images to produce relative pose estimation with associated uncertainties. The -EKF, which is proved to overcome the inconsistency issues of VIO, propagates inertial measurement unit kinematics-based states, and fuses relative measurements and uncertainties from the VO network in its update step. The extensive experimental results on the KITTI data set and the EuRoC data set demonstrate the superior performance of the proposed method compared to other related methods.
期刊介绍:
The Journal of Field Robotics seeks to promote scholarly publications dealing with the fundamentals of robotics in unstructured and dynamic environments.
The Journal focuses on experimental robotics and encourages publication of work that has both theoretical and practical significance.