基于深度图和语义分割的室外单目视觉里程增强

Jee-seong Kim, Chul-hong Kim, Yong-Min Shin, Ilsoo Cho, D. Cho
{"title":"基于深度图和语义分割的室外单目视觉里程增强","authors":"Jee-seong Kim, Chul-hong Kim, Yong-Min Shin, Ilsoo Cho, D. Cho","doi":"10.23919/ICCAS50221.2020.9268347","DOIUrl":null,"url":null,"abstract":"An outdoor environment is challenging for the localization of a mobile robot. For robust visual odometry, accurate feature matching and triangulation are essential. The features extracted from the windows of buildings and car surfaces lead to wrong triangulation results due to reflective features. The landmarks at short-distances affect the feature matching performance and the landmarks at long-distances cause triangulation errors. Inaccurate feature matching and triangulation error lead to the localization error of the robot pose. In this paper, an outdoor monocular visual odometry using the pre-trained depth estimation network and semantic segmentation network is proposed. By using the pre-trained semantic segmentation network, a semantic label is predicted for every pixel. Also, by using the pre-trained depth map estimation network, the depth of every pixel is predicted. Using semantic constraints for feature matching and depth constraint for triangulation, the accuracy of these procedures is enhanced. Additionally, pose graph optimization is performed on every estimated robot pose and landmark position. The performance of the proposed method is evaluated using dataset-based experiments. The experiments showed that the proposed algorithm is more accurate than the visual odometry algorithm that uses Oriented FAST and rotated BRIEF (ORB) features.","PeriodicalId":6732,"journal":{"name":"2020 20th International Conference on Control, Automation and Systems (ICCAS)","volume":"171 1","pages":"1040-1045"},"PeriodicalIF":0.0000,"publicationDate":"2020-10-13","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":"{\"title\":\"Outdoor Monocular Visual Odometry Enhancement Using Depth Map and Semantic Segmentation\",\"authors\":\"Jee-seong Kim, Chul-hong Kim, Yong-Min Shin, Ilsoo Cho, D. Cho\",\"doi\":\"10.23919/ICCAS50221.2020.9268347\",\"DOIUrl\":null,\"url\":null,\"abstract\":\"An outdoor environment is challenging for the localization of a mobile robot. For robust visual odometry, accurate feature matching and triangulation are essential. The features extracted from the windows of buildings and car surfaces lead to wrong triangulation results due to reflective features. The landmarks at short-distances affect the feature matching performance and the landmarks at long-distances cause triangulation errors. Inaccurate feature matching and triangulation error lead to the localization error of the robot pose. In this paper, an outdoor monocular visual odometry using the pre-trained depth estimation network and semantic segmentation network is proposed. By using the pre-trained semantic segmentation network, a semantic label is predicted for every pixel. Also, by using the pre-trained depth map estimation network, the depth of every pixel is predicted. Using semantic constraints for feature matching and depth constraint for triangulation, the accuracy of these procedures is enhanced. Additionally, pose graph optimization is performed on every estimated robot pose and landmark position. The performance of the proposed method is evaluated using dataset-based experiments. The experiments showed that the proposed algorithm is more accurate than the visual odometry algorithm that uses Oriented FAST and rotated BRIEF (ORB) features.\",\"PeriodicalId\":6732,\"journal\":{\"name\":\"2020 20th International Conference on Control, Automation and Systems (ICCAS)\",\"volume\":\"171 1\",\"pages\":\"1040-1045\"},\"PeriodicalIF\":0.0000,\"publicationDate\":\"2020-10-13\",\"publicationTypes\":\"Journal Article\",\"fieldsOfStudy\":null,\"isOpenAccess\":false,\"openAccessPdf\":\"\",\"citationCount\":\"0\",\"resultStr\":null,\"platform\":\"Semanticscholar\",\"paperid\":null,\"PeriodicalName\":\"2020 20th International Conference on Control, Automation and Systems (ICCAS)\",\"FirstCategoryId\":\"1085\",\"ListUrlMain\":\"https://doi.org/10.23919/ICCAS50221.2020.9268347\",\"RegionNum\":0,\"RegionCategory\":null,\"ArticlePicture\":[],\"TitleCN\":null,\"AbstractTextCN\":null,\"PMCID\":null,\"EPubDate\":\"\",\"PubModel\":\"\",\"JCR\":\"\",\"JCRName\":\"\",\"Score\":null,\"Total\":0}","platform":"Semanticscholar","paperid":null,"PeriodicalName":"2020 20th International Conference on Control, Automation and Systems (ICCAS)","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/10.23919/ICCAS50221.2020.9268347","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
引用次数: 0

摘要

户外环境对移动机器人的定位具有挑战性。对于稳健的视觉里程计,精确的特征匹配和三角测量是必不可少的。从建筑物窗户和汽车表面提取的特征由于反射特征导致三角剖分结果错误。距离较近的特征点影响特征匹配性能,距离较远的特征点导致三角测量误差。特征匹配不准确和三角测量误差导致机器人姿态定位误差。本文提出了一种基于预训练深度估计网络和语义分割网络的室外单目视觉里程测量方法。利用预训练的语义分割网络,对每个像素预测一个语义标签。利用预训练的深度图估计网络,对每个像素点的深度进行预测。利用语义约束进行特征匹配,利用深度约束进行三角剖分,提高了算法的精度。此外,对每个估计的机器人姿态和地标位置进行姿态图优化。通过基于数据集的实验对该方法的性能进行了评价。实验表明,该算法比使用定向快速和旋转简短(ORB)特征的视觉里程计算法精度更高。
本文章由计算机程序翻译,如有差异,请以英文原文为准。
Outdoor Monocular Visual Odometry Enhancement Using Depth Map and Semantic Segmentation
An outdoor environment is challenging for the localization of a mobile robot. For robust visual odometry, accurate feature matching and triangulation are essential. The features extracted from the windows of buildings and car surfaces lead to wrong triangulation results due to reflective features. The landmarks at short-distances affect the feature matching performance and the landmarks at long-distances cause triangulation errors. Inaccurate feature matching and triangulation error lead to the localization error of the robot pose. In this paper, an outdoor monocular visual odometry using the pre-trained depth estimation network and semantic segmentation network is proposed. By using the pre-trained semantic segmentation network, a semantic label is predicted for every pixel. Also, by using the pre-trained depth map estimation network, the depth of every pixel is predicted. Using semantic constraints for feature matching and depth constraint for triangulation, the accuracy of these procedures is enhanced. Additionally, pose graph optimization is performed on every estimated robot pose and landmark position. The performance of the proposed method is evaluated using dataset-based experiments. The experiments showed that the proposed algorithm is more accurate than the visual odometry algorithm that uses Oriented FAST and rotated BRIEF (ORB) features.
求助全文
通过发布文献求助,成功后即可免费获取论文全文。 去求助
来源期刊
自引率
0.00%
发文量
0
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
确定
请完成安全验证×
copy
已复制链接
快去分享给好友吧!
我知道了
右上角分享
点击右上角分享
0
联系我们:info@booksci.cn Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。 Copyright © 2023 布克学术 All rights reserved.
京ICP备2023020795号-1
ghs 京公网安备 11010802042870号
Book学术文献互助
Book学术文献互助群
群 号:481959085
Book学术官方微信