Jee-seong Kim, Chul-hong Kim, Yong-Min Shin, Ilsoo Cho, D. Cho
{"title":"基于深度图和语义分割的室外单目视觉里程增强","authors":"Jee-seong Kim, Chul-hong Kim, Yong-Min Shin, Ilsoo Cho, D. Cho","doi":"10.23919/ICCAS50221.2020.9268347","DOIUrl":null,"url":null,"abstract":"An outdoor environment is challenging for the localization of a mobile robot. For robust visual odometry, accurate feature matching and triangulation are essential. The features extracted from the windows of buildings and car surfaces lead to wrong triangulation results due to reflective features. The landmarks at short-distances affect the feature matching performance and the landmarks at long-distances cause triangulation errors. Inaccurate feature matching and triangulation error lead to the localization error of the robot pose. In this paper, an outdoor monocular visual odometry using the pre-trained depth estimation network and semantic segmentation network is proposed. By using the pre-trained semantic segmentation network, a semantic label is predicted for every pixel. Also, by using the pre-trained depth map estimation network, the depth of every pixel is predicted. Using semantic constraints for feature matching and depth constraint for triangulation, the accuracy of these procedures is enhanced. Additionally, pose graph optimization is performed on every estimated robot pose and landmark position. The performance of the proposed method is evaluated using dataset-based experiments. The experiments showed that the proposed algorithm is more accurate than the visual odometry algorithm that uses Oriented FAST and rotated BRIEF (ORB) features.","PeriodicalId":6732,"journal":{"name":"2020 20th International Conference on Control, Automation and Systems (ICCAS)","volume":"171 1","pages":"1040-1045"},"PeriodicalIF":0.0000,"publicationDate":"2020-10-13","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":"{\"title\":\"Outdoor Monocular Visual Odometry Enhancement Using Depth Map and Semantic Segmentation\",\"authors\":\"Jee-seong Kim, Chul-hong Kim, Yong-Min Shin, Ilsoo Cho, D. Cho\",\"doi\":\"10.23919/ICCAS50221.2020.9268347\",\"DOIUrl\":null,\"url\":null,\"abstract\":\"An outdoor environment is challenging for the localization of a mobile robot. For robust visual odometry, accurate feature matching and triangulation are essential. The features extracted from the windows of buildings and car surfaces lead to wrong triangulation results due to reflective features. The landmarks at short-distances affect the feature matching performance and the landmarks at long-distances cause triangulation errors. Inaccurate feature matching and triangulation error lead to the localization error of the robot pose. In this paper, an outdoor monocular visual odometry using the pre-trained depth estimation network and semantic segmentation network is proposed. By using the pre-trained semantic segmentation network, a semantic label is predicted for every pixel. Also, by using the pre-trained depth map estimation network, the depth of every pixel is predicted. Using semantic constraints for feature matching and depth constraint for triangulation, the accuracy of these procedures is enhanced. Additionally, pose graph optimization is performed on every estimated robot pose and landmark position. The performance of the proposed method is evaluated using dataset-based experiments. The experiments showed that the proposed algorithm is more accurate than the visual odometry algorithm that uses Oriented FAST and rotated BRIEF (ORB) features.\",\"PeriodicalId\":6732,\"journal\":{\"name\":\"2020 20th International Conference on Control, Automation and Systems (ICCAS)\",\"volume\":\"171 1\",\"pages\":\"1040-1045\"},\"PeriodicalIF\":0.0000,\"publicationDate\":\"2020-10-13\",\"publicationTypes\":\"Journal Article\",\"fieldsOfStudy\":null,\"isOpenAccess\":false,\"openAccessPdf\":\"\",\"citationCount\":\"0\",\"resultStr\":null,\"platform\":\"Semanticscholar\",\"paperid\":null,\"PeriodicalName\":\"2020 20th International Conference on Control, Automation and Systems (ICCAS)\",\"FirstCategoryId\":\"1085\",\"ListUrlMain\":\"https://doi.org/10.23919/ICCAS50221.2020.9268347\",\"RegionNum\":0,\"RegionCategory\":null,\"ArticlePicture\":[],\"TitleCN\":null,\"AbstractTextCN\":null,\"PMCID\":null,\"EPubDate\":\"\",\"PubModel\":\"\",\"JCR\":\"\",\"JCRName\":\"\",\"Score\":null,\"Total\":0}","platform":"Semanticscholar","paperid":null,"PeriodicalName":"2020 20th International Conference on Control, Automation and Systems (ICCAS)","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/10.23919/ICCAS50221.2020.9268347","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
Outdoor Monocular Visual Odometry Enhancement Using Depth Map and Semantic Segmentation
An outdoor environment is challenging for the localization of a mobile robot. For robust visual odometry, accurate feature matching and triangulation are essential. The features extracted from the windows of buildings and car surfaces lead to wrong triangulation results due to reflective features. The landmarks at short-distances affect the feature matching performance and the landmarks at long-distances cause triangulation errors. Inaccurate feature matching and triangulation error lead to the localization error of the robot pose. In this paper, an outdoor monocular visual odometry using the pre-trained depth estimation network and semantic segmentation network is proposed. By using the pre-trained semantic segmentation network, a semantic label is predicted for every pixel. Also, by using the pre-trained depth map estimation network, the depth of every pixel is predicted. Using semantic constraints for feature matching and depth constraint for triangulation, the accuracy of these procedures is enhanced. Additionally, pose graph optimization is performed on every estimated robot pose and landmark position. The performance of the proposed method is evaluated using dataset-based experiments. The experiments showed that the proposed algorithm is more accurate than the visual odometry algorithm that uses Oriented FAST and rotated BRIEF (ORB) features.