Bo Chen, Kun Yan, Rongchuan Cao, Tianqi Zhang, Xiaoli Zhang
{"title":"A new visual odometry algorithm based on multi-path deep fully convolutional neural networks","authors":"Bo Chen, Kun Yan, Rongchuan Cao, Tianqi Zhang, Xiaoli Zhang","doi":"10.1117/12.2653846","DOIUrl":null,"url":null,"abstract":"Visual odometry is one of the key core technologies in the field of autonomous driving. However, images captured in lowlight or unevenly-illuminated scenes still cannot guarantee good performance due to low image contrast and lack of detail features. Therefore, we propose an end-to-end visual odometry method based on image fusion and FCNN-LSTM in the paper. The brightness image of the source image sequence is obtained by gray-scale transformation, and an image fusion algorithm based on spectral residual theory is designed to combine the image sequence and its brightness image to enhance the contrast of the image and provide more detailed information. In order to improve the accuracy of image feature extraction and reduce the error in the pose estimation process, we design a feature extraction algorithm based on skipfusion-FCNN. The traditional fully convolutional neural network (FCNN) is improved, a skip-fusion-FCNN network model is proposed, and three different paths are constructed for feature extraction. In each path, the prediction results of different depths are fused by downsampling to obtain a feature map. Merge three different feature maps to obtain feature fusion information, taking into account the structural information and detail information of the image. Experiments show that this algorithm is superior to the state-of-the-art algorithms.","PeriodicalId":32903,"journal":{"name":"JITeCS Journal of Information Technology and Computer Science","volume":null,"pages":null},"PeriodicalIF":0.0000,"publicationDate":"2022-12-08","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"JITeCS Journal of Information Technology and Computer Science","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/10.1117/12.2653846","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
引用次数: 0
Abstract
Visual odometry is one of the key core technologies in the field of autonomous driving. However, images captured in lowlight or unevenly-illuminated scenes still cannot guarantee good performance due to low image contrast and lack of detail features. Therefore, we propose an end-to-end visual odometry method based on image fusion and FCNN-LSTM in the paper. The brightness image of the source image sequence is obtained by gray-scale transformation, and an image fusion algorithm based on spectral residual theory is designed to combine the image sequence and its brightness image to enhance the contrast of the image and provide more detailed information. In order to improve the accuracy of image feature extraction and reduce the error in the pose estimation process, we design a feature extraction algorithm based on skipfusion-FCNN. The traditional fully convolutional neural network (FCNN) is improved, a skip-fusion-FCNN network model is proposed, and three different paths are constructed for feature extraction. In each path, the prediction results of different depths are fused by downsampling to obtain a feature map. Merge three different feature maps to obtain feature fusion information, taking into account the structural information and detail information of the image. Experiments show that this algorithm is superior to the state-of-the-art algorithms.