{"title":"利用基于事件摄像机的视觉里程计方法和机器学习特征进行轨迹重建","authors":"S. Chiodini","doi":"10.21741/9781644902813-150","DOIUrl":null,"url":null,"abstract":"Abstract. This paper presents a machine learned feature detector targeted to event-camera based visual odometry methods for unmanned aerial vehicles trajectory reconstruction. The proposed method uses machine-learned features to enhance the accuracy of the trajectory reconstruction. Traditional visual odometry methods suffer from poor performance in low light conditions and high-speed motion. The event-camera-based approach overcomes these limitations by detecting and processing only the changes in the visual scene. The machine-learned features are crafted to capture the unique characteristics of the event-camera data, enhancing the accuracy of the trajectory reconstruction. The inference pipeline is composed of a module repeated twice in sequence, formed by a Squeeze-and-Excite block and a ConvLSTM block with residual connection; it is followed by a final convolutional layer that provides the trajectories of the corners as a sequence of heatmaps. In the experimental part, a sequence of images was collected using an event-camera in outdoor environments for training and test.","PeriodicalId":87445,"journal":{"name":"Materials Research Society symposia proceedings. Materials Research Society","volume":null,"pages":null},"PeriodicalIF":0.0000,"publicationDate":"2023-11-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":"{\"title\":\"Trajectory reconstruction by means of an event-camera-based visual odometry method and machine learned features\",\"authors\":\"S. Chiodini\",\"doi\":\"10.21741/9781644902813-150\",\"DOIUrl\":null,\"url\":null,\"abstract\":\"Abstract. This paper presents a machine learned feature detector targeted to event-camera based visual odometry methods for unmanned aerial vehicles trajectory reconstruction. The proposed method uses machine-learned features to enhance the accuracy of the trajectory reconstruction. Traditional visual odometry methods suffer from poor performance in low light conditions and high-speed motion. The event-camera-based approach overcomes these limitations by detecting and processing only the changes in the visual scene. The machine-learned features are crafted to capture the unique characteristics of the event-camera data, enhancing the accuracy of the trajectory reconstruction. The inference pipeline is composed of a module repeated twice in sequence, formed by a Squeeze-and-Excite block and a ConvLSTM block with residual connection; it is followed by a final convolutional layer that provides the trajectories of the corners as a sequence of heatmaps. In the experimental part, a sequence of images was collected using an event-camera in outdoor environments for training and test.\",\"PeriodicalId\":87445,\"journal\":{\"name\":\"Materials Research Society symposia proceedings. Materials Research Society\",\"volume\":null,\"pages\":null},\"PeriodicalIF\":0.0000,\"publicationDate\":\"2023-11-01\",\"publicationTypes\":\"Journal Article\",\"fieldsOfStudy\":null,\"isOpenAccess\":false,\"openAccessPdf\":\"\",\"citationCount\":\"0\",\"resultStr\":null,\"platform\":\"Semanticscholar\",\"paperid\":null,\"PeriodicalName\":\"Materials Research Society symposia proceedings. Materials Research Society\",\"FirstCategoryId\":\"1085\",\"ListUrlMain\":\"https://doi.org/10.21741/9781644902813-150\",\"RegionNum\":0,\"RegionCategory\":null,\"ArticlePicture\":[],\"TitleCN\":null,\"AbstractTextCN\":null,\"PMCID\":null,\"EPubDate\":\"\",\"PubModel\":\"\",\"JCR\":\"\",\"JCRName\":\"\",\"Score\":null,\"Total\":0}","platform":"Semanticscholar","paperid":null,"PeriodicalName":"Materials Research Society symposia proceedings. Materials Research Society","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/10.21741/9781644902813-150","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
Trajectory reconstruction by means of an event-camera-based visual odometry method and machine learned features
Abstract. This paper presents a machine learned feature detector targeted to event-camera based visual odometry methods for unmanned aerial vehicles trajectory reconstruction. The proposed method uses machine-learned features to enhance the accuracy of the trajectory reconstruction. Traditional visual odometry methods suffer from poor performance in low light conditions and high-speed motion. The event-camera-based approach overcomes these limitations by detecting and processing only the changes in the visual scene. The machine-learned features are crafted to capture the unique characteristics of the event-camera data, enhancing the accuracy of the trajectory reconstruction. The inference pipeline is composed of a module repeated twice in sequence, formed by a Squeeze-and-Excite block and a ConvLSTM block with residual connection; it is followed by a final convolutional layer that provides the trajectories of the corners as a sequence of heatmaps. In the experimental part, a sequence of images was collected using an event-camera in outdoor environments for training and test.