Narmada M. Balasooriya, O. de Silva, Awantha Jayasiri, G. Mann
{"title":"基于人工智能的着陆区检测,用于垂直起飞和陆地激光雷达定位及管道测绘","authors":"Narmada M. Balasooriya, O. de Silva, Awantha Jayasiri, G. Mann","doi":"10.1139/dsa-2022-0038","DOIUrl":null,"url":null,"abstract":"This paper proposed a novel point-based neural network landing zone detection architecture that can operate with a VLOAM navigation pipeline and investigates the accuracy-runtime trade-offs of the method for real-time applications. Based on the Semantic3D benchmark leaderboard, ConvPoint architecture was selected as the target model for the task. The work investigated different combinations of hyperparameters, i.e., batch size and sampling size, in terms of the performance metrics, i.e., inference time, throughput, and accuracy. Validation of the method was performed using custom datasets captured on a DJI M600 drone and a Bell 412 aircraft to generate the LZ module's maps at a target update rate (~ 1 Hz) while operating within a VLOAM navigation pipeline. Accurate detection of water bodies, marshlands, and low vegetation as non-landable is crucial for VTOL operations. From the results described in this paper, it is evident that to get a comparatively accurate detection of water areas in the given dataset, a larger sampling size should be set, which also can lead to lower throughput (higher inference time). This bottleneck can be resolved by fusing the semantic labels generated by the point cloud segmentation with the pixel labels generated by the color image semantic segmentation of the same region and by using a broader range of datasets to train the neural network model.","PeriodicalId":202289,"journal":{"name":"Drone Systems and Applications","volume":"53 40","pages":""},"PeriodicalIF":0.0000,"publicationDate":"2024-02-12","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":"{\"title\":\"AI-based Landing Zone Detection for Vertical Takeoff and Land LiDAR Localization and Mapping Pipelines\",\"authors\":\"Narmada M. Balasooriya, O. de Silva, Awantha Jayasiri, G. Mann\",\"doi\":\"10.1139/dsa-2022-0038\",\"DOIUrl\":null,\"url\":null,\"abstract\":\"This paper proposed a novel point-based neural network landing zone detection architecture that can operate with a VLOAM navigation pipeline and investigates the accuracy-runtime trade-offs of the method for real-time applications. Based on the Semantic3D benchmark leaderboard, ConvPoint architecture was selected as the target model for the task. The work investigated different combinations of hyperparameters, i.e., batch size and sampling size, in terms of the performance metrics, i.e., inference time, throughput, and accuracy. Validation of the method was performed using custom datasets captured on a DJI M600 drone and a Bell 412 aircraft to generate the LZ module's maps at a target update rate (~ 1 Hz) while operating within a VLOAM navigation pipeline. Accurate detection of water bodies, marshlands, and low vegetation as non-landable is crucial for VTOL operations. From the results described in this paper, it is evident that to get a comparatively accurate detection of water areas in the given dataset, a larger sampling size should be set, which also can lead to lower throughput (higher inference time). This bottleneck can be resolved by fusing the semantic labels generated by the point cloud segmentation with the pixel labels generated by the color image semantic segmentation of the same region and by using a broader range of datasets to train the neural network model.\",\"PeriodicalId\":202289,\"journal\":{\"name\":\"Drone Systems and Applications\",\"volume\":\"53 40\",\"pages\":\"\"},\"PeriodicalIF\":0.0000,\"publicationDate\":\"2024-02-12\",\"publicationTypes\":\"Journal Article\",\"fieldsOfStudy\":null,\"isOpenAccess\":false,\"openAccessPdf\":\"\",\"citationCount\":\"0\",\"resultStr\":null,\"platform\":\"Semanticscholar\",\"paperid\":null,\"PeriodicalName\":\"Drone Systems and Applications\",\"FirstCategoryId\":\"1085\",\"ListUrlMain\":\"https://doi.org/10.1139/dsa-2022-0038\",\"RegionNum\":0,\"RegionCategory\":null,\"ArticlePicture\":[],\"TitleCN\":null,\"AbstractTextCN\":null,\"PMCID\":null,\"EPubDate\":\"\",\"PubModel\":\"\",\"JCR\":\"\",\"JCRName\":\"\",\"Score\":null,\"Total\":0}","platform":"Semanticscholar","paperid":null,"PeriodicalName":"Drone Systems and Applications","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/10.1139/dsa-2022-0038","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
AI-based Landing Zone Detection for Vertical Takeoff and Land LiDAR Localization and Mapping Pipelines
This paper proposed a novel point-based neural network landing zone detection architecture that can operate with a VLOAM navigation pipeline and investigates the accuracy-runtime trade-offs of the method for real-time applications. Based on the Semantic3D benchmark leaderboard, ConvPoint architecture was selected as the target model for the task. The work investigated different combinations of hyperparameters, i.e., batch size and sampling size, in terms of the performance metrics, i.e., inference time, throughput, and accuracy. Validation of the method was performed using custom datasets captured on a DJI M600 drone and a Bell 412 aircraft to generate the LZ module's maps at a target update rate (~ 1 Hz) while operating within a VLOAM navigation pipeline. Accurate detection of water bodies, marshlands, and low vegetation as non-landable is crucial for VTOL operations. From the results described in this paper, it is evident that to get a comparatively accurate detection of water areas in the given dataset, a larger sampling size should be set, which also can lead to lower throughput (higher inference time). This bottleneck can be resolved by fusing the semantic labels generated by the point cloud segmentation with the pixel labels generated by the color image semantic segmentation of the same region and by using a broader range of datasets to train the neural network model.