{"title":"基于语义分割的自动驾驶汽车BEV检测与定位","authors":"Ashwin Nayak U, Nachiket Naganure, S. S.","doi":"10.1109/CONECCT52877.2021.9622702","DOIUrl":null,"url":null,"abstract":"In autonomous vehicles, the perception system plays an important role in environment modeling and object detection in 3D space. Existing perception systems use various sensors to localize and track the surrounding obstacles, but have some limitations. Most existing end-to-end autonomous systems are computationally heavy as they are built on multiple deep networks that are trained to detect and localize objects, thus requiring custom, high-end computation devices with high compute power. To address this issue, we propose and experiment with different semantic segmentation-based models for Birds Eye View (BEV) detection and localization of surrounding objects like vehicles and pedestrians from LiDAR (light detection, and ranging) point clouds. Voxelisation techniques are used to transform 3D LiDAR point clouds to 2D RGB images. The semantic segmentation models are trained from the ground up on the Lyft Level 5 dataset. During experimental evaluation, the proposed approach achieved a mean average precision score of 0.044 for UNET, 0.041 for SegNet and 0.033 for FCN, while being significantly less compute-intensive when compared to the state-of-the-art approaches.","PeriodicalId":164499,"journal":{"name":"2021 IEEE International Conference on Electronics, Computing and Communication Technologies (CONECCT)","volume":"36 1","pages":"0"},"PeriodicalIF":0.0000,"publicationDate":"2021-07-09","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":"{\"title\":\"BEV Detection and Localisation using Semantic Segmentation in Autonomous Car Driving Systems\",\"authors\":\"Ashwin Nayak U, Nachiket Naganure, S. S.\",\"doi\":\"10.1109/CONECCT52877.2021.9622702\",\"DOIUrl\":null,\"url\":null,\"abstract\":\"In autonomous vehicles, the perception system plays an important role in environment modeling and object detection in 3D space. Existing perception systems use various sensors to localize and track the surrounding obstacles, but have some limitations. Most existing end-to-end autonomous systems are computationally heavy as they are built on multiple deep networks that are trained to detect and localize objects, thus requiring custom, high-end computation devices with high compute power. To address this issue, we propose and experiment with different semantic segmentation-based models for Birds Eye View (BEV) detection and localization of surrounding objects like vehicles and pedestrians from LiDAR (light detection, and ranging) point clouds. Voxelisation techniques are used to transform 3D LiDAR point clouds to 2D RGB images. The semantic segmentation models are trained from the ground up on the Lyft Level 5 dataset. During experimental evaluation, the proposed approach achieved a mean average precision score of 0.044 for UNET, 0.041 for SegNet and 0.033 for FCN, while being significantly less compute-intensive when compared to the state-of-the-art approaches.\",\"PeriodicalId\":164499,\"journal\":{\"name\":\"2021 IEEE International Conference on Electronics, Computing and Communication Technologies (CONECCT)\",\"volume\":\"36 1\",\"pages\":\"0\"},\"PeriodicalIF\":0.0000,\"publicationDate\":\"2021-07-09\",\"publicationTypes\":\"Journal Article\",\"fieldsOfStudy\":null,\"isOpenAccess\":false,\"openAccessPdf\":\"\",\"citationCount\":\"0\",\"resultStr\":null,\"platform\":\"Semanticscholar\",\"paperid\":null,\"PeriodicalName\":\"2021 IEEE International Conference on Electronics, Computing and Communication Technologies (CONECCT)\",\"FirstCategoryId\":\"1085\",\"ListUrlMain\":\"https://doi.org/10.1109/CONECCT52877.2021.9622702\",\"RegionNum\":0,\"RegionCategory\":null,\"ArticlePicture\":[],\"TitleCN\":null,\"AbstractTextCN\":null,\"PMCID\":null,\"EPubDate\":\"\",\"PubModel\":\"\",\"JCR\":\"\",\"JCRName\":\"\",\"Score\":null,\"Total\":0}","platform":"Semanticscholar","paperid":null,"PeriodicalName":"2021 IEEE International Conference on Electronics, Computing and Communication Technologies (CONECCT)","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/10.1109/CONECCT52877.2021.9622702","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
BEV Detection and Localisation using Semantic Segmentation in Autonomous Car Driving Systems
In autonomous vehicles, the perception system plays an important role in environment modeling and object detection in 3D space. Existing perception systems use various sensors to localize and track the surrounding obstacles, but have some limitations. Most existing end-to-end autonomous systems are computationally heavy as they are built on multiple deep networks that are trained to detect and localize objects, thus requiring custom, high-end computation devices with high compute power. To address this issue, we propose and experiment with different semantic segmentation-based models for Birds Eye View (BEV) detection and localization of surrounding objects like vehicles and pedestrians from LiDAR (light detection, and ranging) point clouds. Voxelisation techniques are used to transform 3D LiDAR point clouds to 2D RGB images. The semantic segmentation models are trained from the ground up on the Lyft Level 5 dataset. During experimental evaluation, the proposed approach achieved a mean average precision score of 0.044 for UNET, 0.041 for SegNet and 0.033 for FCN, while being significantly less compute-intensive when compared to the state-of-the-art approaches.