改进激光雷达小类语义分割及自动驾驶泛化能力

Chiao-Hua Tseng, Yu-Ting Lin, Wen-Chieh Lin, Chieh-Chih Wang
{"title":"改进激光雷达小类语义分割及自动驾驶泛化能力","authors":"Chiao-Hua Tseng, Yu-Ting Lin, Wen-Chieh Lin, Chieh-Chih Wang","doi":"10.1109/TAAI57707.2022.00032","DOIUrl":null,"url":null,"abstract":"LiDARs have emerged as an important sensor in autonomous driving systems because they offer more accurate geometric measurements than cameras and radars. Therefore, LiDARs have been commonly combined with cameras or radars to tackle many perception problems in autonomous driving, such as object detection, semantic segmentation, or navigation. For semantic segmentation of LiDAR data, due to the class imbalance issue of large-scale scene, there is a performance gap between majority classes and minority classes of large-scale dataset. The minority classes usually include the crucial classes to the autonomous driving, such as “person”, “motorcyclist”, “traffic-sign”. To improve the performance of minority classes, we adopt U-Net++ as the architecture, KPConv as convolution operator, and use both dice loss and cross entropy as loss functions. We get 5.1% mIoU improvement on SemanticKITTI of all classes and 9.5% mIoU improvement of minority classes. Moreover, due to the different resolution of LiDAR sensors, we show the generalization capability of our model by training it on 64-beam dataset and testing on 32-beam and 128-beam dataset. We get 3.3% mIoU improvement on 128-beam dataset and 1.9% mIoU improvement on 32-beam dataset.","PeriodicalId":111620,"journal":{"name":"2022 International Conference on Technologies and Applications of Artificial Intelligence (TAAI)","volume":"208 1","pages":"0"},"PeriodicalIF":0.0000,"publicationDate":"2022-12-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":"{\"title\":\"Improving LiDAR Semantic Segmentation on Minority Classes and Generalization Capability for Autonomous Driving\",\"authors\":\"Chiao-Hua Tseng, Yu-Ting Lin, Wen-Chieh Lin, Chieh-Chih Wang\",\"doi\":\"10.1109/TAAI57707.2022.00032\",\"DOIUrl\":null,\"url\":null,\"abstract\":\"LiDARs have emerged as an important sensor in autonomous driving systems because they offer more accurate geometric measurements than cameras and radars. Therefore, LiDARs have been commonly combined with cameras or radars to tackle many perception problems in autonomous driving, such as object detection, semantic segmentation, or navigation. For semantic segmentation of LiDAR data, due to the class imbalance issue of large-scale scene, there is a performance gap between majority classes and minority classes of large-scale dataset. The minority classes usually include the crucial classes to the autonomous driving, such as “person”, “motorcyclist”, “traffic-sign”. To improve the performance of minority classes, we adopt U-Net++ as the architecture, KPConv as convolution operator, and use both dice loss and cross entropy as loss functions. We get 5.1% mIoU improvement on SemanticKITTI of all classes and 9.5% mIoU improvement of minority classes. Moreover, due to the different resolution of LiDAR sensors, we show the generalization capability of our model by training it on 64-beam dataset and testing on 32-beam and 128-beam dataset. We get 3.3% mIoU improvement on 128-beam dataset and 1.9% mIoU improvement on 32-beam dataset.\",\"PeriodicalId\":111620,\"journal\":{\"name\":\"2022 International Conference on Technologies and Applications of Artificial Intelligence (TAAI)\",\"volume\":\"208 1\",\"pages\":\"0\"},\"PeriodicalIF\":0.0000,\"publicationDate\":\"2022-12-01\",\"publicationTypes\":\"Journal Article\",\"fieldsOfStudy\":null,\"isOpenAccess\":false,\"openAccessPdf\":\"\",\"citationCount\":\"0\",\"resultStr\":null,\"platform\":\"Semanticscholar\",\"paperid\":null,\"PeriodicalName\":\"2022 International Conference on Technologies and Applications of Artificial Intelligence (TAAI)\",\"FirstCategoryId\":\"1085\",\"ListUrlMain\":\"https://doi.org/10.1109/TAAI57707.2022.00032\",\"RegionNum\":0,\"RegionCategory\":null,\"ArticlePicture\":[],\"TitleCN\":null,\"AbstractTextCN\":null,\"PMCID\":null,\"EPubDate\":\"\",\"PubModel\":\"\",\"JCR\":\"\",\"JCRName\":\"\",\"Score\":null,\"Total\":0}","platform":"Semanticscholar","paperid":null,"PeriodicalName":"2022 International Conference on Technologies and Applications of Artificial Intelligence (TAAI)","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/10.1109/TAAI57707.2022.00032","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
引用次数: 0

摘要

激光雷达已经成为自动驾驶系统中重要的传感器,因为它们比摄像头和雷达提供更精确的几何测量。因此,激光雷达通常与摄像头或雷达相结合,以解决自动驾驶中的许多感知问题,如物体检测、语义分割或导航。对于激光雷达数据的语义分割,由于大规模场景的类不平衡问题,大规模数据集的多数类和少数类之间存在性能差距。少数类通常包括对自动驾驶至关重要的类,如“人”、“摩托车手”、“交通标志”。为了提高少数类的性能,我们采用U-Net++作为体系结构,KPConv作为卷积算子,同时使用骰子损失和交叉熵作为损失函数。所有类别的SemanticKITTI提高了5.1%,少数类别的SemanticKITTI提高了9.5%。此外,由于激光雷达传感器的分辨率不同,我们通过在64波束数据集上进行训练,并在32波束和128波束数据集上进行测试,证明了该模型的泛化能力。在128波束数据集上得到3.3%的mIoU改进,在32波束数据集上得到1.9%的mIoU改进。
本文章由计算机程序翻译,如有差异,请以英文原文为准。
Improving LiDAR Semantic Segmentation on Minority Classes and Generalization Capability for Autonomous Driving
LiDARs have emerged as an important sensor in autonomous driving systems because they offer more accurate geometric measurements than cameras and radars. Therefore, LiDARs have been commonly combined with cameras or radars to tackle many perception problems in autonomous driving, such as object detection, semantic segmentation, or navigation. For semantic segmentation of LiDAR data, due to the class imbalance issue of large-scale scene, there is a performance gap between majority classes and minority classes of large-scale dataset. The minority classes usually include the crucial classes to the autonomous driving, such as “person”, “motorcyclist”, “traffic-sign”. To improve the performance of minority classes, we adopt U-Net++ as the architecture, KPConv as convolution operator, and use both dice loss and cross entropy as loss functions. We get 5.1% mIoU improvement on SemanticKITTI of all classes and 9.5% mIoU improvement of minority classes. Moreover, due to the different resolution of LiDAR sensors, we show the generalization capability of our model by training it on 64-beam dataset and testing on 32-beam and 128-beam dataset. We get 3.3% mIoU improvement on 128-beam dataset and 1.9% mIoU improvement on 32-beam dataset.
求助全文
通过发布文献求助,成功后即可免费获取论文全文。 去求助
来源期刊
自引率
0.00%
发文量
0
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
确定
请完成安全验证×
copy
已复制链接
快去分享给好友吧!
我知道了
右上角分享
点击右上角分享
0
联系我们:info@booksci.cn Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。 Copyright © 2023 布克学术 All rights reserved.
京ICP备2023020795号-1
ghs 京公网安备 11010802042870号
Book学术文献互助
Book学术文献互助群
群 号:604180095
Book学术官方微信