{"title":"GAN Based Method for Labeled Image Augmentation in Autonomous Driving","authors":"Wenbo Yu, Yong Sun, Ruilin Zhou, Xingjian Liu","doi":"10.1109/ICCVE45908.2019.8964902","DOIUrl":null,"url":null,"abstract":"Deep learning models in Autonomous Driving perception tasks commonly use supervised learning methods and thus highly depend on labeled data. Training with more labeled data tends to bring better results, which highlights the meaning of data augmentation. Currently there are two difficulties when doing data augmentation. Firstly, it is time consuming to manually label the collected raw data. The second issue is that the diversity of a dataset is limited by the collection environment and time. In this paper, we proposed to use the current state of the art Multimodal Unsupervised Image-to-Image Translation (MUNIT) to generate synthesized images from labeled data. One of the benefits is that the generated data are automated labeled since they share the same ground truth with the raw data. Then we used the augmentation dataset to do different tasks including drivable area detection and object detection to prove that the data could be used to improve the performance of convolution neural networks (CNNs). We also designed an auto labelling tool that people could do labelling with the help of the improved CNN. The whole process is like a close loop that finishes labelling tasks while making progresses by itself. Generally speaking, our approach introduces an auto labelling pipeline based on unsupervised image-to-image translation to increase the amount and diversity of labeled data.","PeriodicalId":384049,"journal":{"name":"2019 IEEE International Conference on Connected Vehicles and Expo (ICCVE)","volume":"238 1","pages":"0"},"PeriodicalIF":0.0000,"publicationDate":"2019-11-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"2","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"2019 IEEE International Conference on Connected Vehicles and Expo (ICCVE)","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/10.1109/ICCVE45908.2019.8964902","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
引用次数: 2
Abstract
Deep learning models in Autonomous Driving perception tasks commonly use supervised learning methods and thus highly depend on labeled data. Training with more labeled data tends to bring better results, which highlights the meaning of data augmentation. Currently there are two difficulties when doing data augmentation. Firstly, it is time consuming to manually label the collected raw data. The second issue is that the diversity of a dataset is limited by the collection environment and time. In this paper, we proposed to use the current state of the art Multimodal Unsupervised Image-to-Image Translation (MUNIT) to generate synthesized images from labeled data. One of the benefits is that the generated data are automated labeled since they share the same ground truth with the raw data. Then we used the augmentation dataset to do different tasks including drivable area detection and object detection to prove that the data could be used to improve the performance of convolution neural networks (CNNs). We also designed an auto labelling tool that people could do labelling with the help of the improved CNN. The whole process is like a close loop that finishes labelling tasks while making progresses by itself. Generally speaking, our approach introduces an auto labelling pipeline based on unsupervised image-to-image translation to increase the amount and diversity of labeled data.