C. Rapson, Boon-Chong Seet, M. Naeem, J. Lee, R. Klette
{"title":"A Performance Comparison of Deep Learning Methods for Real-time Localisation of Vehicle Lights in Video Frames","authors":"C. Rapson, Boon-Chong Seet, M. Naeem, J. Lee, R. Klette","doi":"10.1109/ITSC.2019.8917087","DOIUrl":null,"url":null,"abstract":"A vehicle’s braking lights can help to infer its future trajectory. Visible light communication using vehicle lights can also transmit other safety information to assist drivers with collision avoidance (whether the drivers be human or autonomous). Both these use cases require accurate localisation of vehicle lights by computer vision. Due to the large variation in lighting conditions (day, night, fog, snow, etc), the shape and brightness of the light itself, as well as difficulties with occlusions and perspectives, conventional methods are challenging and deep learning is a promising strategy. This paper presents a comparison of deep learning methods which are selected based on their potential to evaluate real-time video. The detection accuracy is shown to have a strong dependence on the size of the vehicle light within the image. A cascading approach is taken, where a downsampled image is used to detect vehicles, and then a second routine searches for vehicle lights at higher resolution within these Regions of Interest. This approach is demonstrated to improve detection, especially for small objects. Using YOLOv3 for the first stage and Tiny_YOLO for the second stage achieves satisfactory results across a wide range of conditions, and can execute at 37 frames per second. The ground truth for training and evaluating the methods is available for other researchers to use and compare their results.","PeriodicalId":6717,"journal":{"name":"2019 IEEE Intelligent Transportation Systems Conference (ITSC)","volume":"153 1","pages":"567-572"},"PeriodicalIF":0.0000,"publicationDate":"2019-10-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"5","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"2019 IEEE Intelligent Transportation Systems Conference (ITSC)","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/10.1109/ITSC.2019.8917087","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
引用次数: 5
Abstract
A vehicle’s braking lights can help to infer its future trajectory. Visible light communication using vehicle lights can also transmit other safety information to assist drivers with collision avoidance (whether the drivers be human or autonomous). Both these use cases require accurate localisation of vehicle lights by computer vision. Due to the large variation in lighting conditions (day, night, fog, snow, etc), the shape and brightness of the light itself, as well as difficulties with occlusions and perspectives, conventional methods are challenging and deep learning is a promising strategy. This paper presents a comparison of deep learning methods which are selected based on their potential to evaluate real-time video. The detection accuracy is shown to have a strong dependence on the size of the vehicle light within the image. A cascading approach is taken, where a downsampled image is used to detect vehicles, and then a second routine searches for vehicle lights at higher resolution within these Regions of Interest. This approach is demonstrated to improve detection, especially for small objects. Using YOLOv3 for the first stage and Tiny_YOLO for the second stage achieves satisfactory results across a wide range of conditions, and can execute at 37 frames per second. The ground truth for training and evaluating the methods is available for other researchers to use and compare their results.