S. Jha, Isaac Brooks, Soumitry J. Ray, R. Narasimha, N. Al-Dhahir, Carlos Busso
{"title":"Seatbelt Segmentation Using Synthetic Images","authors":"S. Jha, Isaac Brooks, Soumitry J. Ray, R. Narasimha, N. Al-Dhahir, Carlos Busso","doi":"10.1109/IV55152.2023.10186571","DOIUrl":null,"url":null,"abstract":"Recent advancement in deep learning has led to an increased interest in image processing and computer vision applications for driver monitoring systems. One of the applications where these techniques can be useful is in segmenting and tracking seatbelts. A seatbelt is an important safety feature in the vehicle that if properly used can save lives. Efficient segmentation of the seatbelts in an image provides important information about the correct use of seatbelts. The challenge in developing deep learning algorithms for seatbelt detection and segmentation is the manual annotations required for this task, which is cumbersome. This paper explores a novel formulation to efficiently train a seatbelt model with minimal supervision. We exploit the textureless and shape characteristics of the seatbelts to programmatically synthesize images. Our proposed method synthetically creates images that resemble seatbelt patterns. After training a model exclusively with synthetic images, we iteratively fine-tune it using naturalistic images extracted from online video-sharing websites. The labels for these images are pseudo-labels assigned by the model to confident predictions. Fine-tuning helps adapt the model to better work on real naturalistic images, improving the performance of the system. We obtain an F1-score of 0.55 in segmenting the seatbelt with this approach. We also experiment with fine-tuning the model with a small number of naturalistic images with annotated labels. After pretraining on synthetic samples and pseudo-labeled naturalistic images, we achieve an F1-score of 0.67 using only 200 annotated images.","PeriodicalId":195148,"journal":{"name":"2023 IEEE Intelligent Vehicles Symposium (IV)","volume":"27 1","pages":"0"},"PeriodicalIF":0.0000,"publicationDate":"2023-06-04","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"1","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"2023 IEEE Intelligent Vehicles Symposium (IV)","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/10.1109/IV55152.2023.10186571","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
引用次数: 1
Abstract
Recent advancement in deep learning has led to an increased interest in image processing and computer vision applications for driver monitoring systems. One of the applications where these techniques can be useful is in segmenting and tracking seatbelts. A seatbelt is an important safety feature in the vehicle that if properly used can save lives. Efficient segmentation of the seatbelts in an image provides important information about the correct use of seatbelts. The challenge in developing deep learning algorithms for seatbelt detection and segmentation is the manual annotations required for this task, which is cumbersome. This paper explores a novel formulation to efficiently train a seatbelt model with minimal supervision. We exploit the textureless and shape characteristics of the seatbelts to programmatically synthesize images. Our proposed method synthetically creates images that resemble seatbelt patterns. After training a model exclusively with synthetic images, we iteratively fine-tune it using naturalistic images extracted from online video-sharing websites. The labels for these images are pseudo-labels assigned by the model to confident predictions. Fine-tuning helps adapt the model to better work on real naturalistic images, improving the performance of the system. We obtain an F1-score of 0.55 in segmenting the seatbelt with this approach. We also experiment with fine-tuning the model with a small number of naturalistic images with annotated labels. After pretraining on synthetic samples and pseudo-labeled naturalistic images, we achieve an F1-score of 0.67 using only 200 annotated images.