{"title":"Performance Evaluation of Real-Time Object Detection for Electric Scooters","authors":"Dong Chen, Arman Hosseini, Arik Smith, Amir Farzin Nikkhah, Arsalan Heydarian, Omid Shoghli, Bradford Campbell","doi":"arxiv-2405.03039","DOIUrl":null,"url":null,"abstract":"Electric scooters (e-scooters) have rapidly emerged as a popular mode of\ntransportation in urban areas, yet they pose significant safety challenges. In\nthe United States, the rise of e-scooters has been marked by a concerning\nincrease in related injuries and fatalities. Recently, while deep-learning\nobject detection holds paramount significance in autonomous vehicles to avoid\npotential collisions, its application in the context of e-scooters remains\nrelatively unexplored. This paper addresses this gap by assessing the\neffectiveness and efficiency of cutting-edge object detectors designed for\ne-scooters. To achieve this, the first comprehensive benchmark involving 22\nstate-of-the-art YOLO object detectors, including five versions (YOLOv3,\nYOLOv5, YOLOv6, YOLOv7, and YOLOv8), has been established for real-time traffic\nobject detection using a self-collected dataset featuring e-scooters. The\ndetection accuracy, measured in terms of mAP@0.5, ranges from 27.4%\n(YOLOv7-E6E) to 86.8% (YOLOv5s). All YOLO models, particularly YOLOv3-tiny,\nhave displayed promising potential for real-time object detection in the\ncontext of e-scooters. Both the traffic scene dataset\n(https://zenodo.org/records/10578641) and software program codes\n(https://github.com/DongChen06/ScooterDet) for model benchmarking in this study\nare publicly available, which will not only improve e-scooter safety with\nadvanced object detection but also lay the groundwork for tailored solutions,\npromising a safer and more sustainable urban micromobility landscape.","PeriodicalId":501062,"journal":{"name":"arXiv - CS - Systems and Control","volume":"27 1","pages":""},"PeriodicalIF":0.0000,"publicationDate":"2024-05-05","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"arXiv - CS - Systems and Control","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/arxiv-2405.03039","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
引用次数: 0
Abstract
Electric scooters (e-scooters) have rapidly emerged as a popular mode of
transportation in urban areas, yet they pose significant safety challenges. In
the United States, the rise of e-scooters has been marked by a concerning
increase in related injuries and fatalities. Recently, while deep-learning
object detection holds paramount significance in autonomous vehicles to avoid
potential collisions, its application in the context of e-scooters remains
relatively unexplored. This paper addresses this gap by assessing the
effectiveness and efficiency of cutting-edge object detectors designed for
e-scooters. To achieve this, the first comprehensive benchmark involving 22
state-of-the-art YOLO object detectors, including five versions (YOLOv3,
YOLOv5, YOLOv6, YOLOv7, and YOLOv8), has been established for real-time traffic
object detection using a self-collected dataset featuring e-scooters. The
detection accuracy, measured in terms of mAP@0.5, ranges from 27.4%
(YOLOv7-E6E) to 86.8% (YOLOv5s). All YOLO models, particularly YOLOv3-tiny,
have displayed promising potential for real-time object detection in the
context of e-scooters. Both the traffic scene dataset
(https://zenodo.org/records/10578641) and software program codes
(https://github.com/DongChen06/ScooterDet) for model benchmarking in this study
are publicly available, which will not only improve e-scooter safety with
advanced object detection but also lay the groundwork for tailored solutions,
promising a safer and more sustainable urban micromobility landscape.