Xiaodong Zhang , Jie Bao , Jianlei Chi , Jun Sun , Zijiang Yang
{"title":"Comprehensively evaluating the perception systems of autonomous vehicles against hazards","authors":"Xiaodong Zhang , Jie Bao , Jianlei Chi , Jun Sun , Zijiang Yang","doi":"10.1016/j.inffus.2025.103729","DOIUrl":null,"url":null,"abstract":"<div><div>Perception systems are vital for the safety of autonomous driving. In complex autonomous driving scenarios, autonomous vehicles must overcome various natural hazards, such as heavy rain or raindrops on the camera lens. Therefore, it is essential to conduct comprehensive testing of the perception systems in autonomous vehicles against these hazards, as demanded by the regulatory agencies of many countries for human drivers. Since there are many hazard scenarios, each of which has multiple configurable parameters, the challenges are (1) how do we systematically and adequately test an autonomous vehicle against these hazard scenarios, with measurable outcome; and (2) how do we efficiently explore the huge search space to identify scenarios that would induce failure?</div><div>In this work, we propose a Hazards Generation and Testing framework (HazGT) to generate a customizable and comprehensive repository of hazard scenarios for evaluating the perception system of autonomous vehicles. HazGT not only allows us to measure how comprehensively an autonomous vehicle (AV) has been tested against different hazards but also supports the identification of important hazards through optimization. HazGT supports a total of 70 kinds of hazards relevant to the visual perception of AVs, which are based on industrial regulations. HazGT automatically optimizes the parameter values to efficiently achieve different testing objectives. We have implemented HazGT based on two popular 3D engines, i.e., Unity and Unreal Engine. For the two mainstream perception models (i.e., YOLO and Faster RCNN), we have evaluated their performance against each hazard through extensive experiments, and the results show that both systems have much room to improve. In addition, our experiments also found that ChatGPT4 performs slightly worse than YOLO. Our optimization-based testing system is effective in finding perceptual errors in the perception models. The hazard images generated by HazGT are instrumental for improving perception models.</div></div>","PeriodicalId":50367,"journal":{"name":"Information Fusion","volume":"127 ","pages":"Article 103729"},"PeriodicalIF":15.5000,"publicationDate":"2025-09-18","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"Information Fusion","FirstCategoryId":"94","ListUrlMain":"https://www.sciencedirect.com/science/article/pii/S1566253525007912","RegionNum":1,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"Q1","JCRName":"COMPUTER SCIENCE, ARTIFICIAL INTELLIGENCE","Score":null,"Total":0}
引用次数: 0
Abstract
Perception systems are vital for the safety of autonomous driving. In complex autonomous driving scenarios, autonomous vehicles must overcome various natural hazards, such as heavy rain or raindrops on the camera lens. Therefore, it is essential to conduct comprehensive testing of the perception systems in autonomous vehicles against these hazards, as demanded by the regulatory agencies of many countries for human drivers. Since there are many hazard scenarios, each of which has multiple configurable parameters, the challenges are (1) how do we systematically and adequately test an autonomous vehicle against these hazard scenarios, with measurable outcome; and (2) how do we efficiently explore the huge search space to identify scenarios that would induce failure?
In this work, we propose a Hazards Generation and Testing framework (HazGT) to generate a customizable and comprehensive repository of hazard scenarios for evaluating the perception system of autonomous vehicles. HazGT not only allows us to measure how comprehensively an autonomous vehicle (AV) has been tested against different hazards but also supports the identification of important hazards through optimization. HazGT supports a total of 70 kinds of hazards relevant to the visual perception of AVs, which are based on industrial regulations. HazGT automatically optimizes the parameter values to efficiently achieve different testing objectives. We have implemented HazGT based on two popular 3D engines, i.e., Unity and Unreal Engine. For the two mainstream perception models (i.e., YOLO and Faster RCNN), we have evaluated their performance against each hazard through extensive experiments, and the results show that both systems have much room to improve. In addition, our experiments also found that ChatGPT4 performs slightly worse than YOLO. Our optimization-based testing system is effective in finding perceptual errors in the perception models. The hazard images generated by HazGT are instrumental for improving perception models.
期刊介绍:
Information Fusion serves as a central platform for showcasing advancements in multi-sensor, multi-source, multi-process information fusion, fostering collaboration among diverse disciplines driving its progress. It is the leading outlet for sharing research and development in this field, focusing on architectures, algorithms, and applications. Papers dealing with fundamental theoretical analyses as well as those demonstrating their application to real-world problems will be welcome.