{"title":"BEV perception for autonomous driving: State of the art and future perspectives","authors":"Junhui Zhao , Jingyue Shi , Li Zhuo","doi":"10.1016/j.eswa.2024.125103","DOIUrl":null,"url":null,"abstract":"<div><p>The remarkable performance of Bird’s Eye View (BEV) in perception tasks has led to its gradual emergence as a focal point of attention in both industry and academia. Environmental information perception technology represents a core challenge in the field of autonomous driving, and traditional autonomous driving perception algorithms typically perform tasks such as detection, segmentation, and tracking from a frontal or specific viewpoint. As the complexity of sensor parameters configured on vehicles increases, it has become crucial to integrate multi-source information from different sensors and present features in a unified view. BEV perception is favored because it is an intuitive and user-friendly way to fuse information about the surrounding environment and provide an ideal object representation for subsequent planning and control modules. However, BEV perception also faces some key challenges. One such challenge is how to convert from a perspective view to a BEV view while reconstructing lost 3D information. The question of how to obtain accurate ground truth annotations in the BEV grid is of great importance. Similarly, the design of effective methods to integrate features from different sources is a crucial aspect of BEV perception. In this paper, we first discuss the inherent advantages of BEV perception and introduce the mainstream datasets and performance evaluation criteria for BEV perception. Furthermore, we present a comprehensive examination of recent research on BEV perception from four distinct perspectives, exploring a range of solutions, including BEV camera, BEV LiDAR, BEV fusion, and V2V multi-vehicle cooperative BEV perception. Finally, we identify prospective research directions and challenges in this field, with the aim of providing inspiration to related researchers.</p></div>","PeriodicalId":50461,"journal":{"name":"Expert Systems with Applications","volume":"258 ","pages":"Article 125103"},"PeriodicalIF":7.5000,"publicationDate":"2024-08-24","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"Expert Systems with Applications","FirstCategoryId":"94","ListUrlMain":"https://www.sciencedirect.com/science/article/pii/S0957417424019705","RegionNum":1,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"Q1","JCRName":"COMPUTER SCIENCE, ARTIFICIAL INTELLIGENCE","Score":null,"Total":0}
引用次数: 0
Abstract
The remarkable performance of Bird’s Eye View (BEV) in perception tasks has led to its gradual emergence as a focal point of attention in both industry and academia. Environmental information perception technology represents a core challenge in the field of autonomous driving, and traditional autonomous driving perception algorithms typically perform tasks such as detection, segmentation, and tracking from a frontal or specific viewpoint. As the complexity of sensor parameters configured on vehicles increases, it has become crucial to integrate multi-source information from different sensors and present features in a unified view. BEV perception is favored because it is an intuitive and user-friendly way to fuse information about the surrounding environment and provide an ideal object representation for subsequent planning and control modules. However, BEV perception also faces some key challenges. One such challenge is how to convert from a perspective view to a BEV view while reconstructing lost 3D information. The question of how to obtain accurate ground truth annotations in the BEV grid is of great importance. Similarly, the design of effective methods to integrate features from different sources is a crucial aspect of BEV perception. In this paper, we first discuss the inherent advantages of BEV perception and introduce the mainstream datasets and performance evaluation criteria for BEV perception. Furthermore, we present a comprehensive examination of recent research on BEV perception from four distinct perspectives, exploring a range of solutions, including BEV camera, BEV LiDAR, BEV fusion, and V2V multi-vehicle cooperative BEV perception. Finally, we identify prospective research directions and challenges in this field, with the aim of providing inspiration to related researchers.
期刊介绍:
Expert Systems With Applications is an international journal dedicated to the exchange of information on expert and intelligent systems used globally in industry, government, and universities. The journal emphasizes original papers covering the design, development, testing, implementation, and management of these systems, offering practical guidelines. It spans various sectors such as finance, engineering, marketing, law, project management, information management, medicine, and more. The journal also welcomes papers on multi-agent systems, knowledge management, neural networks, knowledge discovery, data mining, and other related areas, excluding applications to military/defense systems.