{"title":"知识引导可控扩散增强自动驾驶场景生成","authors":"Ce Shan, Lulu Guo, Hong Chen","doi":"10.1016/j.neucom.2025.130616","DOIUrl":null,"url":null,"abstract":"<div><div>Considering the shortcomings of the natural driving datasets composed by typical scenarios in the verification and evaluation of autonomous driving, how to improve the diversity, realism and controllability to satisfy the complex driving scenarios need is an urgent problem to be solved. Therefore, a knowledge-guided controllable diffusion (KGCD) scenarios generation framework is proposed, which allows for the generation behaviors that conform to expected attributes through a differentiable cost designed based on prior driving knowledge during the inference stage. Specifically, factorized attention is utilized for capturing and fusion of the spatio-temporal interaction features for the scene embeddings, based on which the accurate scene-level interactive joint motion prediction is achieved. In the inference stage, diversified guidance is designed based on prior knowledge and embedded into a generation framework to guide the generation of controllable and realistic driving scenarios, filling the gap in the existing scenario library. In response to the slow inference speed of typical diffusion model, a learnable motion pattern estimator (MPE) module is proposed to improve inference speed while ensuring generation quality. Based on the experiments with different datasets, KGCD proposed in this paper can satisfy the requirements of diverse scene generation. Compared with the best performance obtained by other baseline, the realism, controllability and stability metrics have been improved on average by 3.57%, 7.66% and 7.71% respectively evaluated by nuScenes dataset.</div></div>","PeriodicalId":19268,"journal":{"name":"Neurocomputing","volume":"648 ","pages":"Article 130616"},"PeriodicalIF":5.5000,"publicationDate":"2025-06-11","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":"{\"title\":\"Knowledge guided controllable diffusion for enhanced autonomous driving scenarios generation\",\"authors\":\"Ce Shan, Lulu Guo, Hong Chen\",\"doi\":\"10.1016/j.neucom.2025.130616\",\"DOIUrl\":null,\"url\":null,\"abstract\":\"<div><div>Considering the shortcomings of the natural driving datasets composed by typical scenarios in the verification and evaluation of autonomous driving, how to improve the diversity, realism and controllability to satisfy the complex driving scenarios need is an urgent problem to be solved. Therefore, a knowledge-guided controllable diffusion (KGCD) scenarios generation framework is proposed, which allows for the generation behaviors that conform to expected attributes through a differentiable cost designed based on prior driving knowledge during the inference stage. Specifically, factorized attention is utilized for capturing and fusion of the spatio-temporal interaction features for the scene embeddings, based on which the accurate scene-level interactive joint motion prediction is achieved. In the inference stage, diversified guidance is designed based on prior knowledge and embedded into a generation framework to guide the generation of controllable and realistic driving scenarios, filling the gap in the existing scenario library. In response to the slow inference speed of typical diffusion model, a learnable motion pattern estimator (MPE) module is proposed to improve inference speed while ensuring generation quality. Based on the experiments with different datasets, KGCD proposed in this paper can satisfy the requirements of diverse scene generation. Compared with the best performance obtained by other baseline, the realism, controllability and stability metrics have been improved on average by 3.57%, 7.66% and 7.71% respectively evaluated by nuScenes dataset.</div></div>\",\"PeriodicalId\":19268,\"journal\":{\"name\":\"Neurocomputing\",\"volume\":\"648 \",\"pages\":\"Article 130616\"},\"PeriodicalIF\":5.5000,\"publicationDate\":\"2025-06-11\",\"publicationTypes\":\"Journal Article\",\"fieldsOfStudy\":null,\"isOpenAccess\":false,\"openAccessPdf\":\"\",\"citationCount\":\"0\",\"resultStr\":null,\"platform\":\"Semanticscholar\",\"paperid\":null,\"PeriodicalName\":\"Neurocomputing\",\"FirstCategoryId\":\"94\",\"ListUrlMain\":\"https://www.sciencedirect.com/science/article/pii/S0925231225012883\",\"RegionNum\":2,\"RegionCategory\":\"计算机科学\",\"ArticlePicture\":[],\"TitleCN\":null,\"AbstractTextCN\":null,\"PMCID\":null,\"EPubDate\":\"\",\"PubModel\":\"\",\"JCR\":\"Q1\",\"JCRName\":\"COMPUTER SCIENCE, ARTIFICIAL INTELLIGENCE\",\"Score\":null,\"Total\":0}","platform":"Semanticscholar","paperid":null,"PeriodicalName":"Neurocomputing","FirstCategoryId":"94","ListUrlMain":"https://www.sciencedirect.com/science/article/pii/S0925231225012883","RegionNum":2,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"Q1","JCRName":"COMPUTER SCIENCE, ARTIFICIAL INTELLIGENCE","Score":null,"Total":0}
Knowledge guided controllable diffusion for enhanced autonomous driving scenarios generation
Considering the shortcomings of the natural driving datasets composed by typical scenarios in the verification and evaluation of autonomous driving, how to improve the diversity, realism and controllability to satisfy the complex driving scenarios need is an urgent problem to be solved. Therefore, a knowledge-guided controllable diffusion (KGCD) scenarios generation framework is proposed, which allows for the generation behaviors that conform to expected attributes through a differentiable cost designed based on prior driving knowledge during the inference stage. Specifically, factorized attention is utilized for capturing and fusion of the spatio-temporal interaction features for the scene embeddings, based on which the accurate scene-level interactive joint motion prediction is achieved. In the inference stage, diversified guidance is designed based on prior knowledge and embedded into a generation framework to guide the generation of controllable and realistic driving scenarios, filling the gap in the existing scenario library. In response to the slow inference speed of typical diffusion model, a learnable motion pattern estimator (MPE) module is proposed to improve inference speed while ensuring generation quality. Based on the experiments with different datasets, KGCD proposed in this paper can satisfy the requirements of diverse scene generation. Compared with the best performance obtained by other baseline, the realism, controllability and stability metrics have been improved on average by 3.57%, 7.66% and 7.71% respectively evaluated by nuScenes dataset.
期刊介绍:
Neurocomputing publishes articles describing recent fundamental contributions in the field of neurocomputing. Neurocomputing theory, practice and applications are the essential topics being covered.