{"title":"基于优化奖励函数的改进Dueling Deep Q-network驱动决策方法","authors":"Jiaqi Cao, Xiaolan Wang, Yansong Wang, Yongxiang Tian","doi":"10.1177/09544070221106037","DOIUrl":null,"url":null,"abstract":"Aiming at poor effects and single consideration factors of traditional driving decision-making algorithm in high-speed and complex environment, a method based on improved deep reinforcement learning (DRL) is proposed in this paper. We innovatively design and optimize the reward function of the Dueling Deep Q network (Dueling DQN), and the factors such as safety, comfort, traffic efficiency and altruism are taken into account. The weight of each influencing factor is determined by the Analytic Hierarchy Process (AHP), which makes the influence of each factor on driving behavior decision-making more acceptable. Subsequently, a decision-making model of autonomous vehicles (AVs) is built by using improved Dueling DQN. Furthermore, the action space is enriched and combined with the trajectory planner, so that AVs can take appropriate behaviors in the longitudinal and lateral directions according to the environment. The output of the decision model can be combined with the underlying controller with a view to make the AVs maneuver reasonably. The driving decision-making method in two different traffic scenarios is simulated. Moreover, the improved method compares with other methods. The results illustrate that the improved Dueling DQN can make the AVs take safe, comfortable, efficient, and altruistic behavior.","PeriodicalId":54568,"journal":{"name":"Proceedings of the Institution of Mechanical Engineers Part D-Journal of Automobile Engineering","volume":"59 1","pages":"2295 - 2309"},"PeriodicalIF":1.5000,"publicationDate":"2022-06-20","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":"{\"title\":\"An improved Dueling Deep Q-network with optimizing reward functions for driving decision method\",\"authors\":\"Jiaqi Cao, Xiaolan Wang, Yansong Wang, Yongxiang Tian\",\"doi\":\"10.1177/09544070221106037\",\"DOIUrl\":null,\"url\":null,\"abstract\":\"Aiming at poor effects and single consideration factors of traditional driving decision-making algorithm in high-speed and complex environment, a method based on improved deep reinforcement learning (DRL) is proposed in this paper. We innovatively design and optimize the reward function of the Dueling Deep Q network (Dueling DQN), and the factors such as safety, comfort, traffic efficiency and altruism are taken into account. The weight of each influencing factor is determined by the Analytic Hierarchy Process (AHP), which makes the influence of each factor on driving behavior decision-making more acceptable. Subsequently, a decision-making model of autonomous vehicles (AVs) is built by using improved Dueling DQN. Furthermore, the action space is enriched and combined with the trajectory planner, so that AVs can take appropriate behaviors in the longitudinal and lateral directions according to the environment. The output of the decision model can be combined with the underlying controller with a view to make the AVs maneuver reasonably. The driving decision-making method in two different traffic scenarios is simulated. Moreover, the improved method compares with other methods. The results illustrate that the improved Dueling DQN can make the AVs take safe, comfortable, efficient, and altruistic behavior.\",\"PeriodicalId\":54568,\"journal\":{\"name\":\"Proceedings of the Institution of Mechanical Engineers Part D-Journal of Automobile Engineering\",\"volume\":\"59 1\",\"pages\":\"2295 - 2309\"},\"PeriodicalIF\":1.5000,\"publicationDate\":\"2022-06-20\",\"publicationTypes\":\"Journal Article\",\"fieldsOfStudy\":null,\"isOpenAccess\":false,\"openAccessPdf\":\"\",\"citationCount\":\"0\",\"resultStr\":null,\"platform\":\"Semanticscholar\",\"paperid\":null,\"PeriodicalName\":\"Proceedings of the Institution of Mechanical Engineers Part D-Journal of Automobile Engineering\",\"FirstCategoryId\":\"5\",\"ListUrlMain\":\"https://doi.org/10.1177/09544070221106037\",\"RegionNum\":4,\"RegionCategory\":\"工程技术\",\"ArticlePicture\":[],\"TitleCN\":null,\"AbstractTextCN\":null,\"PMCID\":null,\"EPubDate\":\"\",\"PubModel\":\"\",\"JCR\":\"Q3\",\"JCRName\":\"ENGINEERING, MECHANICAL\",\"Score\":null,\"Total\":0}","platform":"Semanticscholar","paperid":null,"PeriodicalName":"Proceedings of the Institution of Mechanical Engineers Part D-Journal of Automobile Engineering","FirstCategoryId":"5","ListUrlMain":"https://doi.org/10.1177/09544070221106037","RegionNum":4,"RegionCategory":"工程技术","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"Q3","JCRName":"ENGINEERING, MECHANICAL","Score":null,"Total":0}
引用次数: 0
摘要
针对传统驾驶决策算法在高速复杂环境下效果差、考虑因素单一的问题,提出了一种基于改进深度强化学习(DRL)的驾驶决策方法。我们创新地设计和优化了Dueling Deep Q网络(Dueling DQN)的奖励函数,并考虑了安全性、舒适性、交通效率和利他性等因素。通过层次分析法确定各影响因素的权重,使各因素对驾驶行为决策的影响更容易被接受。随后,利用改进的Dueling DQN建立了自动驾驶汽车的决策模型。进一步丰富动作空间并与轨迹规划器相结合,使自动驾驶汽车能够根据环境在纵向和横向上采取适当的行为。决策模型的输出可以与底层控制器相结合,使自动驾驶汽车合理机动。仿真了两种不同交通场景下的驾驶决策方法。并与其他方法进行了比较。结果表明,改进的Dueling DQN能使自动驾驶汽车采取安全、舒适、高效和利他的行为。
An improved Dueling Deep Q-network with optimizing reward functions for driving decision method
Aiming at poor effects and single consideration factors of traditional driving decision-making algorithm in high-speed and complex environment, a method based on improved deep reinforcement learning (DRL) is proposed in this paper. We innovatively design and optimize the reward function of the Dueling Deep Q network (Dueling DQN), and the factors such as safety, comfort, traffic efficiency and altruism are taken into account. The weight of each influencing factor is determined by the Analytic Hierarchy Process (AHP), which makes the influence of each factor on driving behavior decision-making more acceptable. Subsequently, a decision-making model of autonomous vehicles (AVs) is built by using improved Dueling DQN. Furthermore, the action space is enriched and combined with the trajectory planner, so that AVs can take appropriate behaviors in the longitudinal and lateral directions according to the environment. The output of the decision model can be combined with the underlying controller with a view to make the AVs maneuver reasonably. The driving decision-making method in two different traffic scenarios is simulated. Moreover, the improved method compares with other methods. The results illustrate that the improved Dueling DQN can make the AVs take safe, comfortable, efficient, and altruistic behavior.
期刊介绍:
The Journal of Automobile Engineering is an established, high quality multi-disciplinary journal which publishes the very best peer-reviewed science and engineering in the field.