基于优化奖励函数的改进Dueling Deep Q-network驱动决策方法

IF 1.5 4区 工程技术 Q3 ENGINEERING, MECHANICAL
Jiaqi Cao, Xiaolan Wang, Yansong Wang, Yongxiang Tian
{"title":"基于优化奖励函数的改进Dueling Deep Q-network驱动决策方法","authors":"Jiaqi Cao, Xiaolan Wang, Yansong Wang, Yongxiang Tian","doi":"10.1177/09544070221106037","DOIUrl":null,"url":null,"abstract":"Aiming at poor effects and single consideration factors of traditional driving decision-making algorithm in high-speed and complex environment, a method based on improved deep reinforcement learning (DRL) is proposed in this paper. We innovatively design and optimize the reward function of the Dueling Deep Q network (Dueling DQN), and the factors such as safety, comfort, traffic efficiency and altruism are taken into account. The weight of each influencing factor is determined by the Analytic Hierarchy Process (AHP), which makes the influence of each factor on driving behavior decision-making more acceptable. Subsequently, a decision-making model of autonomous vehicles (AVs) is built by using improved Dueling DQN. Furthermore, the action space is enriched and combined with the trajectory planner, so that AVs can take appropriate behaviors in the longitudinal and lateral directions according to the environment. The output of the decision model can be combined with the underlying controller with a view to make the AVs maneuver reasonably. The driving decision-making method in two different traffic scenarios is simulated. Moreover, the improved method compares with other methods. The results illustrate that the improved Dueling DQN can make the AVs take safe, comfortable, efficient, and altruistic behavior.","PeriodicalId":54568,"journal":{"name":"Proceedings of the Institution of Mechanical Engineers Part D-Journal of Automobile Engineering","volume":"59 1","pages":"2295 - 2309"},"PeriodicalIF":1.5000,"publicationDate":"2022-06-20","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":"{\"title\":\"An improved Dueling Deep Q-network with optimizing reward functions for driving decision method\",\"authors\":\"Jiaqi Cao, Xiaolan Wang, Yansong Wang, Yongxiang Tian\",\"doi\":\"10.1177/09544070221106037\",\"DOIUrl\":null,\"url\":null,\"abstract\":\"Aiming at poor effects and single consideration factors of traditional driving decision-making algorithm in high-speed and complex environment, a method based on improved deep reinforcement learning (DRL) is proposed in this paper. We innovatively design and optimize the reward function of the Dueling Deep Q network (Dueling DQN), and the factors such as safety, comfort, traffic efficiency and altruism are taken into account. The weight of each influencing factor is determined by the Analytic Hierarchy Process (AHP), which makes the influence of each factor on driving behavior decision-making more acceptable. Subsequently, a decision-making model of autonomous vehicles (AVs) is built by using improved Dueling DQN. Furthermore, the action space is enriched and combined with the trajectory planner, so that AVs can take appropriate behaviors in the longitudinal and lateral directions according to the environment. The output of the decision model can be combined with the underlying controller with a view to make the AVs maneuver reasonably. The driving decision-making method in two different traffic scenarios is simulated. Moreover, the improved method compares with other methods. The results illustrate that the improved Dueling DQN can make the AVs take safe, comfortable, efficient, and altruistic behavior.\",\"PeriodicalId\":54568,\"journal\":{\"name\":\"Proceedings of the Institution of Mechanical Engineers Part D-Journal of Automobile Engineering\",\"volume\":\"59 1\",\"pages\":\"2295 - 2309\"},\"PeriodicalIF\":1.5000,\"publicationDate\":\"2022-06-20\",\"publicationTypes\":\"Journal Article\",\"fieldsOfStudy\":null,\"isOpenAccess\":false,\"openAccessPdf\":\"\",\"citationCount\":\"0\",\"resultStr\":null,\"platform\":\"Semanticscholar\",\"paperid\":null,\"PeriodicalName\":\"Proceedings of the Institution of Mechanical Engineers Part D-Journal of Automobile Engineering\",\"FirstCategoryId\":\"5\",\"ListUrlMain\":\"https://doi.org/10.1177/09544070221106037\",\"RegionNum\":4,\"RegionCategory\":\"工程技术\",\"ArticlePicture\":[],\"TitleCN\":null,\"AbstractTextCN\":null,\"PMCID\":null,\"EPubDate\":\"\",\"PubModel\":\"\",\"JCR\":\"Q3\",\"JCRName\":\"ENGINEERING, MECHANICAL\",\"Score\":null,\"Total\":0}","platform":"Semanticscholar","paperid":null,"PeriodicalName":"Proceedings of the Institution of Mechanical Engineers Part D-Journal of Automobile Engineering","FirstCategoryId":"5","ListUrlMain":"https://doi.org/10.1177/09544070221106037","RegionNum":4,"RegionCategory":"工程技术","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"Q3","JCRName":"ENGINEERING, MECHANICAL","Score":null,"Total":0}
引用次数: 0

摘要

针对传统驾驶决策算法在高速复杂环境下效果差、考虑因素单一的问题,提出了一种基于改进深度强化学习(DRL)的驾驶决策方法。我们创新地设计和优化了Dueling Deep Q网络(Dueling DQN)的奖励函数,并考虑了安全性、舒适性、交通效率和利他性等因素。通过层次分析法确定各影响因素的权重,使各因素对驾驶行为决策的影响更容易被接受。随后,利用改进的Dueling DQN建立了自动驾驶汽车的决策模型。进一步丰富动作空间并与轨迹规划器相结合,使自动驾驶汽车能够根据环境在纵向和横向上采取适当的行为。决策模型的输出可以与底层控制器相结合,使自动驾驶汽车合理机动。仿真了两种不同交通场景下的驾驶决策方法。并与其他方法进行了比较。结果表明,改进的Dueling DQN能使自动驾驶汽车采取安全、舒适、高效和利他的行为。
本文章由计算机程序翻译,如有差异,请以英文原文为准。
An improved Dueling Deep Q-network with optimizing reward functions for driving decision method
Aiming at poor effects and single consideration factors of traditional driving decision-making algorithm in high-speed and complex environment, a method based on improved deep reinforcement learning (DRL) is proposed in this paper. We innovatively design and optimize the reward function of the Dueling Deep Q network (Dueling DQN), and the factors such as safety, comfort, traffic efficiency and altruism are taken into account. The weight of each influencing factor is determined by the Analytic Hierarchy Process (AHP), which makes the influence of each factor on driving behavior decision-making more acceptable. Subsequently, a decision-making model of autonomous vehicles (AVs) is built by using improved Dueling DQN. Furthermore, the action space is enriched and combined with the trajectory planner, so that AVs can take appropriate behaviors in the longitudinal and lateral directions according to the environment. The output of the decision model can be combined with the underlying controller with a view to make the AVs maneuver reasonably. The driving decision-making method in two different traffic scenarios is simulated. Moreover, the improved method compares with other methods. The results illustrate that the improved Dueling DQN can make the AVs take safe, comfortable, efficient, and altruistic behavior.
求助全文
通过发布文献求助,成功后即可免费获取论文全文。 去求助
来源期刊
CiteScore
4.40
自引率
17.60%
发文量
263
审稿时长
3.5 months
期刊介绍: The Journal of Automobile Engineering is an established, high quality multi-disciplinary journal which publishes the very best peer-reviewed science and engineering in the field.
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
确定
请完成安全验证×
copy
已复制链接
快去分享给好友吧!
我知道了
右上角分享
点击右上角分享
0
联系我们:info@booksci.cn Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。 Copyright © 2023 布克学术 All rights reserved.
京ICP备2023020795号-1
ghs 京公网安备 11010802042870号
Book学术文献互助
Book学术文献互助群
群 号:604180095
Book学术官方微信