基于逆强化学习的仿人公路轨迹建模

Ruoyu Sun, Shaochi Hu, Huijing Zhao, M. Moze, F. Aioun, F. Guillemard
{"title":"基于逆强化学习的仿人公路轨迹建模","authors":"Ruoyu Sun, Shaochi Hu, Huijing Zhao, M. Moze, F. Aioun, F. Guillemard","doi":"10.1109/ITSC.2019.8916970","DOIUrl":null,"url":null,"abstract":"Autonomous driving is one of the current cutting edge technologies. For autonomous cars, their driving actions and trajectories should not only achieve autonomy and safety, but also obey human drivers’ behavior patterns, when sharing the roads with other human drivers on the highway. Traditional methods, though robust and interpretable, demands much human labor in engineering the complex mapping from current driving situation to vehicle’s future control. For newly developed deep-learning methods, though they can automatically learn such complex mapping from data and demands fewer humans’ engineering, they mostly act like black-box, and are less interpretable. We proposed a new combined method based on inverse reinforcement learning to harness the advantages of both. Experimental validations on lane-change prediction and human-like trajectory planning show that the proposed method approximates the state-of-the-art performance in modeling human trajectories, and is both interpretable and data-driven.","PeriodicalId":6717,"journal":{"name":"2019 IEEE Intelligent Transportation Systems Conference (ITSC)","volume":"25 1","pages":"1482-1489"},"PeriodicalIF":0.0000,"publicationDate":"2019-10-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"4","resultStr":"{\"title\":\"Human-like Highway Trajectory Modeling based on Inverse Reinforcement Learning\",\"authors\":\"Ruoyu Sun, Shaochi Hu, Huijing Zhao, M. Moze, F. Aioun, F. Guillemard\",\"doi\":\"10.1109/ITSC.2019.8916970\",\"DOIUrl\":null,\"url\":null,\"abstract\":\"Autonomous driving is one of the current cutting edge technologies. For autonomous cars, their driving actions and trajectories should not only achieve autonomy and safety, but also obey human drivers’ behavior patterns, when sharing the roads with other human drivers on the highway. Traditional methods, though robust and interpretable, demands much human labor in engineering the complex mapping from current driving situation to vehicle’s future control. For newly developed deep-learning methods, though they can automatically learn such complex mapping from data and demands fewer humans’ engineering, they mostly act like black-box, and are less interpretable. We proposed a new combined method based on inverse reinforcement learning to harness the advantages of both. Experimental validations on lane-change prediction and human-like trajectory planning show that the proposed method approximates the state-of-the-art performance in modeling human trajectories, and is both interpretable and data-driven.\",\"PeriodicalId\":6717,\"journal\":{\"name\":\"2019 IEEE Intelligent Transportation Systems Conference (ITSC)\",\"volume\":\"25 1\",\"pages\":\"1482-1489\"},\"PeriodicalIF\":0.0000,\"publicationDate\":\"2019-10-01\",\"publicationTypes\":\"Journal Article\",\"fieldsOfStudy\":null,\"isOpenAccess\":false,\"openAccessPdf\":\"\",\"citationCount\":\"4\",\"resultStr\":null,\"platform\":\"Semanticscholar\",\"paperid\":null,\"PeriodicalName\":\"2019 IEEE Intelligent Transportation Systems Conference (ITSC)\",\"FirstCategoryId\":\"1085\",\"ListUrlMain\":\"https://doi.org/10.1109/ITSC.2019.8916970\",\"RegionNum\":0,\"RegionCategory\":null,\"ArticlePicture\":[],\"TitleCN\":null,\"AbstractTextCN\":null,\"PMCID\":null,\"EPubDate\":\"\",\"PubModel\":\"\",\"JCR\":\"\",\"JCRName\":\"\",\"Score\":null,\"Total\":0}","platform":"Semanticscholar","paperid":null,"PeriodicalName":"2019 IEEE Intelligent Transportation Systems Conference (ITSC)","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/10.1109/ITSC.2019.8916970","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
引用次数: 4

摘要

自动驾驶是当前的前沿技术之一。对于自动驾驶汽车来说,在高速公路上与其他人类驾驶员共享道路时,其驾驶动作和轨迹不仅要实现自主性和安全性,而且要服从人类驾驶员的行为模式。传统方法虽然具有鲁棒性和可解释性,但在从当前驾驶状况到车辆未来控制的复杂映射过程中,需要耗费大量人力。对于新开发的深度学习方法,尽管它们可以从数据中自动学习如此复杂的映射,并且需要更少的人工工程,但它们大多像黑箱一样,难以解释。我们提出了一种新的基于逆强化学习的组合方法来利用两者的优点。对变道预测和仿人轨迹规划的实验验证表明,该方法接近人类轨迹建模的最新性能,并且具有可解释性和数据驱动性。
本文章由计算机程序翻译,如有差异,请以英文原文为准。
Human-like Highway Trajectory Modeling based on Inverse Reinforcement Learning
Autonomous driving is one of the current cutting edge technologies. For autonomous cars, their driving actions and trajectories should not only achieve autonomy and safety, but also obey human drivers’ behavior patterns, when sharing the roads with other human drivers on the highway. Traditional methods, though robust and interpretable, demands much human labor in engineering the complex mapping from current driving situation to vehicle’s future control. For newly developed deep-learning methods, though they can automatically learn such complex mapping from data and demands fewer humans’ engineering, they mostly act like black-box, and are less interpretable. We proposed a new combined method based on inverse reinforcement learning to harness the advantages of both. Experimental validations on lane-change prediction and human-like trajectory planning show that the proposed method approximates the state-of-the-art performance in modeling human trajectories, and is both interpretable and data-driven.
求助全文
通过发布文献求助,成功后即可免费获取论文全文。 去求助
来源期刊
自引率
0.00%
发文量
0
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
确定
请完成安全验证×
copy
已复制链接
快去分享给好友吧!
我知道了
右上角分享
点击右上角分享
0
联系我们:info@booksci.cn Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。 Copyright © 2023 布克学术 All rights reserved.
京ICP备2023020795号-1
ghs 京公网安备 11010802042870号
Book学术文献互助
Book学术文献互助群
群 号:481959085
Book学术官方微信