{"title":"基于CNN和LSTM网络的强化学习在高速公路上驾驶","authors":"Lászlo Szőke, S. Aradi, Tamás Bécsi, P. Gáspár","doi":"10.1109/INES49302.2020.9147185","DOIUrl":null,"url":null,"abstract":"This work presents a powerful and intelligent driver agent, designed to operate in a preset highway situation using Policy Gradient Reinforcement Learning (RL) agent. Our goal is to create an agent that is capable of navigating safely in changing highway traffic and successfully accomplish to get through the defined section keeping the reference speed. Meanwhile, creating a state representation that is capable of extracting information from images based on the actual highway situation. The algorithm uses Convolutional Neural Network (CNN) with Long-Short Term Memory (LSTM) layers as a function approximator for the agent with discrete action space on the control level, e.g., acceleration and lane change. Simulation of Urban MObility (SUMO), an open-source microscopic traffic simulator is chosen as our simulation environment. It is integrated with an open interface to interact with the agent in real-time. The agent can learn from numerous driving and highway situations that are created and fed to it. The representation becomes more general by randomizing and customizing the behavior of the other road users in the simulation, thus the experience of the agent can be much more diverse. The article briefly describes the modeling environment, the details on the learning agent, and the rewarding scheme. After evaluating the experiences gained from the training, some further plans and optimization ideas are briefed.","PeriodicalId":175830,"journal":{"name":"2020 IEEE 24th International Conference on Intelligent Engineering Systems (INES)","volume":"59 6 1","pages":"0"},"PeriodicalIF":0.0000,"publicationDate":"2020-07-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"7","resultStr":"{\"title\":\"Driving on Highway by Using Reinforcement Learning with CNN and LSTM Networks\",\"authors\":\"Lászlo Szőke, S. Aradi, Tamás Bécsi, P. Gáspár\",\"doi\":\"10.1109/INES49302.2020.9147185\",\"DOIUrl\":null,\"url\":null,\"abstract\":\"This work presents a powerful and intelligent driver agent, designed to operate in a preset highway situation using Policy Gradient Reinforcement Learning (RL) agent. Our goal is to create an agent that is capable of navigating safely in changing highway traffic and successfully accomplish to get through the defined section keeping the reference speed. Meanwhile, creating a state representation that is capable of extracting information from images based on the actual highway situation. The algorithm uses Convolutional Neural Network (CNN) with Long-Short Term Memory (LSTM) layers as a function approximator for the agent with discrete action space on the control level, e.g., acceleration and lane change. Simulation of Urban MObility (SUMO), an open-source microscopic traffic simulator is chosen as our simulation environment. It is integrated with an open interface to interact with the agent in real-time. The agent can learn from numerous driving and highway situations that are created and fed to it. The representation becomes more general by randomizing and customizing the behavior of the other road users in the simulation, thus the experience of the agent can be much more diverse. The article briefly describes the modeling environment, the details on the learning agent, and the rewarding scheme. After evaluating the experiences gained from the training, some further plans and optimization ideas are briefed.\",\"PeriodicalId\":175830,\"journal\":{\"name\":\"2020 IEEE 24th International Conference on Intelligent Engineering Systems (INES)\",\"volume\":\"59 6 1\",\"pages\":\"0\"},\"PeriodicalIF\":0.0000,\"publicationDate\":\"2020-07-01\",\"publicationTypes\":\"Journal Article\",\"fieldsOfStudy\":null,\"isOpenAccess\":false,\"openAccessPdf\":\"\",\"citationCount\":\"7\",\"resultStr\":null,\"platform\":\"Semanticscholar\",\"paperid\":null,\"PeriodicalName\":\"2020 IEEE 24th International Conference on Intelligent Engineering Systems (INES)\",\"FirstCategoryId\":\"1085\",\"ListUrlMain\":\"https://doi.org/10.1109/INES49302.2020.9147185\",\"RegionNum\":0,\"RegionCategory\":null,\"ArticlePicture\":[],\"TitleCN\":null,\"AbstractTextCN\":null,\"PMCID\":null,\"EPubDate\":\"\",\"PubModel\":\"\",\"JCR\":\"\",\"JCRName\":\"\",\"Score\":null,\"Total\":0}","platform":"Semanticscholar","paperid":null,"PeriodicalName":"2020 IEEE 24th International Conference on Intelligent Engineering Systems (INES)","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/10.1109/INES49302.2020.9147185","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
Driving on Highway by Using Reinforcement Learning with CNN and LSTM Networks
This work presents a powerful and intelligent driver agent, designed to operate in a preset highway situation using Policy Gradient Reinforcement Learning (RL) agent. Our goal is to create an agent that is capable of navigating safely in changing highway traffic and successfully accomplish to get through the defined section keeping the reference speed. Meanwhile, creating a state representation that is capable of extracting information from images based on the actual highway situation. The algorithm uses Convolutional Neural Network (CNN) with Long-Short Term Memory (LSTM) layers as a function approximator for the agent with discrete action space on the control level, e.g., acceleration and lane change. Simulation of Urban MObility (SUMO), an open-source microscopic traffic simulator is chosen as our simulation environment. It is integrated with an open interface to interact with the agent in real-time. The agent can learn from numerous driving and highway situations that are created and fed to it. The representation becomes more general by randomizing and customizing the behavior of the other road users in the simulation, thus the experience of the agent can be much more diverse. The article briefly describes the modeling environment, the details on the learning agent, and the rewarding scheme. After evaluating the experiences gained from the training, some further plans and optimization ideas are briefed.