Jianping Cui , Liang Yuan , Wendong Xiao , Teng Ran , Li He , Jianbo Zhang
{"title":"SGLPER:一种安全的端到端自动驾驶决策框架,通过优先体验回放和Gipps模型,将深度强化学习和专家演示相结合","authors":"Jianping Cui , Liang Yuan , Wendong Xiao , Teng Ran , Li He , Jianbo Zhang","doi":"10.1016/j.displa.2025.103041","DOIUrl":null,"url":null,"abstract":"<div><div>Despite significant advancements in deep reinforcement learning (DRL), existing methods for autonomous driving often need to overcome the cold-start problem, requiring extensive training to converge and fail to fully address safety concerns in dynamic driving environments. To address these limitations, we propose an efficient DRL framework, SGLPER, which integrates Prioritized Experience Replay (PER), expert demonstrations, and a safe speed calculation model to improve learning efficiency and decision-making safety. Specifically, PER mitigates the cold-start problem by prioritizing high-value experiences and accelerating training convergence. The Long Short-Term Memory (LSTM) method also captures spatiotemporal information from observed states, enabling the agent to make informed decisions based on past experiences in complex, dynamic traffic scenarios. The safety strategy incorporates the Gipps model, introducing relatively safe speed limits into the reinforcement learning (RL) process to enhance driving safety. Moreover, Kullback–Leibler (KL) divergence combines RL with expert demonstrations, enabling the agent to learn human-like driving behaviors effectively. Experimental results in two simulated driving scenarios validate the robustness and effectiveness of the proposed framework. Compared to traditional DRL methods, SGLPER demonstrates safer strategies, higher success rates, and faster convergence. This study presents a promising approach for developing safer, more efficient autonomous driving systems.</div></div>","PeriodicalId":50570,"journal":{"name":"Displays","volume":"88 ","pages":"Article 103041"},"PeriodicalIF":3.7000,"publicationDate":"2025-04-07","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":"{\"title\":\"SGLPER: A safe end-to-end autonomous driving decision framework combining deep reinforcement learning and expert demonstrations via prioritized experience replay and the Gipps model\",\"authors\":\"Jianping Cui , Liang Yuan , Wendong Xiao , Teng Ran , Li He , Jianbo Zhang\",\"doi\":\"10.1016/j.displa.2025.103041\",\"DOIUrl\":null,\"url\":null,\"abstract\":\"<div><div>Despite significant advancements in deep reinforcement learning (DRL), existing methods for autonomous driving often need to overcome the cold-start problem, requiring extensive training to converge and fail to fully address safety concerns in dynamic driving environments. To address these limitations, we propose an efficient DRL framework, SGLPER, which integrates Prioritized Experience Replay (PER), expert demonstrations, and a safe speed calculation model to improve learning efficiency and decision-making safety. Specifically, PER mitigates the cold-start problem by prioritizing high-value experiences and accelerating training convergence. The Long Short-Term Memory (LSTM) method also captures spatiotemporal information from observed states, enabling the agent to make informed decisions based on past experiences in complex, dynamic traffic scenarios. The safety strategy incorporates the Gipps model, introducing relatively safe speed limits into the reinforcement learning (RL) process to enhance driving safety. Moreover, Kullback–Leibler (KL) divergence combines RL with expert demonstrations, enabling the agent to learn human-like driving behaviors effectively. Experimental results in two simulated driving scenarios validate the robustness and effectiveness of the proposed framework. Compared to traditional DRL methods, SGLPER demonstrates safer strategies, higher success rates, and faster convergence. This study presents a promising approach for developing safer, more efficient autonomous driving systems.</div></div>\",\"PeriodicalId\":50570,\"journal\":{\"name\":\"Displays\",\"volume\":\"88 \",\"pages\":\"Article 103041\"},\"PeriodicalIF\":3.7000,\"publicationDate\":\"2025-04-07\",\"publicationTypes\":\"Journal Article\",\"fieldsOfStudy\":null,\"isOpenAccess\":false,\"openAccessPdf\":\"\",\"citationCount\":\"0\",\"resultStr\":null,\"platform\":\"Semanticscholar\",\"paperid\":null,\"PeriodicalName\":\"Displays\",\"FirstCategoryId\":\"5\",\"ListUrlMain\":\"https://www.sciencedirect.com/science/article/pii/S0141938225000782\",\"RegionNum\":2,\"RegionCategory\":\"工程技术\",\"ArticlePicture\":[],\"TitleCN\":null,\"AbstractTextCN\":null,\"PMCID\":null,\"EPubDate\":\"\",\"PubModel\":\"\",\"JCR\":\"Q1\",\"JCRName\":\"COMPUTER SCIENCE, HARDWARE & ARCHITECTURE\",\"Score\":null,\"Total\":0}","platform":"Semanticscholar","paperid":null,"PeriodicalName":"Displays","FirstCategoryId":"5","ListUrlMain":"https://www.sciencedirect.com/science/article/pii/S0141938225000782","RegionNum":2,"RegionCategory":"工程技术","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"Q1","JCRName":"COMPUTER SCIENCE, HARDWARE & ARCHITECTURE","Score":null,"Total":0}
SGLPER: A safe end-to-end autonomous driving decision framework combining deep reinforcement learning and expert demonstrations via prioritized experience replay and the Gipps model
Despite significant advancements in deep reinforcement learning (DRL), existing methods for autonomous driving often need to overcome the cold-start problem, requiring extensive training to converge and fail to fully address safety concerns in dynamic driving environments. To address these limitations, we propose an efficient DRL framework, SGLPER, which integrates Prioritized Experience Replay (PER), expert demonstrations, and a safe speed calculation model to improve learning efficiency and decision-making safety. Specifically, PER mitigates the cold-start problem by prioritizing high-value experiences and accelerating training convergence. The Long Short-Term Memory (LSTM) method also captures spatiotemporal information from observed states, enabling the agent to make informed decisions based on past experiences in complex, dynamic traffic scenarios. The safety strategy incorporates the Gipps model, introducing relatively safe speed limits into the reinforcement learning (RL) process to enhance driving safety. Moreover, Kullback–Leibler (KL) divergence combines RL with expert demonstrations, enabling the agent to learn human-like driving behaviors effectively. Experimental results in two simulated driving scenarios validate the robustness and effectiveness of the proposed framework. Compared to traditional DRL methods, SGLPER demonstrates safer strategies, higher success rates, and faster convergence. This study presents a promising approach for developing safer, more efficient autonomous driving systems.
期刊介绍:
Displays is the international journal covering the research and development of display technology, its effective presentation and perception of information, and applications and systems including display-human interface.
Technical papers on practical developments in Displays technology provide an effective channel to promote greater understanding and cross-fertilization across the diverse disciplines of the Displays community. Original research papers solving ergonomics issues at the display-human interface advance effective presentation of information. Tutorial papers covering fundamentals intended for display technologies and human factor engineers new to the field will also occasionally featured.