El houssine Amraouy , Ali Yahyaouy , Hamid Gualous , Hicham Chaoui , Sanaa Faquir
{"title":"基于强化学习模型的自动驾驶汽车城市驾驶车道合并","authors":"El houssine Amraouy , Ali Yahyaouy , Hamid Gualous , Hicham Chaoui , Sanaa Faquir","doi":"10.1016/j.urbmob.2025.100150","DOIUrl":null,"url":null,"abstract":"<div><div>Autonomous vehicle lane merging is a critical task in urban driving, requiring precise navigation through complex and dynamic traffic environments. Challenges such as roadworks, lane reductions, merging from gas stations, low-visibility conditions, and crowded highway on-ramps demand continuous improvements in autonomous driving systems. Effective navigation in these situations, particularly at multi-lane junctions, merging onto high-speed roads, avoiding obstacles, and managing emergency vehicle lanes, requires robust decision-making that can adapt to changing road conditions. This paper compares three popular reinforcement learning (RL) algorithms—Proximal Policy Optimization (PPO), Advantage Actor-Critic (A2C), and Deep Q-Learning (DQL)—to address these challenges. Our findings show that in environments with specific decision points, DQL excels in tasks like lane reduction and obstacle avoidance due to its value-based approach. The A2C model, as an actor-critic policy, is particularly effective in dynamic environments, enabling the optimization of urban traffic control and merging at roundabouts. PPO, known for its policy optimization capabilities, offers a robust solution by balancing safety, efficiency, and adaptability, particularly in complex situations such as high-speed merging and low-visibility conditions. The simulation results confirm that DQL, A2C, and PPO collectively enhance autonomous vehicle performance by improving navigation capabilities, increasing safety, and adapting more effectively to the complexities of urban traffic environments. This work contributes valuable insights into the application of RL for real-world autonomous driving, providing a detailed comparative evaluation that supports the selection of algorithms tailored to specific driving tasks.</div></div>","PeriodicalId":100852,"journal":{"name":"Journal of Urban Mobility","volume":"8 ","pages":"Article 100150"},"PeriodicalIF":6.1000,"publicationDate":"2025-09-18","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":"{\"title\":\"Lane merging in autonomous vehicle urban driving using reinforcement learning models\",\"authors\":\"El houssine Amraouy , Ali Yahyaouy , Hamid Gualous , Hicham Chaoui , Sanaa Faquir\",\"doi\":\"10.1016/j.urbmob.2025.100150\",\"DOIUrl\":null,\"url\":null,\"abstract\":\"<div><div>Autonomous vehicle lane merging is a critical task in urban driving, requiring precise navigation through complex and dynamic traffic environments. Challenges such as roadworks, lane reductions, merging from gas stations, low-visibility conditions, and crowded highway on-ramps demand continuous improvements in autonomous driving systems. Effective navigation in these situations, particularly at multi-lane junctions, merging onto high-speed roads, avoiding obstacles, and managing emergency vehicle lanes, requires robust decision-making that can adapt to changing road conditions. This paper compares three popular reinforcement learning (RL) algorithms—Proximal Policy Optimization (PPO), Advantage Actor-Critic (A2C), and Deep Q-Learning (DQL)—to address these challenges. Our findings show that in environments with specific decision points, DQL excels in tasks like lane reduction and obstacle avoidance due to its value-based approach. The A2C model, as an actor-critic policy, is particularly effective in dynamic environments, enabling the optimization of urban traffic control and merging at roundabouts. PPO, known for its policy optimization capabilities, offers a robust solution by balancing safety, efficiency, and adaptability, particularly in complex situations such as high-speed merging and low-visibility conditions. The simulation results confirm that DQL, A2C, and PPO collectively enhance autonomous vehicle performance by improving navigation capabilities, increasing safety, and adapting more effectively to the complexities of urban traffic environments. This work contributes valuable insights into the application of RL for real-world autonomous driving, providing a detailed comparative evaluation that supports the selection of algorithms tailored to specific driving tasks.</div></div>\",\"PeriodicalId\":100852,\"journal\":{\"name\":\"Journal of Urban Mobility\",\"volume\":\"8 \",\"pages\":\"Article 100150\"},\"PeriodicalIF\":6.1000,\"publicationDate\":\"2025-09-18\",\"publicationTypes\":\"Journal Article\",\"fieldsOfStudy\":null,\"isOpenAccess\":false,\"openAccessPdf\":\"\",\"citationCount\":\"0\",\"resultStr\":null,\"platform\":\"Semanticscholar\",\"paperid\":null,\"PeriodicalName\":\"Journal of Urban Mobility\",\"FirstCategoryId\":\"1085\",\"ListUrlMain\":\"https://www.sciencedirect.com/science/article/pii/S2667091725000524\",\"RegionNum\":0,\"RegionCategory\":null,\"ArticlePicture\":[],\"TitleCN\":null,\"AbstractTextCN\":null,\"PMCID\":null,\"EPubDate\":\"\",\"PubModel\":\"\",\"JCR\":\"Q1\",\"JCRName\":\"GEOGRAPHY\",\"Score\":null,\"Total\":0}","platform":"Semanticscholar","paperid":null,"PeriodicalName":"Journal of Urban Mobility","FirstCategoryId":"1085","ListUrlMain":"https://www.sciencedirect.com/science/article/pii/S2667091725000524","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"Q1","JCRName":"GEOGRAPHY","Score":null,"Total":0}
Lane merging in autonomous vehicle urban driving using reinforcement learning models
Autonomous vehicle lane merging is a critical task in urban driving, requiring precise navigation through complex and dynamic traffic environments. Challenges such as roadworks, lane reductions, merging from gas stations, low-visibility conditions, and crowded highway on-ramps demand continuous improvements in autonomous driving systems. Effective navigation in these situations, particularly at multi-lane junctions, merging onto high-speed roads, avoiding obstacles, and managing emergency vehicle lanes, requires robust decision-making that can adapt to changing road conditions. This paper compares three popular reinforcement learning (RL) algorithms—Proximal Policy Optimization (PPO), Advantage Actor-Critic (A2C), and Deep Q-Learning (DQL)—to address these challenges. Our findings show that in environments with specific decision points, DQL excels in tasks like lane reduction and obstacle avoidance due to its value-based approach. The A2C model, as an actor-critic policy, is particularly effective in dynamic environments, enabling the optimization of urban traffic control and merging at roundabouts. PPO, known for its policy optimization capabilities, offers a robust solution by balancing safety, efficiency, and adaptability, particularly in complex situations such as high-speed merging and low-visibility conditions. The simulation results confirm that DQL, A2C, and PPO collectively enhance autonomous vehicle performance by improving navigation capabilities, increasing safety, and adapting more effectively to the complexities of urban traffic environments. This work contributes valuable insights into the application of RL for real-world autonomous driving, providing a detailed comparative evaluation that supports the selection of algorithms tailored to specific driving tasks.