Nuo Lei , Hao Zhang , Jingjing Hu , Zunyan Hu , Zhi Wang
{"title":"基于强化学习的燃料电池电动汽车能量管理策略仿真设计与开发","authors":"Nuo Lei , Hao Zhang , Jingjing Hu , Zunyan Hu , Zhi Wang","doi":"10.1016/j.apenergy.2025.126030","DOIUrl":null,"url":null,"abstract":"<div><div>The application of reinforcement learning (RL) algorithms in energy management strategies (EMSs) for fuel cell electric vehicles (FCEVs) has shown promising results in simulations. However, transitioning these strategies to real-vehicle implementation remains challenging due to the complexities of vehicle dynamics and system integration. Based on the General Optimal control Problems Solver (GOPS) platform, this paper establishes an RL-based EMS development toolchain that integrates advanced algorithms with high-fidelity vehicle models, leveraging Python-MATLAB/Simulink co-simulation for agent training across Model-in-the-Loop (MiL), Hardware-in-the-Loop (HiL), and Vehicle-in-the-Loop (ViL) stages. Besides, the distributional soft actor-critic algorithm (DSAC) is applied to energy management for the first time, embedding the return distribution function into maximum entropy RL. This approach adapts the Q-value function update step size, significantly enhancing strategy performance. Additionally, two RL-based EMS frameworks are investigated: one where the agent directly outputs fuel cell power commands, and another where the agent generates equivalent factors (EF) for the equivalent consumption minimization strategy (ECMS). Simulation and experimental results validate that both RL frameworks achieve superior fuel economy, reducing hydrogen consumption by approximately 4.35 % to 5.73 % compared to benchmarks. By combining Python's algorithmic flexibility and scalability with MATLAB/Simulink's high-fidelity vehicle models, the proposed toolchain provides a robust foundation for real-vehicle applications of RL-based EMSs.</div></div>","PeriodicalId":246,"journal":{"name":"Applied Energy","volume":"393 ","pages":"Article 126030"},"PeriodicalIF":10.1000,"publicationDate":"2025-05-12","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":"{\"title\":\"Sim-to-real design and development of reinforcement learning-based energy management strategies for fuel cell electric vehicles\",\"authors\":\"Nuo Lei , Hao Zhang , Jingjing Hu , Zunyan Hu , Zhi Wang\",\"doi\":\"10.1016/j.apenergy.2025.126030\",\"DOIUrl\":null,\"url\":null,\"abstract\":\"<div><div>The application of reinforcement learning (RL) algorithms in energy management strategies (EMSs) for fuel cell electric vehicles (FCEVs) has shown promising results in simulations. However, transitioning these strategies to real-vehicle implementation remains challenging due to the complexities of vehicle dynamics and system integration. Based on the General Optimal control Problems Solver (GOPS) platform, this paper establishes an RL-based EMS development toolchain that integrates advanced algorithms with high-fidelity vehicle models, leveraging Python-MATLAB/Simulink co-simulation for agent training across Model-in-the-Loop (MiL), Hardware-in-the-Loop (HiL), and Vehicle-in-the-Loop (ViL) stages. Besides, the distributional soft actor-critic algorithm (DSAC) is applied to energy management for the first time, embedding the return distribution function into maximum entropy RL. This approach adapts the Q-value function update step size, significantly enhancing strategy performance. Additionally, two RL-based EMS frameworks are investigated: one where the agent directly outputs fuel cell power commands, and another where the agent generates equivalent factors (EF) for the equivalent consumption minimization strategy (ECMS). Simulation and experimental results validate that both RL frameworks achieve superior fuel economy, reducing hydrogen consumption by approximately 4.35 % to 5.73 % compared to benchmarks. By combining Python's algorithmic flexibility and scalability with MATLAB/Simulink's high-fidelity vehicle models, the proposed toolchain provides a robust foundation for real-vehicle applications of RL-based EMSs.</div></div>\",\"PeriodicalId\":246,\"journal\":{\"name\":\"Applied Energy\",\"volume\":\"393 \",\"pages\":\"Article 126030\"},\"PeriodicalIF\":10.1000,\"publicationDate\":\"2025-05-12\",\"publicationTypes\":\"Journal Article\",\"fieldsOfStudy\":null,\"isOpenAccess\":false,\"openAccessPdf\":\"\",\"citationCount\":\"0\",\"resultStr\":null,\"platform\":\"Semanticscholar\",\"paperid\":null,\"PeriodicalName\":\"Applied Energy\",\"FirstCategoryId\":\"5\",\"ListUrlMain\":\"https://www.sciencedirect.com/science/article/pii/S0306261925007603\",\"RegionNum\":1,\"RegionCategory\":\"工程技术\",\"ArticlePicture\":[],\"TitleCN\":null,\"AbstractTextCN\":null,\"PMCID\":null,\"EPubDate\":\"\",\"PubModel\":\"\",\"JCR\":\"Q1\",\"JCRName\":\"ENERGY & FUELS\",\"Score\":null,\"Total\":0}","platform":"Semanticscholar","paperid":null,"PeriodicalName":"Applied Energy","FirstCategoryId":"5","ListUrlMain":"https://www.sciencedirect.com/science/article/pii/S0306261925007603","RegionNum":1,"RegionCategory":"工程技术","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"Q1","JCRName":"ENERGY & FUELS","Score":null,"Total":0}
Sim-to-real design and development of reinforcement learning-based energy management strategies for fuel cell electric vehicles
The application of reinforcement learning (RL) algorithms in energy management strategies (EMSs) for fuel cell electric vehicles (FCEVs) has shown promising results in simulations. However, transitioning these strategies to real-vehicle implementation remains challenging due to the complexities of vehicle dynamics and system integration. Based on the General Optimal control Problems Solver (GOPS) platform, this paper establishes an RL-based EMS development toolchain that integrates advanced algorithms with high-fidelity vehicle models, leveraging Python-MATLAB/Simulink co-simulation for agent training across Model-in-the-Loop (MiL), Hardware-in-the-Loop (HiL), and Vehicle-in-the-Loop (ViL) stages. Besides, the distributional soft actor-critic algorithm (DSAC) is applied to energy management for the first time, embedding the return distribution function into maximum entropy RL. This approach adapts the Q-value function update step size, significantly enhancing strategy performance. Additionally, two RL-based EMS frameworks are investigated: one where the agent directly outputs fuel cell power commands, and another where the agent generates equivalent factors (EF) for the equivalent consumption minimization strategy (ECMS). Simulation and experimental results validate that both RL frameworks achieve superior fuel economy, reducing hydrogen consumption by approximately 4.35 % to 5.73 % compared to benchmarks. By combining Python's algorithmic flexibility and scalability with MATLAB/Simulink's high-fidelity vehicle models, the proposed toolchain provides a robust foundation for real-vehicle applications of RL-based EMSs.
期刊介绍:
Applied Energy serves as a platform for sharing innovations, research, development, and demonstrations in energy conversion, conservation, and sustainable energy systems. The journal covers topics such as optimal energy resource use, environmental pollutant mitigation, and energy process analysis. It welcomes original papers, review articles, technical notes, and letters to the editor. Authors are encouraged to submit manuscripts that bridge the gap between research, development, and implementation. The journal addresses a wide spectrum of topics, including fossil and renewable energy technologies, energy economics, and environmental impacts. Applied Energy also explores modeling and forecasting, conservation strategies, and the social and economic implications of energy policies, including climate change mitigation. It is complemented by the open-access journal Advances in Applied Energy.