Imitation reinforcement learning energy management for electric vehicles with hybrid energy storage system

IF 10.1 1区 工程技术 Q1 ENERGY & FUELS
Weirong Liu , Pengfei Yao , Yue Wu , Lijun Duan , Heng Li , Jun Peng
{"title":"Imitation reinforcement learning energy management for electric vehicles with hybrid energy storage system","authors":"Weirong Liu ,&nbsp;Pengfei Yao ,&nbsp;Yue Wu ,&nbsp;Lijun Duan ,&nbsp;Heng Li ,&nbsp;Jun Peng","doi":"10.1016/j.apenergy.2024.124832","DOIUrl":null,"url":null,"abstract":"<div><div>Deep reinforcement learning has become a promising method for the energy management of electric vehicles. However, deep reinforcement learning relies on a large amount of trial-and-error training to acquire near-optimal performance. An adversarial imitation reinforcement learning energy management strategy is proposed for electric vehicles with hybrid energy storage system to minimize the cost of battery capacity loss. Firstly, the reinforcement learning exploration is guided by expert knowledge, which is generated by dynamic programming under various standard driving conditions. The expert knowledge is represented as the optimal power allocation mapping. Secondly, at the early imitation stage, the action of the reinforcement learning agent approaches the optimal power allocation mapping rapidly by using adversarial networks. Thirdly, a dynamic imitation weight is developed according to the Discriminator of adversarial networks, making the agent transit to self-explore the near-optimal power allocation under online driving conditions. Results demonstrate that the proposed strategy can accelerate the training by 42.60% while enhancing the reward by 15.79% compared with traditional reinforcement learning. Under different test driving cycles, the proposed method can further reduce the battery capacity loss cost by 5.1%–12.4%.</div></div>","PeriodicalId":246,"journal":{"name":"Applied Energy","volume":"378 ","pages":"Article 124832"},"PeriodicalIF":10.1000,"publicationDate":"2024-11-11","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"Applied Energy","FirstCategoryId":"5","ListUrlMain":"https://www.sciencedirect.com/science/article/pii/S0306261924022153","RegionNum":1,"RegionCategory":"工程技术","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"Q1","JCRName":"ENERGY & FUELS","Score":null,"Total":0}
引用次数: 0

Abstract

Deep reinforcement learning has become a promising method for the energy management of electric vehicles. However, deep reinforcement learning relies on a large amount of trial-and-error training to acquire near-optimal performance. An adversarial imitation reinforcement learning energy management strategy is proposed for electric vehicles with hybrid energy storage system to minimize the cost of battery capacity loss. Firstly, the reinforcement learning exploration is guided by expert knowledge, which is generated by dynamic programming under various standard driving conditions. The expert knowledge is represented as the optimal power allocation mapping. Secondly, at the early imitation stage, the action of the reinforcement learning agent approaches the optimal power allocation mapping rapidly by using adversarial networks. Thirdly, a dynamic imitation weight is developed according to the Discriminator of adversarial networks, making the agent transit to self-explore the near-optimal power allocation under online driving conditions. Results demonstrate that the proposed strategy can accelerate the training by 42.60% while enhancing the reward by 15.79% compared with traditional reinforcement learning. Under different test driving cycles, the proposed method can further reduce the battery capacity loss cost by 5.1%–12.4%.
混合动力储能系统电动汽车的仿真强化学习能源管理
深度强化学习已成为电动汽车能源管理的一种前景广阔的方法。然而,深度强化学习需要依赖大量的试错训练才能获得接近最优的性能。本文提出了一种对抗式模仿强化学习能源管理策略,适用于采用混合储能系统的电动汽车,以最大限度地降低电池容量损失的成本。首先,强化学习探索由专家知识指导,专家知识是在各种标准驾驶条件下通过动态编程生成的。专家知识表现为最优功率分配映射。其次,在早期模仿阶段,强化学习代理的行动通过对抗网络迅速接近最优功率分配映射。第三,根据对抗网络的判别器(Discriminator)开发动态模仿权重,使代理能够在在线驾驶条件下自我探索接近最优的功率分配。结果表明,与传统的强化学习相比,所提出的策略可以加快 42.60% 的训练速度,同时提高 15.79% 的奖励。在不同的测试驾驶周期下,所提出的方法可进一步降低电池容量损耗成本 5.1%-12.4%。
本文章由计算机程序翻译,如有差异,请以英文原文为准。
求助全文
约1分钟内获得全文 求助全文
来源期刊
Applied Energy
Applied Energy 工程技术-工程:化工
CiteScore
21.20
自引率
10.70%
发文量
1830
审稿时长
41 days
期刊介绍: Applied Energy serves as a platform for sharing innovations, research, development, and demonstrations in energy conversion, conservation, and sustainable energy systems. The journal covers topics such as optimal energy resource use, environmental pollutant mitigation, and energy process analysis. It welcomes original papers, review articles, technical notes, and letters to the editor. Authors are encouraged to submit manuscripts that bridge the gap between research, development, and implementation. The journal addresses a wide spectrum of topics, including fossil and renewable energy technologies, energy economics, and environmental impacts. Applied Energy also explores modeling and forecasting, conservation strategies, and the social and economic implications of energy policies, including climate change mitigation. It is complemented by the open-access journal Advances in Applied Energy.
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
确定
请完成安全验证×
copy
已复制链接
快去分享给好友吧!
我知道了
右上角分享
点击右上角分享
0
联系我们:info@booksci.cn Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。 Copyright © 2023 布克学术 All rights reserved.
京ICP备2023020795号-1
ghs 京公网安备 11010802042870号
Book学术文献互助
Book学术文献互助群
群 号:481959085
Book学术官方微信