基于多代理强化学习方法的电动汽车可扩展协调能源管理

IF 1.9 4区 工程技术 Q3 ENGINEERING, ELECTRICAL & ELECTRONIC
Ruien Bian, Xiuchen Jiang, Guoying Zhao, Yadong Liu, Zhou Dai
{"title":"基于多代理强化学习方法的电动汽车可扩展协调能源管理","authors":"Ruien Bian,&nbsp;Xiuchen Jiang,&nbsp;Guoying Zhao,&nbsp;Yadong Liu,&nbsp;Zhou Dai","doi":"10.1155/2024/7765710","DOIUrl":null,"url":null,"abstract":"<div>\n <p>The electric vehicle (EV) has been popular in recent years, which also brings huge challenges to the distribution network due to its energy instability. In order to consider the economic factors of dispatching these distributed renewable resources, the voltage variation is also important. A novel model-free method is put forward for collaborative management of EV resources of aggregators in the distribution network. The economic costs and physical network constraints for this energy management issue are considered at the same time. A Multiagent Deep Deterministic Policy Gradient (MADDPG) algorithm is applied to learn the cooperative energy control strategies. A transfer learning technique is used to fine-tune the trained policy when more aggregators join in the network. The proposed method can achieve close results to the traditional optimization methods, while it takes less than one second to take control actions, making it is more suitable for real-time online energy management. Compared to other advanced reinforcement learning (RL) models, numerical simulations conducted on IEEE test cases greatly illustrate the effectiveness and superiority of the proposed method.</p>\n </div>","PeriodicalId":51293,"journal":{"name":"International Transactions on Electrical Energy Systems","volume":null,"pages":null},"PeriodicalIF":1.9000,"publicationDate":"2024-10-30","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://onlinelibrary.wiley.com/doi/epdf/10.1155/2024/7765710","citationCount":"0","resultStr":"{\"title\":\"A Scalable and Coordinated Energy Management for Electric Vehicles Based on Multiagent Reinforcement Learning Method\",\"authors\":\"Ruien Bian,&nbsp;Xiuchen Jiang,&nbsp;Guoying Zhao,&nbsp;Yadong Liu,&nbsp;Zhou Dai\",\"doi\":\"10.1155/2024/7765710\",\"DOIUrl\":null,\"url\":null,\"abstract\":\"<div>\\n <p>The electric vehicle (EV) has been popular in recent years, which also brings huge challenges to the distribution network due to its energy instability. In order to consider the economic factors of dispatching these distributed renewable resources, the voltage variation is also important. A novel model-free method is put forward for collaborative management of EV resources of aggregators in the distribution network. The economic costs and physical network constraints for this energy management issue are considered at the same time. A Multiagent Deep Deterministic Policy Gradient (MADDPG) algorithm is applied to learn the cooperative energy control strategies. A transfer learning technique is used to fine-tune the trained policy when more aggregators join in the network. The proposed method can achieve close results to the traditional optimization methods, while it takes less than one second to take control actions, making it is more suitable for real-time online energy management. Compared to other advanced reinforcement learning (RL) models, numerical simulations conducted on IEEE test cases greatly illustrate the effectiveness and superiority of the proposed method.</p>\\n </div>\",\"PeriodicalId\":51293,\"journal\":{\"name\":\"International Transactions on Electrical Energy Systems\",\"volume\":null,\"pages\":null},\"PeriodicalIF\":1.9000,\"publicationDate\":\"2024-10-30\",\"publicationTypes\":\"Journal Article\",\"fieldsOfStudy\":null,\"isOpenAccess\":false,\"openAccessPdf\":\"https://onlinelibrary.wiley.com/doi/epdf/10.1155/2024/7765710\",\"citationCount\":\"0\",\"resultStr\":null,\"platform\":\"Semanticscholar\",\"paperid\":null,\"PeriodicalName\":\"International Transactions on Electrical Energy Systems\",\"FirstCategoryId\":\"5\",\"ListUrlMain\":\"https://onlinelibrary.wiley.com/doi/10.1155/2024/7765710\",\"RegionNum\":4,\"RegionCategory\":\"工程技术\",\"ArticlePicture\":[],\"TitleCN\":null,\"AbstractTextCN\":null,\"PMCID\":null,\"EPubDate\":\"\",\"PubModel\":\"\",\"JCR\":\"Q3\",\"JCRName\":\"ENGINEERING, ELECTRICAL & ELECTRONIC\",\"Score\":null,\"Total\":0}","platform":"Semanticscholar","paperid":null,"PeriodicalName":"International Transactions on Electrical Energy Systems","FirstCategoryId":"5","ListUrlMain":"https://onlinelibrary.wiley.com/doi/10.1155/2024/7765710","RegionNum":4,"RegionCategory":"工程技术","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"Q3","JCRName":"ENGINEERING, ELECTRICAL & ELECTRONIC","Score":null,"Total":0}
引用次数: 0

摘要

近年来,电动汽车(EV)大行其道,由于其能量的不稳定性,也给配电网络带来了巨大的挑战。为了考虑这些分布式可再生资源调度的经济因素,电压变化也很重要。本文提出了一种新型的无模型方法,用于配电网中聚合器电动汽车资源的协同管理。同时考虑了该能源管理问题的经济成本和物理网络约束。应用多代理深度确定性策略梯度(MADDPG)算法来学习合作能源控制策略。当更多聚合器加入网络时,采用迁移学习技术对训练好的策略进行微调。所提出的方法能达到与传统优化方法接近的效果,而采取控制行动的时间不到一秒,因此更适合实时在线能源管理。与其他先进的强化学习(RL)模型相比,在 IEEE 测试用例上进行的数值模拟极大地说明了所提方法的有效性和优越性。
本文章由计算机程序翻译,如有差异,请以英文原文为准。

A Scalable and Coordinated Energy Management for Electric Vehicles Based on Multiagent Reinforcement Learning Method

A Scalable and Coordinated Energy Management for Electric Vehicles Based on Multiagent Reinforcement Learning Method

The electric vehicle (EV) has been popular in recent years, which also brings huge challenges to the distribution network due to its energy instability. In order to consider the economic factors of dispatching these distributed renewable resources, the voltage variation is also important. A novel model-free method is put forward for collaborative management of EV resources of aggregators in the distribution network. The economic costs and physical network constraints for this energy management issue are considered at the same time. A Multiagent Deep Deterministic Policy Gradient (MADDPG) algorithm is applied to learn the cooperative energy control strategies. A transfer learning technique is used to fine-tune the trained policy when more aggregators join in the network. The proposed method can achieve close results to the traditional optimization methods, while it takes less than one second to take control actions, making it is more suitable for real-time online energy management. Compared to other advanced reinforcement learning (RL) models, numerical simulations conducted on IEEE test cases greatly illustrate the effectiveness and superiority of the proposed method.

求助全文
通过发布文献求助,成功后即可免费获取论文全文。 去求助
来源期刊
International Transactions on Electrical Energy Systems
International Transactions on Electrical Energy Systems ENGINEERING, ELECTRICAL & ELECTRONIC-
CiteScore
6.70
自引率
8.70%
发文量
342
期刊介绍: International Transactions on Electrical Energy Systems publishes original research results on key advances in the generation, transmission, and distribution of electrical energy systems. Of particular interest are submissions concerning the modeling, analysis, optimization and control of advanced electric power systems. Manuscripts on topics of economics, finance, policies, insulation materials, low-voltage power electronics, plasmas, and magnetics will generally not be considered for review.
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
确定
请完成安全验证×
copy
已复制链接
快去分享给好友吧!
我知道了
右上角分享
点击右上角分享
0
联系我们:info@booksci.cn Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。 Copyright © 2023 布克学术 All rights reserved.
京ICP备2023020795号-1
ghs 京公网安备 11010802042870号
Book学术文献互助
Book学术文献互助群
群 号:481959085
Book学术官方微信