基于深度确定性策略梯度的分布式能源优化协调

Avijit Das, Di Wu
{"title":"基于深度确定性策略梯度的分布式能源优化协调","authors":"Avijit Das, Di Wu","doi":"10.1109/EESAT55007.2022.9998046","DOIUrl":null,"url":null,"abstract":"Recent studies have shown that reinforcement learning (RL) is a promising approach for coordination and control of distributed energy resources (DERS) under uncertainties. Many existing RL approaches, including Q-learning and approximate dynamic programming, are based on lookup table methods, which become inefficient when the problem size is large and infeasible when continuous states and actions are involved. In addition, when modeling battery energy storage systems (BESSS), the loss of life is not reasonably considered in the decision-making process. This paper proposes an innovative deep RL method for DER coordination considering BESS degradation. The proposed deep RL is designed based on an adaptive actor-critic architecture and employs an off-policy deterministic policy gradient method for determining the dispatch operation that minimizes the operation cost and BESS loss of life. Case studies were performed to validate the proposed method and demonstrate the effects of incorporating degradation models into control design.","PeriodicalId":310250,"journal":{"name":"2022 IEEE Electrical Energy Storage Application and Technologies Conference (EESAT)","volume":"54 1","pages":"0"},"PeriodicalIF":0.0000,"publicationDate":"2022-11-08","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":"{\"title\":\"Optimal Coordination of Distributed Energy Resources Using Deep Deterministic Policy Gradient\",\"authors\":\"Avijit Das, Di Wu\",\"doi\":\"10.1109/EESAT55007.2022.9998046\",\"DOIUrl\":null,\"url\":null,\"abstract\":\"Recent studies have shown that reinforcement learning (RL) is a promising approach for coordination and control of distributed energy resources (DERS) under uncertainties. Many existing RL approaches, including Q-learning and approximate dynamic programming, are based on lookup table methods, which become inefficient when the problem size is large and infeasible when continuous states and actions are involved. In addition, when modeling battery energy storage systems (BESSS), the loss of life is not reasonably considered in the decision-making process. This paper proposes an innovative deep RL method for DER coordination considering BESS degradation. The proposed deep RL is designed based on an adaptive actor-critic architecture and employs an off-policy deterministic policy gradient method for determining the dispatch operation that minimizes the operation cost and BESS loss of life. Case studies were performed to validate the proposed method and demonstrate the effects of incorporating degradation models into control design.\",\"PeriodicalId\":310250,\"journal\":{\"name\":\"2022 IEEE Electrical Energy Storage Application and Technologies Conference (EESAT)\",\"volume\":\"54 1\",\"pages\":\"0\"},\"PeriodicalIF\":0.0000,\"publicationDate\":\"2022-11-08\",\"publicationTypes\":\"Journal Article\",\"fieldsOfStudy\":null,\"isOpenAccess\":false,\"openAccessPdf\":\"\",\"citationCount\":\"0\",\"resultStr\":null,\"platform\":\"Semanticscholar\",\"paperid\":null,\"PeriodicalName\":\"2022 IEEE Electrical Energy Storage Application and Technologies Conference (EESAT)\",\"FirstCategoryId\":\"1085\",\"ListUrlMain\":\"https://doi.org/10.1109/EESAT55007.2022.9998046\",\"RegionNum\":0,\"RegionCategory\":null,\"ArticlePicture\":[],\"TitleCN\":null,\"AbstractTextCN\":null,\"PMCID\":null,\"EPubDate\":\"\",\"PubModel\":\"\",\"JCR\":\"\",\"JCRName\":\"\",\"Score\":null,\"Total\":0}","platform":"Semanticscholar","paperid":null,"PeriodicalName":"2022 IEEE Electrical Energy Storage Application and Technologies Conference (EESAT)","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/10.1109/EESAT55007.2022.9998046","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
引用次数: 0

摘要

近年来的研究表明,强化学习(RL)是一种很有前途的不确定条件下分布式能源(DERS)协调与控制方法。许多现有的强化学习方法,包括q学习和近似动态规划,都是基于查找表方法,当问题规模很大时,这种方法效率低下,当涉及连续状态和动作时,这种方法不可行。此外,在对电池储能系统(BESSS)进行建模时,在决策过程中没有合理考虑寿命损失。提出了一种考虑BESS退化的深度强化学习DER协调方法。所提出的深度强化学习是基于自适应行为者批判体系结构设计的,并采用非策略确定性策略梯度方法来确定调度操作,从而使运行成本和BESS生命损失最小化。案例研究验证了所提出的方法,并证明了将退化模型纳入控制设计的效果。
本文章由计算机程序翻译,如有差异,请以英文原文为准。
Optimal Coordination of Distributed Energy Resources Using Deep Deterministic Policy Gradient
Recent studies have shown that reinforcement learning (RL) is a promising approach for coordination and control of distributed energy resources (DERS) under uncertainties. Many existing RL approaches, including Q-learning and approximate dynamic programming, are based on lookup table methods, which become inefficient when the problem size is large and infeasible when continuous states and actions are involved. In addition, when modeling battery energy storage systems (BESSS), the loss of life is not reasonably considered in the decision-making process. This paper proposes an innovative deep RL method for DER coordination considering BESS degradation. The proposed deep RL is designed based on an adaptive actor-critic architecture and employs an off-policy deterministic policy gradient method for determining the dispatch operation that minimizes the operation cost and BESS loss of life. Case studies were performed to validate the proposed method and demonstrate the effects of incorporating degradation models into control design.
求助全文
通过发布文献求助,成功后即可免费获取论文全文。 去求助
来源期刊
自引率
0.00%
发文量
0
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
确定
请完成安全验证×
copy
已复制链接
快去分享给好友吧!
我知道了
右上角分享
点击右上角分享
0
联系我们:info@booksci.cn Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。 Copyright © 2023 布克学术 All rights reserved.
京ICP备2023020795号-1
ghs 京公网安备 11010802042870号
Book学术文献互助
Book学术文献互助群
群 号:481959085
Book学术官方微信