{"title":"Behind-the-Meter Energy Storage System Operation via Hindsight Experience Replay","authors":"Zhenhuan Ding, Wentao Li, Xiaoyu Zhang, Qianqian Zhang","doi":"10.1109/CIEEC58067.2023.10166167","DOIUrl":null,"url":null,"abstract":"A typical use case for Behind-the-Meter Battery Energy Storage Systems (BESS) is to manage demand charges. However, the sparse nature of demand charges presents a challenge when using reinforcement learning algorithms. To address this issue, we propose a method based on Hindsight Experience Replay which overcomes the problems associated with sparse rewards in BESS operation. The main idea of the Hindsight Experience Replay is to develop an additional reward counting process from the existing trajectory which already could handle some non-sparse rewards. And the developed reward provides a direction guidance to the final operation destination, which demolish the difficulties on sparse reward training compared to conventional reinforcement learning. Our proposed method has been verified for its feasibility and effectiveness through simulations conducted in a charge station modeled after the IEEE 13-node test feeder.","PeriodicalId":185921,"journal":{"name":"2023 IEEE 6th International Electrical and Energy Conference (CIEEC)","volume":"110 1","pages":"0"},"PeriodicalIF":0.0000,"publicationDate":"2023-05-12","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"2023 IEEE 6th International Electrical and Energy Conference (CIEEC)","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/10.1109/CIEEC58067.2023.10166167","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
引用次数: 0
Abstract
A typical use case for Behind-the-Meter Battery Energy Storage Systems (BESS) is to manage demand charges. However, the sparse nature of demand charges presents a challenge when using reinforcement learning algorithms. To address this issue, we propose a method based on Hindsight Experience Replay which overcomes the problems associated with sparse rewards in BESS operation. The main idea of the Hindsight Experience Replay is to develop an additional reward counting process from the existing trajectory which already could handle some non-sparse rewards. And the developed reward provides a direction guidance to the final operation destination, which demolish the difficulties on sparse reward training compared to conventional reinforcement learning. Our proposed method has been verified for its feasibility and effectiveness through simulations conducted in a charge station modeled after the IEEE 13-node test feeder.