Bidding Strategy for Two-Sided Electricity Markets: A Reinforcement Learning based Framework

Bala Suraj Pedasingu, E. Subramanian, Y. Bichpuriya, V. Sarangan, Nidhisha Mahilong
{"title":"Bidding Strategy for Two-Sided Electricity Markets: A Reinforcement Learning based Framework","authors":"Bala Suraj Pedasingu, E. Subramanian, Y. Bichpuriya, V. Sarangan, Nidhisha Mahilong","doi":"10.1145/3408308.3427976","DOIUrl":null,"url":null,"abstract":"We aim to increase the revenue or reduce the purchase cost of a given market participant in a double-sided, day-ahead, wholesale electricity market serving a smart city. Using an operations research based market clearing mechanism and attention based time series forecaster as sub-modules, we build a holistic interactive system. Through this system, we discover better bidding strategies for a market participant using reinforcement learning (RL). We relax several assumptions made in existing literature in order to make the problem setting more relevant to real life. Our Markov Decision Process (MDP) formulation enables us to tackle action space explosion and also compute optimal actions across time-steps in parallel. Our RL framework is generic enough to be used by either a generator or a consumer participating in the electricity market. We study the efficacy of the proposed RL based bidding framework from the perspective of a generator as well as a buyer on real world day-ahead electricity market data obtained from the European Power Exchange (EPEX). We compare the performance of our RL based bidding framework against three baselines: (a) an ideal but un-realizable bidding strategy; (b) a realizable approximate version of the ideal strategy; and (c) historical performance as found from the logs. Under both perspectives, we find that our RL based framework is more closer to the ideal strategy than other baselines. Further, the RL based framework improves the average daily revenue of the generator by nearly €7,200 (€2.64 M per year) and €9,000 (€3.28 M per year) over the realizable ideal and historical strategies respectively. When used on behalf of a buyer, it reduces average daily procurement cost by nearly €2,700 (€0.97 M per year) and €57,200 (€52.63 M per year) over the realizable ideal and historical strategies respectively. We also observe that our RL based framework automatically adapts its actions to changes in the market power of the participant.","PeriodicalId":287030,"journal":{"name":"Proceedings of the 7th ACM International Conference on Systems for Energy-Efficient Buildings, Cities, and Transportation","volume":"14 1","pages":"0"},"PeriodicalIF":0.0000,"publicationDate":"2020-11-18","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"6","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"Proceedings of the 7th ACM International Conference on Systems for Energy-Efficient Buildings, Cities, and Transportation","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/10.1145/3408308.3427976","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
引用次数: 6

Abstract

We aim to increase the revenue or reduce the purchase cost of a given market participant in a double-sided, day-ahead, wholesale electricity market serving a smart city. Using an operations research based market clearing mechanism and attention based time series forecaster as sub-modules, we build a holistic interactive system. Through this system, we discover better bidding strategies for a market participant using reinforcement learning (RL). We relax several assumptions made in existing literature in order to make the problem setting more relevant to real life. Our Markov Decision Process (MDP) formulation enables us to tackle action space explosion and also compute optimal actions across time-steps in parallel. Our RL framework is generic enough to be used by either a generator or a consumer participating in the electricity market. We study the efficacy of the proposed RL based bidding framework from the perspective of a generator as well as a buyer on real world day-ahead electricity market data obtained from the European Power Exchange (EPEX). We compare the performance of our RL based bidding framework against three baselines: (a) an ideal but un-realizable bidding strategy; (b) a realizable approximate version of the ideal strategy; and (c) historical performance as found from the logs. Under both perspectives, we find that our RL based framework is more closer to the ideal strategy than other baselines. Further, the RL based framework improves the average daily revenue of the generator by nearly €7,200 (€2.64 M per year) and €9,000 (€3.28 M per year) over the realizable ideal and historical strategies respectively. When used on behalf of a buyer, it reduces average daily procurement cost by nearly €2,700 (€0.97 M per year) and €57,200 (€52.63 M per year) over the realizable ideal and historical strategies respectively. We also observe that our RL based framework automatically adapts its actions to changes in the market power of the participant.
双边电力市场竞价策略:基于强化学习的框架
我们的目标是在一个为智慧城市服务的双边、提前一天的批发电力市场中,增加某一特定市场参与者的收入或降低其购买成本。以基于运筹学的市场出清机制和基于关注的时间序列预测为子模块,构建了一个整体的交互系统。通过这个系统,我们使用强化学习(RL)为市场参与者发现更好的竞标策略。为了使问题设置更贴近现实生活,我们放宽了现有文献中的几个假设。我们的马尔可夫决策过程(MDP)公式使我们能够处理行动空间爆炸,并并行计算跨时间步的最佳行动。我们的RL框架是通用的,无论是发电机还是参与电力市场的消费者都可以使用。本文以欧洲电力交易所(EPEX)的实时电力市场数据为基础,从发电商和买方的角度研究了基于RL的竞价框架的有效性。我们将基于RL的投标框架的性能与三个基线进行比较:(a)理想但无法实现的投标策略;(b)理想战略的可实现的近似版本;(c)从日志中发现的历史性能。在这两个角度下,我们发现基于强化学习的框架比其他基线更接近理想策略。此外,基于RL的框架比可实现的理想和历史策略分别提高了近7200欧元(每年264万欧元)和9000欧元(每年328万欧元)的发电机平均日收入。当代表买方使用时,它比可实现的理想策略和历史策略分别减少了近2700欧元(每年97万欧元)和57,200欧元(每年5263万欧元)的平均每日采购成本。我们还观察到,基于强化学习的框架会自动调整其行为以适应参与者市场力量的变化。
本文章由计算机程序翻译,如有差异,请以英文原文为准。
求助全文
约1分钟内获得全文 求助全文
来源期刊
自引率
0.00%
发文量
0
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
确定
请完成安全验证×
copy
已复制链接
快去分享给好友吧!
我知道了
右上角分享
点击右上角分享
0
联系我们:info@booksci.cn Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。 Copyright © 2023 布克学术 All rights reserved.
京ICP备2023020795号-1
ghs 京公网安备 11010802042870号
Book学术文献互助
Book学术文献互助群
群 号:481959085
Book学术官方微信