Zhanhong Huang , Tao Yu , Zhenning Pan , Bairong Deng , Xuehan Zhang , Yufeng Wu , Qiaoyi Ding
{"title":"通过上下文元图强化学习实现具有高泛化和少量适应性的随机动态电力调度","authors":"Zhanhong Huang , Tao Yu , Zhenning Pan , Bairong Deng , Xuehan Zhang , Yufeng Wu , Qiaoyi Ding","doi":"10.1016/j.ijepes.2024.110272","DOIUrl":null,"url":null,"abstract":"<div><div>Reinforcement learning, as an efficient method for solving uncertainty decision making in power systems, is widely used in multi-stage stochastic power dispatch and dynamic optimization. However, the low generalization and practicality of traditional reinforcement learning algorithms limit their online application. The dispatch strategy learned offline can only adapt to specific scenarios, and its policy performance degrades significantly if the sample drastically change or the topology variation. To fill these gaps, a novel contextual meta graph reinforcement learning (Meta-GRL) method a more general contextual Markov decision process (CMDP) modeling are proposed. The proposed Meta-GRL adopts CMDP scheme and graph representation, extracts and encodes the differentiated scene context, and can be extended to various scene changes. The upper meta-learner embedded in context in Meta-GRL is proposed to realize scene recognition. While the lower base-earner is guided to learn generalized context-specified policy. The test results in IEEE39 and open environment show that the Meta-GRL achieves more than 90% optimization and entire period applicability under the premise of saving computing resources.</div></div>","PeriodicalId":50326,"journal":{"name":"International Journal of Electrical Power & Energy Systems","volume":"162 ","pages":""},"PeriodicalIF":5.0000,"publicationDate":"2024-10-16","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":"{\"title\":\"Stochastic dynamic power dispatch with high generalization and few-shot adaption via contextual meta graph reinforcement learning\",\"authors\":\"Zhanhong Huang , Tao Yu , Zhenning Pan , Bairong Deng , Xuehan Zhang , Yufeng Wu , Qiaoyi Ding\",\"doi\":\"10.1016/j.ijepes.2024.110272\",\"DOIUrl\":null,\"url\":null,\"abstract\":\"<div><div>Reinforcement learning, as an efficient method for solving uncertainty decision making in power systems, is widely used in multi-stage stochastic power dispatch and dynamic optimization. However, the low generalization and practicality of traditional reinforcement learning algorithms limit their online application. The dispatch strategy learned offline can only adapt to specific scenarios, and its policy performance degrades significantly if the sample drastically change or the topology variation. To fill these gaps, a novel contextual meta graph reinforcement learning (Meta-GRL) method a more general contextual Markov decision process (CMDP) modeling are proposed. The proposed Meta-GRL adopts CMDP scheme and graph representation, extracts and encodes the differentiated scene context, and can be extended to various scene changes. The upper meta-learner embedded in context in Meta-GRL is proposed to realize scene recognition. While the lower base-earner is guided to learn generalized context-specified policy. The test results in IEEE39 and open environment show that the Meta-GRL achieves more than 90% optimization and entire period applicability under the premise of saving computing resources.</div></div>\",\"PeriodicalId\":50326,\"journal\":{\"name\":\"International Journal of Electrical Power & Energy Systems\",\"volume\":\"162 \",\"pages\":\"\"},\"PeriodicalIF\":5.0000,\"publicationDate\":\"2024-10-16\",\"publicationTypes\":\"Journal Article\",\"fieldsOfStudy\":null,\"isOpenAccess\":false,\"openAccessPdf\":\"\",\"citationCount\":\"0\",\"resultStr\":null,\"platform\":\"Semanticscholar\",\"paperid\":null,\"PeriodicalName\":\"International Journal of Electrical Power & Energy Systems\",\"FirstCategoryId\":\"5\",\"ListUrlMain\":\"https://www.sciencedirect.com/science/article/pii/S0142061524004940\",\"RegionNum\":2,\"RegionCategory\":\"工程技术\",\"ArticlePicture\":[],\"TitleCN\":null,\"AbstractTextCN\":null,\"PMCID\":null,\"EPubDate\":\"\",\"PubModel\":\"\",\"JCR\":\"Q1\",\"JCRName\":\"ENGINEERING, ELECTRICAL & ELECTRONIC\",\"Score\":null,\"Total\":0}","platform":"Semanticscholar","paperid":null,"PeriodicalName":"International Journal of Electrical Power & Energy Systems","FirstCategoryId":"5","ListUrlMain":"https://www.sciencedirect.com/science/article/pii/S0142061524004940","RegionNum":2,"RegionCategory":"工程技术","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"Q1","JCRName":"ENGINEERING, ELECTRICAL & ELECTRONIC","Score":null,"Total":0}
Stochastic dynamic power dispatch with high generalization and few-shot adaption via contextual meta graph reinforcement learning
Reinforcement learning, as an efficient method for solving uncertainty decision making in power systems, is widely used in multi-stage stochastic power dispatch and dynamic optimization. However, the low generalization and practicality of traditional reinforcement learning algorithms limit their online application. The dispatch strategy learned offline can only adapt to specific scenarios, and its policy performance degrades significantly if the sample drastically change or the topology variation. To fill these gaps, a novel contextual meta graph reinforcement learning (Meta-GRL) method a more general contextual Markov decision process (CMDP) modeling are proposed. The proposed Meta-GRL adopts CMDP scheme and graph representation, extracts and encodes the differentiated scene context, and can be extended to various scene changes. The upper meta-learner embedded in context in Meta-GRL is proposed to realize scene recognition. While the lower base-earner is guided to learn generalized context-specified policy. The test results in IEEE39 and open environment show that the Meta-GRL achieves more than 90% optimization and entire period applicability under the premise of saving computing resources.
期刊介绍:
The journal covers theoretical developments in electrical power and energy systems and their applications. The coverage embraces: generation and network planning; reliability; long and short term operation; expert systems; neural networks; object oriented systems; system control centres; database and information systems; stock and parameter estimation; system security and adequacy; network theory, modelling and computation; small and large system dynamics; dynamic model identification; on-line control including load and switching control; protection; distribution systems; energy economics; impact of non-conventional systems; and man-machine interfaces.
As well as original research papers, the journal publishes short contributions, book reviews and conference reports. All papers are peer-reviewed by at least two referees.