{"title":"基于q -学习的lcl耦合参数偏差逆变器有限控制集模型预测控制","authors":"Lei Zhang, Yunjian Peng, Weijie Sun, Jinze Li","doi":"10.1109/DDCLS58216.2023.10167014","DOIUrl":null,"url":null,"abstract":"Finite Control Set (FCS) Model Predictive Control (MPC), as an efficient method used for current tracking of LCL-Coupled three-phase inverters, runs into high computational complexity while finding its optimal version with a long predictive interval. For such a difficult problem we take a value function with discounted factors as an indicator to measure the pros and cons of control and propose a novel alternative method based on Q-learning algorithm. In the control scheme, the value function is approximated by reinforcement learning(RL) algorithm and furthermore, the long horizons prediction is transformed into an iterative multi-step matrix calculation. At the same time, the optimal switching position is directly obtained without a modulation link, which greatly reduces the computational complexity. Accordingly, a data-driven Q-learning algorithm is designed with a proof of convergence. Last, the proposed algorithm's performance in the case of complete deviation from the (unknown) system parameters is verified by simulations.","PeriodicalId":415532,"journal":{"name":"2023 IEEE 12th Data Driven Control and Learning Systems Conference (DDCLS)","volume":"156 1 1","pages":"0"},"PeriodicalIF":0.0000,"publicationDate":"2023-05-12","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":"{\"title\":\"Q-Learning-based Finite Control Set Model Predictive Control for LCL-Coupled Inverters with Deviated Parameters\",\"authors\":\"Lei Zhang, Yunjian Peng, Weijie Sun, Jinze Li\",\"doi\":\"10.1109/DDCLS58216.2023.10167014\",\"DOIUrl\":null,\"url\":null,\"abstract\":\"Finite Control Set (FCS) Model Predictive Control (MPC), as an efficient method used for current tracking of LCL-Coupled three-phase inverters, runs into high computational complexity while finding its optimal version with a long predictive interval. For such a difficult problem we take a value function with discounted factors as an indicator to measure the pros and cons of control and propose a novel alternative method based on Q-learning algorithm. In the control scheme, the value function is approximated by reinforcement learning(RL) algorithm and furthermore, the long horizons prediction is transformed into an iterative multi-step matrix calculation. At the same time, the optimal switching position is directly obtained without a modulation link, which greatly reduces the computational complexity. Accordingly, a data-driven Q-learning algorithm is designed with a proof of convergence. Last, the proposed algorithm's performance in the case of complete deviation from the (unknown) system parameters is verified by simulations.\",\"PeriodicalId\":415532,\"journal\":{\"name\":\"2023 IEEE 12th Data Driven Control and Learning Systems Conference (DDCLS)\",\"volume\":\"156 1 1\",\"pages\":\"0\"},\"PeriodicalIF\":0.0000,\"publicationDate\":\"2023-05-12\",\"publicationTypes\":\"Journal Article\",\"fieldsOfStudy\":null,\"isOpenAccess\":false,\"openAccessPdf\":\"\",\"citationCount\":\"0\",\"resultStr\":null,\"platform\":\"Semanticscholar\",\"paperid\":null,\"PeriodicalName\":\"2023 IEEE 12th Data Driven Control and Learning Systems Conference (DDCLS)\",\"FirstCategoryId\":\"1085\",\"ListUrlMain\":\"https://doi.org/10.1109/DDCLS58216.2023.10167014\",\"RegionNum\":0,\"RegionCategory\":null,\"ArticlePicture\":[],\"TitleCN\":null,\"AbstractTextCN\":null,\"PMCID\":null,\"EPubDate\":\"\",\"PubModel\":\"\",\"JCR\":\"\",\"JCRName\":\"\",\"Score\":null,\"Total\":0}","platform":"Semanticscholar","paperid":null,"PeriodicalName":"2023 IEEE 12th Data Driven Control and Learning Systems Conference (DDCLS)","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/10.1109/DDCLS58216.2023.10167014","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
Q-Learning-based Finite Control Set Model Predictive Control for LCL-Coupled Inverters with Deviated Parameters
Finite Control Set (FCS) Model Predictive Control (MPC), as an efficient method used for current tracking of LCL-Coupled three-phase inverters, runs into high computational complexity while finding its optimal version with a long predictive interval. For such a difficult problem we take a value function with discounted factors as an indicator to measure the pros and cons of control and propose a novel alternative method based on Q-learning algorithm. In the control scheme, the value function is approximated by reinforcement learning(RL) algorithm and furthermore, the long horizons prediction is transformed into an iterative multi-step matrix calculation. At the same time, the optimal switching position is directly obtained without a modulation link, which greatly reduces the computational complexity. Accordingly, a data-driven Q-learning algorithm is designed with a proof of convergence. Last, the proposed algorithm's performance in the case of complete deviation from the (unknown) system parameters is verified by simulations.