{"title":"基于多智能体深度强化学习的污水处理过程多变量控制","authors":"Shengli Du, Rui Sun, Peixi Chen","doi":"10.1049/cth2.70021","DOIUrl":null,"url":null,"abstract":"<p>This paper investigates the multivariable control of wastewater treatment processes (WWTP). This paper integrates deep reinforcement learning (DRL) with PID control and proposes a multivariable adaptive PID control strategy based on multi-agent DRL (MADRL) for WWTP. The approach begins with the construction of a MADRL-PID controller structure, consisting of an agent and a PID controller module. The agent adjusts the PID controller values while the PID module calculates the control signal. To enhance the agent's ability to cooperatively tune multiple PID controllers, the algorithm's components–reward function, action space, environment, and state space–are designed according to the BSM1 simulation platform principles and the MADRL framework requirements. Additionally, to handle WWTP's non-linearities, uncertainties, and parameter coupling, the multi-agent deep deterministic policy gradient algorithm is selected as the foundation for training the agents. Experimental results demonstrate that the proposed algorithm exhibits greater adaptability than traditional PID control and achieves superior control performance.</p>","PeriodicalId":50382,"journal":{"name":"IET Control Theory and Applications","volume":"19 1","pages":""},"PeriodicalIF":2.2000,"publicationDate":"2025-04-20","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://onlinelibrary.wiley.com/doi/epdf/10.1049/cth2.70021","citationCount":"0","resultStr":"{\"title\":\"Multivariable Control of Wastewater Treatment Process Based on Multi-Agent Deep Reinforcement Learning\",\"authors\":\"Shengli Du, Rui Sun, Peixi Chen\",\"doi\":\"10.1049/cth2.70021\",\"DOIUrl\":null,\"url\":null,\"abstract\":\"<p>This paper investigates the multivariable control of wastewater treatment processes (WWTP). This paper integrates deep reinforcement learning (DRL) with PID control and proposes a multivariable adaptive PID control strategy based on multi-agent DRL (MADRL) for WWTP. The approach begins with the construction of a MADRL-PID controller structure, consisting of an agent and a PID controller module. The agent adjusts the PID controller values while the PID module calculates the control signal. To enhance the agent's ability to cooperatively tune multiple PID controllers, the algorithm's components–reward function, action space, environment, and state space–are designed according to the BSM1 simulation platform principles and the MADRL framework requirements. Additionally, to handle WWTP's non-linearities, uncertainties, and parameter coupling, the multi-agent deep deterministic policy gradient algorithm is selected as the foundation for training the agents. Experimental results demonstrate that the proposed algorithm exhibits greater adaptability than traditional PID control and achieves superior control performance.</p>\",\"PeriodicalId\":50382,\"journal\":{\"name\":\"IET Control Theory and Applications\",\"volume\":\"19 1\",\"pages\":\"\"},\"PeriodicalIF\":2.2000,\"publicationDate\":\"2025-04-20\",\"publicationTypes\":\"Journal Article\",\"fieldsOfStudy\":null,\"isOpenAccess\":false,\"openAccessPdf\":\"https://onlinelibrary.wiley.com/doi/epdf/10.1049/cth2.70021\",\"citationCount\":\"0\",\"resultStr\":null,\"platform\":\"Semanticscholar\",\"paperid\":null,\"PeriodicalName\":\"IET Control Theory and Applications\",\"FirstCategoryId\":\"94\",\"ListUrlMain\":\"https://onlinelibrary.wiley.com/doi/10.1049/cth2.70021\",\"RegionNum\":4,\"RegionCategory\":\"计算机科学\",\"ArticlePicture\":[],\"TitleCN\":null,\"AbstractTextCN\":null,\"PMCID\":null,\"EPubDate\":\"\",\"PubModel\":\"\",\"JCR\":\"Q2\",\"JCRName\":\"AUTOMATION & CONTROL SYSTEMS\",\"Score\":null,\"Total\":0}","platform":"Semanticscholar","paperid":null,"PeriodicalName":"IET Control Theory and Applications","FirstCategoryId":"94","ListUrlMain":"https://onlinelibrary.wiley.com/doi/10.1049/cth2.70021","RegionNum":4,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"Q2","JCRName":"AUTOMATION & CONTROL SYSTEMS","Score":null,"Total":0}
Multivariable Control of Wastewater Treatment Process Based on Multi-Agent Deep Reinforcement Learning
This paper investigates the multivariable control of wastewater treatment processes (WWTP). This paper integrates deep reinforcement learning (DRL) with PID control and proposes a multivariable adaptive PID control strategy based on multi-agent DRL (MADRL) for WWTP. The approach begins with the construction of a MADRL-PID controller structure, consisting of an agent and a PID controller module. The agent adjusts the PID controller values while the PID module calculates the control signal. To enhance the agent's ability to cooperatively tune multiple PID controllers, the algorithm's components–reward function, action space, environment, and state space–are designed according to the BSM1 simulation platform principles and the MADRL framework requirements. Additionally, to handle WWTP's non-linearities, uncertainties, and parameter coupling, the multi-agent deep deterministic policy gradient algorithm is selected as the foundation for training the agents. Experimental results demonstrate that the proposed algorithm exhibits greater adaptability than traditional PID control and achieves superior control performance.
期刊介绍:
IET Control Theory & Applications is devoted to control systems in the broadest sense, covering new theoretical results and the applications of new and established control methods. Among the topics of interest are system modelling, identification and simulation, the analysis and design of control systems (including computer-aided design), and practical implementation. The scope encompasses technological, economic, physiological (biomedical) and other systems, including man-machine interfaces.
Most of the papers published deal with original work from industrial and government laboratories and universities, but subject reviews and tutorial expositions of current methods are welcomed. Correspondence discussing published papers is also welcomed.
Applications papers need not necessarily involve new theory. Papers which describe new realisations of established methods, or control techniques applied in a novel situation, or practical studies which compare various designs, would be of interest. Of particular value are theoretical papers which discuss the applicability of new work or applications which engender new theoretical applications.