Chao Wang , Yaofei Zhang , Sherong Zhang , Xiaohua Wang , Zhiyong Zhao
{"title":"A multi-agent reinforcement learning model incorporating historical operational data for pump station operational scheduling","authors":"Chao Wang , Yaofei Zhang , Sherong Zhang , Xiaohua Wang , Zhiyong Zhao","doi":"10.1016/j.cor.2025.107221","DOIUrl":null,"url":null,"abstract":"<div><div>The development of operational scheduling for pump units is a critical focus in pump station management. This study introduces a pump station operation scheduling model based on multi-agent reinforcement learning that integrates historical operational data, addressing inefficiencies, poor generalization, and operational complexity encountered with traditional evolutionary algorithms. Utilizing a graph neural network, the model incorporates historical data and prior knowledge about pump station performance curves to establish a performance computation model for multi-unit operation combinations. The scheduling of multi-unit operations at a pump station is conceptualized as a parallel decision-making problem, incorporating rules aimed at cost reduction, efficiency improvement, and meeting water delivery volume requirements while minimizing operational complexity. A MARL model is developed, taking into account the variability in initial operating conditions to enhance generalization capabilities. The study compares the performance of various reinforcement learning models with evolutionary algorithms. Results indicate that the trained MARL model adapts effectively to dynamic water delivery conditions and exhibits strong generalization capabilities. Compared to actual operational scheduling, it achieves significant savings of over 419,000 units in operational costs and over 390,000 kWh in energy consumption. Furthermore, compared to evolutionary algorithms, the decision-making solutions generated by the reinforcement learning model align more closely with operational logic and are more efficient, achieving a planning speed increase of over 70 times.</div></div>","PeriodicalId":10542,"journal":{"name":"Computers & Operations Research","volume":"184 ","pages":"Article 107221"},"PeriodicalIF":4.3000,"publicationDate":"2025-07-26","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"Computers & Operations Research","FirstCategoryId":"5","ListUrlMain":"https://www.sciencedirect.com/science/article/pii/S0305054825002503","RegionNum":2,"RegionCategory":"工程技术","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"Q2","JCRName":"COMPUTER SCIENCE, INTERDISCIPLINARY APPLICATIONS","Score":null,"Total":0}
引用次数: 0
Abstract
The development of operational scheduling for pump units is a critical focus in pump station management. This study introduces a pump station operation scheduling model based on multi-agent reinforcement learning that integrates historical operational data, addressing inefficiencies, poor generalization, and operational complexity encountered with traditional evolutionary algorithms. Utilizing a graph neural network, the model incorporates historical data and prior knowledge about pump station performance curves to establish a performance computation model for multi-unit operation combinations. The scheduling of multi-unit operations at a pump station is conceptualized as a parallel decision-making problem, incorporating rules aimed at cost reduction, efficiency improvement, and meeting water delivery volume requirements while minimizing operational complexity. A MARL model is developed, taking into account the variability in initial operating conditions to enhance generalization capabilities. The study compares the performance of various reinforcement learning models with evolutionary algorithms. Results indicate that the trained MARL model adapts effectively to dynamic water delivery conditions and exhibits strong generalization capabilities. Compared to actual operational scheduling, it achieves significant savings of over 419,000 units in operational costs and over 390,000 kWh in energy consumption. Furthermore, compared to evolutionary algorithms, the decision-making solutions generated by the reinforcement learning model align more closely with operational logic and are more efficient, achieving a planning speed increase of over 70 times.
期刊介绍:
Operations research and computers meet in a large number of scientific fields, many of which are of vital current concern to our troubled society. These include, among others, ecology, transportation, safety, reliability, urban planning, economics, inventory control, investment strategy and logistics (including reverse logistics). Computers & Operations Research provides an international forum for the application of computers and operations research techniques to problems in these and related fields.