{"title":"Adaptive transfer reinforcement learning (TRL) for cooling water systems with uniform agent design and multi-agent coordination","authors":"Zhechao Wang, Zhihong Pang","doi":"10.1016/j.enbuild.2025.116071","DOIUrl":null,"url":null,"abstract":"<div><div>Transfer Reinforcement Learning (TRL) offers a promising approach to optimizing building cooling water systems by improving both energy efficiency and operational effectiveness. This study introduces a novel TRL framework designed to accelerate the learning process of Reinforcement Learning (RL) agents by systematically leveraging prior experience from analogous systems. Unlike conventional RL approaches that start from scratch, our framework enhances initial performance while mitigating negative transfer through an adaptive multi-agent supervision mechanism. Our methodology involves three key innovations. First, we collect field data from three cooling water systems and train individual data-driven models to enable realistic energy and control simulations. Next, we design a uniform RL model with ratio-based inputs and outputs, ensuring transferability across systems with varying characteristics. This model includes two controllers—one for cooling towers and the other for cooling water pumps. Third, we introduce a structured TRL process in which a pre-trained RL model from a source system is transferred to two target systems. Within this transfer framework, we integrate three types of agents: one completely new agent and two variants of the trained RL model. A key feature of the framework is a supervision mechanism that coordinates these agents for positive transfer. It dynamically adjusts the selection probability of each agent through a constantly updated preference function and modifies learning objectives at different training stages. Various scenarios are tested to evaluate the framework’s performance with different transferred agents and learning stages. Simulation results demonstrate a 10 % improvement in energy savings—both initially and at convergence—compared to learning from scratch. Moreover, the proposed TRL framework effectively mitigates negative transfer and avoids converging to the suboptimal performance of a transferred agent. More importantly, it significantly reduces the effort required to select appropriate source systems, highlighting its practical applicability and potential for widespread adoption in building cooling water system optimization.</div></div>","PeriodicalId":11641,"journal":{"name":"Energy and Buildings","volume":"345 ","pages":"Article 116071"},"PeriodicalIF":7.1000,"publicationDate":"2025-07-05","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"Energy and Buildings","FirstCategoryId":"5","ListUrlMain":"https://www.sciencedirect.com/science/article/pii/S0378778825008011","RegionNum":2,"RegionCategory":"工程技术","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"Q1","JCRName":"CONSTRUCTION & BUILDING TECHNOLOGY","Score":null,"Total":0}
引用次数: 0
Abstract
Transfer Reinforcement Learning (TRL) offers a promising approach to optimizing building cooling water systems by improving both energy efficiency and operational effectiveness. This study introduces a novel TRL framework designed to accelerate the learning process of Reinforcement Learning (RL) agents by systematically leveraging prior experience from analogous systems. Unlike conventional RL approaches that start from scratch, our framework enhances initial performance while mitigating negative transfer through an adaptive multi-agent supervision mechanism. Our methodology involves three key innovations. First, we collect field data from three cooling water systems and train individual data-driven models to enable realistic energy and control simulations. Next, we design a uniform RL model with ratio-based inputs and outputs, ensuring transferability across systems with varying characteristics. This model includes two controllers—one for cooling towers and the other for cooling water pumps. Third, we introduce a structured TRL process in which a pre-trained RL model from a source system is transferred to two target systems. Within this transfer framework, we integrate three types of agents: one completely new agent and two variants of the trained RL model. A key feature of the framework is a supervision mechanism that coordinates these agents for positive transfer. It dynamically adjusts the selection probability of each agent through a constantly updated preference function and modifies learning objectives at different training stages. Various scenarios are tested to evaluate the framework’s performance with different transferred agents and learning stages. Simulation results demonstrate a 10 % improvement in energy savings—both initially and at convergence—compared to learning from scratch. Moreover, the proposed TRL framework effectively mitigates negative transfer and avoids converging to the suboptimal performance of a transferred agent. More importantly, it significantly reduces the effort required to select appropriate source systems, highlighting its practical applicability and potential for widespread adoption in building cooling water system optimization.
期刊介绍:
An international journal devoted to investigations of energy use and efficiency in buildings
Energy and Buildings is an international journal publishing articles with explicit links to energy use in buildings. The aim is to present new research results, and new proven practice aimed at reducing the energy needs of a building and improving indoor environment quality.