Tianhao Liu;Chunhua Yang;Can Zhou;Yonggang Li;Bei Sun
{"title":"A Reinforcement Learning Control Method for Process Industry Based on Implicit and Explicit Knowledge Extraction and Embedding","authors":"Tianhao Liu;Chunhua Yang;Can Zhou;Yonggang Li;Bei Sun","doi":"10.1109/TSMC.2025.3559766","DOIUrl":null,"url":null,"abstract":"The process industry is a key manufacturing process that consumes a vast amount of energy consumption. On the premise of ensuring process stability, controlling process variables to operate the process close to the optimal working condition plays a critical role in reducing energy consumption. Reinforcement learning (RL), using trial and error to learn control strategies, has received much attention. However, the substantial fluctuations of process variables and the switching delay gap of the process industry result in a high-dimension state-action space, making it difficult to learn control strategies efficiently, and there is no guarantee of control stability. To get around these issues, first, a generic knowledge-extracted method for process industry RL control is proposed. It does not require laborious expert knowledge acquisition processes. Second, to improve learning efficiency, the implicit knowledge is extracted using decision trees from operation trajectory data and embedded into agent controllers. Third, an explicit knowledge-oriented reward constructing method is designed to guarantee control stability. A case of the zinc electrowinning process is provided to validate its superiority. The result shows that it can reduce power consumption while stabilizing process variables within the spec limits, without a laborious expert knowledge acquisition process.","PeriodicalId":48915,"journal":{"name":"IEEE Transactions on Systems Man Cybernetics-Systems","volume":"55 8","pages":"5152-5165"},"PeriodicalIF":8.6000,"publicationDate":"2025-04-30","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"IEEE Transactions on Systems Man Cybernetics-Systems","FirstCategoryId":"94","ListUrlMain":"https://ieeexplore.ieee.org/document/10980644/","RegionNum":1,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"Q1","JCRName":"AUTOMATION & CONTROL SYSTEMS","Score":null,"Total":0}
引用次数: 0
Abstract
The process industry is a key manufacturing process that consumes a vast amount of energy consumption. On the premise of ensuring process stability, controlling process variables to operate the process close to the optimal working condition plays a critical role in reducing energy consumption. Reinforcement learning (RL), using trial and error to learn control strategies, has received much attention. However, the substantial fluctuations of process variables and the switching delay gap of the process industry result in a high-dimension state-action space, making it difficult to learn control strategies efficiently, and there is no guarantee of control stability. To get around these issues, first, a generic knowledge-extracted method for process industry RL control is proposed. It does not require laborious expert knowledge acquisition processes. Second, to improve learning efficiency, the implicit knowledge is extracted using decision trees from operation trajectory data and embedded into agent controllers. Third, an explicit knowledge-oriented reward constructing method is designed to guarantee control stability. A case of the zinc electrowinning process is provided to validate its superiority. The result shows that it can reduce power consumption while stabilizing process variables within the spec limits, without a laborious expert knowledge acquisition process.
期刊介绍:
The IEEE Transactions on Systems, Man, and Cybernetics: Systems encompasses the fields of systems engineering, covering issue formulation, analysis, and modeling throughout the systems engineering lifecycle phases. It addresses decision-making, issue interpretation, systems management, processes, and various methods such as optimization, modeling, and simulation in the development and deployment of large systems.