{"title":"基于深度强化学习的分时电价节能调度","authors":"Caixia Wang;Wenxuan Fang;Xin Dai;Renchu He;Wei Du;Yang Tang","doi":"10.1109/TASE.2025.3536198","DOIUrl":null,"url":null,"abstract":"In the context of carbon peaking and carbon neutrality, green production scheduling that considers energy has attracted increasing attention. Reinforcement learning (RL) has emerged as a topic for developing efficient algorithms for solving complex combinatorial optimization problems, such as real-world shop scheduling problems. In this paper, we investigate the energy-conscious scheduling problem (ECSP) under time-of-use (TOU) electricity price with the goal of minimizing both the waiting time and extra electricity cost. A mathematical model is formulated, and an efficient deep RL (DRL)-based optimization method is proposed to solve the problem effectively. We design a novel ECSP network (ECSPNet) tailored to handle various ECSP scales based on the characteristics of the problem. Moreover, the Tchebycheff decomposition method is used to solve the multi-objective optimization problems, complemented by the application of the policy gradient method from reinforcement learning to train the ECSPNet without size limitations. Experiments verify that the proposed ECSPNet outperforms state-of-the-art methods and is computationally efficient, even on instances of larger scales unseen in training. Real-world case studies reveal that the proposed method can reduce annual total electricity cost by approximately 30% while effectively maximizing production efficiency. Note to Practitioners —Enhancing energy consciousness alongside improving production efficiency in manufacturing systems has increasingly become a focal point for both academia and industry, especially with the growing emphasis on environmental concerns and advancing industrialization. Time-of-use (TOU) electricity price policy is widely implemented to effectively balance electricity supply and demand. For business managers, appropriately responding to this policy by optimizing scheduling tasks can significantly reduce energy costs. This paper addresses a novel energy-conscious scheduling problem (ECSP) under TOU electricity price, specifically arising from the electrode graphitization production process in graphite material manufacturing. We propose a deep reinforcement learning-based energy-saving scheduling optimization method, which integrates considerations for both production efficiency and energy cost indicators. By combining the unique characteristics of ECSP with the decision-making capabilities of reinforcement learning and the perception capabilities of deep learning, our method excels in solution speed, efficiency, and adaptability. The effectiveness and practical applicability of our method have been demonstrated through experiments on actual enterprise cases of various scales. The rapidly generated optimization scheduling plans can assist enterprises in rationally arranging scheduling tasks while minimizing electricity cost. Looking ahead, our proposed method has the potential to address a wide range of energy-conscious scheduling problems under TOU electricity price policy.","PeriodicalId":51060,"journal":{"name":"IEEE Transactions on Automation Science and Engineering","volume":"22 ","pages":"11490-11504"},"PeriodicalIF":6.4000,"publicationDate":"2025-01-29","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":"{\"title\":\"Deep Reinforcement Learning-Based Energy-Conscious Scheduling Under Time-of-Use Electricity Price\",\"authors\":\"Caixia Wang;Wenxuan Fang;Xin Dai;Renchu He;Wei Du;Yang Tang\",\"doi\":\"10.1109/TASE.2025.3536198\",\"DOIUrl\":null,\"url\":null,\"abstract\":\"In the context of carbon peaking and carbon neutrality, green production scheduling that considers energy has attracted increasing attention. Reinforcement learning (RL) has emerged as a topic for developing efficient algorithms for solving complex combinatorial optimization problems, such as real-world shop scheduling problems. In this paper, we investigate the energy-conscious scheduling problem (ECSP) under time-of-use (TOU) electricity price with the goal of minimizing both the waiting time and extra electricity cost. A mathematical model is formulated, and an efficient deep RL (DRL)-based optimization method is proposed to solve the problem effectively. We design a novel ECSP network (ECSPNet) tailored to handle various ECSP scales based on the characteristics of the problem. Moreover, the Tchebycheff decomposition method is used to solve the multi-objective optimization problems, complemented by the application of the policy gradient method from reinforcement learning to train the ECSPNet without size limitations. Experiments verify that the proposed ECSPNet outperforms state-of-the-art methods and is computationally efficient, even on instances of larger scales unseen in training. Real-world case studies reveal that the proposed method can reduce annual total electricity cost by approximately 30% while effectively maximizing production efficiency. Note to Practitioners —Enhancing energy consciousness alongside improving production efficiency in manufacturing systems has increasingly become a focal point for both academia and industry, especially with the growing emphasis on environmental concerns and advancing industrialization. Time-of-use (TOU) electricity price policy is widely implemented to effectively balance electricity supply and demand. For business managers, appropriately responding to this policy by optimizing scheduling tasks can significantly reduce energy costs. This paper addresses a novel energy-conscious scheduling problem (ECSP) under TOU electricity price, specifically arising from the electrode graphitization production process in graphite material manufacturing. We propose a deep reinforcement learning-based energy-saving scheduling optimization method, which integrates considerations for both production efficiency and energy cost indicators. By combining the unique characteristics of ECSP with the decision-making capabilities of reinforcement learning and the perception capabilities of deep learning, our method excels in solution speed, efficiency, and adaptability. The effectiveness and practical applicability of our method have been demonstrated through experiments on actual enterprise cases of various scales. The rapidly generated optimization scheduling plans can assist enterprises in rationally arranging scheduling tasks while minimizing electricity cost. Looking ahead, our proposed method has the potential to address a wide range of energy-conscious scheduling problems under TOU electricity price policy.\",\"PeriodicalId\":51060,\"journal\":{\"name\":\"IEEE Transactions on Automation Science and Engineering\",\"volume\":\"22 \",\"pages\":\"11490-11504\"},\"PeriodicalIF\":6.4000,\"publicationDate\":\"2025-01-29\",\"publicationTypes\":\"Journal Article\",\"fieldsOfStudy\":null,\"isOpenAccess\":false,\"openAccessPdf\":\"\",\"citationCount\":\"0\",\"resultStr\":null,\"platform\":\"Semanticscholar\",\"paperid\":null,\"PeriodicalName\":\"IEEE Transactions on Automation Science and Engineering\",\"FirstCategoryId\":\"94\",\"ListUrlMain\":\"https://ieeexplore.ieee.org/document/10857376/\",\"RegionNum\":2,\"RegionCategory\":\"计算机科学\",\"ArticlePicture\":[],\"TitleCN\":null,\"AbstractTextCN\":null,\"PMCID\":null,\"EPubDate\":\"\",\"PubModel\":\"\",\"JCR\":\"Q1\",\"JCRName\":\"AUTOMATION & CONTROL SYSTEMS\",\"Score\":null,\"Total\":0}","platform":"Semanticscholar","paperid":null,"PeriodicalName":"IEEE Transactions on Automation Science and Engineering","FirstCategoryId":"94","ListUrlMain":"https://ieeexplore.ieee.org/document/10857376/","RegionNum":2,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"Q1","JCRName":"AUTOMATION & CONTROL SYSTEMS","Score":null,"Total":0}
Deep Reinforcement Learning-Based Energy-Conscious Scheduling Under Time-of-Use Electricity Price
In the context of carbon peaking and carbon neutrality, green production scheduling that considers energy has attracted increasing attention. Reinforcement learning (RL) has emerged as a topic for developing efficient algorithms for solving complex combinatorial optimization problems, such as real-world shop scheduling problems. In this paper, we investigate the energy-conscious scheduling problem (ECSP) under time-of-use (TOU) electricity price with the goal of minimizing both the waiting time and extra electricity cost. A mathematical model is formulated, and an efficient deep RL (DRL)-based optimization method is proposed to solve the problem effectively. We design a novel ECSP network (ECSPNet) tailored to handle various ECSP scales based on the characteristics of the problem. Moreover, the Tchebycheff decomposition method is used to solve the multi-objective optimization problems, complemented by the application of the policy gradient method from reinforcement learning to train the ECSPNet without size limitations. Experiments verify that the proposed ECSPNet outperforms state-of-the-art methods and is computationally efficient, even on instances of larger scales unseen in training. Real-world case studies reveal that the proposed method can reduce annual total electricity cost by approximately 30% while effectively maximizing production efficiency. Note to Practitioners —Enhancing energy consciousness alongside improving production efficiency in manufacturing systems has increasingly become a focal point for both academia and industry, especially with the growing emphasis on environmental concerns and advancing industrialization. Time-of-use (TOU) electricity price policy is widely implemented to effectively balance electricity supply and demand. For business managers, appropriately responding to this policy by optimizing scheduling tasks can significantly reduce energy costs. This paper addresses a novel energy-conscious scheduling problem (ECSP) under TOU electricity price, specifically arising from the electrode graphitization production process in graphite material manufacturing. We propose a deep reinforcement learning-based energy-saving scheduling optimization method, which integrates considerations for both production efficiency and energy cost indicators. By combining the unique characteristics of ECSP with the decision-making capabilities of reinforcement learning and the perception capabilities of deep learning, our method excels in solution speed, efficiency, and adaptability. The effectiveness and practical applicability of our method have been demonstrated through experiments on actual enterprise cases of various scales. The rapidly generated optimization scheduling plans can assist enterprises in rationally arranging scheduling tasks while minimizing electricity cost. Looking ahead, our proposed method has the potential to address a wide range of energy-conscious scheduling problems under TOU electricity price policy.
期刊介绍:
The IEEE Transactions on Automation Science and Engineering (T-ASE) publishes fundamental papers on Automation, emphasizing scientific results that advance efficiency, quality, productivity, and reliability. T-ASE encourages interdisciplinary approaches from computer science, control systems, electrical engineering, mathematics, mechanical engineering, operations research, and other fields. T-ASE welcomes results relevant to industries such as agriculture, biotechnology, healthcare, home automation, maintenance, manufacturing, pharmaceuticals, retail, security, service, supply chains, and transportation. T-ASE addresses a research community willing to integrate knowledge across disciplines and industries. For this purpose, each paper includes a Note to Practitioners that summarizes how its results can be applied or how they might be extended to apply in practice.