Hairun Xu , Ao Zhang , Qingle Wang , Yang Hu , Fang Fang , Long Cheng
{"title":"Quantum Reinforcement Learning for real-time optimization in Electric Vehicle charging systems","authors":"Hairun Xu , Ao Zhang , Qingle Wang , Yang Hu , Fang Fang , Long Cheng","doi":"10.1016/j.apenergy.2025.125279","DOIUrl":null,"url":null,"abstract":"<div><div>The rapid growth of electric vehicles (EVs) presents new challenges for EV charging scheduling, particularly due to the unpredictable nature of charging demand and the dynamic availability of resources. Currently, Deep Reinforcement Learning (DRL) has become a critical technology for improving scheduling efficiency. At the same time, advancements in quantum computing have led to Quantum Neural Networks (QNNs), which use the superposition states of quantum bits for more efficient information encoding. Building on these advancements, this study explores Quantum Reinforcement Learning (QRL) for EV charging systems. We propose a method called QRL-based Electric Vehicle Charging Scheduling (Q-EVCS) to optimize charging resource allocation based on real-time user demand. This approach aims to reduce average charging service times, increase the service success rate, and lower operational costs. We provide the detailed design and implementation of our approach, and our experimental results demonstrate that Q-EVCS maintains performance levels comparable to the DRL-based method while significantly reducing the number of model parameters.</div></div>","PeriodicalId":246,"journal":{"name":"Applied Energy","volume":"383 ","pages":"Article 125279"},"PeriodicalIF":10.1000,"publicationDate":"2025-01-28","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"Applied Energy","FirstCategoryId":"5","ListUrlMain":"https://www.sciencedirect.com/science/article/pii/S0306261925000091","RegionNum":1,"RegionCategory":"工程技术","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"Q1","JCRName":"ENERGY & FUELS","Score":null,"Total":0}
引用次数: 0
Abstract
The rapid growth of electric vehicles (EVs) presents new challenges for EV charging scheduling, particularly due to the unpredictable nature of charging demand and the dynamic availability of resources. Currently, Deep Reinforcement Learning (DRL) has become a critical technology for improving scheduling efficiency. At the same time, advancements in quantum computing have led to Quantum Neural Networks (QNNs), which use the superposition states of quantum bits for more efficient information encoding. Building on these advancements, this study explores Quantum Reinforcement Learning (QRL) for EV charging systems. We propose a method called QRL-based Electric Vehicle Charging Scheduling (Q-EVCS) to optimize charging resource allocation based on real-time user demand. This approach aims to reduce average charging service times, increase the service success rate, and lower operational costs. We provide the detailed design and implementation of our approach, and our experimental results demonstrate that Q-EVCS maintains performance levels comparable to the DRL-based method while significantly reducing the number of model parameters.
期刊介绍:
Applied Energy serves as a platform for sharing innovations, research, development, and demonstrations in energy conversion, conservation, and sustainable energy systems. The journal covers topics such as optimal energy resource use, environmental pollutant mitigation, and energy process analysis. It welcomes original papers, review articles, technical notes, and letters to the editor. Authors are encouraged to submit manuscripts that bridge the gap between research, development, and implementation. The journal addresses a wide spectrum of topics, including fossil and renewable energy technologies, energy economics, and environmental impacts. Applied Energy also explores modeling and forecasting, conservation strategies, and the social and economic implications of energy policies, including climate change mitigation. It is complemented by the open-access journal Advances in Applied Energy.