Distributed multi-agent reinforcement learning approach for energy-saving optimization under disturbance conditions

IF 8.3 1区 工程技术 Q1 ECONOMICS
Dahan Wang , Jianjun Wu , Ximing Chang , Haodong Yin
{"title":"Distributed multi-agent reinforcement learning approach for energy-saving optimization under disturbance conditions","authors":"Dahan Wang ,&nbsp;Jianjun Wu ,&nbsp;Ximing Chang ,&nbsp;Haodong Yin","doi":"10.1016/j.tre.2025.104180","DOIUrl":null,"url":null,"abstract":"<div><div>Urban rail transit systems exhibit substantial energy consumption, underpinning the significance of energy-saving optimization strategies for train timetables. Conventionally, trains operate according to an energy-efficient timetable formulated offline. However, station incidents and disturbances often result in deviations from the planned schedule, leading to additional energy expenditure. To address this challenge, the current study introduces a distributed multi-agent reinforcement learning approach(DMARL) for real-time energy-efficient optimization of train timetables. Initially, trains are conceptualized as agents, adopting the Actor-Critic network structure as the learning paradigm, with a distributed deployment architecture facilitating the training of the model. During the interaction phase between agents and the subway system, a progressive reward mechanism is designed to encourage efficient exploratory actions by the agents. In the final case study, data from Shanghai Metro Line 1(SML1) was utilized to demonstrate the effectiveness of the proposed method. The results indicate that when disturbances occur at stations, necessitating extended stop times, the method presented in this paper exhibited stable performance and faster convergence rates in both two-train and three-train systems. Compared to the energy consumption without any action, the energy savings were enhanced by 14.11 % and 11 %, respectively. The timetable updates were completed within milliseconds, confirming the efficacy of the method and its compliance with real-time updating requirements.</div></div>","PeriodicalId":49418,"journal":{"name":"Transportation Research Part E-Logistics and Transportation Review","volume":"200 ","pages":"Article 104180"},"PeriodicalIF":8.3000,"publicationDate":"2025-05-21","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"Transportation Research Part E-Logistics and Transportation Review","FirstCategoryId":"5","ListUrlMain":"https://www.sciencedirect.com/science/article/pii/S1366554525002212","RegionNum":1,"RegionCategory":"工程技术","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"Q1","JCRName":"ECONOMICS","Score":null,"Total":0}
引用次数: 0

Abstract

Urban rail transit systems exhibit substantial energy consumption, underpinning the significance of energy-saving optimization strategies for train timetables. Conventionally, trains operate according to an energy-efficient timetable formulated offline. However, station incidents and disturbances often result in deviations from the planned schedule, leading to additional energy expenditure. To address this challenge, the current study introduces a distributed multi-agent reinforcement learning approach(DMARL) for real-time energy-efficient optimization of train timetables. Initially, trains are conceptualized as agents, adopting the Actor-Critic network structure as the learning paradigm, with a distributed deployment architecture facilitating the training of the model. During the interaction phase between agents and the subway system, a progressive reward mechanism is designed to encourage efficient exploratory actions by the agents. In the final case study, data from Shanghai Metro Line 1(SML1) was utilized to demonstrate the effectiveness of the proposed method. The results indicate that when disturbances occur at stations, necessitating extended stop times, the method presented in this paper exhibited stable performance and faster convergence rates in both two-train and three-train systems. Compared to the energy consumption without any action, the energy savings were enhanced by 14.11 % and 11 %, respectively. The timetable updates were completed within milliseconds, confirming the efficacy of the method and its compliance with real-time updating requirements.
扰动条件下分布式多智能体强化学习节能优化方法
城市轨道交通系统能耗巨大,因此列车时刻表的节能优化策略具有重要意义。传统上,列车是根据离线制定的节能时间表运行的。然而,车站事件和干扰往往导致偏离计划的时间表,导致额外的能源消耗。为了应对这一挑战,本研究引入了一种分布式多智能体强化学习方法(DMARL),用于列车时刻表的实时节能优化。最初,训练被概念化为代理,采用Actor-Critic网络结构作为学习范式,采用分布式部署架构促进模型的训练。在agent与地铁系统交互阶段,设计了累进奖励机制,鼓励agent进行有效的探索性行为。在最后的案例研究中,使用上海地铁1号线(SML1)的数据来证明所提出方法的有效性。结果表明,当车站发生干扰,需要延长停车时间时,本文提出的方法在两列和三列系统中都表现出稳定的性能和更快的收敛速度。与不采取任何行动的能耗相比,节能效果分别提高了14.11%和11%。时间表更新在几毫秒内完成,证实了该方法的有效性及其符合实时更新要求。
本文章由计算机程序翻译,如有差异,请以英文原文为准。
求助全文
约1分钟内获得全文 求助全文
来源期刊
CiteScore
16.20
自引率
16.00%
发文量
285
审稿时长
62 days
期刊介绍: Transportation Research Part E: Logistics and Transportation Review is a reputable journal that publishes high-quality articles covering a wide range of topics in the field of logistics and transportation research. The journal welcomes submissions on various subjects, including transport economics, transport infrastructure and investment appraisal, evaluation of public policies related to transportation, empirical and analytical studies of logistics management practices and performance, logistics and operations models, and logistics and supply chain management. Part E aims to provide informative and well-researched articles that contribute to the understanding and advancement of the field. The content of the journal is complementary to other prestigious journals in transportation research, such as Transportation Research Part A: Policy and Practice, Part B: Methodological, Part C: Emerging Technologies, Part D: Transport and Environment, and Part F: Traffic Psychology and Behaviour. Together, these journals form a comprehensive and cohesive reference for current research in transportation science.
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
确定
请完成安全验证×
copy
已复制链接
快去分享给好友吧!
我知道了
右上角分享
点击右上角分享
0
联系我们:info@booksci.cn Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。 Copyright © 2023 布克学术 All rights reserved.
京ICP备2023020795号-1
ghs 京公网安备 11010802042870号
Book学术文献互助
Book学术文献互助群
群 号:481959085
Book学术官方微信