Yiwen Huang;Wentao Huang;Ran Li;Tao Huang;Canbing Li;Nengling Tai
{"title":"集装箱港口物流-能源时空协调调度策略生成的自适应MARL大模型","authors":"Yiwen Huang;Wentao Huang;Ran Li;Tao Huang;Canbing Li;Nengling Tai","doi":"10.1109/TSG.2025.3547830","DOIUrl":null,"url":null,"abstract":"Logistics-energy coordination significantly enhances energy efficiency in electrified seaports. However, daily changes in environment data necessitate the re-implementation of optimization procedures, causing huge computational burdens. This paper proposes an adaptive multi-agent reinforcement learning (MARL) large model for logistics-energy spatiotemporal coordination of container seaports. The well-trained large model can directly generate optimal policy for each operating day from environment data without re-solving. To achieve this, a comprehensive logistics-energy coordination model is first established considering the spatial and temporal constraints of all-electric ships (AESs), quay cranes (QCs), auto guided vehicles (AGVs), and the seaport power distribution network (SPDN). The model is formulated as a Markov Decision Process (MDP). Then a MARL large model is developed, involving a hypernetwork mapping environment data to optimal policy, and special structures for both hypernetwork and agent policy networks to adapt to any number of daily arrival AESs. Additionally, a cascading action modification layer is designed to ensure correct action outputs within complex spatiotemporal constraints. A tailored training method with two acceleration strategies are developed for the large model. Case studies illustrate that the large model after training can automatically generate optimal policies with little to no fine-tuning, outperforming existing methods that require extensive solution time.","PeriodicalId":13331,"journal":{"name":"IEEE Transactions on Smart Grid","volume":"16 3","pages":"2261-2277"},"PeriodicalIF":8.6000,"publicationDate":"2025-03-06","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":"{\"title\":\"An Adaptive MARL Large Model for Dispatch Strategy Generation in Logistics-Energy Spatiotemporal Coordination of Container Seaports\",\"authors\":\"Yiwen Huang;Wentao Huang;Ran Li;Tao Huang;Canbing Li;Nengling Tai\",\"doi\":\"10.1109/TSG.2025.3547830\",\"DOIUrl\":null,\"url\":null,\"abstract\":\"Logistics-energy coordination significantly enhances energy efficiency in electrified seaports. However, daily changes in environment data necessitate the re-implementation of optimization procedures, causing huge computational burdens. This paper proposes an adaptive multi-agent reinforcement learning (MARL) large model for logistics-energy spatiotemporal coordination of container seaports. The well-trained large model can directly generate optimal policy for each operating day from environment data without re-solving. To achieve this, a comprehensive logistics-energy coordination model is first established considering the spatial and temporal constraints of all-electric ships (AESs), quay cranes (QCs), auto guided vehicles (AGVs), and the seaport power distribution network (SPDN). The model is formulated as a Markov Decision Process (MDP). Then a MARL large model is developed, involving a hypernetwork mapping environment data to optimal policy, and special structures for both hypernetwork and agent policy networks to adapt to any number of daily arrival AESs. Additionally, a cascading action modification layer is designed to ensure correct action outputs within complex spatiotemporal constraints. A tailored training method with two acceleration strategies are developed for the large model. Case studies illustrate that the large model after training can automatically generate optimal policies with little to no fine-tuning, outperforming existing methods that require extensive solution time.\",\"PeriodicalId\":13331,\"journal\":{\"name\":\"IEEE Transactions on Smart Grid\",\"volume\":\"16 3\",\"pages\":\"2261-2277\"},\"PeriodicalIF\":8.6000,\"publicationDate\":\"2025-03-06\",\"publicationTypes\":\"Journal Article\",\"fieldsOfStudy\":null,\"isOpenAccess\":false,\"openAccessPdf\":\"\",\"citationCount\":\"0\",\"resultStr\":null,\"platform\":\"Semanticscholar\",\"paperid\":null,\"PeriodicalName\":\"IEEE Transactions on Smart Grid\",\"FirstCategoryId\":\"5\",\"ListUrlMain\":\"https://ieeexplore.ieee.org/document/10915539/\",\"RegionNum\":1,\"RegionCategory\":\"工程技术\",\"ArticlePicture\":[],\"TitleCN\":null,\"AbstractTextCN\":null,\"PMCID\":null,\"EPubDate\":\"\",\"PubModel\":\"\",\"JCR\":\"Q1\",\"JCRName\":\"ENGINEERING, ELECTRICAL & ELECTRONIC\",\"Score\":null,\"Total\":0}","platform":"Semanticscholar","paperid":null,"PeriodicalName":"IEEE Transactions on Smart Grid","FirstCategoryId":"5","ListUrlMain":"https://ieeexplore.ieee.org/document/10915539/","RegionNum":1,"RegionCategory":"工程技术","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"Q1","JCRName":"ENGINEERING, ELECTRICAL & ELECTRONIC","Score":null,"Total":0}
An Adaptive MARL Large Model for Dispatch Strategy Generation in Logistics-Energy Spatiotemporal Coordination of Container Seaports
Logistics-energy coordination significantly enhances energy efficiency in electrified seaports. However, daily changes in environment data necessitate the re-implementation of optimization procedures, causing huge computational burdens. This paper proposes an adaptive multi-agent reinforcement learning (MARL) large model for logistics-energy spatiotemporal coordination of container seaports. The well-trained large model can directly generate optimal policy for each operating day from environment data without re-solving. To achieve this, a comprehensive logistics-energy coordination model is first established considering the spatial and temporal constraints of all-electric ships (AESs), quay cranes (QCs), auto guided vehicles (AGVs), and the seaport power distribution network (SPDN). The model is formulated as a Markov Decision Process (MDP). Then a MARL large model is developed, involving a hypernetwork mapping environment data to optimal policy, and special structures for both hypernetwork and agent policy networks to adapt to any number of daily arrival AESs. Additionally, a cascading action modification layer is designed to ensure correct action outputs within complex spatiotemporal constraints. A tailored training method with two acceleration strategies are developed for the large model. Case studies illustrate that the large model after training can automatically generate optimal policies with little to no fine-tuning, outperforming existing methods that require extensive solution time.
期刊介绍:
The IEEE Transactions on Smart Grid is a multidisciplinary journal that focuses on research and development in the field of smart grid technology. It covers various aspects of the smart grid, including energy networks, prosumers (consumers who also produce energy), electric transportation, distributed energy resources, and communications. The journal also addresses the integration of microgrids and active distribution networks with transmission systems. It publishes original research on smart grid theories and principles, including technologies and systems for demand response, Advance Metering Infrastructure, cyber-physical systems, multi-energy systems, transactive energy, data analytics, and electric vehicle integration. Additionally, the journal considers surveys of existing work on the smart grid that propose new perspectives on the history and future of intelligent and active grids.