Yan Lin;Liqin Xiao;Yiyu Tao;Yijin Zhang;Feng Shu;Jun Li
{"title":"多智能体计算-车辆边缘计算中的能效优化:非合作与合作解决方案","authors":"Yan Lin;Liqin Xiao;Yiyu Tao;Yijin Zhang;Feng Shu;Jun Li","doi":"10.1109/TWC.2025.3547377","DOIUrl":null,"url":null,"abstract":"Vehicular edge computing (VEC) has driven the proliferation of computation-intensive and delay-sensitive vehicular services by deploying computing and energy resources at the edge. However, the exploitation of edge resources faces challenges due to unpredictable environmental dynamics and partial observability. To this end, this paper investigates the computing energy efficiency (CEE) problem in twin-timescale VEC scenarios by dynamically adjusting the offloading policy. Based upon modeling the problem as a decentralized partially observable Markov decision process (Dec-POMDP), a pair of non-cooperative and cooperative offloading solutions are proposed relying on multi-agent reinforcement learning (MARL), respectively. Specifically, the non-cooperative solution employs multi-agent independent proximal policy optimization (IPPO) to enable vehicular user equipments (VUEs) to learn their policies in a fully distributed manner without any information sharing. By contrast, the cooperative solution combines the multi-agent shared PPO with graph attention networks (MAPPO-GAT), where the relationship among agents is learned cooperatively and the historical learning experience is shared. Additionally, we compare the computational complexity and analyze the convergence. Simulation results show that in terms of the trade-off between offloading delay and offloading energy consumption, the proposed cooperative solution is superior to the non-cooperative counterpart with the cost of moderate training overhead for cooperative learning.","PeriodicalId":13431,"journal":{"name":"IEEE Transactions on Wireless Communications","volume":"24 7","pages":"5461-5476"},"PeriodicalIF":10.7000,"publicationDate":"2025-03-11","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":"{\"title\":\"Multi-Agent Computing-Energy-Efficiency Optimization in Vehicular Edge Computing: Non-Cooperative Versus Cooperative Solutions\",\"authors\":\"Yan Lin;Liqin Xiao;Yiyu Tao;Yijin Zhang;Feng Shu;Jun Li\",\"doi\":\"10.1109/TWC.2025.3547377\",\"DOIUrl\":null,\"url\":null,\"abstract\":\"Vehicular edge computing (VEC) has driven the proliferation of computation-intensive and delay-sensitive vehicular services by deploying computing and energy resources at the edge. However, the exploitation of edge resources faces challenges due to unpredictable environmental dynamics and partial observability. To this end, this paper investigates the computing energy efficiency (CEE) problem in twin-timescale VEC scenarios by dynamically adjusting the offloading policy. Based upon modeling the problem as a decentralized partially observable Markov decision process (Dec-POMDP), a pair of non-cooperative and cooperative offloading solutions are proposed relying on multi-agent reinforcement learning (MARL), respectively. Specifically, the non-cooperative solution employs multi-agent independent proximal policy optimization (IPPO) to enable vehicular user equipments (VUEs) to learn their policies in a fully distributed manner without any information sharing. By contrast, the cooperative solution combines the multi-agent shared PPO with graph attention networks (MAPPO-GAT), where the relationship among agents is learned cooperatively and the historical learning experience is shared. Additionally, we compare the computational complexity and analyze the convergence. Simulation results show that in terms of the trade-off between offloading delay and offloading energy consumption, the proposed cooperative solution is superior to the non-cooperative counterpart with the cost of moderate training overhead for cooperative learning.\",\"PeriodicalId\":13431,\"journal\":{\"name\":\"IEEE Transactions on Wireless Communications\",\"volume\":\"24 7\",\"pages\":\"5461-5476\"},\"PeriodicalIF\":10.7000,\"publicationDate\":\"2025-03-11\",\"publicationTypes\":\"Journal Article\",\"fieldsOfStudy\":null,\"isOpenAccess\":false,\"openAccessPdf\":\"\",\"citationCount\":\"0\",\"resultStr\":null,\"platform\":\"Semanticscholar\",\"paperid\":null,\"PeriodicalName\":\"IEEE Transactions on Wireless Communications\",\"FirstCategoryId\":\"94\",\"ListUrlMain\":\"https://ieeexplore.ieee.org/document/10923641/\",\"RegionNum\":1,\"RegionCategory\":\"计算机科学\",\"ArticlePicture\":[],\"TitleCN\":null,\"AbstractTextCN\":null,\"PMCID\":null,\"EPubDate\":\"\",\"PubModel\":\"\",\"JCR\":\"Q1\",\"JCRName\":\"ENGINEERING, ELECTRICAL & ELECTRONIC\",\"Score\":null,\"Total\":0}","platform":"Semanticscholar","paperid":null,"PeriodicalName":"IEEE Transactions on Wireless Communications","FirstCategoryId":"94","ListUrlMain":"https://ieeexplore.ieee.org/document/10923641/","RegionNum":1,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"Q1","JCRName":"ENGINEERING, ELECTRICAL & ELECTRONIC","Score":null,"Total":0}
Multi-Agent Computing-Energy-Efficiency Optimization in Vehicular Edge Computing: Non-Cooperative Versus Cooperative Solutions
Vehicular edge computing (VEC) has driven the proliferation of computation-intensive and delay-sensitive vehicular services by deploying computing and energy resources at the edge. However, the exploitation of edge resources faces challenges due to unpredictable environmental dynamics and partial observability. To this end, this paper investigates the computing energy efficiency (CEE) problem in twin-timescale VEC scenarios by dynamically adjusting the offloading policy. Based upon modeling the problem as a decentralized partially observable Markov decision process (Dec-POMDP), a pair of non-cooperative and cooperative offloading solutions are proposed relying on multi-agent reinforcement learning (MARL), respectively. Specifically, the non-cooperative solution employs multi-agent independent proximal policy optimization (IPPO) to enable vehicular user equipments (VUEs) to learn their policies in a fully distributed manner without any information sharing. By contrast, the cooperative solution combines the multi-agent shared PPO with graph attention networks (MAPPO-GAT), where the relationship among agents is learned cooperatively and the historical learning experience is shared. Additionally, we compare the computational complexity and analyze the convergence. Simulation results show that in terms of the trade-off between offloading delay and offloading energy consumption, the proposed cooperative solution is superior to the non-cooperative counterpart with the cost of moderate training overhead for cooperative learning.
期刊介绍:
The IEEE Transactions on Wireless Communications is a prestigious publication that showcases cutting-edge advancements in wireless communications. It welcomes both theoretical and practical contributions in various areas. The scope of the Transactions encompasses a wide range of topics, including modulation and coding, detection and estimation, propagation and channel characterization, and diversity techniques. The journal also emphasizes the physical and link layer communication aspects of network architectures and protocols.
The journal is open to papers on specific topics or non-traditional topics related to specific application areas. This includes simulation tools and methodologies, orthogonal frequency division multiplexing, MIMO systems, and wireless over optical technologies.
Overall, the IEEE Transactions on Wireless Communications serves as a platform for high-quality manuscripts that push the boundaries of wireless communications and contribute to advancements in the field.