{"title":"面向多自动驾驶辅助边缘计算的个性化联邦强化学习","authors":"Jinwei Chai;Zhe Wang;Chuan Ma;Guanyu Gao;Long Shi","doi":"10.1109/LWC.2025.3562611","DOIUrl":null,"url":null,"abstract":"In this letter, we propose an innovative autonomous aerial vehicle (AAV)-assisted computation offloading scheme, where multiple AAVs with heterogeneous computational capabilities collaborate to provide task offloading services for mobile users (MUs). Our objective is to minimize the weighted sum of the MUs’ maximum task processing delay and the AAVs’ average energy consumption by jointly optimizing the AAVs’ service locations, velocities, user association and offloading policies. We model this problem as a fully decentralized partially observable Markov decision process (Dec-POMDP) and propose a personalized federated reinforcement learning (FRL) algorithm, named federated knowledge distillation based parameterized deep Q-network (FKD-PDQN), to collaboratively optimize the AAVs’ policies in a decentralized and cost-efficient manner. In this approach, each AAV optimizes its local policy using PDQN, and updates its model by aggregating the distilled knowledge from the neighboring AAVs through an attention mechanism. Simulation results reveal that the proposed algorithm effectively reduces the average delay, energy consumption, and communication overhead, offering significant improvements over baseline algorithms such as Independent PDQN (IPDQN), Federated PDQN (FPDQN), and Krum.","PeriodicalId":13343,"journal":{"name":"IEEE Wireless Communications Letters","volume":"14 7","pages":"2074-2078"},"PeriodicalIF":5.5000,"publicationDate":"2025-04-21","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":"{\"title\":\"Personalized Federated Reinforcement Learning for Multi-AAV Assisted Edge Computing\",\"authors\":\"Jinwei Chai;Zhe Wang;Chuan Ma;Guanyu Gao;Long Shi\",\"doi\":\"10.1109/LWC.2025.3562611\",\"DOIUrl\":null,\"url\":null,\"abstract\":\"In this letter, we propose an innovative autonomous aerial vehicle (AAV)-assisted computation offloading scheme, where multiple AAVs with heterogeneous computational capabilities collaborate to provide task offloading services for mobile users (MUs). Our objective is to minimize the weighted sum of the MUs’ maximum task processing delay and the AAVs’ average energy consumption by jointly optimizing the AAVs’ service locations, velocities, user association and offloading policies. We model this problem as a fully decentralized partially observable Markov decision process (Dec-POMDP) and propose a personalized federated reinforcement learning (FRL) algorithm, named federated knowledge distillation based parameterized deep Q-network (FKD-PDQN), to collaboratively optimize the AAVs’ policies in a decentralized and cost-efficient manner. In this approach, each AAV optimizes its local policy using PDQN, and updates its model by aggregating the distilled knowledge from the neighboring AAVs through an attention mechanism. Simulation results reveal that the proposed algorithm effectively reduces the average delay, energy consumption, and communication overhead, offering significant improvements over baseline algorithms such as Independent PDQN (IPDQN), Federated PDQN (FPDQN), and Krum.\",\"PeriodicalId\":13343,\"journal\":{\"name\":\"IEEE Wireless Communications Letters\",\"volume\":\"14 7\",\"pages\":\"2074-2078\"},\"PeriodicalIF\":5.5000,\"publicationDate\":\"2025-04-21\",\"publicationTypes\":\"Journal Article\",\"fieldsOfStudy\":null,\"isOpenAccess\":false,\"openAccessPdf\":\"\",\"citationCount\":\"0\",\"resultStr\":null,\"platform\":\"Semanticscholar\",\"paperid\":null,\"PeriodicalName\":\"IEEE Wireless Communications Letters\",\"FirstCategoryId\":\"94\",\"ListUrlMain\":\"https://ieeexplore.ieee.org/document/10970733/\",\"RegionNum\":3,\"RegionCategory\":\"计算机科学\",\"ArticlePicture\":[],\"TitleCN\":null,\"AbstractTextCN\":null,\"PMCID\":null,\"EPubDate\":\"\",\"PubModel\":\"\",\"JCR\":\"Q1\",\"JCRName\":\"COMPUTER SCIENCE, INFORMATION SYSTEMS\",\"Score\":null,\"Total\":0}","platform":"Semanticscholar","paperid":null,"PeriodicalName":"IEEE Wireless Communications Letters","FirstCategoryId":"94","ListUrlMain":"https://ieeexplore.ieee.org/document/10970733/","RegionNum":3,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"Q1","JCRName":"COMPUTER SCIENCE, INFORMATION SYSTEMS","Score":null,"Total":0}
Personalized Federated Reinforcement Learning for Multi-AAV Assisted Edge Computing
In this letter, we propose an innovative autonomous aerial vehicle (AAV)-assisted computation offloading scheme, where multiple AAVs with heterogeneous computational capabilities collaborate to provide task offloading services for mobile users (MUs). Our objective is to minimize the weighted sum of the MUs’ maximum task processing delay and the AAVs’ average energy consumption by jointly optimizing the AAVs’ service locations, velocities, user association and offloading policies. We model this problem as a fully decentralized partially observable Markov decision process (Dec-POMDP) and propose a personalized federated reinforcement learning (FRL) algorithm, named federated knowledge distillation based parameterized deep Q-network (FKD-PDQN), to collaboratively optimize the AAVs’ policies in a decentralized and cost-efficient manner. In this approach, each AAV optimizes its local policy using PDQN, and updates its model by aggregating the distilled knowledge from the neighboring AAVs through an attention mechanism. Simulation results reveal that the proposed algorithm effectively reduces the average delay, energy consumption, and communication overhead, offering significant improvements over baseline algorithms such as Independent PDQN (IPDQN), Federated PDQN (FPDQN), and Krum.
期刊介绍:
IEEE Wireless Communications Letters publishes short papers in a rapid publication cycle on advances in the state-of-the-art of wireless communications. Both theoretical contributions (including new techniques, concepts, and analyses) and practical contributions (including system experiments and prototypes, and new applications) are encouraged. This journal focuses on the physical layer and the link layer of wireless communication systems.