Cong Wang , Ke Liu , Ying Yuan , Sancheng Peng , Guorui Li
{"title":"基于联合多智能体强化学习和势场的无人机辅助MEC联合轨迹与卸载优化","authors":"Cong Wang , Ke Liu , Ying Yuan , Sancheng Peng , Guorui Li","doi":"10.1016/j.comnet.2025.111681","DOIUrl":null,"url":null,"abstract":"<div><div>Unmanned Aerial Vehicles (UAVs) assisted mobile edge computing (MEC) is characterized by flexible deployment, high mobility, and dynamic coverage. It facilitates an efficient execution of latency-sensitive tasks in scenarios such as emergency rescue and dynamic computility support, thereby demonstrating significant application prospect. However, joint scheduling of computility and task is still an open issue in optimizing task efficiency and UAVs’ energy consumption. To address this problem, we propose an UAV-assisted MEC framework based on federated multi-agent reinforcement learning (MARL) and potential fields (PF), which jointly optimizes UAV trajectories and task offloading strategies to minimize age of information (AoI) under latency and energy constraints. The decision-making process of multiple UAVs is to be modeled as a Partially Observable Markov Decision Process (POMDP) and to be solved by using a distributed federated MARL architecture. An adaptive federated collaboration model is designed for periodic parameter sharing based on credit allocation to enhance UAV collaboration and to alleviate partial observability. Additionally, a deep reinforcement learning (DRL) trajectory planning algorithm based on PF to enhance agents’ environment perception and decision-making ability. Experimental results show the effectiveness and feasibility of our proposed framework. It outperforms several existing RL-based approaches in terms of data freshness, task efficiency, and other key metrics while demonstrating strong adaptability in dynamic and complex MEC environments.</div></div>","PeriodicalId":50637,"journal":{"name":"Computer Networks","volume":"272 ","pages":"Article 111681"},"PeriodicalIF":4.6000,"publicationDate":"2025-09-10","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":"{\"title\":\"Joint trajectory and offloading optimization in UAV-assisted MEC via federated multi-agent reinforcement learning and potential fields\",\"authors\":\"Cong Wang , Ke Liu , Ying Yuan , Sancheng Peng , Guorui Li\",\"doi\":\"10.1016/j.comnet.2025.111681\",\"DOIUrl\":null,\"url\":null,\"abstract\":\"<div><div>Unmanned Aerial Vehicles (UAVs) assisted mobile edge computing (MEC) is characterized by flexible deployment, high mobility, and dynamic coverage. It facilitates an efficient execution of latency-sensitive tasks in scenarios such as emergency rescue and dynamic computility support, thereby demonstrating significant application prospect. However, joint scheduling of computility and task is still an open issue in optimizing task efficiency and UAVs’ energy consumption. To address this problem, we propose an UAV-assisted MEC framework based on federated multi-agent reinforcement learning (MARL) and potential fields (PF), which jointly optimizes UAV trajectories and task offloading strategies to minimize age of information (AoI) under latency and energy constraints. The decision-making process of multiple UAVs is to be modeled as a Partially Observable Markov Decision Process (POMDP) and to be solved by using a distributed federated MARL architecture. An adaptive federated collaboration model is designed for periodic parameter sharing based on credit allocation to enhance UAV collaboration and to alleviate partial observability. Additionally, a deep reinforcement learning (DRL) trajectory planning algorithm based on PF to enhance agents’ environment perception and decision-making ability. Experimental results show the effectiveness and feasibility of our proposed framework. It outperforms several existing RL-based approaches in terms of data freshness, task efficiency, and other key metrics while demonstrating strong adaptability in dynamic and complex MEC environments.</div></div>\",\"PeriodicalId\":50637,\"journal\":{\"name\":\"Computer Networks\",\"volume\":\"272 \",\"pages\":\"Article 111681\"},\"PeriodicalIF\":4.6000,\"publicationDate\":\"2025-09-10\",\"publicationTypes\":\"Journal Article\",\"fieldsOfStudy\":null,\"isOpenAccess\":false,\"openAccessPdf\":\"\",\"citationCount\":\"0\",\"resultStr\":null,\"platform\":\"Semanticscholar\",\"paperid\":null,\"PeriodicalName\":\"Computer Networks\",\"FirstCategoryId\":\"94\",\"ListUrlMain\":\"https://www.sciencedirect.com/science/article/pii/S1389128625006486\",\"RegionNum\":2,\"RegionCategory\":\"计算机科学\",\"ArticlePicture\":[],\"TitleCN\":null,\"AbstractTextCN\":null,\"PMCID\":null,\"EPubDate\":\"\",\"PubModel\":\"\",\"JCR\":\"Q1\",\"JCRName\":\"COMPUTER SCIENCE, HARDWARE & ARCHITECTURE\",\"Score\":null,\"Total\":0}","platform":"Semanticscholar","paperid":null,"PeriodicalName":"Computer Networks","FirstCategoryId":"94","ListUrlMain":"https://www.sciencedirect.com/science/article/pii/S1389128625006486","RegionNum":2,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"Q1","JCRName":"COMPUTER SCIENCE, HARDWARE & ARCHITECTURE","Score":null,"Total":0}
Joint trajectory and offloading optimization in UAV-assisted MEC via federated multi-agent reinforcement learning and potential fields
Unmanned Aerial Vehicles (UAVs) assisted mobile edge computing (MEC) is characterized by flexible deployment, high mobility, and dynamic coverage. It facilitates an efficient execution of latency-sensitive tasks in scenarios such as emergency rescue and dynamic computility support, thereby demonstrating significant application prospect. However, joint scheduling of computility and task is still an open issue in optimizing task efficiency and UAVs’ energy consumption. To address this problem, we propose an UAV-assisted MEC framework based on federated multi-agent reinforcement learning (MARL) and potential fields (PF), which jointly optimizes UAV trajectories and task offloading strategies to minimize age of information (AoI) under latency and energy constraints. The decision-making process of multiple UAVs is to be modeled as a Partially Observable Markov Decision Process (POMDP) and to be solved by using a distributed federated MARL architecture. An adaptive federated collaboration model is designed for periodic parameter sharing based on credit allocation to enhance UAV collaboration and to alleviate partial observability. Additionally, a deep reinforcement learning (DRL) trajectory planning algorithm based on PF to enhance agents’ environment perception and decision-making ability. Experimental results show the effectiveness and feasibility of our proposed framework. It outperforms several existing RL-based approaches in terms of data freshness, task efficiency, and other key metrics while demonstrating strong adaptability in dynamic and complex MEC environments.
期刊介绍:
Computer Networks is an international, archival journal providing a publication vehicle for complete coverage of all topics of interest to those involved in the computer communications networking area. The audience includes researchers, managers and operators of networks as well as designers and implementors. The Editorial Board will consider any material for publication that is of interest to those groups.