Che Chen;Shimin Gong;Wenjie Zhang;Yifeng Zheng;Yeo Chai Kiat
{"title":"基于 DRL 的无线供电和无人机辅助反向散射 MEC 系统合同激励机制","authors":"Che Chen;Shimin Gong;Wenjie Zhang;Yifeng Zheng;Yeo Chai Kiat","doi":"10.1109/TCC.2024.3360443","DOIUrl":null,"url":null,"abstract":"Mobile edge computing (MEC) is viewed as a promising technology to address the challenges of intensive computing demands in hotspots (HSs). In this article, we consider a unmanned aerial vehicle (UAV)-assisted backscattering MEC system. The UAVs can fly from parking aprons to HSs, providing energy to HSs via RF beamforming and collecting data from wireless users in HSs through backscattering. We aim to maximize the long-term utility of all HSs, subject to the stability of the HSs’ energy queues. This problem is a joint optimization of the data offloading decision and contract design that should be adaptive to the users’ random task demands and the time-varying wireless channel conditions. A deep reinforcement learning based contract incentive (DRLCI) strategy is proposed to solve this problem in two steps. First, we use deep Q-network (DQN) algorithm to update the HSs’ offloading decisions according to the changing network environment. Second, to motivate the UAVs to participate in resource sharing, a contract specific to each type of UAVs has been designed, utilizing Lagrangian multiplier method to approach the optimal contract. Simulation results show the feasibility and efficiency of the proposed strategy, demonstrating a better performance than the natural DQN and Double-DQN algorithms.","PeriodicalId":13202,"journal":{"name":"IEEE Transactions on Cloud Computing","volume":"12 1","pages":"264-276"},"PeriodicalIF":5.3000,"publicationDate":"2024-01-31","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":"{\"title\":\"DRL-Based Contract Incentive for Wireless-Powered and UAV-Assisted Backscattering MEC System\",\"authors\":\"Che Chen;Shimin Gong;Wenjie Zhang;Yifeng Zheng;Yeo Chai Kiat\",\"doi\":\"10.1109/TCC.2024.3360443\",\"DOIUrl\":null,\"url\":null,\"abstract\":\"Mobile edge computing (MEC) is viewed as a promising technology to address the challenges of intensive computing demands in hotspots (HSs). In this article, we consider a unmanned aerial vehicle (UAV)-assisted backscattering MEC system. The UAVs can fly from parking aprons to HSs, providing energy to HSs via RF beamforming and collecting data from wireless users in HSs through backscattering. We aim to maximize the long-term utility of all HSs, subject to the stability of the HSs’ energy queues. This problem is a joint optimization of the data offloading decision and contract design that should be adaptive to the users’ random task demands and the time-varying wireless channel conditions. A deep reinforcement learning based contract incentive (DRLCI) strategy is proposed to solve this problem in two steps. First, we use deep Q-network (DQN) algorithm to update the HSs’ offloading decisions according to the changing network environment. Second, to motivate the UAVs to participate in resource sharing, a contract specific to each type of UAVs has been designed, utilizing Lagrangian multiplier method to approach the optimal contract. Simulation results show the feasibility and efficiency of the proposed strategy, demonstrating a better performance than the natural DQN and Double-DQN algorithms.\",\"PeriodicalId\":13202,\"journal\":{\"name\":\"IEEE Transactions on Cloud Computing\",\"volume\":\"12 1\",\"pages\":\"264-276\"},\"PeriodicalIF\":5.3000,\"publicationDate\":\"2024-01-31\",\"publicationTypes\":\"Journal Article\",\"fieldsOfStudy\":null,\"isOpenAccess\":false,\"openAccessPdf\":\"\",\"citationCount\":\"0\",\"resultStr\":null,\"platform\":\"Semanticscholar\",\"paperid\":null,\"PeriodicalName\":\"IEEE Transactions on Cloud Computing\",\"FirstCategoryId\":\"94\",\"ListUrlMain\":\"https://ieeexplore.ieee.org/document/10417719/\",\"RegionNum\":2,\"RegionCategory\":\"计算机科学\",\"ArticlePicture\":[],\"TitleCN\":null,\"AbstractTextCN\":null,\"PMCID\":null,\"EPubDate\":\"\",\"PubModel\":\"\",\"JCR\":\"Q1\",\"JCRName\":\"COMPUTER SCIENCE, INFORMATION SYSTEMS\",\"Score\":null,\"Total\":0}","platform":"Semanticscholar","paperid":null,"PeriodicalName":"IEEE Transactions on Cloud Computing","FirstCategoryId":"94","ListUrlMain":"https://ieeexplore.ieee.org/document/10417719/","RegionNum":2,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"Q1","JCRName":"COMPUTER SCIENCE, INFORMATION SYSTEMS","Score":null,"Total":0}
DRL-Based Contract Incentive for Wireless-Powered and UAV-Assisted Backscattering MEC System
Mobile edge computing (MEC) is viewed as a promising technology to address the challenges of intensive computing demands in hotspots (HSs). In this article, we consider a unmanned aerial vehicle (UAV)-assisted backscattering MEC system. The UAVs can fly from parking aprons to HSs, providing energy to HSs via RF beamforming and collecting data from wireless users in HSs through backscattering. We aim to maximize the long-term utility of all HSs, subject to the stability of the HSs’ energy queues. This problem is a joint optimization of the data offloading decision and contract design that should be adaptive to the users’ random task demands and the time-varying wireless channel conditions. A deep reinforcement learning based contract incentive (DRLCI) strategy is proposed to solve this problem in two steps. First, we use deep Q-network (DQN) algorithm to update the HSs’ offloading decisions according to the changing network environment. Second, to motivate the UAVs to participate in resource sharing, a contract specific to each type of UAVs has been designed, utilizing Lagrangian multiplier method to approach the optimal contract. Simulation results show the feasibility and efficiency of the proposed strategy, demonstrating a better performance than the natural DQN and Double-DQN algorithms.
期刊介绍:
The IEEE Transactions on Cloud Computing (TCC) is dedicated to the multidisciplinary field of cloud computing. It is committed to the publication of articles that present innovative research ideas, application results, and case studies in cloud computing, focusing on key technical issues related to theory, algorithms, systems, applications, and performance.