Computation Offloading in Energy Harvesting Systems via Continuous Deep Reinforcement Learning

Jing Zhang, Jun Du, Chunxiao Jiang, Yuan Shen, Jian Wang
{"title":"Computation Offloading in Energy Harvesting Systems via Continuous Deep Reinforcement Learning","authors":"Jing Zhang, Jun Du, Chunxiao Jiang, Yuan Shen, Jian Wang","doi":"10.1109/ICC40277.2020.9148938","DOIUrl":null,"url":null,"abstract":"As a promising technology to improve the computation experience for mobile devices, mobile edge computing (MEC) is becoming an emerging paradigm to meet the tremendous increasing computation demands. In this paper, a mobile edge computing system consisting of multiple mobile devices with energy harvesting and an edge server is considered. Specifically, multiple devices decide the offloading ratio and local computation capacity, which are both in continuous values. Each device equips a task load queue and energy harvesting, which increases the system dynamics and leads to the time-dependence of the optimal offloading decision. In order to minimize the sum cost of the execution time and energy consumption in the long-term, we develop a continuous control based deep reinforcement learning algorithm for computation offloading. Utilizing the actor-critic learning approach, we propose a centralized learning policy for each device. By incorporating the states of other devices with centralized learning, the proposed method learns to coordinate among all devices. Simulation results validate the effectiveness of our proposed algorithm, which demonstrates superior generalization ability and achieves a better performance compared with discrete decision based deep reinforcement learning methods.","PeriodicalId":106560,"journal":{"name":"ICC 2020 - 2020 IEEE International Conference on Communications (ICC)","volume":"9 1","pages":"0"},"PeriodicalIF":0.0000,"publicationDate":"2020-06-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"7","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"ICC 2020 - 2020 IEEE International Conference on Communications (ICC)","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/10.1109/ICC40277.2020.9148938","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
引用次数: 7

Abstract

As a promising technology to improve the computation experience for mobile devices, mobile edge computing (MEC) is becoming an emerging paradigm to meet the tremendous increasing computation demands. In this paper, a mobile edge computing system consisting of multiple mobile devices with energy harvesting and an edge server is considered. Specifically, multiple devices decide the offloading ratio and local computation capacity, which are both in continuous values. Each device equips a task load queue and energy harvesting, which increases the system dynamics and leads to the time-dependence of the optimal offloading decision. In order to minimize the sum cost of the execution time and energy consumption in the long-term, we develop a continuous control based deep reinforcement learning algorithm for computation offloading. Utilizing the actor-critic learning approach, we propose a centralized learning policy for each device. By incorporating the states of other devices with centralized learning, the proposed method learns to coordinate among all devices. Simulation results validate the effectiveness of our proposed algorithm, which demonstrates superior generalization ability and achieves a better performance compared with discrete decision based deep reinforcement learning methods.
基于连续深度强化学习的能量收集系统计算卸载
移动边缘计算(MEC)作为改善移动设备计算体验的一种有前景的技术,正在成为满足日益增长的计算需求的新兴范式。本文研究了一种由多个具有能量收集功能的移动设备和一个边缘服务器组成的移动边缘计算系统。具体来说,多个设备决定了卸载比和本地计算能力,两者都是连续值。每个设备配备一个任务负载队列和能量收集,这增加了系统的动态性,并导致最优卸载决策的时间依赖性。为了使长期执行时间和能量消耗的总和成本最小化,我们开发了一种基于连续控制的深度强化学习算法来进行计算卸载。利用行动者-评论家学习方法,我们为每个设备提出了一个集中的学习策略。通过将其他设备的状态与集中学习相结合,该方法学会了在所有设备之间进行协调。仿真结果验证了该算法的有效性,与基于离散决策的深度强化学习方法相比,该算法具有优越的泛化能力,取得了更好的性能。
本文章由计算机程序翻译,如有差异,请以英文原文为准。
求助全文
约1分钟内获得全文 求助全文
来源期刊
自引率
0.00%
发文量
0
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
确定
请完成安全验证×
copy
已复制链接
快去分享给好友吧!
我知道了
右上角分享
点击右上角分享
0
联系我们:info@booksci.cn Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。 Copyright © 2023 布克学术 All rights reserved.
京ICP备2023020795号-1
ghs 京公网安备 11010802042870号
Book学术文献互助
Book学术文献互助群
群 号:481959085
Book学术官方微信