Optimal Energy Allocation and Multiuser Scheduling in SWIPT Systems with Hybrid Power Supply

Delin Guo, Lan Tang, Xinggan Zhang
{"title":"Optimal Energy Allocation and Multiuser Scheduling in SWIPT Systems with Hybrid Power Supply","authors":"Delin Guo, Lan Tang, Xinggan Zhang","doi":"10.1109/GCWkshps45667.2019.9024326","DOIUrl":null,"url":null,"abstract":"This paper studies the utilization and transfer of renewable green energy in a multiuser downlink communication network. In the considered multiuser system, the base station (BS) is powered by both harvested energy and grid. When the BS transmits data to one user terminal, other terminals can replenish energy opportunistically from received radio-frequency (RF) signals, which is called simultaneous wireless information and power transfer (SWIPT). Our objective is to maximize the average throughput by multiuser scheduling and energy allocation utilizing causal channel state information while satisfying the requirement for harvested energy and the average power constraint of the grid. With channel dynamics and energy arrival modeled as Markov processes, we characterize the problem as a Markov decision process (MDP). The standard reinforcement learning framework is considered as an effective solution to MDP. If the transition probability of MDP is known, the policy iteration (PI) algorithm is used to solve the problem, otherwise, the R-learning algorithm is adopted. Simulation results show that the proposed algorithm can improve the average throughput of the system and increase the energy harvested by idle user terminals compared with the existing work. And R-learning can achieve performance close to the PI algorithm under the condition that the channel transition probability is unknown.","PeriodicalId":210825,"journal":{"name":"2019 IEEE Globecom Workshops (GC Wkshps)","volume":null,"pages":null},"PeriodicalIF":0.0000,"publicationDate":"2019-12-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"1","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"2019 IEEE Globecom Workshops (GC Wkshps)","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/10.1109/GCWkshps45667.2019.9024326","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
引用次数: 1

Abstract

This paper studies the utilization and transfer of renewable green energy in a multiuser downlink communication network. In the considered multiuser system, the base station (BS) is powered by both harvested energy and grid. When the BS transmits data to one user terminal, other terminals can replenish energy opportunistically from received radio-frequency (RF) signals, which is called simultaneous wireless information and power transfer (SWIPT). Our objective is to maximize the average throughput by multiuser scheduling and energy allocation utilizing causal channel state information while satisfying the requirement for harvested energy and the average power constraint of the grid. With channel dynamics and energy arrival modeled as Markov processes, we characterize the problem as a Markov decision process (MDP). The standard reinforcement learning framework is considered as an effective solution to MDP. If the transition probability of MDP is known, the policy iteration (PI) algorithm is used to solve the problem, otherwise, the R-learning algorithm is adopted. Simulation results show that the proposed algorithm can improve the average throughput of the system and increase the energy harvested by idle user terminals compared with the existing work. And R-learning can achieve performance close to the PI algorithm under the condition that the channel transition probability is unknown.
混合电源下SWIPT系统的最优能量分配与多用户调度
本文研究了多用户下行通信网络中可再生绿色能源的利用与传输问题。在考虑的多用户系统中,基站(BS)由收集的能量和电网供电。当BS向一个用户终端传输数据时,其他终端可以根据接收到的射频(RF)信号适时地补充能量,这被称为同步无线信息和电力传输(SWIPT)。我们的目标是通过多用户调度和利用因果通道状态信息的能量分配来最大化平均吞吐量,同时满足收获能量的要求和电网的平均功率约束。将通道动态和能量到达建模为马尔可夫过程,我们将问题描述为马尔可夫决策过程(MDP)。标准强化学习框架被认为是MDP的有效解决方案。若MDP的转移概率已知,则采用策略迭代(PI)算法求解,否则采用r -学习算法求解。仿真结果表明,与现有算法相比,该算法可以提高系统的平均吞吐量,增加闲置用户终端的能量收获。在信道转移概率未知的情况下,r学习可以达到接近PI算法的性能。
本文章由计算机程序翻译,如有差异,请以英文原文为准。
求助全文
约1分钟内获得全文 求助全文
来源期刊
自引率
0.00%
发文量
0
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
确定
请完成安全验证×
copy
已复制链接
快去分享给好友吧!
我知道了
右上角分享
点击右上角分享
0
联系我们:info@booksci.cn Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。 Copyright © 2023 布克学术 All rights reserved.
京ICP备2023020795号-1
ghs 京公网安备 11010802042870号
Book学术文献互助
Book学术文献互助群
群 号:481959085
Book学术官方微信