Deep Reinforcement Learning-Based Dynamic MultiChannel Access for Heterogeneous Wireless Networks with DenseNet

K. Zong
{"title":"Deep Reinforcement Learning-Based Dynamic MultiChannel Access for Heterogeneous Wireless Networks with DenseNet","authors":"K. Zong","doi":"10.1109/ICCCWorkshops52231.2021.9538886","DOIUrl":null,"url":null,"abstract":"In this paper, we consider the problem of dynamic multi-channel access in the heterogeneous wireless networks, where multiple independent channels are shared by multiple nodes which have different types. The objective is to find a strategy to maximize the expected long-term probability of successful transmissions. The problem of dynamic multi-channel access can be formulated as a partially observable Markov decision process (POMDP). In order to deal with this problem, we apply the deep reinforcement learning (DRL) approach to provide a model-free access method, where the nodes don’t have a prior knowledge of the wireless networks or the ability to exchange messages with other nodes. Specially, we take advantage of the double deep Q-network (DDQN) with DenseNet to learn the wireless network environment and to select the optimal channel at the beginning of each time slot. We investigate the proposed DDQN approach in different environments for both the fixed-pattern scenarios and the time-varying scenarios. The experimental results show that the proposed DDQN with DenseNet can efficiently learn the pattern of channel switch and choose the near optimal action to avoid the collision for every slot. Besides, the proposed DDQN approach can also achieve satisfactory performance to adapt the time-varying scenarios.","PeriodicalId":335240,"journal":{"name":"2021 IEEE/CIC International Conference on Communications in China (ICCC Workshops)","volume":"212 1","pages":"0"},"PeriodicalIF":0.0000,"publicationDate":"2021-07-28","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"2","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"2021 IEEE/CIC International Conference on Communications in China (ICCC Workshops)","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/10.1109/ICCCWorkshops52231.2021.9538886","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
引用次数: 2

Abstract

In this paper, we consider the problem of dynamic multi-channel access in the heterogeneous wireless networks, where multiple independent channels are shared by multiple nodes which have different types. The objective is to find a strategy to maximize the expected long-term probability of successful transmissions. The problem of dynamic multi-channel access can be formulated as a partially observable Markov decision process (POMDP). In order to deal with this problem, we apply the deep reinforcement learning (DRL) approach to provide a model-free access method, where the nodes don’t have a prior knowledge of the wireless networks or the ability to exchange messages with other nodes. Specially, we take advantage of the double deep Q-network (DDQN) with DenseNet to learn the wireless network environment and to select the optimal channel at the beginning of each time slot. We investigate the proposed DDQN approach in different environments for both the fixed-pattern scenarios and the time-varying scenarios. The experimental results show that the proposed DDQN with DenseNet can efficiently learn the pattern of channel switch and choose the near optimal action to avoid the collision for every slot. Besides, the proposed DDQN approach can also achieve satisfactory performance to adapt the time-varying scenarios.
基于深度强化学习的DenseNet异构无线网络动态多通道接入
本文研究了异构无线网络中多个独立信道由不同类型的多个节点共享的动态多信道接入问题。目标是找到一种策略,以最大限度地提高预期的成功传输的长期概率。动态多通道接入问题可以表示为部分可观察马尔可夫决策过程(POMDP)。为了解决这个问题,我们应用深度强化学习(DRL)方法来提供一种无模型访问方法,其中节点不具有无线网络的先验知识或与其他节点交换消息的能力。特别地,我们利用DenseNet的双深度q网络(DDQN)来学习无线网络环境,并在每个时隙的开始选择最优信道。在固定模式和时变场景下,研究了不同环境下提出的DDQN方法。实验结果表明,基于DenseNet的DDQN可以有效地学习信道切换模式,并选择接近最优的动作来避免每个槽的碰撞。此外,所提出的DDQN方法在适应时变场景时也能取得令人满意的性能。
本文章由计算机程序翻译,如有差异,请以英文原文为准。
求助全文
约1分钟内获得全文 求助全文
来源期刊
自引率
0.00%
发文量
0
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
确定
请完成安全验证×
copy
已复制链接
快去分享给好友吧!
我知道了
右上角分享
点击右上角分享
0
联系我们:info@booksci.cn Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。 Copyright © 2023 布克学术 All rights reserved.
京ICP备2023020795号-1
ghs 京公网安备 11010802042870号
Book学术文献互助
Book学术文献互助群
群 号:481959085
Book学术官方微信