Contextual Deep Reinforcement Learning for Flow and Energy Management in Wireless Sensor and IoT Networks

IF 5.3 2区 计算机科学 Q1 TELECOMMUNICATIONS
Hrishikesh Dutta;Amit Kumar Bhuyan;Subir Biswas
{"title":"Contextual Deep Reinforcement Learning for Flow and Energy Management in Wireless Sensor and IoT Networks","authors":"Hrishikesh Dutta;Amit Kumar Bhuyan;Subir Biswas","doi":"10.1109/TGCN.2024.3358230","DOIUrl":null,"url":null,"abstract":"Efficient slot allocation and transmit-sleep scheduling is an effective access control mechanism for improving communication performance and network lifetime in resource-constrained wireless networks. In this paper, a decentralized and multi-tier framework is presented for joint slot allocation and transmit-sleep scheduling in wireless network nodes with thin energy budget. The key learning objectives of this architecture are: collision-free transmission scheduling, reducing energy consumption, and improving network performance. This is achieved using a cooperative and decentralized learning behavior of multiple Reinforcement Learning (RL) agents. The resulting architecture provides throughput-sustainable support for data flows while minimizing energy expenditure and sleep-induced packet losses. To achieve this, a concept of Context is introduced to the RL framework in order to capture network traffic dynamics. The resulting Contextual Deep Q-Learning (CDQL) model makes the system adaptive to dynamic and heterogeneous network load. It also improves energy efficiency when compared with the traditional tabular Q-learning-based approaches. The results demonstrate how this framework can be used for prioritizing application-specific requirements, namely, energy saving and communication reliability. The trade-offs among packet drop, energy expenditure, and learning convergence are studied, and an application-specific solution is proposed for managing them. The performance is compared against an existing state-of-the-art scheduling approach. Moreover, an analytical model of the system dynamics is developed and validated using simulation for arbitrary mesh topologies and traffic patterns.","PeriodicalId":13052,"journal":{"name":"IEEE Transactions on Green Communications and Networking","volume":"8 3","pages":"1233-1244"},"PeriodicalIF":5.3000,"publicationDate":"2024-01-24","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"IEEE Transactions on Green Communications and Networking","FirstCategoryId":"94","ListUrlMain":"https://ieeexplore.ieee.org/document/10413520/","RegionNum":2,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"Q1","JCRName":"TELECOMMUNICATIONS","Score":null,"Total":0}
引用次数: 0

Abstract

Efficient slot allocation and transmit-sleep scheduling is an effective access control mechanism for improving communication performance and network lifetime in resource-constrained wireless networks. In this paper, a decentralized and multi-tier framework is presented for joint slot allocation and transmit-sleep scheduling in wireless network nodes with thin energy budget. The key learning objectives of this architecture are: collision-free transmission scheduling, reducing energy consumption, and improving network performance. This is achieved using a cooperative and decentralized learning behavior of multiple Reinforcement Learning (RL) agents. The resulting architecture provides throughput-sustainable support for data flows while minimizing energy expenditure and sleep-induced packet losses. To achieve this, a concept of Context is introduced to the RL framework in order to capture network traffic dynamics. The resulting Contextual Deep Q-Learning (CDQL) model makes the system adaptive to dynamic and heterogeneous network load. It also improves energy efficiency when compared with the traditional tabular Q-learning-based approaches. The results demonstrate how this framework can be used for prioritizing application-specific requirements, namely, energy saving and communication reliability. The trade-offs among packet drop, energy expenditure, and learning convergence are studied, and an application-specific solution is proposed for managing them. The performance is compared against an existing state-of-the-art scheduling approach. Moreover, an analytical model of the system dynamics is developed and validated using simulation for arbitrary mesh topologies and traffic patterns.
针对无线传感器和物联网网络中流量和能量管理的情境深度强化学习
在资源受限的无线网络中,高效的时隙分配和发送-休眠调度是提高通信性能和网络寿命的有效访问控制机制。本文提出了一种去中心化的多层框架,用于在能量预算较低的无线网络节点中进行联合时隙分配和发送-休眠调度。该架构的主要学习目标是:无碰撞传输调度、降低能耗和提高网络性能。这是通过多个强化学习(RL)代理的合作和分散学习行为来实现的。由此产生的架构可为数据流提供吞吐量可持续的支持,同时最大限度地减少能源消耗和睡眠引起的数据包丢失。为此,RL 框架引入了 "情境 "概念,以捕捉网络流量动态。由此产生的上下文深度 Q 学习(CDQL)模型使系统能够适应动态和异构网络负载。与传统的基于表格的 Q 学习方法相比,它还提高了能效。研究结果表明,该框架可用于优先满足特定应用的要求,即节能和通信可靠性。研究了丢包、能量消耗和学习收敛之间的权衡,并提出了管理这些问题的特定应用解决方案。将其性能与现有的最先进调度方法进行了比较。此外,还开发了一个系统动态分析模型,并针对任意网状拓扑和流量模式进行了仿真验证。
本文章由计算机程序翻译,如有差异,请以英文原文为准。
求助全文
约1分钟内获得全文 求助全文
来源期刊
IEEE Transactions on Green Communications and Networking
IEEE Transactions on Green Communications and Networking Computer Science-Computer Networks and Communications
CiteScore
9.30
自引率
6.20%
发文量
181
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
确定
请完成安全验证×
copy
已复制链接
快去分享给好友吧!
我知道了
右上角分享
点击右上角分享
0
联系我们:info@booksci.cn Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。 Copyright © 2023 布克学术 All rights reserved.
京ICP备2023020795号-1
ghs 京公网安备 11010802042870号
Book学术文献互助
Book学术文献互助群
群 号:481959085
Book学术官方微信