{"title":"Partially Observable Multi-Agent Deep Reinforcement Learning for Cognitive Resource Management","authors":"Ning Yang, Haijun Zhang, R. Berry","doi":"10.1109/GLOBECOM42002.2020.9322150","DOIUrl":null,"url":null,"abstract":"In this paper, the problem of dynamic resource management in a cognitive radio network (CRN) with multiple primary users (PUs), multiple secondary users (SUs), and multiple channels is investigated. An optimization problem is formulated as a multi-agent partially observable Markov decision process (POMDP) problem in a dynamic and not fully observable environment. We consider using deep reinforcement learning (DRL) to address this problem. Based on the channel occupancy of PUs, a multi-agent deep Q-network (DQN)-based dynamic joint spectrum access and mode selection (SAMS) scheme is proposed for the SUs in the partially observable environment. The current observation of each SU is mapped to a suitable action. Each secondary user (SU) takes its own decision without exchanging information with other SUs. It seeks to maximize the total sum rate. Simulation results verify the effectiveness of our proposed schemes.","PeriodicalId":12759,"journal":{"name":"GLOBECOM 2020 - 2020 IEEE Global Communications Conference","volume":"1 1","pages":"1-6"},"PeriodicalIF":0.0000,"publicationDate":"2020-12-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"9","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"GLOBECOM 2020 - 2020 IEEE Global Communications Conference","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/10.1109/GLOBECOM42002.2020.9322150","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
引用次数: 9
Abstract
In this paper, the problem of dynamic resource management in a cognitive radio network (CRN) with multiple primary users (PUs), multiple secondary users (SUs), and multiple channels is investigated. An optimization problem is formulated as a multi-agent partially observable Markov decision process (POMDP) problem in a dynamic and not fully observable environment. We consider using deep reinforcement learning (DRL) to address this problem. Based on the channel occupancy of PUs, a multi-agent deep Q-network (DQN)-based dynamic joint spectrum access and mode selection (SAMS) scheme is proposed for the SUs in the partially observable environment. The current observation of each SU is mapped to a suitable action. Each secondary user (SU) takes its own decision without exchanging information with other SUs. It seeks to maximize the total sum rate. Simulation results verify the effectiveness of our proposed schemes.