{"title":"基于频谱共存的NextG无线接入网切片的深度强化学习","authors":"Yi Shi;Maice Costa;Tugba Erpek;Yalin E. Sagduyu","doi":"10.1109/LNET.2023.3284665","DOIUrl":null,"url":null,"abstract":"Reinforcement learning (RL) is applied for dynamic admission control and resource allocation in NextG radio access network slicing. When sharing the spectrum with an incumbent user (that dynamically occupies frequency-time blocks), communication and computational resources are allocated to slicing requests, each with priority (weight), throughput, latency, and computational requirements. RL maximizes the total weight of granted requests over time beyond myopic, greedy, random, and first come, first served solutions. As the state-action space grows, Deep Q-network effectively admits requests and allocates resources as a low-complexity solution that is robust to sensing errors in detecting the incumbent user activity.","PeriodicalId":100628,"journal":{"name":"IEEE Networking Letters","volume":"5 3","pages":"149-153"},"PeriodicalIF":0.0000,"publicationDate":"2023-06-09","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"5","resultStr":"{\"title\":\"Deep Reinforcement Learning for NextG Radio Access Network Slicing With Spectrum Coexistence\",\"authors\":\"Yi Shi;Maice Costa;Tugba Erpek;Yalin E. Sagduyu\",\"doi\":\"10.1109/LNET.2023.3284665\",\"DOIUrl\":null,\"url\":null,\"abstract\":\"Reinforcement learning (RL) is applied for dynamic admission control and resource allocation in NextG radio access network slicing. When sharing the spectrum with an incumbent user (that dynamically occupies frequency-time blocks), communication and computational resources are allocated to slicing requests, each with priority (weight), throughput, latency, and computational requirements. RL maximizes the total weight of granted requests over time beyond myopic, greedy, random, and first come, first served solutions. As the state-action space grows, Deep Q-network effectively admits requests and allocates resources as a low-complexity solution that is robust to sensing errors in detecting the incumbent user activity.\",\"PeriodicalId\":100628,\"journal\":{\"name\":\"IEEE Networking Letters\",\"volume\":\"5 3\",\"pages\":\"149-153\"},\"PeriodicalIF\":0.0000,\"publicationDate\":\"2023-06-09\",\"publicationTypes\":\"Journal Article\",\"fieldsOfStudy\":null,\"isOpenAccess\":false,\"openAccessPdf\":\"\",\"citationCount\":\"5\",\"resultStr\":null,\"platform\":\"Semanticscholar\",\"paperid\":null,\"PeriodicalName\":\"IEEE Networking Letters\",\"FirstCategoryId\":\"1085\",\"ListUrlMain\":\"https://ieeexplore.ieee.org/document/10147283/\",\"RegionNum\":0,\"RegionCategory\":null,\"ArticlePicture\":[],\"TitleCN\":null,\"AbstractTextCN\":null,\"PMCID\":null,\"EPubDate\":\"\",\"PubModel\":\"\",\"JCR\":\"\",\"JCRName\":\"\",\"Score\":null,\"Total\":0}","platform":"Semanticscholar","paperid":null,"PeriodicalName":"IEEE Networking Letters","FirstCategoryId":"1085","ListUrlMain":"https://ieeexplore.ieee.org/document/10147283/","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
Deep Reinforcement Learning for NextG Radio Access Network Slicing With Spectrum Coexistence
Reinforcement learning (RL) is applied for dynamic admission control and resource allocation in NextG radio access network slicing. When sharing the spectrum with an incumbent user (that dynamically occupies frequency-time blocks), communication and computational resources are allocated to slicing requests, each with priority (weight), throughput, latency, and computational requirements. RL maximizes the total weight of granted requests over time beyond myopic, greedy, random, and first come, first served solutions. As the state-action space grows, Deep Q-network effectively admits requests and allocates resources as a low-complexity solution that is robust to sensing errors in detecting the incumbent user activity.