{"title":"小细胞切换策略:一种强化学习方法","authors":"Luyang Wang, Xinxin Feng, Xiaoying Gan, Jing Liu, Hui Yu","doi":"10.1109/WCSP.2014.6992126","DOIUrl":null,"url":null,"abstract":"Small cell is a flexible solution to satisfy the continuously increasing wireless traffic demand. In this paper, we focus on on-off switch operation on small cell base stations (SBS) in heterogeneous networks. In our scenario, the users can either choose SBS when it is active or macro cell base station (MBS) to transmit data. Start-up energy cost is considered when SBS switches on. The whole network acts as a queueing system, and network latency is also under consideration. The network traffic is modeled by a Markov Modulated Poisson Process (MMPP) whose parameters are unknown to the network control center. To maximize the system reward, we introduce a reinforcement learning approach to obtain the optimal on-off switch policy. The learning procedure is defined as a Markov Decision Process (MDP). An estimation method is proposed to measure the load of the network. A single-agent Q-learning algorithm is proposed afterwards. The convergence of this algorithm is proved. Simulation results are given to evaluate the performance of the proposed algorithm.","PeriodicalId":412971,"journal":{"name":"2014 Sixth International Conference on Wireless Communications and Signal Processing (WCSP)","volume":"47 1","pages":"0"},"PeriodicalIF":0.0000,"publicationDate":"2014-12-22","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"4","resultStr":"{\"title\":\"Small cell switch policy: A reinforcement learning approach\",\"authors\":\"Luyang Wang, Xinxin Feng, Xiaoying Gan, Jing Liu, Hui Yu\",\"doi\":\"10.1109/WCSP.2014.6992126\",\"DOIUrl\":null,\"url\":null,\"abstract\":\"Small cell is a flexible solution to satisfy the continuously increasing wireless traffic demand. In this paper, we focus on on-off switch operation on small cell base stations (SBS) in heterogeneous networks. In our scenario, the users can either choose SBS when it is active or macro cell base station (MBS) to transmit data. Start-up energy cost is considered when SBS switches on. The whole network acts as a queueing system, and network latency is also under consideration. The network traffic is modeled by a Markov Modulated Poisson Process (MMPP) whose parameters are unknown to the network control center. To maximize the system reward, we introduce a reinforcement learning approach to obtain the optimal on-off switch policy. The learning procedure is defined as a Markov Decision Process (MDP). An estimation method is proposed to measure the load of the network. A single-agent Q-learning algorithm is proposed afterwards. The convergence of this algorithm is proved. Simulation results are given to evaluate the performance of the proposed algorithm.\",\"PeriodicalId\":412971,\"journal\":{\"name\":\"2014 Sixth International Conference on Wireless Communications and Signal Processing (WCSP)\",\"volume\":\"47 1\",\"pages\":\"0\"},\"PeriodicalIF\":0.0000,\"publicationDate\":\"2014-12-22\",\"publicationTypes\":\"Journal Article\",\"fieldsOfStudy\":null,\"isOpenAccess\":false,\"openAccessPdf\":\"\",\"citationCount\":\"4\",\"resultStr\":null,\"platform\":\"Semanticscholar\",\"paperid\":null,\"PeriodicalName\":\"2014 Sixth International Conference on Wireless Communications and Signal Processing (WCSP)\",\"FirstCategoryId\":\"1085\",\"ListUrlMain\":\"https://doi.org/10.1109/WCSP.2014.6992126\",\"RegionNum\":0,\"RegionCategory\":null,\"ArticlePicture\":[],\"TitleCN\":null,\"AbstractTextCN\":null,\"PMCID\":null,\"EPubDate\":\"\",\"PubModel\":\"\",\"JCR\":\"\",\"JCRName\":\"\",\"Score\":null,\"Total\":0}","platform":"Semanticscholar","paperid":null,"PeriodicalName":"2014 Sixth International Conference on Wireless Communications and Signal Processing (WCSP)","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/10.1109/WCSP.2014.6992126","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
Small cell switch policy: A reinforcement learning approach
Small cell is a flexible solution to satisfy the continuously increasing wireless traffic demand. In this paper, we focus on on-off switch operation on small cell base stations (SBS) in heterogeneous networks. In our scenario, the users can either choose SBS when it is active or macro cell base station (MBS) to transmit data. Start-up energy cost is considered when SBS switches on. The whole network acts as a queueing system, and network latency is also under consideration. The network traffic is modeled by a Markov Modulated Poisson Process (MMPP) whose parameters are unknown to the network control center. To maximize the system reward, we introduce a reinforcement learning approach to obtain the optimal on-off switch policy. The learning procedure is defined as a Markov Decision Process (MDP). An estimation method is proposed to measure the load of the network. A single-agent Q-learning algorithm is proposed afterwards. The convergence of this algorithm is proved. Simulation results are given to evaluate the performance of the proposed algorithm.