{"title":"A Hybrid Sensing and Reinforcement Learning Scheduler for Vehicle-to-Vehicle Communications","authors":"T. Şahin, Mate Boban, R. Khalili, A. Wolisz","doi":"10.1109/IEEECONF44664.2019.9048691","DOIUrl":null,"url":null,"abstract":"Vehicle-to-vehicle (V2V) communications performance depends significantly on the approach taken to schedule the radio resources. When the infrastructure is available, so far the best performing V2V scheduling algorithms are based on centralized approach. In case there is no infrastructure, sensing the resources in a distributed manner to determine whether a specific resource is free performs well. We propose a hybrid solution, where a centralized reinforcement learning (RL) algorithm provides a candidate subset of resources, whereas a distributed sensing mechanism, running on each vehicle, makes the final resource selection. We evaluate the performance of the proposed approach in an out-of-coverage setting and show that it outperforms the state-of-the-art algorithms in highly dynamic scenarios by using the best of both worlds: RL agent provides optimized long-term resource allocations, while the distributed sensing handles temporary and unforeseen network conditions that can not be predicted effectively.","PeriodicalId":6684,"journal":{"name":"2019 53rd Asilomar Conference on Signals, Systems, and Computers","volume":"72 1","pages":"1136-1143"},"PeriodicalIF":0.0000,"publicationDate":"2019-11-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"1","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"2019 53rd Asilomar Conference on Signals, Systems, and Computers","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/10.1109/IEEECONF44664.2019.9048691","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
引用次数: 1
Abstract
Vehicle-to-vehicle (V2V) communications performance depends significantly on the approach taken to schedule the radio resources. When the infrastructure is available, so far the best performing V2V scheduling algorithms are based on centralized approach. In case there is no infrastructure, sensing the resources in a distributed manner to determine whether a specific resource is free performs well. We propose a hybrid solution, where a centralized reinforcement learning (RL) algorithm provides a candidate subset of resources, whereas a distributed sensing mechanism, running on each vehicle, makes the final resource selection. We evaluate the performance of the proposed approach in an out-of-coverage setting and show that it outperforms the state-of-the-art algorithms in highly dynamic scenarios by using the best of both worlds: RL agent provides optimized long-term resource allocations, while the distributed sensing handles temporary and unforeseen network conditions that can not be predicted effectively.