Shubhangi Bhadauria, L. Ravichandran, Elke Roth-Mandutz, Georg Fischer
{"title":"基于QoS的V2X资源分配的多agent与单agent深度强化学习","authors":"Shubhangi Bhadauria, L. Ravichandran, Elke Roth-Mandutz, Georg Fischer","doi":"10.1109/SOFTT54252.2021.9673150","DOIUrl":null,"url":null,"abstract":"Autonomous driving requires Vehicle-to-Everything (V2X) communication as standardized in the 3rd generation partnership project (3GPP). Diverse use cases and service types are foreseen to be supported, including safety-critical use cases, e.g., lane merging and cooperative collision avoidance. Each service type's quality of service (QoS) requirements vary enormously regarding latency, reliability, data rates, and positioning accuracy. In this paper, we analyze and evaluate the performance of a QoS-aware decentralized resource allocation scheme using first, a single-agent reinforcement learning (SARL) and second, a multi-agent reinforcement learning (MARL) approach. In addition, the impact of multiple vehicular user equipments (V-UEs) supporting one and multiple services are considered. The QoS parameter considered here is the latency and the relative distance between the communicating V-UEs, which is mapped on the Priority to reflect the required packet delay budget (PDB). The goal is to maximize the throughput of all V2N links while meeting the V2V link's latency constraint of the supported service. The results based on a system-level simulation for an urban scenario show that MARL improves the throughput for V-UEs set up for single and multiple services compared to SARL. However, for latency SARL indicates advantages at least when multiple services per V-UE apply.","PeriodicalId":443155,"journal":{"name":"2021 IEEE Symposium On Future Telecommunication Technologies (SOFTT)","volume":"90 2","pages":"0"},"PeriodicalIF":0.0000,"publicationDate":"2021-12-06","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":"{\"title\":\"QoS based Multi-Agent vs. Single-Agent Deep Reinforcement Learning for V2X Resource Allocation\",\"authors\":\"Shubhangi Bhadauria, L. Ravichandran, Elke Roth-Mandutz, Georg Fischer\",\"doi\":\"10.1109/SOFTT54252.2021.9673150\",\"DOIUrl\":null,\"url\":null,\"abstract\":\"Autonomous driving requires Vehicle-to-Everything (V2X) communication as standardized in the 3rd generation partnership project (3GPP). Diverse use cases and service types are foreseen to be supported, including safety-critical use cases, e.g., lane merging and cooperative collision avoidance. Each service type's quality of service (QoS) requirements vary enormously regarding latency, reliability, data rates, and positioning accuracy. In this paper, we analyze and evaluate the performance of a QoS-aware decentralized resource allocation scheme using first, a single-agent reinforcement learning (SARL) and second, a multi-agent reinforcement learning (MARL) approach. In addition, the impact of multiple vehicular user equipments (V-UEs) supporting one and multiple services are considered. The QoS parameter considered here is the latency and the relative distance between the communicating V-UEs, which is mapped on the Priority to reflect the required packet delay budget (PDB). The goal is to maximize the throughput of all V2N links while meeting the V2V link's latency constraint of the supported service. The results based on a system-level simulation for an urban scenario show that MARL improves the throughput for V-UEs set up for single and multiple services compared to SARL. However, for latency SARL indicates advantages at least when multiple services per V-UE apply.\",\"PeriodicalId\":443155,\"journal\":{\"name\":\"2021 IEEE Symposium On Future Telecommunication Technologies (SOFTT)\",\"volume\":\"90 2\",\"pages\":\"0\"},\"PeriodicalIF\":0.0000,\"publicationDate\":\"2021-12-06\",\"publicationTypes\":\"Journal Article\",\"fieldsOfStudy\":null,\"isOpenAccess\":false,\"openAccessPdf\":\"\",\"citationCount\":\"0\",\"resultStr\":null,\"platform\":\"Semanticscholar\",\"paperid\":null,\"PeriodicalName\":\"2021 IEEE Symposium On Future Telecommunication Technologies (SOFTT)\",\"FirstCategoryId\":\"1085\",\"ListUrlMain\":\"https://doi.org/10.1109/SOFTT54252.2021.9673150\",\"RegionNum\":0,\"RegionCategory\":null,\"ArticlePicture\":[],\"TitleCN\":null,\"AbstractTextCN\":null,\"PMCID\":null,\"EPubDate\":\"\",\"PubModel\":\"\",\"JCR\":\"\",\"JCRName\":\"\",\"Score\":null,\"Total\":0}","platform":"Semanticscholar","paperid":null,"PeriodicalName":"2021 IEEE Symposium On Future Telecommunication Technologies (SOFTT)","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/10.1109/SOFTT54252.2021.9673150","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
QoS based Multi-Agent vs. Single-Agent Deep Reinforcement Learning for V2X Resource Allocation
Autonomous driving requires Vehicle-to-Everything (V2X) communication as standardized in the 3rd generation partnership project (3GPP). Diverse use cases and service types are foreseen to be supported, including safety-critical use cases, e.g., lane merging and cooperative collision avoidance. Each service type's quality of service (QoS) requirements vary enormously regarding latency, reliability, data rates, and positioning accuracy. In this paper, we analyze and evaluate the performance of a QoS-aware decentralized resource allocation scheme using first, a single-agent reinforcement learning (SARL) and second, a multi-agent reinforcement learning (MARL) approach. In addition, the impact of multiple vehicular user equipments (V-UEs) supporting one and multiple services are considered. The QoS parameter considered here is the latency and the relative distance between the communicating V-UEs, which is mapped on the Priority to reflect the required packet delay budget (PDB). The goal is to maximize the throughput of all V2N links while meeting the V2V link's latency constraint of the supported service. The results based on a system-level simulation for an urban scenario show that MARL improves the throughput for V-UEs set up for single and multiple services compared to SARL. However, for latency SARL indicates advantages at least when multiple services per V-UE apply.