{"title":"分布式机器人多类型资源分配公平性的深度强化学习","authors":"Qinyun Zhu, J. Oh","doi":"10.1109/ICMLA.2018.00075","DOIUrl":null,"url":null,"abstract":"As autonomous robots are becoming a reality, we discover new challenges in coordination among these robots. We present a unique new problem that each robot makes decisions in achieving tasks that require multiple robots with different capabilities. Fair resource allocation is essential to ensure that all resource requesters acquire adequate robot resources and accomplish their tasks. We propose solutions to the fairness problem in multi-type resource allocation for multi-robot systems that have multiple resource requesters requiring heterogeneous robots with different capabilities to accomplish tasks. In particular, this work focuses on systems of single-tasking robots with multi-robot tasks (STR-MRT). In STR-MRT, the capability of a robot is the resource for accomplishing a specific task. In this problem, 1) each robot can perform only one task at a time, 2) tasks are divisible, and 3) accomplishing each task requires resources from one or more robots. We model the decentralized resource allocation in STR-MRT as a coordination game between the robots. Each robot strategically selects one resource requester. Then a consensus-based algorithm conducts formation of a robotic team for each task. We leverage the Dominant Resource Fairness (DRF) and Deep Q-learning Network (DQN) to support requester selection. The results suggest that the DQN outperforms the common Q-learning.","PeriodicalId":6533,"journal":{"name":"2018 17th IEEE International Conference on Machine Learning and Applications (ICMLA)","volume":"387 1","pages":"460-466"},"PeriodicalIF":0.0000,"publicationDate":"2018-12-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"7","resultStr":"{\"title\":\"Deep Reinforcement Learning for Fairness in Distributed Robotic Multi-type Resource Allocation\",\"authors\":\"Qinyun Zhu, J. Oh\",\"doi\":\"10.1109/ICMLA.2018.00075\",\"DOIUrl\":null,\"url\":null,\"abstract\":\"As autonomous robots are becoming a reality, we discover new challenges in coordination among these robots. We present a unique new problem that each robot makes decisions in achieving tasks that require multiple robots with different capabilities. Fair resource allocation is essential to ensure that all resource requesters acquire adequate robot resources and accomplish their tasks. We propose solutions to the fairness problem in multi-type resource allocation for multi-robot systems that have multiple resource requesters requiring heterogeneous robots with different capabilities to accomplish tasks. In particular, this work focuses on systems of single-tasking robots with multi-robot tasks (STR-MRT). In STR-MRT, the capability of a robot is the resource for accomplishing a specific task. In this problem, 1) each robot can perform only one task at a time, 2) tasks are divisible, and 3) accomplishing each task requires resources from one or more robots. We model the decentralized resource allocation in STR-MRT as a coordination game between the robots. Each robot strategically selects one resource requester. Then a consensus-based algorithm conducts formation of a robotic team for each task. We leverage the Dominant Resource Fairness (DRF) and Deep Q-learning Network (DQN) to support requester selection. The results suggest that the DQN outperforms the common Q-learning.\",\"PeriodicalId\":6533,\"journal\":{\"name\":\"2018 17th IEEE International Conference on Machine Learning and Applications (ICMLA)\",\"volume\":\"387 1\",\"pages\":\"460-466\"},\"PeriodicalIF\":0.0000,\"publicationDate\":\"2018-12-01\",\"publicationTypes\":\"Journal Article\",\"fieldsOfStudy\":null,\"isOpenAccess\":false,\"openAccessPdf\":\"\",\"citationCount\":\"7\",\"resultStr\":null,\"platform\":\"Semanticscholar\",\"paperid\":null,\"PeriodicalName\":\"2018 17th IEEE International Conference on Machine Learning and Applications (ICMLA)\",\"FirstCategoryId\":\"1085\",\"ListUrlMain\":\"https://doi.org/10.1109/ICMLA.2018.00075\",\"RegionNum\":0,\"RegionCategory\":null,\"ArticlePicture\":[],\"TitleCN\":null,\"AbstractTextCN\":null,\"PMCID\":null,\"EPubDate\":\"\",\"PubModel\":\"\",\"JCR\":\"\",\"JCRName\":\"\",\"Score\":null,\"Total\":0}","platform":"Semanticscholar","paperid":null,"PeriodicalName":"2018 17th IEEE International Conference on Machine Learning and Applications (ICMLA)","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/10.1109/ICMLA.2018.00075","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
Deep Reinforcement Learning for Fairness in Distributed Robotic Multi-type Resource Allocation
As autonomous robots are becoming a reality, we discover new challenges in coordination among these robots. We present a unique new problem that each robot makes decisions in achieving tasks that require multiple robots with different capabilities. Fair resource allocation is essential to ensure that all resource requesters acquire adequate robot resources and accomplish their tasks. We propose solutions to the fairness problem in multi-type resource allocation for multi-robot systems that have multiple resource requesters requiring heterogeneous robots with different capabilities to accomplish tasks. In particular, this work focuses on systems of single-tasking robots with multi-robot tasks (STR-MRT). In STR-MRT, the capability of a robot is the resource for accomplishing a specific task. In this problem, 1) each robot can perform only one task at a time, 2) tasks are divisible, and 3) accomplishing each task requires resources from one or more robots. We model the decentralized resource allocation in STR-MRT as a coordination game between the robots. Each robot strategically selects one resource requester. Then a consensus-based algorithm conducts formation of a robotic team for each task. We leverage the Dominant Resource Fairness (DRF) and Deep Q-learning Network (DQN) to support requester selection. The results suggest that the DQN outperforms the common Q-learning.