Derrik E. Asher, Michael Garber-Barron, Sebastian S. Rodriguez, Erin G. Zaroukian, Nicholas R. Waytowich
{"title":"状态空间扰动下的多智能体协调曲线","authors":"Derrik E. Asher, Michael Garber-Barron, Sebastian S. Rodriguez, Erin G. Zaroukian, Nicholas R. Waytowich","doi":"10.1109/CSCI49370.2019.00051","DOIUrl":null,"url":null,"abstract":"The current work utilized a multi-agent reinforcement learning (MARL) algorithm embedded in a continuous predator-prey pursuit simulation environment to measure and evaluate coordination between cooperating agents. In this simulation environment, it is generally assumed that successful performance for cooperative agents necessarily results in the emergence of coordination, but a clear quantitative demonstration of coordination in this environment still does not exist. The current work focuses on 1) detecting emergent coordination between cooperating agents in a multi-agent predator-prey simulation environment, and 2) showing coordination profiles between cooperating agents extracted from systematic state perturbations. This work introduces a method for detecting and comparing the typically 'black-box' behavioral solutions that result from emergent coordination in multi-agent learning spatial tasks with a shared goal. Comparing coordination profiles can provide insights into overlapping patterns that define how agents learn to interact in cooperative multi-agent environments. Similarly, this approach provides an avenue for measuring and training agents to coordinate with humans. In this way, the present work looks towards understanding and creating artificial team-mates that will strive to coordinate optimally.","PeriodicalId":103662,"journal":{"name":"2019 International Conference on Computational Science and Computational Intelligence (CSCI)","volume":"11 1","pages":"0"},"PeriodicalIF":0.0000,"publicationDate":"2019-12-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"6","resultStr":"{\"title\":\"Multi-Agent Coordination Profiles through State Space Perturbations\",\"authors\":\"Derrik E. Asher, Michael Garber-Barron, Sebastian S. Rodriguez, Erin G. Zaroukian, Nicholas R. Waytowich\",\"doi\":\"10.1109/CSCI49370.2019.00051\",\"DOIUrl\":null,\"url\":null,\"abstract\":\"The current work utilized a multi-agent reinforcement learning (MARL) algorithm embedded in a continuous predator-prey pursuit simulation environment to measure and evaluate coordination between cooperating agents. In this simulation environment, it is generally assumed that successful performance for cooperative agents necessarily results in the emergence of coordination, but a clear quantitative demonstration of coordination in this environment still does not exist. The current work focuses on 1) detecting emergent coordination between cooperating agents in a multi-agent predator-prey simulation environment, and 2) showing coordination profiles between cooperating agents extracted from systematic state perturbations. This work introduces a method for detecting and comparing the typically 'black-box' behavioral solutions that result from emergent coordination in multi-agent learning spatial tasks with a shared goal. Comparing coordination profiles can provide insights into overlapping patterns that define how agents learn to interact in cooperative multi-agent environments. Similarly, this approach provides an avenue for measuring and training agents to coordinate with humans. In this way, the present work looks towards understanding and creating artificial team-mates that will strive to coordinate optimally.\",\"PeriodicalId\":103662,\"journal\":{\"name\":\"2019 International Conference on Computational Science and Computational Intelligence (CSCI)\",\"volume\":\"11 1\",\"pages\":\"0\"},\"PeriodicalIF\":0.0000,\"publicationDate\":\"2019-12-01\",\"publicationTypes\":\"Journal Article\",\"fieldsOfStudy\":null,\"isOpenAccess\":false,\"openAccessPdf\":\"\",\"citationCount\":\"6\",\"resultStr\":null,\"platform\":\"Semanticscholar\",\"paperid\":null,\"PeriodicalName\":\"2019 International Conference on Computational Science and Computational Intelligence (CSCI)\",\"FirstCategoryId\":\"1085\",\"ListUrlMain\":\"https://doi.org/10.1109/CSCI49370.2019.00051\",\"RegionNum\":0,\"RegionCategory\":null,\"ArticlePicture\":[],\"TitleCN\":null,\"AbstractTextCN\":null,\"PMCID\":null,\"EPubDate\":\"\",\"PubModel\":\"\",\"JCR\":\"\",\"JCRName\":\"\",\"Score\":null,\"Total\":0}","platform":"Semanticscholar","paperid":null,"PeriodicalName":"2019 International Conference on Computational Science and Computational Intelligence (CSCI)","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/10.1109/CSCI49370.2019.00051","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
Multi-Agent Coordination Profiles through State Space Perturbations
The current work utilized a multi-agent reinforcement learning (MARL) algorithm embedded in a continuous predator-prey pursuit simulation environment to measure and evaluate coordination between cooperating agents. In this simulation environment, it is generally assumed that successful performance for cooperative agents necessarily results in the emergence of coordination, but a clear quantitative demonstration of coordination in this environment still does not exist. The current work focuses on 1) detecting emergent coordination between cooperating agents in a multi-agent predator-prey simulation environment, and 2) showing coordination profiles between cooperating agents extracted from systematic state perturbations. This work introduces a method for detecting and comparing the typically 'black-box' behavioral solutions that result from emergent coordination in multi-agent learning spatial tasks with a shared goal. Comparing coordination profiles can provide insights into overlapping patterns that define how agents learn to interact in cooperative multi-agent environments. Similarly, this approach provides an avenue for measuring and training agents to coordinate with humans. In this way, the present work looks towards understanding and creating artificial team-mates that will strive to coordinate optimally.