Rihab Gorsane, Omayma Mahjoub, Ruan de Kock, Roland Dubb, Siddarth Singh, Arnu Pretorius
{"title":"合作MARL的标准化绩效评估协议研究","authors":"Rihab Gorsane, Omayma Mahjoub, Ruan de Kock, Roland Dubb, Siddarth Singh, Arnu Pretorius","doi":"arxiv-2209.10485","DOIUrl":null,"url":null,"abstract":"Multi-agent reinforcement learning (MARL) has emerged as a useful approach to\nsolving decentralised decision-making problems at scale. Research in the field\nhas been growing steadily with many breakthrough algorithms proposed in recent\nyears. In this work, we take a closer look at this rapid development with a\nfocus on evaluation methodologies employed across a large body of research in\ncooperative MARL. By conducting a detailed meta-analysis of prior work,\nspanning 75 papers accepted for publication from 2016 to 2022, we bring to\nlight worrying trends that put into question the true rate of progress. We\nfurther consider these trends in a wider context and take inspiration from\nsingle-agent RL literature on similar issues with recommendations that remain\napplicable to MARL. Combining these recommendations, with novel insights from\nour analysis, we propose a standardised performance evaluation protocol for\ncooperative MARL. We argue that such a standard protocol, if widely adopted,\nwould greatly improve the validity and credibility of future research, make\nreplication and reproducibility easier, as well as improve the ability of the\nfield to accurately gauge the rate of progress over time by being able to make\nsound comparisons across different works. Finally, we release our meta-analysis\ndata publicly on our project website for future research on evaluation:\nhttps://sites.google.com/view/marl-standard-protocol","PeriodicalId":501533,"journal":{"name":"arXiv - CS - General Literature","volume":"76 1","pages":""},"PeriodicalIF":0.0000,"publicationDate":"2022-09-21","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":"{\"title\":\"Towards a Standardised Performance Evaluation Protocol for Cooperative MARL\",\"authors\":\"Rihab Gorsane, Omayma Mahjoub, Ruan de Kock, Roland Dubb, Siddarth Singh, Arnu Pretorius\",\"doi\":\"arxiv-2209.10485\",\"DOIUrl\":null,\"url\":null,\"abstract\":\"Multi-agent reinforcement learning (MARL) has emerged as a useful approach to\\nsolving decentralised decision-making problems at scale. Research in the field\\nhas been growing steadily with many breakthrough algorithms proposed in recent\\nyears. In this work, we take a closer look at this rapid development with a\\nfocus on evaluation methodologies employed across a large body of research in\\ncooperative MARL. By conducting a detailed meta-analysis of prior work,\\nspanning 75 papers accepted for publication from 2016 to 2022, we bring to\\nlight worrying trends that put into question the true rate of progress. We\\nfurther consider these trends in a wider context and take inspiration from\\nsingle-agent RL literature on similar issues with recommendations that remain\\napplicable to MARL. Combining these recommendations, with novel insights from\\nour analysis, we propose a standardised performance evaluation protocol for\\ncooperative MARL. We argue that such a standard protocol, if widely adopted,\\nwould greatly improve the validity and credibility of future research, make\\nreplication and reproducibility easier, as well as improve the ability of the\\nfield to accurately gauge the rate of progress over time by being able to make\\nsound comparisons across different works. Finally, we release our meta-analysis\\ndata publicly on our project website for future research on evaluation:\\nhttps://sites.google.com/view/marl-standard-protocol\",\"PeriodicalId\":501533,\"journal\":{\"name\":\"arXiv - CS - General Literature\",\"volume\":\"76 1\",\"pages\":\"\"},\"PeriodicalIF\":0.0000,\"publicationDate\":\"2022-09-21\",\"publicationTypes\":\"Journal Article\",\"fieldsOfStudy\":null,\"isOpenAccess\":false,\"openAccessPdf\":\"\",\"citationCount\":\"0\",\"resultStr\":null,\"platform\":\"Semanticscholar\",\"paperid\":null,\"PeriodicalName\":\"arXiv - CS - General Literature\",\"FirstCategoryId\":\"1085\",\"ListUrlMain\":\"https://doi.org/arxiv-2209.10485\",\"RegionNum\":0,\"RegionCategory\":null,\"ArticlePicture\":[],\"TitleCN\":null,\"AbstractTextCN\":null,\"PMCID\":null,\"EPubDate\":\"\",\"PubModel\":\"\",\"JCR\":\"\",\"JCRName\":\"\",\"Score\":null,\"Total\":0}","platform":"Semanticscholar","paperid":null,"PeriodicalName":"arXiv - CS - General Literature","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/arxiv-2209.10485","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
Towards a Standardised Performance Evaluation Protocol for Cooperative MARL
Multi-agent reinforcement learning (MARL) has emerged as a useful approach to
solving decentralised decision-making problems at scale. Research in the field
has been growing steadily with many breakthrough algorithms proposed in recent
years. In this work, we take a closer look at this rapid development with a
focus on evaluation methodologies employed across a large body of research in
cooperative MARL. By conducting a detailed meta-analysis of prior work,
spanning 75 papers accepted for publication from 2016 to 2022, we bring to
light worrying trends that put into question the true rate of progress. We
further consider these trends in a wider context and take inspiration from
single-agent RL literature on similar issues with recommendations that remain
applicable to MARL. Combining these recommendations, with novel insights from
our analysis, we propose a standardised performance evaluation protocol for
cooperative MARL. We argue that such a standard protocol, if widely adopted,
would greatly improve the validity and credibility of future research, make
replication and reproducibility easier, as well as improve the ability of the
field to accurately gauge the rate of progress over time by being able to make
sound comparisons across different works. Finally, we release our meta-analysis
data publicly on our project website for future research on evaluation:
https://sites.google.com/view/marl-standard-protocol