Dongkun Huo, Huateng Zhang, Yixue Hao, Yuanlin Ye, Long Hu, Rui Wang, Min Chen
{"title":"DCMAC: Demand-aware Customized Multi-Agent Communication via Upper Bound Training","authors":"Dongkun Huo, Huateng Zhang, Yixue Hao, Yuanlin Ye, Long Hu, Rui Wang, Min Chen","doi":"arxiv-2409.07127","DOIUrl":null,"url":null,"abstract":"Efficient communication can enhance the overall performance of collaborative\nmulti-agent reinforcement learning. A common approach is to share observations\nthrough full communication, leading to significant communication overhead.\nExisting work attempts to perceive the global state by conducting teammate\nmodel based on local information. However, they ignore that the uncertainty\ngenerated by prediction may lead to difficult training. To address this\nproblem, we propose a Demand-aware Customized Multi-Agent Communication (DCMAC)\nprotocol, which use an upper bound training to obtain the ideal policy. By\nutilizing the demand parsing module, agent can interpret the gain of sending\nlocal message on teammate, and generate customized messages via compute the\ncorrelation between demands and local observation using cross-attention\nmechanism. Moreover, our method can adapt to the communication resources of\nagents and accelerate the training progress by appropriating the ideal policy\nwhich is trained with joint observation. Experimental results reveal that DCMAC\nsignificantly outperforms the baseline algorithms in both unconstrained and\ncommunication constrained scenarios.","PeriodicalId":501315,"journal":{"name":"arXiv - CS - Multiagent Systems","volume":"8 1","pages":""},"PeriodicalIF":0.0000,"publicationDate":"2024-09-11","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"arXiv - CS - Multiagent Systems","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/arxiv-2409.07127","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
引用次数: 0
Abstract
Efficient communication can enhance the overall performance of collaborative
multi-agent reinforcement learning. A common approach is to share observations
through full communication, leading to significant communication overhead.
Existing work attempts to perceive the global state by conducting teammate
model based on local information. However, they ignore that the uncertainty
generated by prediction may lead to difficult training. To address this
problem, we propose a Demand-aware Customized Multi-Agent Communication (DCMAC)
protocol, which use an upper bound training to obtain the ideal policy. By
utilizing the demand parsing module, agent can interpret the gain of sending
local message on teammate, and generate customized messages via compute the
correlation between demands and local observation using cross-attention
mechanism. Moreover, our method can adapt to the communication resources of
agents and accelerate the training progress by appropriating the ideal policy
which is trained with joint observation. Experimental results reveal that DCMAC
significantly outperforms the baseline algorithms in both unconstrained and
communication constrained scenarios.