{"title":"QTypeMix: Enhancing Multi-Agent Cooperative Strategies through Heterogeneous and Homogeneous Value Decomposition","authors":"Songchen Fu, Shaojing Zhao, Ta Li, YongHong Yan","doi":"arxiv-2408.07098","DOIUrl":null,"url":null,"abstract":"In multi-agent cooperative tasks, the presence of heterogeneous agents is\nfamiliar. Compared to cooperation among homogeneous agents, collaboration\nrequires considering the best-suited sub-tasks for each agent. However, the\noperation of multi-agent systems often involves a large amount of complex\ninteraction information, making it more challenging to learn heterogeneous\nstrategies. Related multi-agent reinforcement learning methods sometimes use\ngrouping mechanisms to form smaller cooperative groups or leverage prior domain\nknowledge to learn strategies for different roles. In contrast, agents should\nlearn deeper role features without relying on additional information.\nTherefore, we propose QTypeMix, which divides the value decomposition process\ninto homogeneous and heterogeneous stages. QTypeMix learns to extract type\nfeatures from local historical observations through the TE loss. In addition,\nwe introduce advanced network structures containing attention mechanisms and\nhypernets to enhance the representation capability and achieve the value\ndecomposition process. The results of testing the proposed method on 14 maps\nfrom SMAC and SMACv2 show that QTypeMix achieves state-of-the-art performance\nin tasks of varying difficulty.","PeriodicalId":501315,"journal":{"name":"arXiv - CS - Multiagent Systems","volume":"31 1","pages":""},"PeriodicalIF":0.0000,"publicationDate":"2024-08-12","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"arXiv - CS - Multiagent Systems","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/arxiv-2408.07098","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
引用次数: 0
Abstract
In multi-agent cooperative tasks, the presence of heterogeneous agents is
familiar. Compared to cooperation among homogeneous agents, collaboration
requires considering the best-suited sub-tasks for each agent. However, the
operation of multi-agent systems often involves a large amount of complex
interaction information, making it more challenging to learn heterogeneous
strategies. Related multi-agent reinforcement learning methods sometimes use
grouping mechanisms to form smaller cooperative groups or leverage prior domain
knowledge to learn strategies for different roles. In contrast, agents should
learn deeper role features without relying on additional information.
Therefore, we propose QTypeMix, which divides the value decomposition process
into homogeneous and heterogeneous stages. QTypeMix learns to extract type
features from local historical observations through the TE loss. In addition,
we introduce advanced network structures containing attention mechanisms and
hypernets to enhance the representation capability and achieve the value
decomposition process. The results of testing the proposed method on 14 maps
from SMAC and SMACv2 show that QTypeMix achieves state-of-the-art performance
in tasks of varying difficulty.