Siying Wang , Hongfei Du , Yang Zhou , Zhitong Zhao , Ruoning Zhang , Wenyu Chen
{"title":"利用相关轨迹加强多代理强化学习中的协作","authors":"Siying Wang , Hongfei Du , Yang Zhou , Zhitong Zhao , Ruoning Zhang , Wenyu Chen","doi":"10.1016/j.knosys.2024.112665","DOIUrl":null,"url":null,"abstract":"<div><div>Collaborative behaviors in human social activities can be modeled with multi-agent reinforcement learning and used to train the collaborative policies of agents to achieve efficient cooperation. In general, agents with similar behaviors have a certain behavioral common cognition and are more likely to understand the intentions of both parties then to form cooperative policies. Traditional approaches focus on the collaborative allocation process between agents, ignoring the effects of similar behaviors and common cognition characteristics in collaborative interactions. In order to better establish collaborative relationships between agents, we propose a novel multi-agent reinforcement learning collaborative algorithm based on the similarity of agents’ behavioral features. In this model, the interactions of agents are established as a graph neural network. Specifically, the Pearson correlation coefficient is proposed to compute the similarity of the history trajectories of the agents as a means of determining their behavioral common cognition, which is used to establish the weights of the edges in the modeled graph neural network. In addition, we design a transformer-encoder structured state information complementation module to enhance the decision representation of the agents. The experimental results on Predator–Prey and StarCraft II show that the proposed method can effectively enhance the collaborative behaviors between agents and improve the training efficiency of collaborative models.</div></div>","PeriodicalId":49939,"journal":{"name":"Knowledge-Based Systems","volume":"305 ","pages":"Article 112665"},"PeriodicalIF":7.2000,"publicationDate":"2024-10-28","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":"{\"title\":\"Enhancing collaboration in multi-agent reinforcement learning with correlated trajectories\",\"authors\":\"Siying Wang , Hongfei Du , Yang Zhou , Zhitong Zhao , Ruoning Zhang , Wenyu Chen\",\"doi\":\"10.1016/j.knosys.2024.112665\",\"DOIUrl\":null,\"url\":null,\"abstract\":\"<div><div>Collaborative behaviors in human social activities can be modeled with multi-agent reinforcement learning and used to train the collaborative policies of agents to achieve efficient cooperation. In general, agents with similar behaviors have a certain behavioral common cognition and are more likely to understand the intentions of both parties then to form cooperative policies. Traditional approaches focus on the collaborative allocation process between agents, ignoring the effects of similar behaviors and common cognition characteristics in collaborative interactions. In order to better establish collaborative relationships between agents, we propose a novel multi-agent reinforcement learning collaborative algorithm based on the similarity of agents’ behavioral features. In this model, the interactions of agents are established as a graph neural network. Specifically, the Pearson correlation coefficient is proposed to compute the similarity of the history trajectories of the agents as a means of determining their behavioral common cognition, which is used to establish the weights of the edges in the modeled graph neural network. In addition, we design a transformer-encoder structured state information complementation module to enhance the decision representation of the agents. The experimental results on Predator–Prey and StarCraft II show that the proposed method can effectively enhance the collaborative behaviors between agents and improve the training efficiency of collaborative models.</div></div>\",\"PeriodicalId\":49939,\"journal\":{\"name\":\"Knowledge-Based Systems\",\"volume\":\"305 \",\"pages\":\"Article 112665\"},\"PeriodicalIF\":7.2000,\"publicationDate\":\"2024-10-28\",\"publicationTypes\":\"Journal Article\",\"fieldsOfStudy\":null,\"isOpenAccess\":false,\"openAccessPdf\":\"\",\"citationCount\":\"0\",\"resultStr\":null,\"platform\":\"Semanticscholar\",\"paperid\":null,\"PeriodicalName\":\"Knowledge-Based Systems\",\"FirstCategoryId\":\"94\",\"ListUrlMain\":\"https://www.sciencedirect.com/science/article/pii/S0950705124012991\",\"RegionNum\":1,\"RegionCategory\":\"计算机科学\",\"ArticlePicture\":[],\"TitleCN\":null,\"AbstractTextCN\":null,\"PMCID\":null,\"EPubDate\":\"\",\"PubModel\":\"\",\"JCR\":\"Q1\",\"JCRName\":\"COMPUTER SCIENCE, ARTIFICIAL INTELLIGENCE\",\"Score\":null,\"Total\":0}","platform":"Semanticscholar","paperid":null,"PeriodicalName":"Knowledge-Based Systems","FirstCategoryId":"94","ListUrlMain":"https://www.sciencedirect.com/science/article/pii/S0950705124012991","RegionNum":1,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"Q1","JCRName":"COMPUTER SCIENCE, ARTIFICIAL INTELLIGENCE","Score":null,"Total":0}
Enhancing collaboration in multi-agent reinforcement learning with correlated trajectories
Collaborative behaviors in human social activities can be modeled with multi-agent reinforcement learning and used to train the collaborative policies of agents to achieve efficient cooperation. In general, agents with similar behaviors have a certain behavioral common cognition and are more likely to understand the intentions of both parties then to form cooperative policies. Traditional approaches focus on the collaborative allocation process between agents, ignoring the effects of similar behaviors and common cognition characteristics in collaborative interactions. In order to better establish collaborative relationships between agents, we propose a novel multi-agent reinforcement learning collaborative algorithm based on the similarity of agents’ behavioral features. In this model, the interactions of agents are established as a graph neural network. Specifically, the Pearson correlation coefficient is proposed to compute the similarity of the history trajectories of the agents as a means of determining their behavioral common cognition, which is used to establish the weights of the edges in the modeled graph neural network. In addition, we design a transformer-encoder structured state information complementation module to enhance the decision representation of the agents. The experimental results on Predator–Prey and StarCraft II show that the proposed method can effectively enhance the collaborative behaviors between agents and improve the training efficiency of collaborative models.
期刊介绍:
Knowledge-Based Systems, an international and interdisciplinary journal in artificial intelligence, publishes original, innovative, and creative research results in the field. It focuses on knowledge-based and other artificial intelligence techniques-based systems. The journal aims to support human prediction and decision-making through data science and computation techniques, provide a balanced coverage of theory and practical study, and encourage the development and implementation of knowledge-based intelligence models, methods, systems, and software tools. Applications in business, government, education, engineering, and healthcare are emphasized.