{"title":"基于动机规则的交互式多智能体强化学习","authors":"T. Yamaguchi, Ryo Marukawa","doi":"10.1109/ICCIMA.2001.970456","DOIUrl":null,"url":null,"abstract":"Presents a new framework of multi-agent reinforcement learning to acquire cooperative behaviors by generating and coordinating each learning goal interactively among agents. One of the main goals of artificial intelligence is to realize an intelligent agent that behaves autonomously by its sense of values. Reinforcement learning (RL) is the major learning mechanism for the agent to adapt itself to various situations of an unknown environment flexibly. However, in a multi-agent system environment that has mutual dependency among agents, it is difficult for a human to set up suitable learning goals for each agent, and, in addition, the existing framework of RL that aims for egoistic optimality of each agent is inadequate. Therefore, an active and interactive learning mechanism is required to generate and coordinate each learning goal among the agents. To realize this, first we propose to treat each learning goal as a reinforcement signal (RS) that can be communicated among the agents. Second, we introduce motivation rules to integrate the RSs communicated among the agents into a reward value for RL of an agent. Then we define cooperative rewards as learning goals with mutual dependency. Learning experiments for two agents with various motivation rules are performed. The experimental results show that several combinations of motivation rules converge to cooperative behaviors.","PeriodicalId":232504,"journal":{"name":"Proceedings Fourth International Conference on Computational Intelligence and Multimedia Applications. ICCIMA 2001","volume":"68 1","pages":"0"},"PeriodicalIF":0.0000,"publicationDate":"2001-10-30","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"2","resultStr":"{\"title\":\"Interactive multiagent reinforcement learning with motivation rules\",\"authors\":\"T. Yamaguchi, Ryo Marukawa\",\"doi\":\"10.1109/ICCIMA.2001.970456\",\"DOIUrl\":null,\"url\":null,\"abstract\":\"Presents a new framework of multi-agent reinforcement learning to acquire cooperative behaviors by generating and coordinating each learning goal interactively among agents. One of the main goals of artificial intelligence is to realize an intelligent agent that behaves autonomously by its sense of values. Reinforcement learning (RL) is the major learning mechanism for the agent to adapt itself to various situations of an unknown environment flexibly. However, in a multi-agent system environment that has mutual dependency among agents, it is difficult for a human to set up suitable learning goals for each agent, and, in addition, the existing framework of RL that aims for egoistic optimality of each agent is inadequate. Therefore, an active and interactive learning mechanism is required to generate and coordinate each learning goal among the agents. To realize this, first we propose to treat each learning goal as a reinforcement signal (RS) that can be communicated among the agents. Second, we introduce motivation rules to integrate the RSs communicated among the agents into a reward value for RL of an agent. Then we define cooperative rewards as learning goals with mutual dependency. Learning experiments for two agents with various motivation rules are performed. The experimental results show that several combinations of motivation rules converge to cooperative behaviors.\",\"PeriodicalId\":232504,\"journal\":{\"name\":\"Proceedings Fourth International Conference on Computational Intelligence and Multimedia Applications. ICCIMA 2001\",\"volume\":\"68 1\",\"pages\":\"0\"},\"PeriodicalIF\":0.0000,\"publicationDate\":\"2001-10-30\",\"publicationTypes\":\"Journal Article\",\"fieldsOfStudy\":null,\"isOpenAccess\":false,\"openAccessPdf\":\"\",\"citationCount\":\"2\",\"resultStr\":null,\"platform\":\"Semanticscholar\",\"paperid\":null,\"PeriodicalName\":\"Proceedings Fourth International Conference on Computational Intelligence and Multimedia Applications. ICCIMA 2001\",\"FirstCategoryId\":\"1085\",\"ListUrlMain\":\"https://doi.org/10.1109/ICCIMA.2001.970456\",\"RegionNum\":0,\"RegionCategory\":null,\"ArticlePicture\":[],\"TitleCN\":null,\"AbstractTextCN\":null,\"PMCID\":null,\"EPubDate\":\"\",\"PubModel\":\"\",\"JCR\":\"\",\"JCRName\":\"\",\"Score\":null,\"Total\":0}","platform":"Semanticscholar","paperid":null,"PeriodicalName":"Proceedings Fourth International Conference on Computational Intelligence and Multimedia Applications. ICCIMA 2001","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/10.1109/ICCIMA.2001.970456","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
Interactive multiagent reinforcement learning with motivation rules
Presents a new framework of multi-agent reinforcement learning to acquire cooperative behaviors by generating and coordinating each learning goal interactively among agents. One of the main goals of artificial intelligence is to realize an intelligent agent that behaves autonomously by its sense of values. Reinforcement learning (RL) is the major learning mechanism for the agent to adapt itself to various situations of an unknown environment flexibly. However, in a multi-agent system environment that has mutual dependency among agents, it is difficult for a human to set up suitable learning goals for each agent, and, in addition, the existing framework of RL that aims for egoistic optimality of each agent is inadequate. Therefore, an active and interactive learning mechanism is required to generate and coordinate each learning goal among the agents. To realize this, first we propose to treat each learning goal as a reinforcement signal (RS) that can be communicated among the agents. Second, we introduce motivation rules to integrate the RSs communicated among the agents into a reward value for RL of an agent. Then we define cooperative rewards as learning goals with mutual dependency. Learning experiments for two agents with various motivation rules are performed. The experimental results show that several combinations of motivation rules converge to cooperative behaviors.