{"title":"集体多目标规划","authors":"A. Mouaddib, M. Juin","doi":"10.1109/DIS.2006.17","DOIUrl":null,"url":null,"abstract":"Multi-objective multi-agent planning addresses the problem of resolving conflicts between individual agent interests and the group interests. In this paper, we address this problem using a vector-valued decentralized Markov decision process (2V-DEC-MDP). The formal framework of a vector-valued MDP considered uses the value function which returns a vector representing the individual and the group interests. Indeed, local decisions of each agent are guided by the local reward of the agent, the opportunity gain to the group and the opportunity cost of its nuisance. An optimal policy in such contexts is not clear but in this approach we develop a regret-based technique to find a good tradeoff between the group and individual interests. To do that, the approach we present uses Egalitarian social welfare orderings that allow an agent to consider during its local optimization the satisfaction of all criteria and reducing their differences. The obtained result is an equilibrium of individual and group satisfactions where the local policies can lead to more global satisfying behaviors in some settings. This result is illustrated in an example and compared to alternate local policies","PeriodicalId":318812,"journal":{"name":"IEEE Workshop on Distributed Intelligent Systems: Collective Intelligence and Its Applications (DIS'06)","volume":"41 1","pages":"0"},"PeriodicalIF":0.0000,"publicationDate":"2006-06-15","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"5","resultStr":"{\"title\":\"Collective Multi-Objective Planning\",\"authors\":\"A. Mouaddib, M. Juin\",\"doi\":\"10.1109/DIS.2006.17\",\"DOIUrl\":null,\"url\":null,\"abstract\":\"Multi-objective multi-agent planning addresses the problem of resolving conflicts between individual agent interests and the group interests. In this paper, we address this problem using a vector-valued decentralized Markov decision process (2V-DEC-MDP). The formal framework of a vector-valued MDP considered uses the value function which returns a vector representing the individual and the group interests. Indeed, local decisions of each agent are guided by the local reward of the agent, the opportunity gain to the group and the opportunity cost of its nuisance. An optimal policy in such contexts is not clear but in this approach we develop a regret-based technique to find a good tradeoff between the group and individual interests. To do that, the approach we present uses Egalitarian social welfare orderings that allow an agent to consider during its local optimization the satisfaction of all criteria and reducing their differences. The obtained result is an equilibrium of individual and group satisfactions where the local policies can lead to more global satisfying behaviors in some settings. This result is illustrated in an example and compared to alternate local policies\",\"PeriodicalId\":318812,\"journal\":{\"name\":\"IEEE Workshop on Distributed Intelligent Systems: Collective Intelligence and Its Applications (DIS'06)\",\"volume\":\"41 1\",\"pages\":\"0\"},\"PeriodicalIF\":0.0000,\"publicationDate\":\"2006-06-15\",\"publicationTypes\":\"Journal Article\",\"fieldsOfStudy\":null,\"isOpenAccess\":false,\"openAccessPdf\":\"\",\"citationCount\":\"5\",\"resultStr\":null,\"platform\":\"Semanticscholar\",\"paperid\":null,\"PeriodicalName\":\"IEEE Workshop on Distributed Intelligent Systems: Collective Intelligence and Its Applications (DIS'06)\",\"FirstCategoryId\":\"1085\",\"ListUrlMain\":\"https://doi.org/10.1109/DIS.2006.17\",\"RegionNum\":0,\"RegionCategory\":null,\"ArticlePicture\":[],\"TitleCN\":null,\"AbstractTextCN\":null,\"PMCID\":null,\"EPubDate\":\"\",\"PubModel\":\"\",\"JCR\":\"\",\"JCRName\":\"\",\"Score\":null,\"Total\":0}","platform":"Semanticscholar","paperid":null,"PeriodicalName":"IEEE Workshop on Distributed Intelligent Systems: Collective Intelligence and Its Applications (DIS'06)","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/10.1109/DIS.2006.17","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
Multi-objective multi-agent planning addresses the problem of resolving conflicts between individual agent interests and the group interests. In this paper, we address this problem using a vector-valued decentralized Markov decision process (2V-DEC-MDP). The formal framework of a vector-valued MDP considered uses the value function which returns a vector representing the individual and the group interests. Indeed, local decisions of each agent are guided by the local reward of the agent, the opportunity gain to the group and the opportunity cost of its nuisance. An optimal policy in such contexts is not clear but in this approach we develop a regret-based technique to find a good tradeoff between the group and individual interests. To do that, the approach we present uses Egalitarian social welfare orderings that allow an agent to consider during its local optimization the satisfaction of all criteria and reducing their differences. The obtained result is an equilibrium of individual and group satisfactions where the local policies can lead to more global satisfying behaviors in some settings. This result is illustrated in an example and compared to alternate local policies