Mohammad Hasan Karami, Hossein Aghababa, A. Keyhanipour
{"title":"机器人导航中协作多智能体q -学习算法的纠缠启发行为选择和知识共享方案","authors":"Mohammad Hasan Karami, Hossein Aghababa, A. Keyhanipour","doi":"10.1109/ICCKE50421.2020.9303636","DOIUrl":null,"url":null,"abstract":"Multi-agent reinforcement learning, especially learning in unknown complex environments, requires new algorithms. In this work, our focus is on adopting the concept of the quantum entanglement phenomena to the action selection procedure of multi-agent Q-learning, aiming to enhance the learning speed, collision avoidance, and also providing full coverage of the environment. The exploration procedure is exclusively induced by a memory-based probabilistic sequential action selection method acting as a knowledge hub, shared among the agents, which is the central pillar of this work. This causes enhancing the parallelism of the learning process, plus, building an effective yet simple communicating-bridge between the learning agents; that is, they can signal and guide one another through sharing their gained experience and knowledge in order not to repeat the same mistake that the other agents have already run into. The simulation results demonstrated the effectiveness of our proposed algorithm in terms of reducing the learning time, significant reduction of collision occurrence, and providing full coverage of big complex clutter environments.","PeriodicalId":402043,"journal":{"name":"2020 10th International Conference on Computer and Knowledge Engineering (ICCKE)","volume":"5 1","pages":"0"},"PeriodicalIF":0.0000,"publicationDate":"2020-10-29","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":"{\"title\":\"An Entanglement-Inspired Action Selection and Knowledge Sharing Scheme for Cooperative Multi-agent Q-Learning Algorithm used in Robot Navigation\",\"authors\":\"Mohammad Hasan Karami, Hossein Aghababa, A. Keyhanipour\",\"doi\":\"10.1109/ICCKE50421.2020.9303636\",\"DOIUrl\":null,\"url\":null,\"abstract\":\"Multi-agent reinforcement learning, especially learning in unknown complex environments, requires new algorithms. In this work, our focus is on adopting the concept of the quantum entanglement phenomena to the action selection procedure of multi-agent Q-learning, aiming to enhance the learning speed, collision avoidance, and also providing full coverage of the environment. The exploration procedure is exclusively induced by a memory-based probabilistic sequential action selection method acting as a knowledge hub, shared among the agents, which is the central pillar of this work. This causes enhancing the parallelism of the learning process, plus, building an effective yet simple communicating-bridge between the learning agents; that is, they can signal and guide one another through sharing their gained experience and knowledge in order not to repeat the same mistake that the other agents have already run into. The simulation results demonstrated the effectiveness of our proposed algorithm in terms of reducing the learning time, significant reduction of collision occurrence, and providing full coverage of big complex clutter environments.\",\"PeriodicalId\":402043,\"journal\":{\"name\":\"2020 10th International Conference on Computer and Knowledge Engineering (ICCKE)\",\"volume\":\"5 1\",\"pages\":\"0\"},\"PeriodicalIF\":0.0000,\"publicationDate\":\"2020-10-29\",\"publicationTypes\":\"Journal Article\",\"fieldsOfStudy\":null,\"isOpenAccess\":false,\"openAccessPdf\":\"\",\"citationCount\":\"0\",\"resultStr\":null,\"platform\":\"Semanticscholar\",\"paperid\":null,\"PeriodicalName\":\"2020 10th International Conference on Computer and Knowledge Engineering (ICCKE)\",\"FirstCategoryId\":\"1085\",\"ListUrlMain\":\"https://doi.org/10.1109/ICCKE50421.2020.9303636\",\"RegionNum\":0,\"RegionCategory\":null,\"ArticlePicture\":[],\"TitleCN\":null,\"AbstractTextCN\":null,\"PMCID\":null,\"EPubDate\":\"\",\"PubModel\":\"\",\"JCR\":\"\",\"JCRName\":\"\",\"Score\":null,\"Total\":0}","platform":"Semanticscholar","paperid":null,"PeriodicalName":"2020 10th International Conference on Computer and Knowledge Engineering (ICCKE)","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/10.1109/ICCKE50421.2020.9303636","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
An Entanglement-Inspired Action Selection and Knowledge Sharing Scheme for Cooperative Multi-agent Q-Learning Algorithm used in Robot Navigation
Multi-agent reinforcement learning, especially learning in unknown complex environments, requires new algorithms. In this work, our focus is on adopting the concept of the quantum entanglement phenomena to the action selection procedure of multi-agent Q-learning, aiming to enhance the learning speed, collision avoidance, and also providing full coverage of the environment. The exploration procedure is exclusively induced by a memory-based probabilistic sequential action selection method acting as a knowledge hub, shared among the agents, which is the central pillar of this work. This causes enhancing the parallelism of the learning process, plus, building an effective yet simple communicating-bridge between the learning agents; that is, they can signal and guide one another through sharing their gained experience and knowledge in order not to repeat the same mistake that the other agents have already run into. The simulation results demonstrated the effectiveness of our proposed algorithm in terms of reducing the learning time, significant reduction of collision occurrence, and providing full coverage of big complex clutter environments.