基于连接智能体强化学习的多智能体动态区域覆盖

IF 2.2 4区 计算机科学 Q2 Computer Science
Fatih Aydemir, Aydın Çetin
{"title":"基于连接智能体强化学习的多智能体动态区域覆盖","authors":"Fatih Aydemir, Aydın Çetin","doi":"10.32604/csse.2023.031116","DOIUrl":null,"url":null,"abstract":"Dynamic area coverage with small unmanned aerial vehicle (UAV) systems is one of the major research topics due to limited payloads and the difficulty of decentralized decision-making process. Collaborative behavior of a group of UAVs in an unknown environment is another hard problem to be solved. In this paper, we propose a method for decentralized execution of multi-UAVs for dynamic area coverage problems. The proposed decentralized decision-making dynamic area coverage (DDMDAC) method utilizes reinforcement learning (RL) where each UAV is represented by an intelligent agent that learns policies to create collaborative behaviors in partially observable environment. Intelligent agents increase their global observations by gathering information about the environment by connecting with other agents. The connectivity provides a consensus for the decision-making process, while each agent takes decisions. At each step, agents acquire all reachable agents’ states, determine the optimum location for maximal area coverage and receive reward using the covered rate on the target area, respectively. The method was tested in a multi-agent actor-critic simulation platform. In the study, it has been considered that each UAV has a certain communication distance as in real applications. The results show that UAVs with limited communication distance can act jointly in the target area and can successfully cover the area without guidance from the central command unit.","PeriodicalId":50634,"journal":{"name":"Computer Systems Science and Engineering","volume":"17 1","pages":"215-230"},"PeriodicalIF":2.2000,"publicationDate":"2023-01-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"4","resultStr":"{\"title\":\"Multi-Agent Dynamic Area Coverage Based on Reinforcement Learning with Connected Agents\",\"authors\":\"Fatih Aydemir, Aydın Çetin\",\"doi\":\"10.32604/csse.2023.031116\",\"DOIUrl\":null,\"url\":null,\"abstract\":\"Dynamic area coverage with small unmanned aerial vehicle (UAV) systems is one of the major research topics due to limited payloads and the difficulty of decentralized decision-making process. Collaborative behavior of a group of UAVs in an unknown environment is another hard problem to be solved. In this paper, we propose a method for decentralized execution of multi-UAVs for dynamic area coverage problems. The proposed decentralized decision-making dynamic area coverage (DDMDAC) method utilizes reinforcement learning (RL) where each UAV is represented by an intelligent agent that learns policies to create collaborative behaviors in partially observable environment. Intelligent agents increase their global observations by gathering information about the environment by connecting with other agents. The connectivity provides a consensus for the decision-making process, while each agent takes decisions. At each step, agents acquire all reachable agents’ states, determine the optimum location for maximal area coverage and receive reward using the covered rate on the target area, respectively. The method was tested in a multi-agent actor-critic simulation platform. In the study, it has been considered that each UAV has a certain communication distance as in real applications. The results show that UAVs with limited communication distance can act jointly in the target area and can successfully cover the area without guidance from the central command unit.\",\"PeriodicalId\":50634,\"journal\":{\"name\":\"Computer Systems Science and Engineering\",\"volume\":\"17 1\",\"pages\":\"215-230\"},\"PeriodicalIF\":2.2000,\"publicationDate\":\"2023-01-01\",\"publicationTypes\":\"Journal Article\",\"fieldsOfStudy\":null,\"isOpenAccess\":false,\"openAccessPdf\":\"\",\"citationCount\":\"4\",\"resultStr\":null,\"platform\":\"Semanticscholar\",\"paperid\":null,\"PeriodicalName\":\"Computer Systems Science and Engineering\",\"FirstCategoryId\":\"94\",\"ListUrlMain\":\"https://doi.org/10.32604/csse.2023.031116\",\"RegionNum\":4,\"RegionCategory\":\"计算机科学\",\"ArticlePicture\":[],\"TitleCN\":null,\"AbstractTextCN\":null,\"PMCID\":null,\"EPubDate\":\"\",\"PubModel\":\"\",\"JCR\":\"Q2\",\"JCRName\":\"Computer Science\",\"Score\":null,\"Total\":0}","platform":"Semanticscholar","paperid":null,"PeriodicalName":"Computer Systems Science and Engineering","FirstCategoryId":"94","ListUrlMain":"https://doi.org/10.32604/csse.2023.031116","RegionNum":4,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"Q2","JCRName":"Computer Science","Score":null,"Total":0}
引用次数: 4

摘要

由于载荷有限和决策过程分散的困难,小型无人机系统的动态区域覆盖一直是研究的热点之一。无人机群在未知环境下的协同行为是另一个难以解决的问题。针对动态区域覆盖问题,提出了一种多无人机分散执行的方法。提出的分散决策动态区域覆盖(DDMDAC)方法利用强化学习(RL),其中每个无人机由一个智能代理代表,智能代理学习策略以在部分可观察环境中创建协作行为。智能代理通过与其他代理连接来收集有关环境的信息,从而增加其全局观察。连接为决策过程提供了共识,而每个代理都进行决策。在每一步中,智能体获取所有可达智能体的状态,确定最大区域覆盖的最佳位置,并分别使用目标区域的覆盖率获得奖励。该方法在多智能体行为者评价仿真平台上进行了验证。在研究中,考虑到每架无人机在实际应用中都有一定的通信距离。结果表明,在通信距离有限的情况下,无人机可以在目标区域内联合行动,并且可以在没有中央指挥单位引导的情况下成功覆盖目标区域。
本文章由计算机程序翻译,如有差异,请以英文原文为准。
Multi-Agent Dynamic Area Coverage Based on Reinforcement Learning with Connected Agents
Dynamic area coverage with small unmanned aerial vehicle (UAV) systems is one of the major research topics due to limited payloads and the difficulty of decentralized decision-making process. Collaborative behavior of a group of UAVs in an unknown environment is another hard problem to be solved. In this paper, we propose a method for decentralized execution of multi-UAVs for dynamic area coverage problems. The proposed decentralized decision-making dynamic area coverage (DDMDAC) method utilizes reinforcement learning (RL) where each UAV is represented by an intelligent agent that learns policies to create collaborative behaviors in partially observable environment. Intelligent agents increase their global observations by gathering information about the environment by connecting with other agents. The connectivity provides a consensus for the decision-making process, while each agent takes decisions. At each step, agents acquire all reachable agents’ states, determine the optimum location for maximal area coverage and receive reward using the covered rate on the target area, respectively. The method was tested in a multi-agent actor-critic simulation platform. In the study, it has been considered that each UAV has a certain communication distance as in real applications. The results show that UAVs with limited communication distance can act jointly in the target area and can successfully cover the area without guidance from the central command unit.
求助全文
通过发布文献求助,成功后即可免费获取论文全文。 去求助
来源期刊
Computer Systems Science and Engineering
Computer Systems Science and Engineering 工程技术-计算机:理论方法
CiteScore
3.10
自引率
13.60%
发文量
308
审稿时长
>12 weeks
期刊介绍: The journal is devoted to the publication of high quality papers on theoretical developments in computer systems science, and their applications in computer systems engineering. Original research papers, state-of-the-art reviews and technical notes are invited for publication. All papers will be refereed by acknowledged experts in the field, and may be (i) accepted without change, (ii) require amendment and subsequent re-refereeing, or (iii) be rejected on the grounds of either relevance or content. The submission of a paper implies that, if accepted for publication, it will not be published elsewhere in the same form, in any language, without the prior consent of the Publisher.
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
确定
请完成安全验证×
copy
已复制链接
快去分享给好友吧!
我知道了
右上角分享
点击右上角分享
0
联系我们:info@booksci.cn Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。 Copyright © 2023 布克学术 All rights reserved.
京ICP备2023020795号-1
ghs 京公网安备 11010802042870号
Book学术文献互助
Book学术文献互助群
群 号:481959085
Book学术官方微信