{"title":"基于威胁分析的无人机多智能体路径规划","authors":"Gang Lei, Min-zhou Dong, Tao Xu, Liang Wang","doi":"10.1109/ISA.2011.5873344","DOIUrl":null,"url":null,"abstract":"This paper focuses on the flight path planning process with multi-agent for Unmanned Aerial Vehicle (UAV) based on threats analysis and path length constraint. Path planner agent searches the path with global view considering path length constraint and information collector agent deals with path planning in the zone of threats. Scoring function is presented based on analysis the threats' attributes. We consider the path planning process as the multi-agent cooperation in a dynamic and non-stationarity environment. In order to perfectly adapt agents to environment changing, we restructure the traditional Q-value learning algorithm into a dynamic reinforcement learning algorithm by introducing current beliefs and recency-based exploration bonus. The simulation results show that the proposed method converges rapidly and can be used in flight path planning¿D","PeriodicalId":128163,"journal":{"name":"2011 3rd International Workshop on Intelligent Systems and Applications","volume":"285 1","pages":"0"},"PeriodicalIF":0.0000,"publicationDate":"2011-05-28","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"11","resultStr":"{\"title\":\"Multi-Agent Path Planning for Unmanned Aerial Vehicle Based on Threats Analysis\",\"authors\":\"Gang Lei, Min-zhou Dong, Tao Xu, Liang Wang\",\"doi\":\"10.1109/ISA.2011.5873344\",\"DOIUrl\":null,\"url\":null,\"abstract\":\"This paper focuses on the flight path planning process with multi-agent for Unmanned Aerial Vehicle (UAV) based on threats analysis and path length constraint. Path planner agent searches the path with global view considering path length constraint and information collector agent deals with path planning in the zone of threats. Scoring function is presented based on analysis the threats' attributes. We consider the path planning process as the multi-agent cooperation in a dynamic and non-stationarity environment. In order to perfectly adapt agents to environment changing, we restructure the traditional Q-value learning algorithm into a dynamic reinforcement learning algorithm by introducing current beliefs and recency-based exploration bonus. The simulation results show that the proposed method converges rapidly and can be used in flight path planning¿D\",\"PeriodicalId\":128163,\"journal\":{\"name\":\"2011 3rd International Workshop on Intelligent Systems and Applications\",\"volume\":\"285 1\",\"pages\":\"0\"},\"PeriodicalIF\":0.0000,\"publicationDate\":\"2011-05-28\",\"publicationTypes\":\"Journal Article\",\"fieldsOfStudy\":null,\"isOpenAccess\":false,\"openAccessPdf\":\"\",\"citationCount\":\"11\",\"resultStr\":null,\"platform\":\"Semanticscholar\",\"paperid\":null,\"PeriodicalName\":\"2011 3rd International Workshop on Intelligent Systems and Applications\",\"FirstCategoryId\":\"1085\",\"ListUrlMain\":\"https://doi.org/10.1109/ISA.2011.5873344\",\"RegionNum\":0,\"RegionCategory\":null,\"ArticlePicture\":[],\"TitleCN\":null,\"AbstractTextCN\":null,\"PMCID\":null,\"EPubDate\":\"\",\"PubModel\":\"\",\"JCR\":\"\",\"JCRName\":\"\",\"Score\":null,\"Total\":0}","platform":"Semanticscholar","paperid":null,"PeriodicalName":"2011 3rd International Workshop on Intelligent Systems and Applications","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/10.1109/ISA.2011.5873344","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
Multi-Agent Path Planning for Unmanned Aerial Vehicle Based on Threats Analysis
This paper focuses on the flight path planning process with multi-agent for Unmanned Aerial Vehicle (UAV) based on threats analysis and path length constraint. Path planner agent searches the path with global view considering path length constraint and information collector agent deals with path planning in the zone of threats. Scoring function is presented based on analysis the threats' attributes. We consider the path planning process as the multi-agent cooperation in a dynamic and non-stationarity environment. In order to perfectly adapt agents to environment changing, we restructure the traditional Q-value learning algorithm into a dynamic reinforcement learning algorithm by introducing current beliefs and recency-based exploration bonus. The simulation results show that the proposed method converges rapidly and can be used in flight path planning¿D