{"title":"基于深度强化学习的复杂三维环境下无人机覆盖路径规划","authors":"Julian Bialas, M. Döller","doi":"10.1109/ROBIO55434.2022.10011936","DOIUrl":null,"url":null,"abstract":"Coverage path planning (CPP) for unmanned aerial vehicles (UAVs) defines a vital role in the automation process of UAV-supported disaster management. While multiple algorithms exist to solve the CPP problem for planar areas, the proposed algorithm is the first to handle complex three-dimensional environments and also account for power constraints and changing environments. By applying proximal policy optimization to an advantage-based actor-critic deep reinforcement learning model, the proposed framework enables an agent to efficiently cover the target area (TA), considering the orientation of the observation sensor, avoiding collisions as well as no-flying zones (NFZ) and reacting to changing environments. Furthermore, a safe landing mechanism, based on the Dijkstra algorithm, expands the framework to guarantee a successful landing in the respective start and landing zone (SLZ) within the power constraints. The model is trained on real data to learn the optimal control policy. Additionally, the framework was tested and validated on hardware in a drone lab to confirm its effectiveness and capability to perform real-time path planning.","PeriodicalId":151112,"journal":{"name":"2022 IEEE International Conference on Robotics and Biomimetics (ROBIO)","volume":"107 ","pages":"0"},"PeriodicalIF":0.0000,"publicationDate":"2022-12-05","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":"{\"title\":\"Coverage Path Planning for Unmanned Aerial Vehicles in Complex 3D Environments with Deep Reinforcement Learning\",\"authors\":\"Julian Bialas, M. Döller\",\"doi\":\"10.1109/ROBIO55434.2022.10011936\",\"DOIUrl\":null,\"url\":null,\"abstract\":\"Coverage path planning (CPP) for unmanned aerial vehicles (UAVs) defines a vital role in the automation process of UAV-supported disaster management. While multiple algorithms exist to solve the CPP problem for planar areas, the proposed algorithm is the first to handle complex three-dimensional environments and also account for power constraints and changing environments. By applying proximal policy optimization to an advantage-based actor-critic deep reinforcement learning model, the proposed framework enables an agent to efficiently cover the target area (TA), considering the orientation of the observation sensor, avoiding collisions as well as no-flying zones (NFZ) and reacting to changing environments. Furthermore, a safe landing mechanism, based on the Dijkstra algorithm, expands the framework to guarantee a successful landing in the respective start and landing zone (SLZ) within the power constraints. The model is trained on real data to learn the optimal control policy. Additionally, the framework was tested and validated on hardware in a drone lab to confirm its effectiveness and capability to perform real-time path planning.\",\"PeriodicalId\":151112,\"journal\":{\"name\":\"2022 IEEE International Conference on Robotics and Biomimetics (ROBIO)\",\"volume\":\"107 \",\"pages\":\"0\"},\"PeriodicalIF\":0.0000,\"publicationDate\":\"2022-12-05\",\"publicationTypes\":\"Journal Article\",\"fieldsOfStudy\":null,\"isOpenAccess\":false,\"openAccessPdf\":\"\",\"citationCount\":\"0\",\"resultStr\":null,\"platform\":\"Semanticscholar\",\"paperid\":null,\"PeriodicalName\":\"2022 IEEE International Conference on Robotics and Biomimetics (ROBIO)\",\"FirstCategoryId\":\"1085\",\"ListUrlMain\":\"https://doi.org/10.1109/ROBIO55434.2022.10011936\",\"RegionNum\":0,\"RegionCategory\":null,\"ArticlePicture\":[],\"TitleCN\":null,\"AbstractTextCN\":null,\"PMCID\":null,\"EPubDate\":\"\",\"PubModel\":\"\",\"JCR\":\"\",\"JCRName\":\"\",\"Score\":null,\"Total\":0}","platform":"Semanticscholar","paperid":null,"PeriodicalName":"2022 IEEE International Conference on Robotics and Biomimetics (ROBIO)","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/10.1109/ROBIO55434.2022.10011936","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
Coverage Path Planning for Unmanned Aerial Vehicles in Complex 3D Environments with Deep Reinforcement Learning
Coverage path planning (CPP) for unmanned aerial vehicles (UAVs) defines a vital role in the automation process of UAV-supported disaster management. While multiple algorithms exist to solve the CPP problem for planar areas, the proposed algorithm is the first to handle complex three-dimensional environments and also account for power constraints and changing environments. By applying proximal policy optimization to an advantage-based actor-critic deep reinforcement learning model, the proposed framework enables an agent to efficiently cover the target area (TA), considering the orientation of the observation sensor, avoiding collisions as well as no-flying zones (NFZ) and reacting to changing environments. Furthermore, a safe landing mechanism, based on the Dijkstra algorithm, expands the framework to guarantee a successful landing in the respective start and landing zone (SLZ) within the power constraints. The model is trained on real data to learn the optimal control policy. Additionally, the framework was tested and validated on hardware in a drone lab to confirm its effectiveness and capability to perform real-time path planning.