{"title":"姿态不确定条件下四轴飞行器避碰互反机动的神经进化学习","authors":"A. Behjat, Krushang Gabani, Souma Chowdhury","doi":"10.1115/detc2019-97975","DOIUrl":null,"url":null,"abstract":"\n This paper focuses on the idea of energy efficient cooperative collision avoidance between two quadcopters. Two strategies for reciprocal online collision-avoiding actions (i.e., coherent maneuvers without requiring any real-time consensus) are proposed. In the first strategy, UAVs change their speed, while in the second strategy they change their heading to avoid a collision. The avoidance actions are parameterized in terms of the time difference between detecting the collision and starting the maneuver and the amount of speed/heading change. These action parameters are used to generate intermediate way-points, subsequently translated into a minimum snap trajectory, to be executed by a PD controller. For realism, the relative pose of the other UAV, estimated by each UAV (at the point of detection), is considered to be uncertain — thereby presenting substantial challenges to undertaking reciprocal actions. Performing supervised learning based on optimization derived labels (as done in prior work) becomes computationally burden-some under these uncertainties. Instead, an (unsupervised) neuroevolution algorithm, called AGENT, is employed to learn a neural network (NN) model that takes the initial (uncertain) pose as state inputs and maps it to a robust optimal action. In neuroevolution, the NN topology and weights are simultaneously optimized using a special evolutionary process, where the fitness of candidate NNs are evaluated over a set of sample (in this case, various collision) scenarios. For further computational tractability, a surrogate model is used to estimate the energy consumption and a classifier is used to identify trajectories where the controller fails. The trained neural network shows encouraging performance for collision avoidance over a large variety of unseen scenarios.","PeriodicalId":365601,"journal":{"name":"Volume 2A: 45th Design Automation Conference","volume":"73 1","pages":"0"},"PeriodicalIF":0.0000,"publicationDate":"2019-11-25","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"2","resultStr":"{\"title\":\"A Neuroevolution-Based Learning of Reciprocal Maneuver for Collision Avoidance in Quadcopters Under Pose Uncertainties\",\"authors\":\"A. Behjat, Krushang Gabani, Souma Chowdhury\",\"doi\":\"10.1115/detc2019-97975\",\"DOIUrl\":null,\"url\":null,\"abstract\":\"\\n This paper focuses on the idea of energy efficient cooperative collision avoidance between two quadcopters. Two strategies for reciprocal online collision-avoiding actions (i.e., coherent maneuvers without requiring any real-time consensus) are proposed. In the first strategy, UAVs change their speed, while in the second strategy they change their heading to avoid a collision. The avoidance actions are parameterized in terms of the time difference between detecting the collision and starting the maneuver and the amount of speed/heading change. These action parameters are used to generate intermediate way-points, subsequently translated into a minimum snap trajectory, to be executed by a PD controller. For realism, the relative pose of the other UAV, estimated by each UAV (at the point of detection), is considered to be uncertain — thereby presenting substantial challenges to undertaking reciprocal actions. Performing supervised learning based on optimization derived labels (as done in prior work) becomes computationally burden-some under these uncertainties. Instead, an (unsupervised) neuroevolution algorithm, called AGENT, is employed to learn a neural network (NN) model that takes the initial (uncertain) pose as state inputs and maps it to a robust optimal action. In neuroevolution, the NN topology and weights are simultaneously optimized using a special evolutionary process, where the fitness of candidate NNs are evaluated over a set of sample (in this case, various collision) scenarios. For further computational tractability, a surrogate model is used to estimate the energy consumption and a classifier is used to identify trajectories where the controller fails. The trained neural network shows encouraging performance for collision avoidance over a large variety of unseen scenarios.\",\"PeriodicalId\":365601,\"journal\":{\"name\":\"Volume 2A: 45th Design Automation Conference\",\"volume\":\"73 1\",\"pages\":\"0\"},\"PeriodicalIF\":0.0000,\"publicationDate\":\"2019-11-25\",\"publicationTypes\":\"Journal Article\",\"fieldsOfStudy\":null,\"isOpenAccess\":false,\"openAccessPdf\":\"\",\"citationCount\":\"2\",\"resultStr\":null,\"platform\":\"Semanticscholar\",\"paperid\":null,\"PeriodicalName\":\"Volume 2A: 45th Design Automation Conference\",\"FirstCategoryId\":\"1085\",\"ListUrlMain\":\"https://doi.org/10.1115/detc2019-97975\",\"RegionNum\":0,\"RegionCategory\":null,\"ArticlePicture\":[],\"TitleCN\":null,\"AbstractTextCN\":null,\"PMCID\":null,\"EPubDate\":\"\",\"PubModel\":\"\",\"JCR\":\"\",\"JCRName\":\"\",\"Score\":null,\"Total\":0}","platform":"Semanticscholar","paperid":null,"PeriodicalName":"Volume 2A: 45th Design Automation Conference","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/10.1115/detc2019-97975","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
A Neuroevolution-Based Learning of Reciprocal Maneuver for Collision Avoidance in Quadcopters Under Pose Uncertainties
This paper focuses on the idea of energy efficient cooperative collision avoidance between two quadcopters. Two strategies for reciprocal online collision-avoiding actions (i.e., coherent maneuvers without requiring any real-time consensus) are proposed. In the first strategy, UAVs change their speed, while in the second strategy they change their heading to avoid a collision. The avoidance actions are parameterized in terms of the time difference between detecting the collision and starting the maneuver and the amount of speed/heading change. These action parameters are used to generate intermediate way-points, subsequently translated into a minimum snap trajectory, to be executed by a PD controller. For realism, the relative pose of the other UAV, estimated by each UAV (at the point of detection), is considered to be uncertain — thereby presenting substantial challenges to undertaking reciprocal actions. Performing supervised learning based on optimization derived labels (as done in prior work) becomes computationally burden-some under these uncertainties. Instead, an (unsupervised) neuroevolution algorithm, called AGENT, is employed to learn a neural network (NN) model that takes the initial (uncertain) pose as state inputs and maps it to a robust optimal action. In neuroevolution, the NN topology and weights are simultaneously optimized using a special evolutionary process, where the fitness of candidate NNs are evaluated over a set of sample (in this case, various collision) scenarios. For further computational tractability, a surrogate model is used to estimate the energy consumption and a classifier is used to identify trajectories where the controller fails. The trained neural network shows encouraging performance for collision avoidance over a large variety of unseen scenarios.