A Neuroevolution-Based Learning of Reciprocal Maneuver for Collision Avoidance in Quadcopters Under Pose Uncertainties

A. Behjat, Krushang Gabani, Souma Chowdhury
{"title":"A Neuroevolution-Based Learning of Reciprocal Maneuver for Collision Avoidance in Quadcopters Under Pose Uncertainties","authors":"A. Behjat, Krushang Gabani, Souma Chowdhury","doi":"10.1115/detc2019-97975","DOIUrl":null,"url":null,"abstract":"\n This paper focuses on the idea of energy efficient cooperative collision avoidance between two quadcopters. Two strategies for reciprocal online collision-avoiding actions (i.e., coherent maneuvers without requiring any real-time consensus) are proposed. In the first strategy, UAVs change their speed, while in the second strategy they change their heading to avoid a collision. The avoidance actions are parameterized in terms of the time difference between detecting the collision and starting the maneuver and the amount of speed/heading change. These action parameters are used to generate intermediate way-points, subsequently translated into a minimum snap trajectory, to be executed by a PD controller. For realism, the relative pose of the other UAV, estimated by each UAV (at the point of detection), is considered to be uncertain — thereby presenting substantial challenges to undertaking reciprocal actions. Performing supervised learning based on optimization derived labels (as done in prior work) becomes computationally burden-some under these uncertainties. Instead, an (unsupervised) neuroevolution algorithm, called AGENT, is employed to learn a neural network (NN) model that takes the initial (uncertain) pose as state inputs and maps it to a robust optimal action. In neuroevolution, the NN topology and weights are simultaneously optimized using a special evolutionary process, where the fitness of candidate NNs are evaluated over a set of sample (in this case, various collision) scenarios. For further computational tractability, a surrogate model is used to estimate the energy consumption and a classifier is used to identify trajectories where the controller fails. The trained neural network shows encouraging performance for collision avoidance over a large variety of unseen scenarios.","PeriodicalId":365601,"journal":{"name":"Volume 2A: 45th Design Automation Conference","volume":"73 1","pages":"0"},"PeriodicalIF":0.0000,"publicationDate":"2019-11-25","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"2","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"Volume 2A: 45th Design Automation Conference","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/10.1115/detc2019-97975","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
引用次数: 2

Abstract

This paper focuses on the idea of energy efficient cooperative collision avoidance between two quadcopters. Two strategies for reciprocal online collision-avoiding actions (i.e., coherent maneuvers without requiring any real-time consensus) are proposed. In the first strategy, UAVs change their speed, while in the second strategy they change their heading to avoid a collision. The avoidance actions are parameterized in terms of the time difference between detecting the collision and starting the maneuver and the amount of speed/heading change. These action parameters are used to generate intermediate way-points, subsequently translated into a minimum snap trajectory, to be executed by a PD controller. For realism, the relative pose of the other UAV, estimated by each UAV (at the point of detection), is considered to be uncertain — thereby presenting substantial challenges to undertaking reciprocal actions. Performing supervised learning based on optimization derived labels (as done in prior work) becomes computationally burden-some under these uncertainties. Instead, an (unsupervised) neuroevolution algorithm, called AGENT, is employed to learn a neural network (NN) model that takes the initial (uncertain) pose as state inputs and maps it to a robust optimal action. In neuroevolution, the NN topology and weights are simultaneously optimized using a special evolutionary process, where the fitness of candidate NNs are evaluated over a set of sample (in this case, various collision) scenarios. For further computational tractability, a surrogate model is used to estimate the energy consumption and a classifier is used to identify trajectories where the controller fails. The trained neural network shows encouraging performance for collision avoidance over a large variety of unseen scenarios.
姿态不确定条件下四轴飞行器避碰互反机动的神经进化学习
本文主要研究了两架四轴飞行器之间高效节能的协同避碰思想。提出了两种互惠在线避碰行动策略(即不需要任何实时共识的连贯机动)。在第一种策略中,无人机改变其速度,而在第二种策略中,无人机改变其航向以避免碰撞。避碰动作的参数化是根据检测到碰撞和开始机动之间的时间差以及速度/航向的变化量。这些动作参数用于生成中间路径点,随后转化为最小snap轨迹,由PD控制器执行。出于现实主义考虑,由每架无人机(在探测点)估计的其他无人机的相对姿态被认为是不确定的——因此对采取相互行动提出了实质性的挑战。在这些不确定性下,基于优化衍生标签执行监督学习(如在先前的工作中所做的那样)变得计算负担很大。相反,一个(无监督的)神经进化算法,称为AGENT,被用来学习一个神经网络(NN)模型,该模型将初始(不确定)姿态作为状态输入,并将其映射到一个鲁棒的最优动作。在神经进化中,使用特殊的进化过程同时优化神经网络的拓扑和权重,其中候选神经网络的适应度在一组样本(在这种情况下是各种碰撞)场景上进行评估。为了进一步的计算可追溯性,使用代理模型来估计能量消耗,并使用分类器来识别控制器失效的轨迹。经过训练的神经网络在各种未知场景中显示出令人鼓舞的避碰性能。
本文章由计算机程序翻译,如有差异,请以英文原文为准。
求助全文
约1分钟内获得全文 求助全文
来源期刊
自引率
0.00%
发文量
0
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
确定
请完成安全验证×
copy
已复制链接
快去分享给好友吧!
我知道了
右上角分享
点击右上角分享
0
联系我们:info@booksci.cn Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。 Copyright © 2023 布克学术 All rights reserved.
京ICP备2023020795号-1
ghs 京公网安备 11010802042870号
Book学术文献互助
Book学术文献互助群
群 号:481959085
Book学术官方微信