{"title":"FERO: Efficient Deep Reinforcement Learning based UAV Obstacle Avoidance at the Edge","authors":"Patrick McEnroe;Shen Wang;Madhusanka Liyanage","doi":"10.1109/OJCS.2025.3600916","DOIUrl":null,"url":null,"abstract":"With the expanding use of unmanned aerial vehicles (UAVs) across various fields, efficient obstacle avoidance has become increasingly crucial. This UAV obstacle avoidance can be achieved through deep reinforcement learning (DRL) algorithms deployed directly on-device (i.e., at the edge). However, practical deployment is constrained by high training time and high inference latency. In this paper, we propose methods to improve DRL-based UAV obstacle avoidance efficiency through improving both training efficiency and inference latency. To reduce inference latency, we employ input dimension reduction, streamlining the state representation to enable faster decision-making. For training time reduction, we leverage transfer learning, allowing the obstacle avoidance models to rapidly adapt to new environments without starting from scratch. To show the generalizability of our methods, we applied them to a discrete action space dueling double deep Q-network (D3QN) model and a continuous action space soft actor critic (SAC) model. Inference results are evaluated on both an NVIDIA Jetson Nano edge device and a NVIDIA Jetson Orin Nano edge device and we propose a combined method called FERO which combines state space reduction, transfer learning, and conversion to TensorRT for optimum deployment on NVIDIA Jetson devices. For our individual methods and combined method, we demonstrate reductions in training and inference times with minimal compromise in obstacle avoidance performance.","PeriodicalId":13205,"journal":{"name":"IEEE Open Journal of the Computer Society","volume":"6 ","pages":"1378-1389"},"PeriodicalIF":0.0000,"publicationDate":"2025-08-20","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://ieeexplore.ieee.org/stamp/stamp.jsp?tp=&arnumber=11130910","citationCount":"0","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"IEEE Open Journal of the Computer Society","FirstCategoryId":"1085","ListUrlMain":"https://ieeexplore.ieee.org/document/11130910/","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
引用次数: 0
Abstract
With the expanding use of unmanned aerial vehicles (UAVs) across various fields, efficient obstacle avoidance has become increasingly crucial. This UAV obstacle avoidance can be achieved through deep reinforcement learning (DRL) algorithms deployed directly on-device (i.e., at the edge). However, practical deployment is constrained by high training time and high inference latency. In this paper, we propose methods to improve DRL-based UAV obstacle avoidance efficiency through improving both training efficiency and inference latency. To reduce inference latency, we employ input dimension reduction, streamlining the state representation to enable faster decision-making. For training time reduction, we leverage transfer learning, allowing the obstacle avoidance models to rapidly adapt to new environments without starting from scratch. To show the generalizability of our methods, we applied them to a discrete action space dueling double deep Q-network (D3QN) model and a continuous action space soft actor critic (SAC) model. Inference results are evaluated on both an NVIDIA Jetson Nano edge device and a NVIDIA Jetson Orin Nano edge device and we propose a combined method called FERO which combines state space reduction, transfer learning, and conversion to TensorRT for optimum deployment on NVIDIA Jetson devices. For our individual methods and combined method, we demonstrate reductions in training and inference times with minimal compromise in obstacle avoidance performance.