{"title":"Autonomous Quadrotor Path Planning Through Deep Reinforcement Learning With Monocular Depth Estimation","authors":"Mahdi Shahbazi Khojasteh;Armin Salimi-Badr","doi":"10.1109/OJVT.2024.3502296","DOIUrl":null,"url":null,"abstract":"Autonomous navigation is a formidable challenge for autonomous aerial vehicles operating in dense or dynamic environments. This paper proposes a path-planning approach based on deep reinforcement learning for a quadrotor equipped with only a monocular camera. The proposed method employs a two-stage structure comprising a depth estimation and a decision-making module. The former module uses a convolutional encoder-decoder network to learn image depth from visual cues self-supervised, with the output serving as input for the latter module. The latter module uses dueling double deep recurrent Q-learning to make decisions in high-dimensional and partially observable state spaces. To reduce meaningless explorations, we introduce the Insight Memory Pool alongside the regular memory pool to provide a rapid boost in learning by emphasizing early sampling from it and relying on the agent's experiences later. Once the agent has gained enough knowledge from the insightful data, we transition to a targeted exploration phase by employing the Boltzmann behavior policy, which relies on the refined Q-value estimates. To validate our approach, we tested the model in three diverse environments simulated with AirSim: a dynamic city street, a downtown, and a pillar world, each with different weather conditions. Experimental results show that our method significantly improves success rates and demonstrates strong generalization across various starting points and environmental transformations.","PeriodicalId":34270,"journal":{"name":"IEEE Open Journal of Vehicular Technology","volume":"6 ","pages":"34-51"},"PeriodicalIF":5.3000,"publicationDate":"2024-11-19","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://ieeexplore.ieee.org/stamp/stamp.jsp?tp=&arnumber=10758436","citationCount":"0","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"IEEE Open Journal of Vehicular Technology","FirstCategoryId":"1085","ListUrlMain":"https://ieeexplore.ieee.org/document/10758436/","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"Q1","JCRName":"ENGINEERING, ELECTRICAL & ELECTRONIC","Score":null,"Total":0}
引用次数: 0
Abstract
Autonomous navigation is a formidable challenge for autonomous aerial vehicles operating in dense or dynamic environments. This paper proposes a path-planning approach based on deep reinforcement learning for a quadrotor equipped with only a monocular camera. The proposed method employs a two-stage structure comprising a depth estimation and a decision-making module. The former module uses a convolutional encoder-decoder network to learn image depth from visual cues self-supervised, with the output serving as input for the latter module. The latter module uses dueling double deep recurrent Q-learning to make decisions in high-dimensional and partially observable state spaces. To reduce meaningless explorations, we introduce the Insight Memory Pool alongside the regular memory pool to provide a rapid boost in learning by emphasizing early sampling from it and relying on the agent's experiences later. Once the agent has gained enough knowledge from the insightful data, we transition to a targeted exploration phase by employing the Boltzmann behavior policy, which relies on the refined Q-value estimates. To validate our approach, we tested the model in three diverse environments simulated with AirSim: a dynamic city street, a downtown, and a pillar world, each with different weather conditions. Experimental results show that our method significantly improves success rates and demonstrates strong generalization across various starting points and environmental transformations.