An Improved Method Based on Deep Reinforcement Learning for Target Searching

Xiaolong Wei, Xiangsong Huang, Tao Lu, Ge Song
{"title":"An Improved Method Based on Deep Reinforcement Learning for Target Searching","authors":"Xiaolong Wei, Xiangsong Huang, Tao Lu, Ge Song","doi":"10.1109/ICRAE48301.2019.9043821","DOIUrl":null,"url":null,"abstract":"Unmanned Aerial Vehicle (UAV), due to their high mobility and the ability to cover areas of different heights and locations at relatively low cost, are increasingly used for disaster monitoring and detecting. However, developing and testing UAVs in real world is an expensive task, especially in the domain of search and rescue, most of the previous systems are developed on the basis of greedy or potential-based heuristics without neural network. On the basis of the recent development of deep neural network architecture and deep reinforcement learning (DRL), in this research we improved the probability of success rate of searching target in an unstructured environment by combining image processing algorithms and reinforcement learning methods (RL). This paper aims at the deficiency of target tracking in unstructured environment, trying to propose an algorithm of stationary target positioning of UAV based on computer vision system. Firstly, a new input source is formed by acquiring depth information image of current environment and combining segmentation image. Secondly, the DQN algorithm is used to regulate the reinforcement learning model, and the specific flight response can be independently selected by the UAV through training. This paper utilizes open-source Microsoft UAV simulator AirSim as training and test environment based with Keras a machine learning framework. The main approach investigated in this research is modifying the network of Deep Q-Network, which designs the moving target tracking experiment of UAV in simulation scene. The experimental results demonstrate that this method has better tracking effect.","PeriodicalId":270665,"journal":{"name":"2019 4th International Conference on Robotics and Automation Engineering (ICRAE)","volume":"1 1","pages":"0"},"PeriodicalIF":0.0000,"publicationDate":"2019-11-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"10","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"2019 4th International Conference on Robotics and Automation Engineering (ICRAE)","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/10.1109/ICRAE48301.2019.9043821","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
引用次数: 10

Abstract

Unmanned Aerial Vehicle (UAV), due to their high mobility and the ability to cover areas of different heights and locations at relatively low cost, are increasingly used for disaster monitoring and detecting. However, developing and testing UAVs in real world is an expensive task, especially in the domain of search and rescue, most of the previous systems are developed on the basis of greedy or potential-based heuristics without neural network. On the basis of the recent development of deep neural network architecture and deep reinforcement learning (DRL), in this research we improved the probability of success rate of searching target in an unstructured environment by combining image processing algorithms and reinforcement learning methods (RL). This paper aims at the deficiency of target tracking in unstructured environment, trying to propose an algorithm of stationary target positioning of UAV based on computer vision system. Firstly, a new input source is formed by acquiring depth information image of current environment and combining segmentation image. Secondly, the DQN algorithm is used to regulate the reinforcement learning model, and the specific flight response can be independently selected by the UAV through training. This paper utilizes open-source Microsoft UAV simulator AirSim as training and test environment based with Keras a machine learning framework. The main approach investigated in this research is modifying the network of Deep Q-Network, which designs the moving target tracking experiment of UAV in simulation scene. The experimental results demonstrate that this method has better tracking effect.
一种基于深度强化学习的目标搜索改进方法
无人机(UAV)由于其高机动性和以相对较低的成本覆盖不同高度和位置的能力,越来越多地用于灾害监测和探测。然而,在现实世界中开发和测试无人机是一项昂贵的任务,特别是在搜索和救援领域,大多数以前的系统是基于贪婪或基于潜在的启发式开发的,没有神经网络。在深度神经网络架构和深度强化学习(DRL)最新发展的基础上,本研究将图像处理算法与强化学习方法(RL)相结合,提高了非结构化环境下目标搜索成功率的概率。本文针对非结构化环境下目标跟踪的不足,尝试提出一种基于计算机视觉系统的无人机静止目标定位算法。首先,通过获取当前环境的深度信息图像,结合分割图像形成新的输入源;其次,采用DQN算法对强化学习模型进行调节,使无人机通过训练自主选择具体的飞行响应;本文利用开源的微软无人机模拟器AirSim作为训练和测试环境,基于Keras机器学习框架。本文研究的主要方法是对Deep Q-Network网络进行改进,设计了无人机在仿真场景下的运动目标跟踪实验。实验结果表明,该方法具有较好的跟踪效果。
本文章由计算机程序翻译,如有差异,请以英文原文为准。
求助全文
约1分钟内获得全文 求助全文
来源期刊
自引率
0.00%
发文量
0
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
确定
请完成安全验证×
copy
已复制链接
快去分享给好友吧!
我知道了
右上角分享
点击右上角分享
0
联系我们:info@booksci.cn Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。 Copyright © 2023 布克学术 All rights reserved.
京ICP备2023020795号-1
ghs 京公网安备 11010802042870号
Book学术文献互助
Book学术文献互助群
群 号:481959085
Book学术官方微信