Obstacle avoidance navigation method for robot based on deep reinforcement learning

X. Ruan, Chenliang Lin, Jing Huang, Yufan Li
{"title":"Obstacle avoidance navigation method for robot based on deep reinforcement learning","authors":"X. Ruan, Chenliang Lin, Jing Huang, Yufan Li","doi":"10.1109/ITOEC53115.2022.9734337","DOIUrl":null,"url":null,"abstract":"Aiming at the navigation problem of mobile robots indoor environment, the traditional navigation algorithm based on D3QN has some problems such as sparse reward and slow training speed of the neural network. This paper proposes a deep reinforcement learning Algorithm (LN-D3QN) based on the D3QN network to realize collision-free autonomous navigation of mobile robots. To improve the efficiency of mobile robot learning and exploration, the Vision Sensor is used to acquire the input data from the environment, and the layer normalization method is used to normalize the input data. An improved reward function is designed, which improves the reward value of the algorithm, optimizes the state space, and alleviates the problem of sparse reward to some extent. The data is stored in a priority experience replay pool, and the network is trained using small batches of data. In addition, we evaluate our method by experiment related to indoor navigation. The experiments show that the robot trained by the improved D3QN algorithm can adapt to the unknown environment more quickly than the basic D3QN algorithm. The network's convergence speed is also improved, and it can complete the obstacle avoidance navigation task more efficiently.","PeriodicalId":127300,"journal":{"name":"2022 IEEE 6th Information Technology and Mechatronics Engineering Conference (ITOEC)","volume":"66 1","pages":"0"},"PeriodicalIF":0.0000,"publicationDate":"2022-03-04","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"1","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"2022 IEEE 6th Information Technology and Mechatronics Engineering Conference (ITOEC)","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/10.1109/ITOEC53115.2022.9734337","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
引用次数: 1

Abstract

Aiming at the navigation problem of mobile robots indoor environment, the traditional navigation algorithm based on D3QN has some problems such as sparse reward and slow training speed of the neural network. This paper proposes a deep reinforcement learning Algorithm (LN-D3QN) based on the D3QN network to realize collision-free autonomous navigation of mobile robots. To improve the efficiency of mobile robot learning and exploration, the Vision Sensor is used to acquire the input data from the environment, and the layer normalization method is used to normalize the input data. An improved reward function is designed, which improves the reward value of the algorithm, optimizes the state space, and alleviates the problem of sparse reward to some extent. The data is stored in a priority experience replay pool, and the network is trained using small batches of data. In addition, we evaluate our method by experiment related to indoor navigation. The experiments show that the robot trained by the improved D3QN algorithm can adapt to the unknown environment more quickly than the basic D3QN algorithm. The network's convergence speed is also improved, and it can complete the obstacle avoidance navigation task more efficiently.
基于深度强化学习的机器人避障导航方法
针对移动机器人室内环境的导航问题,传统的基于D3QN的导航算法存在奖励稀疏、神经网络训练速度慢等问题。本文提出了一种基于D3QN网络的深度强化学习算法(LN-D3QN)来实现移动机器人的无碰撞自主导航。为了提高移动机器人学习和探索的效率,采用视觉传感器从环境中获取输入数据,并采用层归一化方法对输入数据进行归一化处理。设计了改进的奖励函数,提高了算法的奖励值,优化了状态空间,在一定程度上缓解了奖励稀疏问题。数据存储在优先级体验重放池中,网络使用小批量数据进行训练。此外,我们还通过室内导航相关的实验对我们的方法进行了验证。实验表明,改进D3QN算法训练的机器人比基本D3QN算法训练的机器人能更快地适应未知环境。网络的收敛速度也得到了提高,能够更高效地完成避障导航任务。
本文章由计算机程序翻译,如有差异,请以英文原文为准。
求助全文
约1分钟内获得全文 求助全文
来源期刊
自引率
0.00%
发文量
0
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
确定
请完成安全验证×
copy
已复制链接
快去分享给好友吧!
我知道了
右上角分享
点击右上角分享
0
联系我们:info@booksci.cn Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。 Copyright © 2023 布克学术 All rights reserved.
京ICP备2023020795号-1
ghs 京公网安备 11010802042870号
Book学术文献互助
Book学术文献互助群
群 号:604180095
Book学术官方微信