Abdullah Ahmed, Yasser F. O. Mohammad, V. Parque, Haitham El-Hussieny, S. Ahmed
{"title":"End-to-End Mobile Robot Navigation using a Residual Deep Reinforcement Learning in Dynamic Human Environments","authors":"Abdullah Ahmed, Yasser F. O. Mohammad, V. Parque, Haitham El-Hussieny, S. Ahmed","doi":"10.1109/MESA55290.2022.10004394","DOIUrl":null,"url":null,"abstract":"Safe navigation through human crowds is key to enabling practical mobility ubiquitously. The Deep Reinforcement Learning (DRL) and the End-to-End (E2E) approaches to goal-oriented robot navigation have the potential to render policies able to tackle localization, path planning, obstacle avoidance, and adaptation to change in unison. In this paper, we report an architecture based on convolutional units and residual blocks being able to enhance adaptability to unseen and dynamic human environments. In particular, our scheme outperformed the state-of-the-art baselines SOADRL and NAVREP by about 13% and 18% on average success rate, respectively, throughout 27 unseen and dynamic navigation instances. Furthermore, our approach avoids the explicit encoding of positions and trajectories of moving humans compared to the standard models. Our results show the potential to render adaptive and generalizable policies for unknown and dynamic human environments.","PeriodicalId":410029,"journal":{"name":"2022 18th IEEE/ASME International Conference on Mechatronic and Embedded Systems and Applications (MESA)","volume":"114 1","pages":"0"},"PeriodicalIF":0.0000,"publicationDate":"2022-11-28","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"1","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"2022 18th IEEE/ASME International Conference on Mechatronic and Embedded Systems and Applications (MESA)","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/10.1109/MESA55290.2022.10004394","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
引用次数: 1
Abstract
Safe navigation through human crowds is key to enabling practical mobility ubiquitously. The Deep Reinforcement Learning (DRL) and the End-to-End (E2E) approaches to goal-oriented robot navigation have the potential to render policies able to tackle localization, path planning, obstacle avoidance, and adaptation to change in unison. In this paper, we report an architecture based on convolutional units and residual blocks being able to enhance adaptability to unseen and dynamic human environments. In particular, our scheme outperformed the state-of-the-art baselines SOADRL and NAVREP by about 13% and 18% on average success rate, respectively, throughout 27 unseen and dynamic navigation instances. Furthermore, our approach avoids the explicit encoding of positions and trajectories of moving humans compared to the standard models. Our results show the potential to render adaptive and generalizable policies for unknown and dynamic human environments.