Mobile robotics planning using abstract Markov decision processes

Pierre Laroche, F. Charpillet, R. Schott
{"title":"Mobile robotics planning using abstract Markov decision processes","authors":"Pierre Laroche, F. Charpillet, R. Schott","doi":"10.1109/TAI.1999.809804","DOIUrl":null,"url":null,"abstract":"Markov decision processes have been successfully used in robotics for indoor robot navigation problems. They allow the computation of optimal sequences of actions in order to achieve a given goal, accounting for actuator uncertainties. However, MDPs are unsatisfactory at avoiding unknown obstacles. On the other hand, reactive navigators are particularly adapted to that, and don't need any prior knowledge about the environment, but they are unable to plan the set of actions that will permit the realization of a given mission. We present a new state aggregation technique for Markov decision processes, such that part of the work usually dedicated to the planner is achieved by a reactive navigator. Thus some characteristics of our environments, such as the width of corridors, have not been considered, which allows to cluster states together, significantly reducing the state space. As a consequence, policies are computed faster and are shown to be at least as efficient as optimal ones.","PeriodicalId":194023,"journal":{"name":"Proceedings 11th International Conference on Tools with Artificial Intelligence","volume":"12 1","pages":"0"},"PeriodicalIF":0.0000,"publicationDate":"1999-11-08","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"13","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"Proceedings 11th International Conference on Tools with Artificial Intelligence","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/10.1109/TAI.1999.809804","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
引用次数: 13

Abstract

Markov decision processes have been successfully used in robotics for indoor robot navigation problems. They allow the computation of optimal sequences of actions in order to achieve a given goal, accounting for actuator uncertainties. However, MDPs are unsatisfactory at avoiding unknown obstacles. On the other hand, reactive navigators are particularly adapted to that, and don't need any prior knowledge about the environment, but they are unable to plan the set of actions that will permit the realization of a given mission. We present a new state aggregation technique for Markov decision processes, such that part of the work usually dedicated to the planner is achieved by a reactive navigator. Thus some characteristics of our environments, such as the width of corridors, have not been considered, which allows to cluster states together, significantly reducing the state space. As a consequence, policies are computed faster and are shown to be at least as efficient as optimal ones.
基于抽象马尔可夫决策过程的移动机器人规划
马尔可夫决策过程已成功地应用于机器人室内机器人导航问题。它们允许计算最佳的动作序列,以实现给定的目标,考虑执行器的不确定性。然而,民主党在避免未知障碍方面并不令人满意。另一方面,反应式导航器特别适应这种情况,不需要任何关于环境的先验知识,但它们无法计划一系列行动,从而允许实现给定的任务。我们提出了一种新的马尔可夫决策过程的状态聚合技术,使得通常由计划器完成的部分工作由响应式导航器完成。因此,我们的环境的一些特征,如走廊的宽度,没有被考虑在内,这允许将状态聚集在一起,大大减少了状态空间。因此,策略的计算速度更快,并且至少与最优策略一样有效。
本文章由计算机程序翻译,如有差异,请以英文原文为准。
求助全文
约1分钟内获得全文 求助全文
来源期刊
自引率
0.00%
发文量
0
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
确定
请完成安全验证×
copy
已复制链接
快去分享给好友吧!
我知道了
右上角分享
点击右上角分享
0
联系我们:info@booksci.cn Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。 Copyright © 2023 布克学术 All rights reserved.
京ICP备2023020795号-1
ghs 京公网安备 11010802042870号
Book学术文献互助
Book学术文献互助群
群 号:604180095
Book学术官方微信