Modeling humans as observation providers using POMDPs

Stephanie Rosenthal, M. Veloso
{"title":"Modeling humans as observation providers using POMDPs","authors":"Stephanie Rosenthal, M. Veloso","doi":"10.1109/ROMAN.2011.6005272","DOIUrl":null,"url":null,"abstract":"The ability to obtain accurate observations while navigating in uncertain environments is a difficult challenge in deploying robots. Robots have relied heavily on human supervisors who are always available to provide additional observations to reduce uncertainty. We are instead interested in taking advantage of humans who are already in the environment to receive observations. The challenge is in modeling these humans' availability and higher costs of interruption to determine when to query them during navigation. In this work, we introduce a Human Observation Provider POMDP framework (HOP-POMDP), and contribute new algorithms for planning and executing with HOP-POMDPs that account for the differences between humans and other probabilistic sensors that provide observations. We compare optimal HOP-POMDP policies that plan for needing humans' observations with oracle POMDP policies that do not take human costs and availability into account. We show in benchmark tests and real-world environments that the oracle policies match the optimal HOP-POMDP policy 60% of the time, and can be used in cases when humans are likely to be available on the shortest paths. However, the HOP-POMDP policies receive higher rewards in general as they take into account the possibility that a human may be unavailable. HOP-POMDP policies only need to be computed once prior to the deployment of the robot, so it is feasible to precompute and use in practice.","PeriodicalId":408015,"journal":{"name":"2011 RO-MAN","volume":"139 1","pages":"0"},"PeriodicalIF":0.0000,"publicationDate":"2011-08-30","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"34","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"2011 RO-MAN","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/10.1109/ROMAN.2011.6005272","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
引用次数: 34

Abstract

The ability to obtain accurate observations while navigating in uncertain environments is a difficult challenge in deploying robots. Robots have relied heavily on human supervisors who are always available to provide additional observations to reduce uncertainty. We are instead interested in taking advantage of humans who are already in the environment to receive observations. The challenge is in modeling these humans' availability and higher costs of interruption to determine when to query them during navigation. In this work, we introduce a Human Observation Provider POMDP framework (HOP-POMDP), and contribute new algorithms for planning and executing with HOP-POMDPs that account for the differences between humans and other probabilistic sensors that provide observations. We compare optimal HOP-POMDP policies that plan for needing humans' observations with oracle POMDP policies that do not take human costs and availability into account. We show in benchmark tests and real-world environments that the oracle policies match the optimal HOP-POMDP policy 60% of the time, and can be used in cases when humans are likely to be available on the shortest paths. However, the HOP-POMDP policies receive higher rewards in general as they take into account the possibility that a human may be unavailable. HOP-POMDP policies only need to be computed once prior to the deployment of the robot, so it is feasible to precompute and use in practice.
使用pomdp将人类建模为观察提供者
在不确定环境中导航时获得准确观测的能力是部署机器人的一个困难挑战。机器人在很大程度上依赖于人类监督者,他们总是可以提供额外的观察,以减少不确定性。相反,我们感兴趣的是利用已经在环境中接受观察的人类。挑战在于如何对这些人的可用性和更高的中断成本进行建模,以确定在导航过程中何时查询他们。在这项工作中,我们引入了一个人类观测提供者POMDP框架(HOP-POMDP),并提供了用于规划和执行HOP-POMDP的新算法,该算法解释了人类和其他提供观测的概率传感器之间的差异。我们将最优的HOP-POMDP策略与不考虑人力成本和可用性的oracle POMDP策略进行了比较。我们在基准测试和实际环境中显示,oracle策略在60%的时间内匹配最优的HOP-POMDP策略,并且可以在人类可能在最短路径上可用的情况下使用。然而,HOP-POMDP策略通常会获得更高的回报,因为它们考虑了人类可能不可用的可能性。在机器人部署之前,HOP-POMDP策略只需要计算一次,因此可以预先计算并在实际中使用。
本文章由计算机程序翻译,如有差异,请以英文原文为准。
求助全文
约1分钟内获得全文 求助全文
来源期刊
自引率
0.00%
发文量
0
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
确定
请完成安全验证×
copy
已复制链接
快去分享给好友吧!
我知道了
右上角分享
点击右上角分享
0
联系我们:info@booksci.cn Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。 Copyright © 2023 布克学术 All rights reserved.
京ICP备2023020795号-1
ghs 京公网安备 11010802042870号
Book学术文献互助
Book学术文献互助群
群 号:604180095
Book学术官方微信