{"title":"Learning assistance by demonstration","authors":"Harold Soh, Y. Demiris","doi":"10.5898/JHRI.4.3.Soh","DOIUrl":null,"url":null,"abstract":"In this paper, we present a framework, probabilistic model, and algorithm for learning shared control policies by observing an assistant. This is a methodology we refer to as Learning Assistance by Demonstration (LAD). As a subset of robot Learning by Demonstration (LbD), LAD focuses on the assistive element by explicitly capturing how and when to help. The latter is especially important in assistive scenarios---such as rehabilitation and training---where there exists multiple and possibly conflicting goals. We formalize these notions in a probabilistic model and develop an efficient online mixture of experts (OME) algorithm, based on sparse Gaussian processes (GPs), for learning the assistive policy. Focusing on smart mobility, we couple the LAD methodology with a novel paired-haptic-controllers setup for helping smart wheelchair users navigate their environment. Experimental results with 15 able-bodied participants demonstrate that our learned shared control policy improved driving performance (as measured in lap seconds) by 43 s (a speedup of 191%). Furthermore, survey results indicate that the participants not only performed better quantitatively, but also qualitatively felt the model assistance helped them complete the task.","PeriodicalId":92076,"journal":{"name":"Journal of human-robot interaction","volume":null,"pages":null},"PeriodicalIF":0.0000,"publicationDate":"2015-12-06","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://sci-hub-pdf.com/10.5898/JHRI.4.3.Soh","citationCount":"18","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"Journal of human-robot interaction","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/10.5898/JHRI.4.3.Soh","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
引用次数: 18
Abstract
In this paper, we present a framework, probabilistic model, and algorithm for learning shared control policies by observing an assistant. This is a methodology we refer to as Learning Assistance by Demonstration (LAD). As a subset of robot Learning by Demonstration (LbD), LAD focuses on the assistive element by explicitly capturing how and when to help. The latter is especially important in assistive scenarios---such as rehabilitation and training---where there exists multiple and possibly conflicting goals. We formalize these notions in a probabilistic model and develop an efficient online mixture of experts (OME) algorithm, based on sparse Gaussian processes (GPs), for learning the assistive policy. Focusing on smart mobility, we couple the LAD methodology with a novel paired-haptic-controllers setup for helping smart wheelchair users navigate their environment. Experimental results with 15 able-bodied participants demonstrate that our learned shared control policy improved driving performance (as measured in lap seconds) by 43 s (a speedup of 191%). Furthermore, survey results indicate that the participants not only performed better quantitatively, but also qualitatively felt the model assistance helped them complete the task.