Combining brain-computer interfaces with deep reinforcement learning for robot training: a feasibility study in a simulation environment

M. Vukelić, Michael Bui, Anna Vorreuther, Katharina Lingelbach
{"title":"Combining brain-computer interfaces with deep reinforcement learning for robot training: a feasibility study in a simulation environment","authors":"M. Vukelić, Michael Bui, Anna Vorreuther, Katharina Lingelbach","doi":"10.3389/fnrgo.2023.1274730","DOIUrl":null,"url":null,"abstract":"Deep reinforcement learning (RL) is used as a strategy to teach robot agents how to autonomously learn complex tasks. While sparsity is a natural way to define a reward in realistic robot scenarios, it provides poor learning signals for the agent, thus making the design of good reward functions challenging. To overcome this challenge learning from human feedback through an implicit brain-computer interface (BCI) is used. We combined a BCI with deep RL for robot training in a 3-D physical realistic simulation environment. In a first study, we compared the feasibility of different electroencephalography (EEG) systems (wet- vs. dry-based electrodes) and its application for automatic classification of perceived errors during a robot task with different machine learning models. In a second study, we compared the performance of the BCI-based deep RL training to feedback explicitly given by participants. Our findings from the first study indicate the use of a high-quality dry-based EEG-system can provide a robust and fast method for automatically assessing robot behavior using a sophisticated convolutional neural network machine learning model. The results of our second study prove that the implicit BCI-based deep RL version in combination with the dry EEG-system can significantly accelerate the learning process in a realistic 3-D robot simulation environment. Performance of the BCI-based trained deep RL model was even comparable to that achieved by the approach with explicit human feedback. Our findings emphasize the usage of BCI-based deep RL methods as a valid alternative in those human-robot applications where no access to cognitive demanding explicit human feedback is available.","PeriodicalId":507972,"journal":{"name":"Frontiers in Neuroergonomics","volume":"44 2","pages":""},"PeriodicalIF":0.0000,"publicationDate":"2023-11-23","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"Frontiers in Neuroergonomics","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/10.3389/fnrgo.2023.1274730","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
引用次数: 0

Abstract

Deep reinforcement learning (RL) is used as a strategy to teach robot agents how to autonomously learn complex tasks. While sparsity is a natural way to define a reward in realistic robot scenarios, it provides poor learning signals for the agent, thus making the design of good reward functions challenging. To overcome this challenge learning from human feedback through an implicit brain-computer interface (BCI) is used. We combined a BCI with deep RL for robot training in a 3-D physical realistic simulation environment. In a first study, we compared the feasibility of different electroencephalography (EEG) systems (wet- vs. dry-based electrodes) and its application for automatic classification of perceived errors during a robot task with different machine learning models. In a second study, we compared the performance of the BCI-based deep RL training to feedback explicitly given by participants. Our findings from the first study indicate the use of a high-quality dry-based EEG-system can provide a robust and fast method for automatically assessing robot behavior using a sophisticated convolutional neural network machine learning model. The results of our second study prove that the implicit BCI-based deep RL version in combination with the dry EEG-system can significantly accelerate the learning process in a realistic 3-D robot simulation environment. Performance of the BCI-based trained deep RL model was even comparable to that achieved by the approach with explicit human feedback. Our findings emphasize the usage of BCI-based deep RL methods as a valid alternative in those human-robot applications where no access to cognitive demanding explicit human feedback is available.
将脑机接口与深度强化学习相结合用于机器人训练:模拟环境中的可行性研究
深度强化学习(RL)作为一种策略,被用于教授机器人代理如何自主学习复杂任务。虽然稀疏性是在现实机器人场景中定义奖励的一种自然方式,但它为代理提供的学习信号很差,因此设计良好的奖励函数具有挑战性。为了克服这一挑战,我们采用了通过隐式脑机接口(BCI)从人类反馈中学习的方法。我们将 BCI 与深度 RL 结合起来,在三维物理仿真环境中进行机器人训练。在第一项研究中,我们比较了不同脑电图(EEG)系统(湿式电极与干式电极)的可行性,以及不同机器学习模型在机器人任务中用于感知错误自动分类的应用。在第二项研究中,我们将基于 BCI 的深度 RL 训练的性能与参与者明确给出的反馈进行了比较。第一项研究的结果表明,使用高质量的干式脑电图系统可以为使用复杂的卷积神经网络机器学习模型自动评估机器人行为提供一种稳健而快速的方法。我们的第二项研究结果证明,基于隐式 BCI 的深度 RL 版本与干式脑电图系统相结合,可以显著加快现实三维机器人模拟环境中的学习过程。基于 BCI 训练的深度 RL 模型的性能甚至可与显式人类反馈方法所达到的性能相媲美。我们的研究结果表明,在那些无法获得认知要求较高的明确人类反馈的人机应用中,基于 BCI 的深度 RL 方法是一种有效的替代方法。
本文章由计算机程序翻译,如有差异,请以英文原文为准。
求助全文
约1分钟内获得全文 求助全文
来源期刊
自引率
0.00%
发文量
0
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
确定
请完成安全验证×
copy
已复制链接
快去分享给好友吧!
我知道了
右上角分享
点击右上角分享
0
联系我们:info@booksci.cn Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。 Copyright © 2023 布克学术 All rights reserved.
京ICP备2023020795号-1
ghs 京公网安备 11010802042870号
Book学术文献互助
Book学术文献互助群
群 号:481959085
Book学术官方微信