Shared autonomy between human electroencephalography and TD3 deep reinforcement learning: A multi-agent copilot approach

IF 4.1 3区 综合性期刊 Q1 MULTIDISCIPLINARY SCIENCES
Chun-Ren Phang, Akimasa Hirata
{"title":"Shared autonomy between human electroencephalography and TD3 deep reinforcement learning: A multi-agent copilot approach","authors":"Chun-Ren Phang,&nbsp;Akimasa Hirata","doi":"10.1111/nyas.15322","DOIUrl":null,"url":null,"abstract":"<p>Deep reinforcement learning (RL) algorithms enable the development of fully autonomous agents that can interact with the environment. Brain–computer interface (BCI) systems decipher human implicit brain signals regardless of the explicit environment. We proposed a novel integration technique between deep RL and BCI to improve beneficial human interventions in autonomous systems and the performance in decoding brain activities by considering environmental factors. Shared autonomy was allowed between the action command decoded from the electroencephalography (EEG) of the human agent and the action generated from the twin delayed DDPG (TD3) agent for a given complex environment. Our proposed copilot control scheme with a full blocker (Co-FB) significantly outperformed the individual EEG (EEG-NB) or TD3 control. The Co-FB model achieved a higher target-approaching score, lower failure rate, and lower human workload than the EEG-NB model. We also proposed a disparity <span></span><math>\n <semantics>\n <mi>d</mi>\n <annotation>$d$</annotation>\n </semantics></math>-index to evaluate the effect of contradicting agent decisions on the control accuracy and authority of the copilot model. We observed that shifting control authority to the TD3 agent improved performance when BCI decoding was not optimal. These findings indicate that the copilot system can effectively handle complex environments and that BCI performance can be improved by considering environmental factors.</p>","PeriodicalId":8250,"journal":{"name":"Annals of the New York Academy of Sciences","volume":"1546 1","pages":"157-172"},"PeriodicalIF":4.1000,"publicationDate":"2025-03-30","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"Annals of the New York Academy of Sciences","FirstCategoryId":"103","ListUrlMain":"https://onlinelibrary.wiley.com/doi/10.1111/nyas.15322","RegionNum":3,"RegionCategory":"综合性期刊","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"Q1","JCRName":"MULTIDISCIPLINARY SCIENCES","Score":null,"Total":0}
引用次数: 0

Abstract

Deep reinforcement learning (RL) algorithms enable the development of fully autonomous agents that can interact with the environment. Brain–computer interface (BCI) systems decipher human implicit brain signals regardless of the explicit environment. We proposed a novel integration technique between deep RL and BCI to improve beneficial human interventions in autonomous systems and the performance in decoding brain activities by considering environmental factors. Shared autonomy was allowed between the action command decoded from the electroencephalography (EEG) of the human agent and the action generated from the twin delayed DDPG (TD3) agent for a given complex environment. Our proposed copilot control scheme with a full blocker (Co-FB) significantly outperformed the individual EEG (EEG-NB) or TD3 control. The Co-FB model achieved a higher target-approaching score, lower failure rate, and lower human workload than the EEG-NB model. We also proposed a disparity d $d$ -index to evaluate the effect of contradicting agent decisions on the control accuracy and authority of the copilot model. We observed that shifting control authority to the TD3 agent improved performance when BCI decoding was not optimal. These findings indicate that the copilot system can effectively handle complex environments and that BCI performance can be improved by considering environmental factors.

Abstract Image

Abstract Image

人类脑电图和TD3深度强化学习之间的共享自主性:一种多智能体副驾驶方法
深度强化学习(RL)算法使得能够与环境交互的完全自主代理的开发成为可能。脑机接口(BCI)系统可以在不受外显环境影响的情况下破译人类内隐脑信号。我们提出了一种新的深度RL和脑机接口之间的集成技术,以提高自主系统中有益的人类干预和在考虑环境因素的情况下解码大脑活动的性能。在给定的复杂环境中,允许从人类智能体的脑电图(EEG)解码的动作命令与双延迟DDPG (TD3)智能体生成的动作之间共享自治。我们提出的具有完全阻滞剂(Co-FB)的副驾驶控制方案显着优于单个EEG (EEG- nb)或TD3控制。与EEG-NB模型相比,Co-FB模型实现了更高的目标接近得分、更低的故障率和更低的人工工作量。我们还提出了一个视差d$d$-指标来评估相互矛盾的智能体决策对副驾驶模型控制精度和权威的影响。我们观察到,当BCI解码不是最优时,将控制权限转移到TD3代理可以提高性能。这些研究结果表明,副驾驶系统能够有效应对复杂环境,考虑环境因素可以提高BCI性能。
本文章由计算机程序翻译,如有差异,请以英文原文为准。
求助全文
约1分钟内获得全文 求助全文
来源期刊
Annals of the New York Academy of Sciences
Annals of the New York Academy of Sciences 综合性期刊-综合性期刊
CiteScore
11.00
自引率
1.90%
发文量
193
审稿时长
2-4 weeks
期刊介绍: Published on behalf of the New York Academy of Sciences, Annals of the New York Academy of Sciences provides multidisciplinary perspectives on research of current scientific interest with far-reaching implications for the wider scientific community and society at large. Each special issue assembles the best thinking of key contributors to a field of investigation at a time when emerging developments offer the promise of new insight. Individually themed, Annals special issues stimulate new ways to think about science by providing a neutral forum for discourse—within and across many institutions and fields.
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
确定
请完成安全验证×
copy
已复制链接
快去分享给好友吧!
我知道了
右上角分享
点击右上角分享
0
联系我们:info@booksci.cn Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。 Copyright © 2023 布克学术 All rights reserved.
京ICP备2023020795号-1
ghs 京公网安备 11010802042870号
Book学术文献互助
Book学术文献互助群
群 号:481959085
Book学术官方微信