Emotion agent: Unsupervised deep reinforcement learning with distribution-prototype reward for continuous emotional EEG analysis

IF 5.5 2区 计算机科学 Q1 COMPUTER SCIENCE, ARTIFICIAL INTELLIGENCE
Zhihao Zhou , Li Zhang , Qile Liu , Gan Huang , Zhuliang Yu , Zhen Liang
{"title":"Emotion agent: Unsupervised deep reinforcement learning with distribution-prototype reward for continuous emotional EEG analysis","authors":"Zhihao Zhou ,&nbsp;Li Zhang ,&nbsp;Qile Liu ,&nbsp;Gan Huang ,&nbsp;Zhuliang Yu ,&nbsp;Zhen Liang","doi":"10.1016/j.neucom.2025.130951","DOIUrl":null,"url":null,"abstract":"<div><div>Continuous Electroencephalography (EEG) signals are widely employed in affective brain-computer interface (aBCI) applications. However, only a subset of the continuously acquired EEG data is truly relevant to emotional processing, while the remainder is often noisy or unrelated. Manual annotation of these key emotional segments is impractical due to their dynamic and individualized nature. To address this challenge, we propose a novel unsupervised deep reinforcement learning framework, termed <em>Emotion Agent</em>, which automatically identifies and extracts the most informative emotional segments from continuous EEG signals. Emotion Agent initially utilizes a heuristic algorithm to perform a global search and generate prototype representations of the EEG signals. These prototypes guide the exploration of the signal space and highlight regions of interest. Furthermore, we design a distribution-prototype-based reward function that evaluates the interaction between samples and prototypes to ensure that the selected segments are both representative and relevant to the underlying emotional states. Finally, the framework is trained using Proximal Policy Optimization (PPO) to achieve stable and efficient convergence. Experimental results on three widely used datasets (covering both discrete and dimensional emotion recognition) show an average improvement of 13.46 % when using the proposed Emotion Agent, demonstrating its significant enhancement of accuracy and robustness in downstream aBCI tasks.</div></div>","PeriodicalId":19268,"journal":{"name":"Neurocomputing","volume":"652 ","pages":"Article 130951"},"PeriodicalIF":5.5000,"publicationDate":"2025-07-17","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"Neurocomputing","FirstCategoryId":"94","ListUrlMain":"https://www.sciencedirect.com/science/article/pii/S0925231225016236","RegionNum":2,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"Q1","JCRName":"COMPUTER SCIENCE, ARTIFICIAL INTELLIGENCE","Score":null,"Total":0}
引用次数: 0

Abstract

Continuous Electroencephalography (EEG) signals are widely employed in affective brain-computer interface (aBCI) applications. However, only a subset of the continuously acquired EEG data is truly relevant to emotional processing, while the remainder is often noisy or unrelated. Manual annotation of these key emotional segments is impractical due to their dynamic and individualized nature. To address this challenge, we propose a novel unsupervised deep reinforcement learning framework, termed Emotion Agent, which automatically identifies and extracts the most informative emotional segments from continuous EEG signals. Emotion Agent initially utilizes a heuristic algorithm to perform a global search and generate prototype representations of the EEG signals. These prototypes guide the exploration of the signal space and highlight regions of interest. Furthermore, we design a distribution-prototype-based reward function that evaluates the interaction between samples and prototypes to ensure that the selected segments are both representative and relevant to the underlying emotional states. Finally, the framework is trained using Proximal Policy Optimization (PPO) to achieve stable and efficient convergence. Experimental results on three widely used datasets (covering both discrete and dimensional emotion recognition) show an average improvement of 13.46 % when using the proposed Emotion Agent, demonstrating its significant enhancement of accuracy and robustness in downstream aBCI tasks.
情绪主体:基于分布原型奖励的无监督深度强化学习用于连续情绪脑电图分析
连续脑电图(EEG)信号在情感脑机接口(aBCI)中得到了广泛的应用。然而,连续获取的脑电图数据中只有一小部分与情绪处理真正相关,而其余的通常是嘈杂的或不相关的。由于这些关键情感片段的动态性和个性化,手工注释是不切实际的。为了解决这一挑战,我们提出了一种新的无监督深度强化学习框架,称为情感代理,它可以自动识别和提取连续脑电图信号中信息量最大的情感片段。情感代理最初利用启发式算法进行全局搜索并生成脑电信号的原型表示。这些原型指导了对信号空间的探索,并突出了感兴趣的区域。此外,我们设计了一个基于分布原型的奖励函数来评估样本和原型之间的相互作用,以确保所选择的部分既具有代表性,又与潜在的情绪状态相关。最后,利用近端策略优化(PPO)对框架进行训练,以实现稳定高效的收敛。在三个广泛使用的数据集(包括离散和维度情绪识别)上的实验结果表明,使用所提出的情绪代理平均提高13.46 %,表明其在下游aBCI任务中的准确性和鲁棒性显著提高。
本文章由计算机程序翻译,如有差异,请以英文原文为准。
求助全文
约1分钟内获得全文 求助全文
来源期刊
Neurocomputing
Neurocomputing 工程技术-计算机:人工智能
CiteScore
13.10
自引率
10.00%
发文量
1382
审稿时长
70 days
期刊介绍: Neurocomputing publishes articles describing recent fundamental contributions in the field of neurocomputing. Neurocomputing theory, practice and applications are the essential topics being covered.
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
确定
请完成安全验证×
copy
已复制链接
快去分享给好友吧!
我知道了
右上角分享
点击右上角分享
0
联系我们:info@booksci.cn Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。 Copyright © 2023 布克学术 All rights reserved.
京ICP备2023020795号-1
ghs 京公网安备 11010802042870号
Book学术文献互助
Book学术文献互助群
群 号:604180095
Book学术官方微信