Zhihao Zhou , Li Zhang , Qile Liu , Gan Huang , Zhuliang Yu , Zhen Liang
{"title":"情绪主体:基于分布原型奖励的无监督深度强化学习用于连续情绪脑电图分析","authors":"Zhihao Zhou , Li Zhang , Qile Liu , Gan Huang , Zhuliang Yu , Zhen Liang","doi":"10.1016/j.neucom.2025.130951","DOIUrl":null,"url":null,"abstract":"<div><div>Continuous Electroencephalography (EEG) signals are widely employed in affective brain-computer interface (aBCI) applications. However, only a subset of the continuously acquired EEG data is truly relevant to emotional processing, while the remainder is often noisy or unrelated. Manual annotation of these key emotional segments is impractical due to their dynamic and individualized nature. To address this challenge, we propose a novel unsupervised deep reinforcement learning framework, termed <em>Emotion Agent</em>, which automatically identifies and extracts the most informative emotional segments from continuous EEG signals. Emotion Agent initially utilizes a heuristic algorithm to perform a global search and generate prototype representations of the EEG signals. These prototypes guide the exploration of the signal space and highlight regions of interest. Furthermore, we design a distribution-prototype-based reward function that evaluates the interaction between samples and prototypes to ensure that the selected segments are both representative and relevant to the underlying emotional states. Finally, the framework is trained using Proximal Policy Optimization (PPO) to achieve stable and efficient convergence. Experimental results on three widely used datasets (covering both discrete and dimensional emotion recognition) show an average improvement of 13.46 % when using the proposed Emotion Agent, demonstrating its significant enhancement of accuracy and robustness in downstream aBCI tasks.</div></div>","PeriodicalId":19268,"journal":{"name":"Neurocomputing","volume":"652 ","pages":"Article 130951"},"PeriodicalIF":5.5000,"publicationDate":"2025-07-17","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":"{\"title\":\"Emotion agent: Unsupervised deep reinforcement learning with distribution-prototype reward for continuous emotional EEG analysis\",\"authors\":\"Zhihao Zhou , Li Zhang , Qile Liu , Gan Huang , Zhuliang Yu , Zhen Liang\",\"doi\":\"10.1016/j.neucom.2025.130951\",\"DOIUrl\":null,\"url\":null,\"abstract\":\"<div><div>Continuous Electroencephalography (EEG) signals are widely employed in affective brain-computer interface (aBCI) applications. However, only a subset of the continuously acquired EEG data is truly relevant to emotional processing, while the remainder is often noisy or unrelated. Manual annotation of these key emotional segments is impractical due to their dynamic and individualized nature. To address this challenge, we propose a novel unsupervised deep reinforcement learning framework, termed <em>Emotion Agent</em>, which automatically identifies and extracts the most informative emotional segments from continuous EEG signals. Emotion Agent initially utilizes a heuristic algorithm to perform a global search and generate prototype representations of the EEG signals. These prototypes guide the exploration of the signal space and highlight regions of interest. Furthermore, we design a distribution-prototype-based reward function that evaluates the interaction between samples and prototypes to ensure that the selected segments are both representative and relevant to the underlying emotional states. Finally, the framework is trained using Proximal Policy Optimization (PPO) to achieve stable and efficient convergence. Experimental results on three widely used datasets (covering both discrete and dimensional emotion recognition) show an average improvement of 13.46 % when using the proposed Emotion Agent, demonstrating its significant enhancement of accuracy and robustness in downstream aBCI tasks.</div></div>\",\"PeriodicalId\":19268,\"journal\":{\"name\":\"Neurocomputing\",\"volume\":\"652 \",\"pages\":\"Article 130951\"},\"PeriodicalIF\":5.5000,\"publicationDate\":\"2025-07-17\",\"publicationTypes\":\"Journal Article\",\"fieldsOfStudy\":null,\"isOpenAccess\":false,\"openAccessPdf\":\"\",\"citationCount\":\"0\",\"resultStr\":null,\"platform\":\"Semanticscholar\",\"paperid\":null,\"PeriodicalName\":\"Neurocomputing\",\"FirstCategoryId\":\"94\",\"ListUrlMain\":\"https://www.sciencedirect.com/science/article/pii/S0925231225016236\",\"RegionNum\":2,\"RegionCategory\":\"计算机科学\",\"ArticlePicture\":[],\"TitleCN\":null,\"AbstractTextCN\":null,\"PMCID\":null,\"EPubDate\":\"\",\"PubModel\":\"\",\"JCR\":\"Q1\",\"JCRName\":\"COMPUTER SCIENCE, ARTIFICIAL INTELLIGENCE\",\"Score\":null,\"Total\":0}","platform":"Semanticscholar","paperid":null,"PeriodicalName":"Neurocomputing","FirstCategoryId":"94","ListUrlMain":"https://www.sciencedirect.com/science/article/pii/S0925231225016236","RegionNum":2,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"Q1","JCRName":"COMPUTER SCIENCE, ARTIFICIAL INTELLIGENCE","Score":null,"Total":0}
Emotion agent: Unsupervised deep reinforcement learning with distribution-prototype reward for continuous emotional EEG analysis
Continuous Electroencephalography (EEG) signals are widely employed in affective brain-computer interface (aBCI) applications. However, only a subset of the continuously acquired EEG data is truly relevant to emotional processing, while the remainder is often noisy or unrelated. Manual annotation of these key emotional segments is impractical due to their dynamic and individualized nature. To address this challenge, we propose a novel unsupervised deep reinforcement learning framework, termed Emotion Agent, which automatically identifies and extracts the most informative emotional segments from continuous EEG signals. Emotion Agent initially utilizes a heuristic algorithm to perform a global search and generate prototype representations of the EEG signals. These prototypes guide the exploration of the signal space and highlight regions of interest. Furthermore, we design a distribution-prototype-based reward function that evaluates the interaction between samples and prototypes to ensure that the selected segments are both representative and relevant to the underlying emotional states. Finally, the framework is trained using Proximal Policy Optimization (PPO) to achieve stable and efficient convergence. Experimental results on three widely used datasets (covering both discrete and dimensional emotion recognition) show an average improvement of 13.46 % when using the proposed Emotion Agent, demonstrating its significant enhancement of accuracy and robustness in downstream aBCI tasks.
期刊介绍:
Neurocomputing publishes articles describing recent fundamental contributions in the field of neurocomputing. Neurocomputing theory, practice and applications are the essential topics being covered.