Xiang Shen, Xiang Zhang, Yifan Huang, Shuhang Chen, Yiwen Wang
{"title":"在脑机接口强化学习框架中建模mPFC活动","authors":"Xiang Shen, Xiang Zhang, Yifan Huang, Shuhang Chen, Yiwen Wang","doi":"10.1109/NER.2019.8717162","DOIUrl":null,"url":null,"abstract":"Reinforcement learning (RL) algorithm interprets the movement intentions in Brain-machine interfaces (BMIs) with a reward signal. This reward can be an external reward (food or water) or an internal representation which links the correct movement with the external reward. Medial prefrontal cortex (mPFC) has been demonstrated to be closely related to the reward-guided learning. In this paper, we propose to model mPFC activities as an internal representation of the reward associated with different actions in a RL framework. Support vector machine (SVM) is adopted to analyze mPFC activities to distinguish the rewarded and unrewarded trials based on mPFC signals considering corresponding actions. Then the discrimination result will be utilized to train a RL decoder. Here we introduce the attention-gated reinforcement learning (AGREL) as the decoder to generate a mapping between motor cortex(M1) and action states. To evaluate our approach, we test on in vivo neural physiological data collected from rats when performing a two-lever discrimination task. The RL decoder using the internal action-reward evaluation achieves a prediction accuracy of 94.8%, which is very close to the one using the external reward. This indicates the potentials of modelling mPFC activities as an internal representation to associate the correct action with the reward.","PeriodicalId":356177,"journal":{"name":"2019 9th International IEEE/EMBS Conference on Neural Engineering (NER)","volume":"118 1","pages":"0"},"PeriodicalIF":0.0000,"publicationDate":"2019-03-20","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"4","resultStr":"{\"title\":\"Modelling mPFC Activities in Reinforcement Learning Framework for Brain-Machine Interfaces\",\"authors\":\"Xiang Shen, Xiang Zhang, Yifan Huang, Shuhang Chen, Yiwen Wang\",\"doi\":\"10.1109/NER.2019.8717162\",\"DOIUrl\":null,\"url\":null,\"abstract\":\"Reinforcement learning (RL) algorithm interprets the movement intentions in Brain-machine interfaces (BMIs) with a reward signal. This reward can be an external reward (food or water) or an internal representation which links the correct movement with the external reward. Medial prefrontal cortex (mPFC) has been demonstrated to be closely related to the reward-guided learning. In this paper, we propose to model mPFC activities as an internal representation of the reward associated with different actions in a RL framework. Support vector machine (SVM) is adopted to analyze mPFC activities to distinguish the rewarded and unrewarded trials based on mPFC signals considering corresponding actions. Then the discrimination result will be utilized to train a RL decoder. Here we introduce the attention-gated reinforcement learning (AGREL) as the decoder to generate a mapping between motor cortex(M1) and action states. To evaluate our approach, we test on in vivo neural physiological data collected from rats when performing a two-lever discrimination task. The RL decoder using the internal action-reward evaluation achieves a prediction accuracy of 94.8%, which is very close to the one using the external reward. This indicates the potentials of modelling mPFC activities as an internal representation to associate the correct action with the reward.\",\"PeriodicalId\":356177,\"journal\":{\"name\":\"2019 9th International IEEE/EMBS Conference on Neural Engineering (NER)\",\"volume\":\"118 1\",\"pages\":\"0\"},\"PeriodicalIF\":0.0000,\"publicationDate\":\"2019-03-20\",\"publicationTypes\":\"Journal Article\",\"fieldsOfStudy\":null,\"isOpenAccess\":false,\"openAccessPdf\":\"\",\"citationCount\":\"4\",\"resultStr\":null,\"platform\":\"Semanticscholar\",\"paperid\":null,\"PeriodicalName\":\"2019 9th International IEEE/EMBS Conference on Neural Engineering (NER)\",\"FirstCategoryId\":\"1085\",\"ListUrlMain\":\"https://doi.org/10.1109/NER.2019.8717162\",\"RegionNum\":0,\"RegionCategory\":null,\"ArticlePicture\":[],\"TitleCN\":null,\"AbstractTextCN\":null,\"PMCID\":null,\"EPubDate\":\"\",\"PubModel\":\"\",\"JCR\":\"\",\"JCRName\":\"\",\"Score\":null,\"Total\":0}","platform":"Semanticscholar","paperid":null,"PeriodicalName":"2019 9th International IEEE/EMBS Conference on Neural Engineering (NER)","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/10.1109/NER.2019.8717162","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
Modelling mPFC Activities in Reinforcement Learning Framework for Brain-Machine Interfaces
Reinforcement learning (RL) algorithm interprets the movement intentions in Brain-machine interfaces (BMIs) with a reward signal. This reward can be an external reward (food or water) or an internal representation which links the correct movement with the external reward. Medial prefrontal cortex (mPFC) has been demonstrated to be closely related to the reward-guided learning. In this paper, we propose to model mPFC activities as an internal representation of the reward associated with different actions in a RL framework. Support vector machine (SVM) is adopted to analyze mPFC activities to distinguish the rewarded and unrewarded trials based on mPFC signals considering corresponding actions. Then the discrimination result will be utilized to train a RL decoder. Here we introduce the attention-gated reinforcement learning (AGREL) as the decoder to generate a mapping between motor cortex(M1) and action states. To evaluate our approach, we test on in vivo neural physiological data collected from rats when performing a two-lever discrimination task. The RL decoder using the internal action-reward evaluation achieves a prediction accuracy of 94.8%, which is very close to the one using the external reward. This indicates the potentials of modelling mPFC activities as an internal representation to associate the correct action with the reward.