基于随机反馈和生物约束的循环网络状态表示在线强化学习。

IF 6.4 1区 生物学 Q1 BIOLOGY
eLife Pub Date : 2025-09-24 DOI:10.7554/eLife.104101
Takayuki Tsurumi, Ayaka Kato, Arvind Kumar, Kenji Morita
{"title":"基于随机反馈和生物约束的循环网络状态表示在线强化学习。","authors":"Takayuki Tsurumi, Ayaka Kato, Arvind Kumar, Kenji Morita","doi":"10.7554/eLife.104101","DOIUrl":null,"url":null,"abstract":"<p><p>Representation of external and internal states in the brain plays a critical role in enabling suitable behavior. Recent studies suggest that state representation and state value can be simultaneously learned through Temporal-Difference-Reinforcement-Learning (TDRL) and Backpropagation-Through-Time (BPTT) in recurrent neural networks (RNNs) and their readout. However, neural implementation of such learning remains unclear as BPTT requires offline update using transported downstream weights, which is suggested to be biologically implausible. We demonstrate that simple online training of RNNs using TD reward prediction error and random feedback, without additional memory or eligibility trace, can still learn the structure of tasks with cue-reward delay and timing variability. This is because TD learning itself is a solution for temporal credit assignment, and feedback alignment, a mechanism originally proposed for supervised learning, enables gradient approximation without weight transport. Furthermore, we show that biologically constraining downstream weights and random feedback to be non-negative not only preserves learning but may even enhance it because the non-negative constraint ensures loose alignment-allowing the downstream and feedback weights to roughly align from the beginning. These results provide insights into the neural mechanisms underlying the learning of state representation and value, highlighting the potential of random feedback and biological constraints.</p>","PeriodicalId":11640,"journal":{"name":"eLife","volume":"14 ","pages":""},"PeriodicalIF":6.4000,"publicationDate":"2025-09-24","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.ncbi.nlm.nih.gov/pmc/articles/PMC12459954/pdf/","citationCount":"0","resultStr":"{\"title\":\"Online reinforcement learning of state representation in recurrent network supported by the power of random feedback and biological constraints.\",\"authors\":\"Takayuki Tsurumi, Ayaka Kato, Arvind Kumar, Kenji Morita\",\"doi\":\"10.7554/eLife.104101\",\"DOIUrl\":null,\"url\":null,\"abstract\":\"<p><p>Representation of external and internal states in the brain plays a critical role in enabling suitable behavior. Recent studies suggest that state representation and state value can be simultaneously learned through Temporal-Difference-Reinforcement-Learning (TDRL) and Backpropagation-Through-Time (BPTT) in recurrent neural networks (RNNs) and their readout. However, neural implementation of such learning remains unclear as BPTT requires offline update using transported downstream weights, which is suggested to be biologically implausible. We demonstrate that simple online training of RNNs using TD reward prediction error and random feedback, without additional memory or eligibility trace, can still learn the structure of tasks with cue-reward delay and timing variability. This is because TD learning itself is a solution for temporal credit assignment, and feedback alignment, a mechanism originally proposed for supervised learning, enables gradient approximation without weight transport. Furthermore, we show that biologically constraining downstream weights and random feedback to be non-negative not only preserves learning but may even enhance it because the non-negative constraint ensures loose alignment-allowing the downstream and feedback weights to roughly align from the beginning. These results provide insights into the neural mechanisms underlying the learning of state representation and value, highlighting the potential of random feedback and biological constraints.</p>\",\"PeriodicalId\":11640,\"journal\":{\"name\":\"eLife\",\"volume\":\"14 \",\"pages\":\"\"},\"PeriodicalIF\":6.4000,\"publicationDate\":\"2025-09-24\",\"publicationTypes\":\"Journal Article\",\"fieldsOfStudy\":null,\"isOpenAccess\":false,\"openAccessPdf\":\"https://www.ncbi.nlm.nih.gov/pmc/articles/PMC12459954/pdf/\",\"citationCount\":\"0\",\"resultStr\":null,\"platform\":\"Semanticscholar\",\"paperid\":null,\"PeriodicalName\":\"eLife\",\"FirstCategoryId\":\"99\",\"ListUrlMain\":\"https://doi.org/10.7554/eLife.104101\",\"RegionNum\":1,\"RegionCategory\":\"生物学\",\"ArticlePicture\":[],\"TitleCN\":null,\"AbstractTextCN\":null,\"PMCID\":null,\"EPubDate\":\"\",\"PubModel\":\"\",\"JCR\":\"Q1\",\"JCRName\":\"BIOLOGY\",\"Score\":null,\"Total\":0}","platform":"Semanticscholar","paperid":null,"PeriodicalName":"eLife","FirstCategoryId":"99","ListUrlMain":"https://doi.org/10.7554/eLife.104101","RegionNum":1,"RegionCategory":"生物学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"Q1","JCRName":"BIOLOGY","Score":null,"Total":0}
引用次数: 0

摘要

大脑中外部和内部状态的表征在适当的行为中起着至关重要的作用。近年来的研究表明,在递归神经网络(rnn)及其读出中,状态表示和状态值可以通过时间差分强化学习(TDRL)和时间反向传播(BPTT)同时学习。然而,这种学习的神经实现仍然不清楚,因为BPTT需要使用传输的下游权重进行离线更新,这在生物学上是不可能的。我们证明了使用TD奖励预测误差和随机反馈的简单rnn在线训练,没有额外的记忆或资格跟踪,仍然可以学习具有提示奖励延迟和时间可变性的任务结构。这是因为TD学习本身是一种时间信用分配的解决方案,而反馈对齐(一种最初为监督学习提出的机制)使梯度逼近不需要权重传输。此外,我们表明,生物约束下游权重和随机反馈非负不仅可以保持学习,甚至可以增强学习,因为非负约束确保松散对齐-允许下游和反馈权重从一开始就大致对齐。这些结果提供了对状态表示和值学习背后的神经机制的见解,突出了随机反馈和生物约束的潜力。
本文章由计算机程序翻译,如有差异,请以英文原文为准。

Online reinforcement learning of state representation in recurrent network supported by the power of random feedback and biological constraints.

Online reinforcement learning of state representation in recurrent network supported by the power of random feedback and biological constraints.

Online reinforcement learning of state representation in recurrent network supported by the power of random feedback and biological constraints.

Online reinforcement learning of state representation in recurrent network supported by the power of random feedback and biological constraints.

Representation of external and internal states in the brain plays a critical role in enabling suitable behavior. Recent studies suggest that state representation and state value can be simultaneously learned through Temporal-Difference-Reinforcement-Learning (TDRL) and Backpropagation-Through-Time (BPTT) in recurrent neural networks (RNNs) and their readout. However, neural implementation of such learning remains unclear as BPTT requires offline update using transported downstream weights, which is suggested to be biologically implausible. We demonstrate that simple online training of RNNs using TD reward prediction error and random feedback, without additional memory or eligibility trace, can still learn the structure of tasks with cue-reward delay and timing variability. This is because TD learning itself is a solution for temporal credit assignment, and feedback alignment, a mechanism originally proposed for supervised learning, enables gradient approximation without weight transport. Furthermore, we show that biologically constraining downstream weights and random feedback to be non-negative not only preserves learning but may even enhance it because the non-negative constraint ensures loose alignment-allowing the downstream and feedback weights to roughly align from the beginning. These results provide insights into the neural mechanisms underlying the learning of state representation and value, highlighting the potential of random feedback and biological constraints.

求助全文
通过发布文献求助,成功后即可免费获取论文全文。 去求助
来源期刊
eLife
eLife BIOLOGY-
CiteScore
12.90
自引率
3.90%
发文量
3122
审稿时长
17 weeks
期刊介绍: eLife is a distinguished, not-for-profit, peer-reviewed open access scientific journal that specializes in the fields of biomedical and life sciences. eLife is known for its selective publication process, which includes a variety of article types such as: Research Articles: Detailed reports of original research findings. Short Reports: Concise presentations of significant findings that do not warrant a full-length research article. Tools and Resources: Descriptions of new tools, technologies, or resources that facilitate scientific research. Research Advances: Brief reports on significant scientific advancements that have immediate implications for the field. Scientific Correspondence: Short communications that comment on or provide additional information related to published articles. Review Articles: Comprehensive overviews of a specific topic or field within the life sciences.
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
确定
请完成安全验证×
copy
已复制链接
快去分享给好友吧!
我知道了
右上角分享
点击右上角分享
0
联系我们:info@booksci.cn Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。 Copyright © 2023 布克学术 All rights reserved.
京ICP备2023020795号-1
ghs 京公网安备 11010802042870号
Book学术文献互助
Book学术文献互助群
群 号:604180095
Book学术官方微信