{"title":"温和地驯服混沌:递归神经网络中的预测对齐学习规则","authors":"Toshitake Asabuki, Claudia Clopath","doi":"10.1038/s41467-025-61309-9","DOIUrl":null,"url":null,"abstract":"<p>Recurrent neural circuits often face inherent complexities in learning and generating their desired outputs, especially when they initially exhibit chaotic spontaneous activity. While the celebrated FORCE learning rule can train chaotic recurrent networks to produce coherent patterns by suppressing chaos, it requires non-local plasticity rules and quick plasticity, raising the question of how synapses adapt on local, biologically plausible timescales to handle potential chaotic dynamics. We propose a novel framework called “predictive alignment”, which tames the chaotic recurrent dynamics to generate a variety of patterned activities via a biologically plausible plasticity rule. Unlike most recurrent learning rules, predictive alignment does not aim to directly minimize output error to train recurrent connections, but rather it tries to efficiently suppress chaos by aligning recurrent prediction with chaotic activity. We show that the proposed learning rule can perform supervised learning of multiple target signals, including complex low-dimensional attractors, delay matching tasks that require short-term temporal memory, and finally even dynamic movie clips with high-dimensional pixels. Our findings shed light on how predictions in recurrent circuits can support learning.</p>","PeriodicalId":19066,"journal":{"name":"Nature Communications","volume":"47 1","pages":""},"PeriodicalIF":15.7000,"publicationDate":"2025-07-23","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":"{\"title\":\"Taming the chaos gently: a predictive alignment learning rule in recurrent neural networks\",\"authors\":\"Toshitake Asabuki, Claudia Clopath\",\"doi\":\"10.1038/s41467-025-61309-9\",\"DOIUrl\":null,\"url\":null,\"abstract\":\"<p>Recurrent neural circuits often face inherent complexities in learning and generating their desired outputs, especially when they initially exhibit chaotic spontaneous activity. While the celebrated FORCE learning rule can train chaotic recurrent networks to produce coherent patterns by suppressing chaos, it requires non-local plasticity rules and quick plasticity, raising the question of how synapses adapt on local, biologically plausible timescales to handle potential chaotic dynamics. We propose a novel framework called “predictive alignment”, which tames the chaotic recurrent dynamics to generate a variety of patterned activities via a biologically plausible plasticity rule. Unlike most recurrent learning rules, predictive alignment does not aim to directly minimize output error to train recurrent connections, but rather it tries to efficiently suppress chaos by aligning recurrent prediction with chaotic activity. We show that the proposed learning rule can perform supervised learning of multiple target signals, including complex low-dimensional attractors, delay matching tasks that require short-term temporal memory, and finally even dynamic movie clips with high-dimensional pixels. Our findings shed light on how predictions in recurrent circuits can support learning.</p>\",\"PeriodicalId\":19066,\"journal\":{\"name\":\"Nature Communications\",\"volume\":\"47 1\",\"pages\":\"\"},\"PeriodicalIF\":15.7000,\"publicationDate\":\"2025-07-23\",\"publicationTypes\":\"Journal Article\",\"fieldsOfStudy\":null,\"isOpenAccess\":false,\"openAccessPdf\":\"\",\"citationCount\":\"0\",\"resultStr\":null,\"platform\":\"Semanticscholar\",\"paperid\":null,\"PeriodicalName\":\"Nature Communications\",\"FirstCategoryId\":\"103\",\"ListUrlMain\":\"https://doi.org/10.1038/s41467-025-61309-9\",\"RegionNum\":1,\"RegionCategory\":\"综合性期刊\",\"ArticlePicture\":[],\"TitleCN\":null,\"AbstractTextCN\":null,\"PMCID\":null,\"EPubDate\":\"\",\"PubModel\":\"\",\"JCR\":\"Q1\",\"JCRName\":\"MULTIDISCIPLINARY SCIENCES\",\"Score\":null,\"Total\":0}","platform":"Semanticscholar","paperid":null,"PeriodicalName":"Nature Communications","FirstCategoryId":"103","ListUrlMain":"https://doi.org/10.1038/s41467-025-61309-9","RegionNum":1,"RegionCategory":"综合性期刊","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"Q1","JCRName":"MULTIDISCIPLINARY SCIENCES","Score":null,"Total":0}
Taming the chaos gently: a predictive alignment learning rule in recurrent neural networks
Recurrent neural circuits often face inherent complexities in learning and generating their desired outputs, especially when they initially exhibit chaotic spontaneous activity. While the celebrated FORCE learning rule can train chaotic recurrent networks to produce coherent patterns by suppressing chaos, it requires non-local plasticity rules and quick plasticity, raising the question of how synapses adapt on local, biologically plausible timescales to handle potential chaotic dynamics. We propose a novel framework called “predictive alignment”, which tames the chaotic recurrent dynamics to generate a variety of patterned activities via a biologically plausible plasticity rule. Unlike most recurrent learning rules, predictive alignment does not aim to directly minimize output error to train recurrent connections, but rather it tries to efficiently suppress chaos by aligning recurrent prediction with chaotic activity. We show that the proposed learning rule can perform supervised learning of multiple target signals, including complex low-dimensional attractors, delay matching tasks that require short-term temporal memory, and finally even dynamic movie clips with high-dimensional pixels. Our findings shed light on how predictions in recurrent circuits can support learning.
期刊介绍:
Nature Communications, an open-access journal, publishes high-quality research spanning all areas of the natural sciences. Papers featured in the journal showcase significant advances relevant to specialists in each respective field. With a 2-year impact factor of 16.6 (2022) and a median time of 8 days from submission to the first editorial decision, Nature Communications is committed to rapid dissemination of research findings. As a multidisciplinary journal, it welcomes contributions from biological, health, physical, chemical, Earth, social, mathematical, applied, and engineering sciences, aiming to highlight important breakthroughs within each domain.