Measuring and Controlling Solution Degeneracy across Task-Trained Recurrent Neural Networks.

ArXiv Pub Date : 2025-05-28
Ann Huang, Satpreet H Singh, Flavio Martinelli, Kanaka Rajan
{"title":"Measuring and Controlling Solution Degeneracy across Task-Trained Recurrent Neural Networks.","authors":"Ann Huang, Satpreet H Singh, Flavio Martinelli, Kanaka Rajan","doi":"","DOIUrl":null,"url":null,"abstract":"<p><p>Task-trained recurrent neural networks (RNNs) are widely used in neuroscience and machine learning to model dynamical computations. To gain mechanistic insight into how neural systems solve tasks, prior work often reverse-engineers individual trained networks. However, different RNNs trained on the same task and achieving similar performance can exhibit strikingly different internal solutions-a phenomenon known as solution degeneracy. Here, we develop a unified framework to systematically quantify and control solution degeneracy across three levels: behavior, neural dynamics, and weight space. We apply this framework to 3,400 RNNs trained on four neuroscience-relevant tasks-flip-flop memory, sine wave generation, delayed discrimination, and path integration-while systematically varying task complexity, learning regime, network size, and regularization. We find that higher task complexity and stronger feature learning reduce degeneracy in neural dynamics but increase it in weight space, with mixed effects on behavior. In contrast, larger networks and structural regularization reduce degeneracy at all three levels. These findings empirically validate the Contravariance Principle and provide practical guidance for researchers aiming to tailor RNN solutions-whether to uncover shared neural mechanisms or to model individual variability observed in biological systems. This work provides a principled framework for quantifying and controlling solution degeneracy in task-trained RNNs, offering new tools for building more interpretable and biologically grounded models of neural computation.</p>","PeriodicalId":93888,"journal":{"name":"ArXiv","volume":" ","pages":""},"PeriodicalIF":0.0000,"publicationDate":"2025-05-28","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.ncbi.nlm.nih.gov/pmc/articles/PMC12148090/pdf/","citationCount":"0","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"ArXiv","FirstCategoryId":"1085","ListUrlMain":"","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
引用次数: 0

Abstract

Task-trained recurrent neural networks (RNNs) are widely used in neuroscience and machine learning to model dynamical computations. To gain mechanistic insight into how neural systems solve tasks, prior work often reverse-engineers individual trained networks. However, different RNNs trained on the same task and achieving similar performance can exhibit strikingly different internal solutions-a phenomenon known as solution degeneracy. Here, we develop a unified framework to systematically quantify and control solution degeneracy across three levels: behavior, neural dynamics, and weight space. We apply this framework to 3,400 RNNs trained on four neuroscience-relevant tasks-flip-flop memory, sine wave generation, delayed discrimination, and path integration-while systematically varying task complexity, learning regime, network size, and regularization. We find that higher task complexity and stronger feature learning reduce degeneracy in neural dynamics but increase it in weight space, with mixed effects on behavior. In contrast, larger networks and structural regularization reduce degeneracy at all three levels. These findings empirically validate the Contravariance Principle and provide practical guidance for researchers aiming to tailor RNN solutions-whether to uncover shared neural mechanisms or to model individual variability observed in biological systems. This work provides a principled framework for quantifying and controlling solution degeneracy in task-trained RNNs, offering new tools for building more interpretable and biologically grounded models of neural computation.

任务训练递归神经网络解退化的测量与控制。
任务训练递归神经网络(RNNs)广泛应用于神经科学和机器学习中,以模拟动态计算。为了深入了解神经系统如何解决任务,之前的工作通常是对单个训练过的网络进行逆向工程。然而,不同的rnn在相同的任务上训练并获得相似的性能,可能会表现出截然不同的内部解决方案——这种现象被称为解决方案退化。在这里,我们开发了一个统一的框架来系统地量化和控制三个层次的解退化:行为、神经动力学和权重空间。我们将该框架应用于3,400个rnn,这些rnn接受了与神经科学相关的四项任务(触发器记忆、正弦波生成、延迟识别和路径集成)的训练,同时系统地改变任务复杂性、学习机制、网络大小和正则化。我们发现更高的任务复杂度和更强的特征学习减少了神经动力学的退化,但增加了权重空间的退化,对行为的影响是混合的。相比之下,更大的网络和结构正则化在所有三个层次上都减少了退化。这些发现从经验上验证了逆变原理,并为旨在定制RNN解决方案的研究人员提供了实用指导-无论是揭示共享的神经机制还是模拟生物系统中观察到的个体可变性。这项工作为量化和控制任务训练rnn中的解退化提供了一个原则性框架,为构建更具可解释性和生物学基础的神经计算模型提供了新的工具。
本文章由计算机程序翻译,如有差异,请以英文原文为准。
求助全文
约1分钟内获得全文 求助全文
来源期刊
自引率
0.00%
发文量
0
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
确定
请完成安全验证×
copy
已复制链接
快去分享给好友吧!
我知道了
右上角分享
点击右上角分享
0
联系我们:info@booksci.cn Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。 Copyright © 2023 布克学术 All rights reserved.
京ICP备2023020795号-1
ghs 京公网安备 11010802042870号
Book学术文献互助
Book学术文献互助群
群 号:604180095
Book学术官方微信