nmODE:神经记忆常微分方程

IF 10.7 2区 计算机科学 Q1 COMPUTER SCIENCE, ARTIFICIAL INTELLIGENCE
Zhang Yi
{"title":"nmODE:神经记忆常微分方程","authors":"Zhang Yi","doi":"10.1007/s10462-023-10496-2","DOIUrl":null,"url":null,"abstract":"<div><p>Brain neural networks are regarded as dynamical systems in neural science, in which memories are interpreted as attractors of the systems. Mathematically, ordinary differential equations (ODEs) can be utilized to describe dynamical systems. Any ODE that is employed to describe the dynamics of a neural network can be called a neuralODE. Inspired by rethinking the nonlinear representation ability of existing artificial neural networks together with the functions of columns in the neocortex, this paper proposes a theory of memory-based neuralODE, which is composed of two novel artificial neural network models: nmODE and <span>\\(\\epsilon\\)</span>-net, and two learning algorithms: nmLA and <span>\\(\\epsilon\\)</span>-LA. The nmODE (neural memory Ordinary Differential Equation) is designed with a special structure that separates learning neurons from memory neurons, making its dynamics clear. Given any external input, the nmODE possesses the global attractor property and is thus embedded with a memory mechanism. The nmODE establishes a nonlinear mapping from the external input to its associated attractor and does not have the problem of learning features homeomorphic to the input data space, as occurs frequently in most existing neuralODEs. The nmLA (neural memory Learning Algorithm) is developed by proposing an interesting three-dimensional inverse ODE (invODE) and has advantages in memory and parameter efficiency. The proposed <span>\\(\\epsilon\\)</span>-net is a discrete version of the nmODE, which is particularly feasible for digital computing. The proposed <span>\\(\\epsilon\\)</span>-LA (<span>\\(\\epsilon\\)</span> learning algorithm) requires no prior knowledge of the number of network layers. Both nmLA and <span>\\(\\epsilon\\)</span>-LA have no problem with gradient vanishing. Experimental results show that the proposed theory is comparable to state-of-the-art methods.</p></div>","PeriodicalId":8449,"journal":{"name":"Artificial Intelligence Review","volume":"56 12","pages":"14403 - 14438"},"PeriodicalIF":10.7000,"publicationDate":"2023-05-22","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://link.springer.com/content/pdf/10.1007/s10462-023-10496-2.pdf","citationCount":"1","resultStr":"{\"title\":\"nmODE: neural memory ordinary differential equation\",\"authors\":\"Zhang Yi\",\"doi\":\"10.1007/s10462-023-10496-2\",\"DOIUrl\":null,\"url\":null,\"abstract\":\"<div><p>Brain neural networks are regarded as dynamical systems in neural science, in which memories are interpreted as attractors of the systems. Mathematically, ordinary differential equations (ODEs) can be utilized to describe dynamical systems. Any ODE that is employed to describe the dynamics of a neural network can be called a neuralODE. Inspired by rethinking the nonlinear representation ability of existing artificial neural networks together with the functions of columns in the neocortex, this paper proposes a theory of memory-based neuralODE, which is composed of two novel artificial neural network models: nmODE and <span>\\\\(\\\\epsilon\\\\)</span>-net, and two learning algorithms: nmLA and <span>\\\\(\\\\epsilon\\\\)</span>-LA. The nmODE (neural memory Ordinary Differential Equation) is designed with a special structure that separates learning neurons from memory neurons, making its dynamics clear. Given any external input, the nmODE possesses the global attractor property and is thus embedded with a memory mechanism. The nmODE establishes a nonlinear mapping from the external input to its associated attractor and does not have the problem of learning features homeomorphic to the input data space, as occurs frequently in most existing neuralODEs. The nmLA (neural memory Learning Algorithm) is developed by proposing an interesting three-dimensional inverse ODE (invODE) and has advantages in memory and parameter efficiency. The proposed <span>\\\\(\\\\epsilon\\\\)</span>-net is a discrete version of the nmODE, which is particularly feasible for digital computing. The proposed <span>\\\\(\\\\epsilon\\\\)</span>-LA (<span>\\\\(\\\\epsilon\\\\)</span> learning algorithm) requires no prior knowledge of the number of network layers. Both nmLA and <span>\\\\(\\\\epsilon\\\\)</span>-LA have no problem with gradient vanishing. Experimental results show that the proposed theory is comparable to state-of-the-art methods.</p></div>\",\"PeriodicalId\":8449,\"journal\":{\"name\":\"Artificial Intelligence Review\",\"volume\":\"56 12\",\"pages\":\"14403 - 14438\"},\"PeriodicalIF\":10.7000,\"publicationDate\":\"2023-05-22\",\"publicationTypes\":\"Journal Article\",\"fieldsOfStudy\":null,\"isOpenAccess\":false,\"openAccessPdf\":\"https://link.springer.com/content/pdf/10.1007/s10462-023-10496-2.pdf\",\"citationCount\":\"1\",\"resultStr\":null,\"platform\":\"Semanticscholar\",\"paperid\":null,\"PeriodicalName\":\"Artificial Intelligence Review\",\"FirstCategoryId\":\"94\",\"ListUrlMain\":\"https://link.springer.com/article/10.1007/s10462-023-10496-2\",\"RegionNum\":2,\"RegionCategory\":\"计算机科学\",\"ArticlePicture\":[],\"TitleCN\":null,\"AbstractTextCN\":null,\"PMCID\":null,\"EPubDate\":\"\",\"PubModel\":\"\",\"JCR\":\"Q1\",\"JCRName\":\"COMPUTER SCIENCE, ARTIFICIAL INTELLIGENCE\",\"Score\":null,\"Total\":0}","platform":"Semanticscholar","paperid":null,"PeriodicalName":"Artificial Intelligence Review","FirstCategoryId":"94","ListUrlMain":"https://link.springer.com/article/10.1007/s10462-023-10496-2","RegionNum":2,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"Q1","JCRName":"COMPUTER SCIENCE, ARTIFICIAL INTELLIGENCE","Score":null,"Total":0}
引用次数: 1

摘要

在神经科学中,大脑神经网络被视为动态系统,记忆被解释为系统的吸引子。在数学上,常微分方程(ode)可以用来描述动力系统。任何用于描述神经网络动态的ODE都可以称为神经ODE。在重新思考现有人工神经网络的非线性表征能力和新皮层中列的功能的基础上,本文提出了一种基于记忆的neuralODE理论,该理论由两种新颖的人工神经网络模型nmODE和\(\epsilon\) -net以及两种学习算法nmLA和\(\epsilon\) -LA组成。nmODE(神经记忆常微分方程)采用了一种特殊的结构,将学习神经元与记忆神经元分开,使其动态清晰。给定任何外部输入,nmODE具有全局吸引子属性,因此嵌入了内存机制。nmODE建立了一个从外部输入到其相关吸引子的非线性映射,并且不存在学习特征与输入数据空间同胚的问题,这在大多数现有的神经ode中经常发生。神经记忆学习算法(nmLA)是通过提出一种有趣的三维逆ODE (invODE)而发展起来的,在内存和参数效率方面具有优势。提出的\(\epsilon\) -net是nmODE的离散版本,特别适用于数字计算。提出的\(\epsilon\) -LA (\(\epsilon\)学习算法)不需要事先知道网络层数。nmLA和\(\epsilon\) -LA都没有梯度消失的问题。实验结果表明,所提出的理论与目前最先进的方法相当。
本文章由计算机程序翻译,如有差异,请以英文原文为准。
nmODE: neural memory ordinary differential equation

Brain neural networks are regarded as dynamical systems in neural science, in which memories are interpreted as attractors of the systems. Mathematically, ordinary differential equations (ODEs) can be utilized to describe dynamical systems. Any ODE that is employed to describe the dynamics of a neural network can be called a neuralODE. Inspired by rethinking the nonlinear representation ability of existing artificial neural networks together with the functions of columns in the neocortex, this paper proposes a theory of memory-based neuralODE, which is composed of two novel artificial neural network models: nmODE and \(\epsilon\)-net, and two learning algorithms: nmLA and \(\epsilon\)-LA. The nmODE (neural memory Ordinary Differential Equation) is designed with a special structure that separates learning neurons from memory neurons, making its dynamics clear. Given any external input, the nmODE possesses the global attractor property and is thus embedded with a memory mechanism. The nmODE establishes a nonlinear mapping from the external input to its associated attractor and does not have the problem of learning features homeomorphic to the input data space, as occurs frequently in most existing neuralODEs. The nmLA (neural memory Learning Algorithm) is developed by proposing an interesting three-dimensional inverse ODE (invODE) and has advantages in memory and parameter efficiency. The proposed \(\epsilon\)-net is a discrete version of the nmODE, which is particularly feasible for digital computing. The proposed \(\epsilon\)-LA (\(\epsilon\) learning algorithm) requires no prior knowledge of the number of network layers. Both nmLA and \(\epsilon\)-LA have no problem with gradient vanishing. Experimental results show that the proposed theory is comparable to state-of-the-art methods.

求助全文
通过发布文献求助,成功后即可免费获取论文全文。 去求助
来源期刊
Artificial Intelligence Review
Artificial Intelligence Review 工程技术-计算机:人工智能
CiteScore
22.00
自引率
3.30%
发文量
194
审稿时长
5.3 months
期刊介绍: Artificial Intelligence Review, a fully open access journal, publishes cutting-edge research in artificial intelligence and cognitive science. It features critical evaluations of applications, techniques, and algorithms, providing a platform for both researchers and application developers. The journal includes refereed survey and tutorial articles, along with reviews and commentary on significant developments in the field.
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
确定
请完成安全验证×
copy
已复制链接
快去分享给好友吧!
我知道了
右上角分享
点击右上角分享
0
联系我们:info@booksci.cn Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。 Copyright © 2023 布克学术 All rights reserved.
京ICP备2023020795号-1
ghs 京公网安备 11010802042870号
Book学术文献互助
Book学术文献互助群
群 号:481959085
Book学术官方微信