Neural Ordinary Differential Equation based Recurrent Neural Network Model

M. Habiba, Barak A. Pearlmutter
{"title":"Neural Ordinary Differential Equation based Recurrent Neural Network Model","authors":"M. Habiba, Barak A. Pearlmutter","doi":"10.1109/ISSC49989.2020.9180182","DOIUrl":null,"url":null,"abstract":"Neural differential equations are a promising new member in the neural network family. They show the potential of differential equations for time-series data analysis. In this paper, the strength of the ordinary differential equation (ODE) is explored with a new extension. The main goal of this work is to answer the following questions: (i) can ODE be used to redefine the existing neural network model? (ii) can Neural ODEs solve the irregular sampling rate challenge of existing neural network models for a continuous time series, i.e., length and dynamic nature, (iii) how to reduce the training and evaluation time of existing Neural ODE systems? This work leverages the mathematical foundation of ODEs to redesign traditional RNNs such as Long Short-Term Memory (LSTM) and Gated Recurrent Unit (GRU). The main contribution of this paper is to illustrate the design of two new ODE-based RNN models (GRU-ODE model and LSTM-ODE) which can compute the hidden state and cell state at any point of time using an ODE solver. These models reduce the computation overhead of hidden state and cell state by a vast amount. The performance evaluation of these two new models for learning continuous time series with irregular sampling rate is then demonstrated. Experiments show that these new ODE based RNN models require less training time than Latent ODEs and conventional Neural ODEs. They can achieve higher accuracy quickly, and the design of the neural network is more straightforward than the previous neural ODE systems.","PeriodicalId":351013,"journal":{"name":"2020 31st Irish Signals and Systems Conference (ISSC)","volume":"9 1","pages":"0"},"PeriodicalIF":0.0000,"publicationDate":"2020-05-20","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"12","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"2020 31st Irish Signals and Systems Conference (ISSC)","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/10.1109/ISSC49989.2020.9180182","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
引用次数: 12

Abstract

Neural differential equations are a promising new member in the neural network family. They show the potential of differential equations for time-series data analysis. In this paper, the strength of the ordinary differential equation (ODE) is explored with a new extension. The main goal of this work is to answer the following questions: (i) can ODE be used to redefine the existing neural network model? (ii) can Neural ODEs solve the irregular sampling rate challenge of existing neural network models for a continuous time series, i.e., length and dynamic nature, (iii) how to reduce the training and evaluation time of existing Neural ODE systems? This work leverages the mathematical foundation of ODEs to redesign traditional RNNs such as Long Short-Term Memory (LSTM) and Gated Recurrent Unit (GRU). The main contribution of this paper is to illustrate the design of two new ODE-based RNN models (GRU-ODE model and LSTM-ODE) which can compute the hidden state and cell state at any point of time using an ODE solver. These models reduce the computation overhead of hidden state and cell state by a vast amount. The performance evaluation of these two new models for learning continuous time series with irregular sampling rate is then demonstrated. Experiments show that these new ODE based RNN models require less training time than Latent ODEs and conventional Neural ODEs. They can achieve higher accuracy quickly, and the design of the neural network is more straightforward than the previous neural ODE systems.
基于神经常微分方程的递归神经网络模型
神经微分方程是神经网络家族中一个很有前途的新成员。它们显示了微分方程在时间序列数据分析中的潜力。本文对常微分方程(ODE)的强度进行了新的扩展。这项工作的主要目标是回答以下问题:(i) ODE可以用来重新定义现有的神经网络模型吗?(ii) Neural ODE能否解决现有神经网络模型对连续时间序列的不规则采样率挑战,即长度和动态性;(iii)如何减少现有Neural ODE系统的训练和评估时间?这项工作利用ode的数学基础来重新设计传统的rnn,如长短期记忆(LSTM)和门控循环单元(GRU)。本文的主要贡献在于阐述了两种新的基于ODE的RNN模型(GRU-ODE模型和LSTM-ODE)的设计,这两种模型可以使用ODE求解器在任何时间点计算隐藏状态和单元状态。这些模型大大减少了隐藏状态和单元状态的计算开销。最后给出了这两种新模型在不规则采样率连续时间序列学习中的性能评价。实验表明,与潜在ODE和传统神经ODE相比,基于ODE的RNN模型所需的训练时间更短。它们可以快速达到更高的精度,并且神经网络的设计比以前的神经ODE系统更直接。
本文章由计算机程序翻译,如有差异,请以英文原文为准。
求助全文
约1分钟内获得全文 求助全文
来源期刊
自引率
0.00%
发文量
0
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
确定
请完成安全验证×
copy
已复制链接
快去分享给好友吧!
我知道了
右上角分享
点击右上角分享
0
联系我们:info@booksci.cn Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。 Copyright © 2023 布克学术 All rights reserved.
京ICP备2023020795号-1
ghs 京公网安备 11010802042870号
Book学术文献互助
Book学术文献互助群
群 号:481959085
Book学术官方微信