连续时间递归神经网络的验证(基准建议)

ARCH@ADHS Pub Date : 2018-09-17 DOI:10.29007/6czp
Patrick Musau, Taylor T. Johnson
{"title":"连续时间递归神经网络的验证(基准建议)","authors":"Patrick Musau, Taylor T. Johnson","doi":"10.29007/6czp","DOIUrl":null,"url":null,"abstract":"This manuscript presents a description and implementation of two benchmark problems for continuous-time recurrent neural network (RNN) verification. The first problem deals with the approximation of a vector field for a fixed point attractor located at the origin, whereas the second problem deals with the system identification of a forced damped pendulum. While the verification of neural networks is complicated and often impenetrable to the majority of verification techniques, continuous-time RNNs represent a class of networks that may be accessible to reachability methods for nonlinear ordinary differential equations (ODEs) derived originally in biology and neuroscience. Thus, an understanding of the behavior of a RNN may be gained by simulating the nonlinear equations from a diverse set of initial conditions and inputs, or considering reachability analysis from a set of initial conditions. The verification of continuous-time RNNs is a research area that has received little attention and if the research community can achieve meaningful results in this domain, then this class of neural networks may prove to be a superior approach in solving complex problems compared to other network architectures. Category: Academic Difficulty: High 1 Context and Origins Artificial Neural Networks have demonstrated an effective and powerful ability to achieve success in numerous contexts, such as adaptive control [43], autonomous vehicles, evolutionary robotics, pattern recognition, image classification, and nonlinear system identification and control [38] [18]. Despite this success, there have been reservations about incorporating them into safety critical systems [23] due to their susceptibility to unexpected and errant behavior from a slight perturbation in their inputs and initial conditions [42] [37]. Typically, neural networks are viewed as \"black boxes\" since the underlying operation of the neuron activations is often indiscernible to the creators of the network [10]. In light of these challenges, there has been significant work towards obtaining formal guarantees about the behavior of neural networks [25]. However, the majority of verification schemes have only been able to deal with neural networks that make use of piecewise-linear activation functions [7]. This is due to the great difficulty exhibited in obtaining formal guarantees for even simple properties of neural networks. In fact, neural network verification has been demonstrated to be an NP-complete problem, and while G. Frehse (ed.), ARCH18 (EPiC Series in Computing, vol. 54), pp. 196–207 Verification of Continuous Time Recurrent Neural Networks (Benchmark Proposal) Musau and Johnson techniques that make use of satisfiability modulo theories [35], mixed integer programming [41], robustness testing [4], and linear programming [13] [37] have been able to deal with small networks, they are incapable of dealing with the complexity and scale of the majority of networks present in real-life applications [23]. Moreover, the majority of verification approaches have dealt only with feed-forward and convolutional neural network architectures. One class of neural networks that has received particularly little attention in the verification literature is the class of recurrent neural networks. Whereas both feed-forward networks and recurrent networks have demonstrated an ability to approximate continuous functions to any accuracy [16], recurrent neural networks have exhibited several advantages over their feed-forward counterparts [26]. By allowing for the presence of feedback connections in their architecture, recurrent neural networks are able to retain information about the past and capture a higher degree of sophisticated dynamics using fewer neurons than their feed-forward counterparts [5]. In fact, recurrent neural networks have demonstrated a higher level of success in solving problems in which there is a temporal relation between events [32] such as capturing the behavior of biological neurons [28], dynamical system identification [22], real time gesture recognition [3], robotics [6, 8, 27, 30] and speech recognition [1]. Therefore, they represent a more attractive framework than feed-forward networks in these domains [47]. However, due to the complexity exhibited by their architecture as well as the non-linear nature of their activation functions the verification approaches currently available in the research literature are incapable of being applied to these networks. Thus, there is an immediate need for methods and advanced software tools that can provide formal guarantees about their operation [23], particularly in the context of the system identification and the control of safety critical systems. In light of this shortcoming, the following paper presents two benchmark problems for the verification of a specific class of recurrent neural networks known as continuous-time recurrent neural networks (CTRNNs). Since the dynamics of CTRNNs can be expressed solely by a set non-linear ordinary differential equations (ODEs), the verification of such systems relies on an ability to reason about the reachable set from a set of initial conditions and inputs [39]. The two CTRNN benchmark problems we present are described as follows: the first is a network without inputs employed for the approximation of a fixed point attractor described in [46], and the second deals with a CTRNN used for the identification of a damped forced pendulum as described in [12]. The problems elucidated in the paper are modeled using Simulink/Stateflow (SLSF), Matlab scripts, and are available in the SpaceEx format1 [15]. We aim to provide a thorough problem description to which the numerous tools and approaches for non-linear systems present in the research community can be evaluated and compared [39]. This paper serves as a first step towards recurrent neural network verification. 2 General Mathematical Model for Continuous Time Recurrent Neural Networks The dynamics of a continuous-time recurrent neural network with n neurons is given by the following system of ordinary differential equations:","PeriodicalId":236469,"journal":{"name":"ARCH@ADHS","volume":"8 1","pages":"0"},"PeriodicalIF":0.0000,"publicationDate":"2018-09-17","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"1","resultStr":"{\"title\":\"Verification of Continuous Time Recurrent Neural Networks (Benchmark Proposal)\",\"authors\":\"Patrick Musau, Taylor T. Johnson\",\"doi\":\"10.29007/6czp\",\"DOIUrl\":null,\"url\":null,\"abstract\":\"This manuscript presents a description and implementation of two benchmark problems for continuous-time recurrent neural network (RNN) verification. The first problem deals with the approximation of a vector field for a fixed point attractor located at the origin, whereas the second problem deals with the system identification of a forced damped pendulum. While the verification of neural networks is complicated and often impenetrable to the majority of verification techniques, continuous-time RNNs represent a class of networks that may be accessible to reachability methods for nonlinear ordinary differential equations (ODEs) derived originally in biology and neuroscience. Thus, an understanding of the behavior of a RNN may be gained by simulating the nonlinear equations from a diverse set of initial conditions and inputs, or considering reachability analysis from a set of initial conditions. The verification of continuous-time RNNs is a research area that has received little attention and if the research community can achieve meaningful results in this domain, then this class of neural networks may prove to be a superior approach in solving complex problems compared to other network architectures. Category: Academic Difficulty: High 1 Context and Origins Artificial Neural Networks have demonstrated an effective and powerful ability to achieve success in numerous contexts, such as adaptive control [43], autonomous vehicles, evolutionary robotics, pattern recognition, image classification, and nonlinear system identification and control [38] [18]. Despite this success, there have been reservations about incorporating them into safety critical systems [23] due to their susceptibility to unexpected and errant behavior from a slight perturbation in their inputs and initial conditions [42] [37]. Typically, neural networks are viewed as \\\"black boxes\\\" since the underlying operation of the neuron activations is often indiscernible to the creators of the network [10]. In light of these challenges, there has been significant work towards obtaining formal guarantees about the behavior of neural networks [25]. However, the majority of verification schemes have only been able to deal with neural networks that make use of piecewise-linear activation functions [7]. This is due to the great difficulty exhibited in obtaining formal guarantees for even simple properties of neural networks. In fact, neural network verification has been demonstrated to be an NP-complete problem, and while G. Frehse (ed.), ARCH18 (EPiC Series in Computing, vol. 54), pp. 196–207 Verification of Continuous Time Recurrent Neural Networks (Benchmark Proposal) Musau and Johnson techniques that make use of satisfiability modulo theories [35], mixed integer programming [41], robustness testing [4], and linear programming [13] [37] have been able to deal with small networks, they are incapable of dealing with the complexity and scale of the majority of networks present in real-life applications [23]. Moreover, the majority of verification approaches have dealt only with feed-forward and convolutional neural network architectures. One class of neural networks that has received particularly little attention in the verification literature is the class of recurrent neural networks. Whereas both feed-forward networks and recurrent networks have demonstrated an ability to approximate continuous functions to any accuracy [16], recurrent neural networks have exhibited several advantages over their feed-forward counterparts [26]. By allowing for the presence of feedback connections in their architecture, recurrent neural networks are able to retain information about the past and capture a higher degree of sophisticated dynamics using fewer neurons than their feed-forward counterparts [5]. In fact, recurrent neural networks have demonstrated a higher level of success in solving problems in which there is a temporal relation between events [32] such as capturing the behavior of biological neurons [28], dynamical system identification [22], real time gesture recognition [3], robotics [6, 8, 27, 30] and speech recognition [1]. Therefore, they represent a more attractive framework than feed-forward networks in these domains [47]. However, due to the complexity exhibited by their architecture as well as the non-linear nature of their activation functions the verification approaches currently available in the research literature are incapable of being applied to these networks. Thus, there is an immediate need for methods and advanced software tools that can provide formal guarantees about their operation [23], particularly in the context of the system identification and the control of safety critical systems. In light of this shortcoming, the following paper presents two benchmark problems for the verification of a specific class of recurrent neural networks known as continuous-time recurrent neural networks (CTRNNs). Since the dynamics of CTRNNs can be expressed solely by a set non-linear ordinary differential equations (ODEs), the verification of such systems relies on an ability to reason about the reachable set from a set of initial conditions and inputs [39]. The two CTRNN benchmark problems we present are described as follows: the first is a network without inputs employed for the approximation of a fixed point attractor described in [46], and the second deals with a CTRNN used for the identification of a damped forced pendulum as described in [12]. The problems elucidated in the paper are modeled using Simulink/Stateflow (SLSF), Matlab scripts, and are available in the SpaceEx format1 [15]. We aim to provide a thorough problem description to which the numerous tools and approaches for non-linear systems present in the research community can be evaluated and compared [39]. This paper serves as a first step towards recurrent neural network verification. 2 General Mathematical Model for Continuous Time Recurrent Neural Networks The dynamics of a continuous-time recurrent neural network with n neurons is given by the following system of ordinary differential equations:\",\"PeriodicalId\":236469,\"journal\":{\"name\":\"ARCH@ADHS\",\"volume\":\"8 1\",\"pages\":\"0\"},\"PeriodicalIF\":0.0000,\"publicationDate\":\"2018-09-17\",\"publicationTypes\":\"Journal Article\",\"fieldsOfStudy\":null,\"isOpenAccess\":false,\"openAccessPdf\":\"\",\"citationCount\":\"1\",\"resultStr\":null,\"platform\":\"Semanticscholar\",\"paperid\":null,\"PeriodicalName\":\"ARCH@ADHS\",\"FirstCategoryId\":\"1085\",\"ListUrlMain\":\"https://doi.org/10.29007/6czp\",\"RegionNum\":0,\"RegionCategory\":null,\"ArticlePicture\":[],\"TitleCN\":null,\"AbstractTextCN\":null,\"PMCID\":null,\"EPubDate\":\"\",\"PubModel\":\"\",\"JCR\":\"\",\"JCRName\":\"\",\"Score\":null,\"Total\":0}","platform":"Semanticscholar","paperid":null,"PeriodicalName":"ARCH@ADHS","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/10.29007/6czp","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
引用次数: 1

摘要

本文介绍了连续时间递归神经网络(RNN)验证的两个基准问题的描述和实现。第一个问题处理的是原点上一个不动点吸引子的矢量场的逼近,而第二个问题处理的是一个受迫阻尼摆的系统辨识。虽然神经网络的验证是复杂的,并且大多数验证技术通常无法穿透,但连续时间rnn代表了一类网络,可以使用最初在生物学和神经科学中导出的非线性常微分方程(ode)的可达性方法进行访问。因此,可以通过模拟来自不同初始条件和输入的非线性方程,或考虑来自一组初始条件的可达性分析,来理解RNN的行为。连续时间rnn的验证是一个很少受到关注的研究领域,如果研究团体能在这一领域取得有意义的成果,那么这类神经网络可能会被证明是解决复杂问题的一种优于其他网络架构的方法。人工神经网络已经证明了在许多环境中取得成功的有效而强大的能力,例如自适应控制[43]、自动驾驶汽车、进化机器人、模式识别、图像分类和非线性系统识别与控制[38][18]。尽管取得了成功,但由于输入和初始条件的轻微扰动对意外和错误行为的敏感性,将它们纳入安全关键系统[23]一直存在保留意见[42][37]。通常,神经网络被视为“黑箱”,因为神经元激活的潜在操作通常对网络的创建者来说是无法分辨的[10]。鉴于这些挑战,人们已经在获得关于神经网络行为的正式保证方面做了大量工作[25]。然而,大多数验证方案只能处理使用分段线性激活函数的神经网络[7]。这是由于即使对神经网络的简单性质也很难得到形式保证。事实上,神经网络验证已经被证明是一个np完全问题,而G. Frehse(主编),ARCH18 (EPiC Series In Computing, vol. 54), pp. 196-207连续时间循环神经网络的验证(基准提案)Musau和Johnson技术利用可满足模理论[35],混合整数规划[41],鲁棒性测试[4]和线性规划[13][37]已经能够处理小型网络,它们无法处理现实应用中大多数网络的复杂性和规模[23]。此外,大多数验证方法只处理前馈和卷积神经网络架构。一类在验证文献中受到特别少关注的神经网络是递归神经网络。虽然前馈网络和循环网络都证明了以任何精度近似连续函数的能力[16],但循环神经网络比前馈网络表现出一些优势[26]。通过允许在其结构中存在反馈连接,循环神经网络能够保留有关过去的信息,并使用比前馈网络更少的神经元捕获更高程度的复杂动态[5]。事实上,递归神经网络在解决事件之间存在时间关系的问题[32]方面表现出了更高的成功水平,例如捕获生物神经元的行为[28]、动态系统识别[22]、实时手势识别[3]、机器人[6,8,27,30]和语音识别[1]。因此,在这些领域,它们代表了比前馈网络更有吸引力的框架[47]。然而,由于其结构的复杂性以及其激活函数的非线性性质,目前研究文献中可用的验证方法无法应用于这些网络。因此,迫切需要能够为其运行提供正式保证的方法和先进的软件工具[23],特别是在系统识别和安全关键系统控制的背景下。鉴于这一缺点,下面的论文提出了两个基准问题,用于验证特定类别的递归神经网络,即连续时间递归神经网络(CTRNNs)。
本文章由计算机程序翻译,如有差异,请以英文原文为准。
Verification of Continuous Time Recurrent Neural Networks (Benchmark Proposal)
This manuscript presents a description and implementation of two benchmark problems for continuous-time recurrent neural network (RNN) verification. The first problem deals with the approximation of a vector field for a fixed point attractor located at the origin, whereas the second problem deals with the system identification of a forced damped pendulum. While the verification of neural networks is complicated and often impenetrable to the majority of verification techniques, continuous-time RNNs represent a class of networks that may be accessible to reachability methods for nonlinear ordinary differential equations (ODEs) derived originally in biology and neuroscience. Thus, an understanding of the behavior of a RNN may be gained by simulating the nonlinear equations from a diverse set of initial conditions and inputs, or considering reachability analysis from a set of initial conditions. The verification of continuous-time RNNs is a research area that has received little attention and if the research community can achieve meaningful results in this domain, then this class of neural networks may prove to be a superior approach in solving complex problems compared to other network architectures. Category: Academic Difficulty: High 1 Context and Origins Artificial Neural Networks have demonstrated an effective and powerful ability to achieve success in numerous contexts, such as adaptive control [43], autonomous vehicles, evolutionary robotics, pattern recognition, image classification, and nonlinear system identification and control [38] [18]. Despite this success, there have been reservations about incorporating them into safety critical systems [23] due to their susceptibility to unexpected and errant behavior from a slight perturbation in their inputs and initial conditions [42] [37]. Typically, neural networks are viewed as "black boxes" since the underlying operation of the neuron activations is often indiscernible to the creators of the network [10]. In light of these challenges, there has been significant work towards obtaining formal guarantees about the behavior of neural networks [25]. However, the majority of verification schemes have only been able to deal with neural networks that make use of piecewise-linear activation functions [7]. This is due to the great difficulty exhibited in obtaining formal guarantees for even simple properties of neural networks. In fact, neural network verification has been demonstrated to be an NP-complete problem, and while G. Frehse (ed.), ARCH18 (EPiC Series in Computing, vol. 54), pp. 196–207 Verification of Continuous Time Recurrent Neural Networks (Benchmark Proposal) Musau and Johnson techniques that make use of satisfiability modulo theories [35], mixed integer programming [41], robustness testing [4], and linear programming [13] [37] have been able to deal with small networks, they are incapable of dealing with the complexity and scale of the majority of networks present in real-life applications [23]. Moreover, the majority of verification approaches have dealt only with feed-forward and convolutional neural network architectures. One class of neural networks that has received particularly little attention in the verification literature is the class of recurrent neural networks. Whereas both feed-forward networks and recurrent networks have demonstrated an ability to approximate continuous functions to any accuracy [16], recurrent neural networks have exhibited several advantages over their feed-forward counterparts [26]. By allowing for the presence of feedback connections in their architecture, recurrent neural networks are able to retain information about the past and capture a higher degree of sophisticated dynamics using fewer neurons than their feed-forward counterparts [5]. In fact, recurrent neural networks have demonstrated a higher level of success in solving problems in which there is a temporal relation between events [32] such as capturing the behavior of biological neurons [28], dynamical system identification [22], real time gesture recognition [3], robotics [6, 8, 27, 30] and speech recognition [1]. Therefore, they represent a more attractive framework than feed-forward networks in these domains [47]. However, due to the complexity exhibited by their architecture as well as the non-linear nature of their activation functions the verification approaches currently available in the research literature are incapable of being applied to these networks. Thus, there is an immediate need for methods and advanced software tools that can provide formal guarantees about their operation [23], particularly in the context of the system identification and the control of safety critical systems. In light of this shortcoming, the following paper presents two benchmark problems for the verification of a specific class of recurrent neural networks known as continuous-time recurrent neural networks (CTRNNs). Since the dynamics of CTRNNs can be expressed solely by a set non-linear ordinary differential equations (ODEs), the verification of such systems relies on an ability to reason about the reachable set from a set of initial conditions and inputs [39]. The two CTRNN benchmark problems we present are described as follows: the first is a network without inputs employed for the approximation of a fixed point attractor described in [46], and the second deals with a CTRNN used for the identification of a damped forced pendulum as described in [12]. The problems elucidated in the paper are modeled using Simulink/Stateflow (SLSF), Matlab scripts, and are available in the SpaceEx format1 [15]. We aim to provide a thorough problem description to which the numerous tools and approaches for non-linear systems present in the research community can be evaluated and compared [39]. This paper serves as a first step towards recurrent neural network verification. 2 General Mathematical Model for Continuous Time Recurrent Neural Networks The dynamics of a continuous-time recurrent neural network with n neurons is given by the following system of ordinary differential equations:
求助全文
通过发布文献求助,成功后即可免费获取论文全文。 去求助
来源期刊
自引率
0.00%
发文量
0
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
确定
请完成安全验证×
copy
已复制链接
快去分享给好友吧!
我知道了
右上角分享
点击右上角分享
0
联系我们:info@booksci.cn Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。 Copyright © 2023 布克学术 All rights reserved.
京ICP备2023020795号-1
ghs 京公网安备 11010802042870号
Book学术文献互助
Book学术文献互助群
群 号:481959085
Book学术官方微信