{"title":"具有s型单元的神经网络的复杂性","authors":"Kai-Yeung Siu, V. Roychowdhury, T. Kailath","doi":"10.1109/NNSP.1992.253711","DOIUrl":null,"url":null,"abstract":"Novel techniques based on classical tools such as rational approximation and harmonic analysis are developed to study the computational properties of neural networks. Using such techniques, one can characterize the class of function whose complexity is almost the same among various models of neural networks with feedforward structures. As a consequence of this characterization, for example, it is proved that any depth-(d+1) network of sigmoidal units computing the parity function of n inputs must have Omega (dn/sup 1/d- in /) units, for any fixed in >0. This lower bound is almost tight since one can compute the parity function with O(dn/sup 1/d/) sigmoidal units in a depth-(d+1) network. The techniques also generalize to networks whose elements can be approximated by piecewise low degree rational functions.<<ETX>>","PeriodicalId":438250,"journal":{"name":"Neural Networks for Signal Processing II Proceedings of the 1992 IEEE Workshop","volume":"106 1","pages":"0"},"PeriodicalIF":0.0000,"publicationDate":"1992-08-31","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"5","resultStr":"{\"title\":\"On the complexity of neural networks with sigmoidal units\",\"authors\":\"Kai-Yeung Siu, V. Roychowdhury, T. Kailath\",\"doi\":\"10.1109/NNSP.1992.253711\",\"DOIUrl\":null,\"url\":null,\"abstract\":\"Novel techniques based on classical tools such as rational approximation and harmonic analysis are developed to study the computational properties of neural networks. Using such techniques, one can characterize the class of function whose complexity is almost the same among various models of neural networks with feedforward structures. As a consequence of this characterization, for example, it is proved that any depth-(d+1) network of sigmoidal units computing the parity function of n inputs must have Omega (dn/sup 1/d- in /) units, for any fixed in >0. This lower bound is almost tight since one can compute the parity function with O(dn/sup 1/d/) sigmoidal units in a depth-(d+1) network. The techniques also generalize to networks whose elements can be approximated by piecewise low degree rational functions.<<ETX>>\",\"PeriodicalId\":438250,\"journal\":{\"name\":\"Neural Networks for Signal Processing II Proceedings of the 1992 IEEE Workshop\",\"volume\":\"106 1\",\"pages\":\"0\"},\"PeriodicalIF\":0.0000,\"publicationDate\":\"1992-08-31\",\"publicationTypes\":\"Journal Article\",\"fieldsOfStudy\":null,\"isOpenAccess\":false,\"openAccessPdf\":\"\",\"citationCount\":\"5\",\"resultStr\":null,\"platform\":\"Semanticscholar\",\"paperid\":null,\"PeriodicalName\":\"Neural Networks for Signal Processing II Proceedings of the 1992 IEEE Workshop\",\"FirstCategoryId\":\"1085\",\"ListUrlMain\":\"https://doi.org/10.1109/NNSP.1992.253711\",\"RegionNum\":0,\"RegionCategory\":null,\"ArticlePicture\":[],\"TitleCN\":null,\"AbstractTextCN\":null,\"PMCID\":null,\"EPubDate\":\"\",\"PubModel\":\"\",\"JCR\":\"\",\"JCRName\":\"\",\"Score\":null,\"Total\":0}","platform":"Semanticscholar","paperid":null,"PeriodicalName":"Neural Networks for Signal Processing II Proceedings of the 1992 IEEE Workshop","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/10.1109/NNSP.1992.253711","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
引用次数: 5
摘要
在有理逼近和谐波分析等经典工具的基础上,提出了研究神经网络计算特性的新方法。利用这种技术,人们可以描述在具有前馈结构的各种神经网络模型中其复杂性几乎相同的函数类。作为这一表征的结果,例如,证明了任何计算n个输入的奇偶函数的s型单元的深度-(d+1)网络必须具有Omega (dn/sup 1/d- in /)单元,对于任何固定在>0。这个下界几乎是紧的,因为在深度-(d+1)网络中可以用O(dn/sup 1/d/) s型单元计算奇偶函数。该技术也推广到网络,其元素可以由分段低次有理函数逼近。
On the complexity of neural networks with sigmoidal units
Novel techniques based on classical tools such as rational approximation and harmonic analysis are developed to study the computational properties of neural networks. Using such techniques, one can characterize the class of function whose complexity is almost the same among various models of neural networks with feedforward structures. As a consequence of this characterization, for example, it is proved that any depth-(d+1) network of sigmoidal units computing the parity function of n inputs must have Omega (dn/sup 1/d- in /) units, for any fixed in >0. This lower bound is almost tight since one can compute the parity function with O(dn/sup 1/d/) sigmoidal units in a depth-(d+1) network. The techniques also generalize to networks whose elements can be approximated by piecewise low degree rational functions.<>