{"title":"反向传播神经网络的基向量分析","authors":"M.-S. Chen, M. Manry","doi":"10.1109/MWSCAS.1991.252222","DOIUrl":null,"url":null,"abstract":"Develops a polynomial basis function approach for modeling BP (backpropagation) neural networks. This method leads directly to a constructive proof of the BP approximation theorem. In addition, the basis vector approach provides a means to synthesize the BP neural network output as a polynomial function. An algorithm for pruning the useless basis vectors is also demonstrated.<<ETX>>","PeriodicalId":6453,"journal":{"name":"[1991] Proceedings of the 34th Midwest Symposium on Circuits and Systems","volume":"171 1","pages":"23-26 vol.1"},"PeriodicalIF":0.0000,"publicationDate":"1991-05-14","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"10","resultStr":"{\"title\":\"Basis vector analyses of back-propagation neural networks\",\"authors\":\"M.-S. Chen, M. Manry\",\"doi\":\"10.1109/MWSCAS.1991.252222\",\"DOIUrl\":null,\"url\":null,\"abstract\":\"Develops a polynomial basis function approach for modeling BP (backpropagation) neural networks. This method leads directly to a constructive proof of the BP approximation theorem. In addition, the basis vector approach provides a means to synthesize the BP neural network output as a polynomial function. An algorithm for pruning the useless basis vectors is also demonstrated.<<ETX>>\",\"PeriodicalId\":6453,\"journal\":{\"name\":\"[1991] Proceedings of the 34th Midwest Symposium on Circuits and Systems\",\"volume\":\"171 1\",\"pages\":\"23-26 vol.1\"},\"PeriodicalIF\":0.0000,\"publicationDate\":\"1991-05-14\",\"publicationTypes\":\"Journal Article\",\"fieldsOfStudy\":null,\"isOpenAccess\":false,\"openAccessPdf\":\"\",\"citationCount\":\"10\",\"resultStr\":null,\"platform\":\"Semanticscholar\",\"paperid\":null,\"PeriodicalName\":\"[1991] Proceedings of the 34th Midwest Symposium on Circuits and Systems\",\"FirstCategoryId\":\"1085\",\"ListUrlMain\":\"https://doi.org/10.1109/MWSCAS.1991.252222\",\"RegionNum\":0,\"RegionCategory\":null,\"ArticlePicture\":[],\"TitleCN\":null,\"AbstractTextCN\":null,\"PMCID\":null,\"EPubDate\":\"\",\"PubModel\":\"\",\"JCR\":\"\",\"JCRName\":\"\",\"Score\":null,\"Total\":0}","platform":"Semanticscholar","paperid":null,"PeriodicalName":"[1991] Proceedings of the 34th Midwest Symposium on Circuits and Systems","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/10.1109/MWSCAS.1991.252222","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
Basis vector analyses of back-propagation neural networks
Develops a polynomial basis function approach for modeling BP (backpropagation) neural networks. This method leads directly to a constructive proof of the BP approximation theorem. In addition, the basis vector approach provides a means to synthesize the BP neural network output as a polynomial function. An algorithm for pruning the useless basis vectors is also demonstrated.<>