{"title":"基于神经复杂度的前馈神经网络节点修剪算法","authors":"Zhaozhao Zhang, J. Qiao","doi":"10.1109/ICICIP.2010.5564272","DOIUrl":null,"url":null,"abstract":"In this paper, a hidden node pruning algorithm based on the neural complexity is proposed, the entropy of neural network can be calculated by the standard covariance matrix of the neural network's connection matrix in the training stage, and the neural complexity can be acquired. In ensuring the information processing capacity of neural network is not reduced, select and delete the least important hidden node, and the simpler neural network architecture is achieved. It is not necessary to train the cost function of the neural network to a local minimal, and the pre-processing neural network weights is avoided before neural network architecture adjustment. The simulation results of the non-linear function approximation shows that the performance of the approximation is ensured and at the same time a simple architecture of neural networks can be achieved.","PeriodicalId":152024,"journal":{"name":"2010 International Conference on Intelligent Control and Information Processing","volume":"157 1","pages":"0"},"PeriodicalIF":0.0000,"publicationDate":"2010-09-09","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"23","resultStr":"{\"title\":\"A node pruning algorithm for feedforward neural network based on neural complexity\",\"authors\":\"Zhaozhao Zhang, J. Qiao\",\"doi\":\"10.1109/ICICIP.2010.5564272\",\"DOIUrl\":null,\"url\":null,\"abstract\":\"In this paper, a hidden node pruning algorithm based on the neural complexity is proposed, the entropy of neural network can be calculated by the standard covariance matrix of the neural network's connection matrix in the training stage, and the neural complexity can be acquired. In ensuring the information processing capacity of neural network is not reduced, select and delete the least important hidden node, and the simpler neural network architecture is achieved. It is not necessary to train the cost function of the neural network to a local minimal, and the pre-processing neural network weights is avoided before neural network architecture adjustment. The simulation results of the non-linear function approximation shows that the performance of the approximation is ensured and at the same time a simple architecture of neural networks can be achieved.\",\"PeriodicalId\":152024,\"journal\":{\"name\":\"2010 International Conference on Intelligent Control and Information Processing\",\"volume\":\"157 1\",\"pages\":\"0\"},\"PeriodicalIF\":0.0000,\"publicationDate\":\"2010-09-09\",\"publicationTypes\":\"Journal Article\",\"fieldsOfStudy\":null,\"isOpenAccess\":false,\"openAccessPdf\":\"\",\"citationCount\":\"23\",\"resultStr\":null,\"platform\":\"Semanticscholar\",\"paperid\":null,\"PeriodicalName\":\"2010 International Conference on Intelligent Control and Information Processing\",\"FirstCategoryId\":\"1085\",\"ListUrlMain\":\"https://doi.org/10.1109/ICICIP.2010.5564272\",\"RegionNum\":0,\"RegionCategory\":null,\"ArticlePicture\":[],\"TitleCN\":null,\"AbstractTextCN\":null,\"PMCID\":null,\"EPubDate\":\"\",\"PubModel\":\"\",\"JCR\":\"\",\"JCRName\":\"\",\"Score\":null,\"Total\":0}","platform":"Semanticscholar","paperid":null,"PeriodicalName":"2010 International Conference on Intelligent Control and Information Processing","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/10.1109/ICICIP.2010.5564272","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
A node pruning algorithm for feedforward neural network based on neural complexity
In this paper, a hidden node pruning algorithm based on the neural complexity is proposed, the entropy of neural network can be calculated by the standard covariance matrix of the neural network's connection matrix in the training stage, and the neural complexity can be acquired. In ensuring the information processing capacity of neural network is not reduced, select and delete the least important hidden node, and the simpler neural network architecture is achieved. It is not necessary to train the cost function of the neural network to a local minimal, and the pre-processing neural network weights is avoided before neural network architecture adjustment. The simulation results of the non-linear function approximation shows that the performance of the approximation is ensured and at the same time a simple architecture of neural networks can be achieved.