{"title":"主成分分析和高斯噪声在MLP神经网络训练中提高了小数据集和不平衡数据集问题的泛化能力","authors":"Icamaan B. Viegas da Silva, P. Adeodato","doi":"10.1109/IJCNN.2011.6033567","DOIUrl":null,"url":null,"abstract":"Machine learning approaches have been successfully applied for automatic decision support in several domains. The quality of these systems, however, degrades severely in classification problems with small and unbalanced data sets for knowledge acquisition. Inherent to several real-world problems, data sets with these characteristics are the reality to be tackled by learning algorithms, but the small amount of data affects the classifiers' generalization power while the imbalance in class distribution makes the classifiers biased towards the larger classes. Previous work had addressed these data constraints with the addition of Gaussian noise to the input patterns' variables during the iterative training process of a MultiLayer perceptron (MLP) neural network (NN). This paper improves the quality of such classifier by decorrelating the input variables via a Principal Component Analysis (PCA) transformation of the original input space before applying additive Gaussian noise to each transformed variable for each input pattern. PCA transformation prevents the conflicting effect of adding decorrelated noise to correlated variables, an effect which increases with the noise level. Three public data sets from a well-known benchmark (Proben1) were used to validate the proposed approach. Experimental results indicate that the proposed methodology improves the performance of the previous approach being statistically better than the traditional training method (95% confidence) in further experimental set-ups.","PeriodicalId":415833,"journal":{"name":"The 2011 International Joint Conference on Neural Networks","volume":"1 1","pages":"0"},"PeriodicalIF":0.0000,"publicationDate":"2011-10-03","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"22","resultStr":"{\"title\":\"PCA and Gaussian noise in MLP neural network training improve generalization in problems with small and unbalanced data sets\",\"authors\":\"Icamaan B. Viegas da Silva, P. Adeodato\",\"doi\":\"10.1109/IJCNN.2011.6033567\",\"DOIUrl\":null,\"url\":null,\"abstract\":\"Machine learning approaches have been successfully applied for automatic decision support in several domains. The quality of these systems, however, degrades severely in classification problems with small and unbalanced data sets for knowledge acquisition. Inherent to several real-world problems, data sets with these characteristics are the reality to be tackled by learning algorithms, but the small amount of data affects the classifiers' generalization power while the imbalance in class distribution makes the classifiers biased towards the larger classes. Previous work had addressed these data constraints with the addition of Gaussian noise to the input patterns' variables during the iterative training process of a MultiLayer perceptron (MLP) neural network (NN). This paper improves the quality of such classifier by decorrelating the input variables via a Principal Component Analysis (PCA) transformation of the original input space before applying additive Gaussian noise to each transformed variable for each input pattern. PCA transformation prevents the conflicting effect of adding decorrelated noise to correlated variables, an effect which increases with the noise level. Three public data sets from a well-known benchmark (Proben1) were used to validate the proposed approach. Experimental results indicate that the proposed methodology improves the performance of the previous approach being statistically better than the traditional training method (95% confidence) in further experimental set-ups.\",\"PeriodicalId\":415833,\"journal\":{\"name\":\"The 2011 International Joint Conference on Neural Networks\",\"volume\":\"1 1\",\"pages\":\"0\"},\"PeriodicalIF\":0.0000,\"publicationDate\":\"2011-10-03\",\"publicationTypes\":\"Journal Article\",\"fieldsOfStudy\":null,\"isOpenAccess\":false,\"openAccessPdf\":\"\",\"citationCount\":\"22\",\"resultStr\":null,\"platform\":\"Semanticscholar\",\"paperid\":null,\"PeriodicalName\":\"The 2011 International Joint Conference on Neural Networks\",\"FirstCategoryId\":\"1085\",\"ListUrlMain\":\"https://doi.org/10.1109/IJCNN.2011.6033567\",\"RegionNum\":0,\"RegionCategory\":null,\"ArticlePicture\":[],\"TitleCN\":null,\"AbstractTextCN\":null,\"PMCID\":null,\"EPubDate\":\"\",\"PubModel\":\"\",\"JCR\":\"\",\"JCRName\":\"\",\"Score\":null,\"Total\":0}","platform":"Semanticscholar","paperid":null,"PeriodicalName":"The 2011 International Joint Conference on Neural Networks","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/10.1109/IJCNN.2011.6033567","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
PCA and Gaussian noise in MLP neural network training improve generalization in problems with small and unbalanced data sets
Machine learning approaches have been successfully applied for automatic decision support in several domains. The quality of these systems, however, degrades severely in classification problems with small and unbalanced data sets for knowledge acquisition. Inherent to several real-world problems, data sets with these characteristics are the reality to be tackled by learning algorithms, but the small amount of data affects the classifiers' generalization power while the imbalance in class distribution makes the classifiers biased towards the larger classes. Previous work had addressed these data constraints with the addition of Gaussian noise to the input patterns' variables during the iterative training process of a MultiLayer perceptron (MLP) neural network (NN). This paper improves the quality of such classifier by decorrelating the input variables via a Principal Component Analysis (PCA) transformation of the original input space before applying additive Gaussian noise to each transformed variable for each input pattern. PCA transformation prevents the conflicting effect of adding decorrelated noise to correlated variables, an effect which increases with the noise level. Three public data sets from a well-known benchmark (Proben1) were used to validate the proposed approach. Experimental results indicate that the proposed methodology improves the performance of the previous approach being statistically better than the traditional training method (95% confidence) in further experimental set-ups.