{"title":"模式识别分类器的容量控制","authors":"S. Solla","doi":"10.1109/NNSP.1992.253687","DOIUrl":null,"url":null,"abstract":"Achieving good performance in statistical pattern recognition requires matching the capacity of the classifier to the size of the available training set. A classifier with too many adjustable parameters (large capacity) is likely to learn the training set without difficulty, but be unable to generalize properly to new patterns. If the capacity is too small, even the training set might not be learned without appreciable error. There is thus an intermediate, optimal classifier capacity which guarantees the best expected generalization for the given training set size. The method of structural risk minimization provides a theoretical tool for tuning the capacity of the classifier to this optimal match. It is noted that the capacity can be controlled through a variety of methods involving not only the structure of the classifier itself, but also properties of the input space that can be modified through preprocessing, as well as modifications of the learning algorithm which regularize the search for solutions to the problem of learning the training set. Experiments performed on a benchmark problem of handwritten digit recognition are discussed.<<ETX>>","PeriodicalId":438250,"journal":{"name":"Neural Networks for Signal Processing II Proceedings of the 1992 IEEE Workshop","volume":"20 1","pages":"0"},"PeriodicalIF":0.0000,"publicationDate":"1992-08-31","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"10","resultStr":"{\"title\":\"Capacity control in classifiers for pattern recognition\",\"authors\":\"S. Solla\",\"doi\":\"10.1109/NNSP.1992.253687\",\"DOIUrl\":null,\"url\":null,\"abstract\":\"Achieving good performance in statistical pattern recognition requires matching the capacity of the classifier to the size of the available training set. A classifier with too many adjustable parameters (large capacity) is likely to learn the training set without difficulty, but be unable to generalize properly to new patterns. If the capacity is too small, even the training set might not be learned without appreciable error. There is thus an intermediate, optimal classifier capacity which guarantees the best expected generalization for the given training set size. The method of structural risk minimization provides a theoretical tool for tuning the capacity of the classifier to this optimal match. It is noted that the capacity can be controlled through a variety of methods involving not only the structure of the classifier itself, but also properties of the input space that can be modified through preprocessing, as well as modifications of the learning algorithm which regularize the search for solutions to the problem of learning the training set. Experiments performed on a benchmark problem of handwritten digit recognition are discussed.<<ETX>>\",\"PeriodicalId\":438250,\"journal\":{\"name\":\"Neural Networks for Signal Processing II Proceedings of the 1992 IEEE Workshop\",\"volume\":\"20 1\",\"pages\":\"0\"},\"PeriodicalIF\":0.0000,\"publicationDate\":\"1992-08-31\",\"publicationTypes\":\"Journal Article\",\"fieldsOfStudy\":null,\"isOpenAccess\":false,\"openAccessPdf\":\"\",\"citationCount\":\"10\",\"resultStr\":null,\"platform\":\"Semanticscholar\",\"paperid\":null,\"PeriodicalName\":\"Neural Networks for Signal Processing II Proceedings of the 1992 IEEE Workshop\",\"FirstCategoryId\":\"1085\",\"ListUrlMain\":\"https://doi.org/10.1109/NNSP.1992.253687\",\"RegionNum\":0,\"RegionCategory\":null,\"ArticlePicture\":[],\"TitleCN\":null,\"AbstractTextCN\":null,\"PMCID\":null,\"EPubDate\":\"\",\"PubModel\":\"\",\"JCR\":\"\",\"JCRName\":\"\",\"Score\":null,\"Total\":0}","platform":"Semanticscholar","paperid":null,"PeriodicalName":"Neural Networks for Signal Processing II Proceedings of the 1992 IEEE Workshop","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/10.1109/NNSP.1992.253687","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
Capacity control in classifiers for pattern recognition
Achieving good performance in statistical pattern recognition requires matching the capacity of the classifier to the size of the available training set. A classifier with too many adjustable parameters (large capacity) is likely to learn the training set without difficulty, but be unable to generalize properly to new patterns. If the capacity is too small, even the training set might not be learned without appreciable error. There is thus an intermediate, optimal classifier capacity which guarantees the best expected generalization for the given training set size. The method of structural risk minimization provides a theoretical tool for tuning the capacity of the classifier to this optimal match. It is noted that the capacity can be controlled through a variety of methods involving not only the structure of the classifier itself, but also properties of the input space that can be modified through preprocessing, as well as modifications of the learning algorithm which regularize the search for solutions to the problem of learning the training set. Experiments performed on a benchmark problem of handwritten digit recognition are discussed.<>