{"title":"可能性空间上无界非负函数集的风险界","authors":"Peng Wang, Chun-Qin Zhang","doi":"10.1109/ICMLC.2011.6016825","DOIUrl":null,"url":null,"abstract":"Statistical learning theory on probability space is an important part of Machine Learning. Based on the key theorem, the bounds of uniform convergence have significant meaning. These bounds determine generalization ability of the learning machines utilizing the empirical risk minimization induction principle. In this paper, the bounds on the risk for sets of unbounded nonnegative functions on possibility space are discussed, and the rate of uniform convergence is estimated.","PeriodicalId":228516,"journal":{"name":"2011 International Conference on Machine Learning and Cybernetics","volume":"78 1","pages":"0"},"PeriodicalIF":0.0000,"publicationDate":"2011-07-10","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"1","resultStr":"{\"title\":\"The bounds on the risk for sets of unbounded nonnegative functions on possibility space\",\"authors\":\"Peng Wang, Chun-Qin Zhang\",\"doi\":\"10.1109/ICMLC.2011.6016825\",\"DOIUrl\":null,\"url\":null,\"abstract\":\"Statistical learning theory on probability space is an important part of Machine Learning. Based on the key theorem, the bounds of uniform convergence have significant meaning. These bounds determine generalization ability of the learning machines utilizing the empirical risk minimization induction principle. In this paper, the bounds on the risk for sets of unbounded nonnegative functions on possibility space are discussed, and the rate of uniform convergence is estimated.\",\"PeriodicalId\":228516,\"journal\":{\"name\":\"2011 International Conference on Machine Learning and Cybernetics\",\"volume\":\"78 1\",\"pages\":\"0\"},\"PeriodicalIF\":0.0000,\"publicationDate\":\"2011-07-10\",\"publicationTypes\":\"Journal Article\",\"fieldsOfStudy\":null,\"isOpenAccess\":false,\"openAccessPdf\":\"\",\"citationCount\":\"1\",\"resultStr\":null,\"platform\":\"Semanticscholar\",\"paperid\":null,\"PeriodicalName\":\"2011 International Conference on Machine Learning and Cybernetics\",\"FirstCategoryId\":\"1085\",\"ListUrlMain\":\"https://doi.org/10.1109/ICMLC.2011.6016825\",\"RegionNum\":0,\"RegionCategory\":null,\"ArticlePicture\":[],\"TitleCN\":null,\"AbstractTextCN\":null,\"PMCID\":null,\"EPubDate\":\"\",\"PubModel\":\"\",\"JCR\":\"\",\"JCRName\":\"\",\"Score\":null,\"Total\":0}","platform":"Semanticscholar","paperid":null,"PeriodicalName":"2011 International Conference on Machine Learning and Cybernetics","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/10.1109/ICMLC.2011.6016825","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
The bounds on the risk for sets of unbounded nonnegative functions on possibility space
Statistical learning theory on probability space is an important part of Machine Learning. Based on the key theorem, the bounds of uniform convergence have significant meaning. These bounds determine generalization ability of the learning machines utilizing the empirical risk minimization induction principle. In this paper, the bounds on the risk for sets of unbounded nonnegative functions on possibility space are discussed, and the rate of uniform convergence is estimated.