{"title":"可信度空间上被噪声破坏的样本学习过程的一致收敛率的界限","authors":"Chun-Qin Zhang, Peng Wang","doi":"10.1109/ANTHOLOGY.2013.6784929","DOIUrl":null,"url":null,"abstract":"The bounds on the rate of convergence of learning processes play an important role in statistical learning theory. However, the researches about them presently only focus on probability measure (additive measure) space. And the samples we deal with are supposed to be noise-free. This paper explores the statistical learning theory on credibility space. The theory of consistency of the empirical risk minimization principle when samples are corrupted by noise is established on credibility space; the bounds on the rate of uniform convergence of learning processes about samples corrupted by noise is proposed and proven on the non-additive measure space.","PeriodicalId":203169,"journal":{"name":"IEEE Conference Anthology","volume":"1992 1","pages":"0"},"PeriodicalIF":0.0000,"publicationDate":"1900-01-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":"{\"title\":\"Bounds on the rate of uniform convergence of learning processes about samples corrupted by noise on credibility space\",\"authors\":\"Chun-Qin Zhang, Peng Wang\",\"doi\":\"10.1109/ANTHOLOGY.2013.6784929\",\"DOIUrl\":null,\"url\":null,\"abstract\":\"The bounds on the rate of convergence of learning processes play an important role in statistical learning theory. However, the researches about them presently only focus on probability measure (additive measure) space. And the samples we deal with are supposed to be noise-free. This paper explores the statistical learning theory on credibility space. The theory of consistency of the empirical risk minimization principle when samples are corrupted by noise is established on credibility space; the bounds on the rate of uniform convergence of learning processes about samples corrupted by noise is proposed and proven on the non-additive measure space.\",\"PeriodicalId\":203169,\"journal\":{\"name\":\"IEEE Conference Anthology\",\"volume\":\"1992 1\",\"pages\":\"0\"},\"PeriodicalIF\":0.0000,\"publicationDate\":\"1900-01-01\",\"publicationTypes\":\"Journal Article\",\"fieldsOfStudy\":null,\"isOpenAccess\":false,\"openAccessPdf\":\"\",\"citationCount\":\"0\",\"resultStr\":null,\"platform\":\"Semanticscholar\",\"paperid\":null,\"PeriodicalName\":\"IEEE Conference Anthology\",\"FirstCategoryId\":\"1085\",\"ListUrlMain\":\"https://doi.org/10.1109/ANTHOLOGY.2013.6784929\",\"RegionNum\":0,\"RegionCategory\":null,\"ArticlePicture\":[],\"TitleCN\":null,\"AbstractTextCN\":null,\"PMCID\":null,\"EPubDate\":\"\",\"PubModel\":\"\",\"JCR\":\"\",\"JCRName\":\"\",\"Score\":null,\"Total\":0}","platform":"Semanticscholar","paperid":null,"PeriodicalName":"IEEE Conference Anthology","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/10.1109/ANTHOLOGY.2013.6784929","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
Bounds on the rate of uniform convergence of learning processes about samples corrupted by noise on credibility space
The bounds on the rate of convergence of learning processes play an important role in statistical learning theory. However, the researches about them presently only focus on probability measure (additive measure) space. And the samples we deal with are supposed to be noise-free. This paper explores the statistical learning theory on credibility space. The theory of consistency of the empirical risk minimization principle when samples are corrupted by noise is established on credibility space; the bounds on the rate of uniform convergence of learning processes about samples corrupted by noise is proposed and proven on the non-additive measure space.