{"title":"四种不确定抽样方法在分类上优于随机抽样方法","authors":"Zhang Guochen","doi":"10.1109/ICAIE53562.2021.00051","DOIUrl":null,"url":null,"abstract":"Active learning has been widely used because it can automatically select the unlabeled samples with the largest amount of information for manual labeling. Therefore, it could solve the problem of knowledge bottleneck. Selective sampling belongs to the active learning approach, reducing labeling costs to supplement training data by requiring labels to provide only the most informative, unlabeled examples. This additional information is added to an original, stochastically selected training set in the expectation of improving the performance of generalization of the learning machine[12]. Uncertainty sampling belonging to selective sampling is one of the key techniques of active learning, which uses a classifier to identify the least reliable unlabeled samples[1]. In addition, active learning is a method applied to classifier. In this paper, four methods under uncertainty sampling are used: Least Confidence Sampling, Margin of Confidence Sampling, Ratio of Confidence Sampling, and Entropy-based Sampling. According to the results, the four methods of uncertain sampling have higher accuracy than the random sampling method.","PeriodicalId":285278,"journal":{"name":"2021 2nd International Conference on Artificial Intelligence and Education (ICAIE)","volume":"101 1","pages":"0"},"PeriodicalIF":0.0000,"publicationDate":"2021-06-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"1","resultStr":"{\"title\":\"Four Uncertain Sampling Methods are Superior to Random Sampling Method in Classification\",\"authors\":\"Zhang Guochen\",\"doi\":\"10.1109/ICAIE53562.2021.00051\",\"DOIUrl\":null,\"url\":null,\"abstract\":\"Active learning has been widely used because it can automatically select the unlabeled samples with the largest amount of information for manual labeling. Therefore, it could solve the problem of knowledge bottleneck. Selective sampling belongs to the active learning approach, reducing labeling costs to supplement training data by requiring labels to provide only the most informative, unlabeled examples. This additional information is added to an original, stochastically selected training set in the expectation of improving the performance of generalization of the learning machine[12]. Uncertainty sampling belonging to selective sampling is one of the key techniques of active learning, which uses a classifier to identify the least reliable unlabeled samples[1]. In addition, active learning is a method applied to classifier. In this paper, four methods under uncertainty sampling are used: Least Confidence Sampling, Margin of Confidence Sampling, Ratio of Confidence Sampling, and Entropy-based Sampling. According to the results, the four methods of uncertain sampling have higher accuracy than the random sampling method.\",\"PeriodicalId\":285278,\"journal\":{\"name\":\"2021 2nd International Conference on Artificial Intelligence and Education (ICAIE)\",\"volume\":\"101 1\",\"pages\":\"0\"},\"PeriodicalIF\":0.0000,\"publicationDate\":\"2021-06-01\",\"publicationTypes\":\"Journal Article\",\"fieldsOfStudy\":null,\"isOpenAccess\":false,\"openAccessPdf\":\"\",\"citationCount\":\"1\",\"resultStr\":null,\"platform\":\"Semanticscholar\",\"paperid\":null,\"PeriodicalName\":\"2021 2nd International Conference on Artificial Intelligence and Education (ICAIE)\",\"FirstCategoryId\":\"1085\",\"ListUrlMain\":\"https://doi.org/10.1109/ICAIE53562.2021.00051\",\"RegionNum\":0,\"RegionCategory\":null,\"ArticlePicture\":[],\"TitleCN\":null,\"AbstractTextCN\":null,\"PMCID\":null,\"EPubDate\":\"\",\"PubModel\":\"\",\"JCR\":\"\",\"JCRName\":\"\",\"Score\":null,\"Total\":0}","platform":"Semanticscholar","paperid":null,"PeriodicalName":"2021 2nd International Conference on Artificial Intelligence and Education (ICAIE)","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/10.1109/ICAIE53562.2021.00051","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
Four Uncertain Sampling Methods are Superior to Random Sampling Method in Classification
Active learning has been widely used because it can automatically select the unlabeled samples with the largest amount of information for manual labeling. Therefore, it could solve the problem of knowledge bottleneck. Selective sampling belongs to the active learning approach, reducing labeling costs to supplement training data by requiring labels to provide only the most informative, unlabeled examples. This additional information is added to an original, stochastically selected training set in the expectation of improving the performance of generalization of the learning machine[12]. Uncertainty sampling belonging to selective sampling is one of the key techniques of active learning, which uses a classifier to identify the least reliable unlabeled samples[1]. In addition, active learning is a method applied to classifier. In this paper, four methods under uncertainty sampling are used: Least Confidence Sampling, Margin of Confidence Sampling, Ratio of Confidence Sampling, and Entropy-based Sampling. According to the results, the four methods of uncertain sampling have higher accuracy than the random sampling method.