{"title":"集成方法","authors":"Jun Liang","doi":"10.1145/3290420.3290454","DOIUrl":null,"url":null,"abstract":"This paper gives an ensemble method called EKNN-RF. Its base classifiers use an enhanced KNN algorithm where an optimal nearest neighbor number and a distance function on a validation set are obtained to make these parameters better reflect the distribution of real data. The feature set of each base classifier is obtained through bootstrap sampling from original feature set, and make the features with higher importance have a better weight. Then the training set of each base classifier is also obtained by bootstrap sampling based original training set and the newly generated feature set. Finally, each base classifier votes to determine the classification result. Experimental results show that compared with Adaboost, Naive Bayes, RandomForest, DCT-KNN [1], LMKNN+DWKNN [2], W-KNN [3], dwh-KNN [4] and LI-KNN [5], the ensemble method EKNN-RF has certain advantages and higher classification accuracy on some datasets.","PeriodicalId":259201,"journal":{"name":"International Conference on Critical Infrastructure Protection","volume":"1 1","pages":"0"},"PeriodicalIF":0.0000,"publicationDate":"2018-11-02","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"10","resultStr":"{\"title\":\"An ensemble method\",\"authors\":\"Jun Liang\",\"doi\":\"10.1145/3290420.3290454\",\"DOIUrl\":null,\"url\":null,\"abstract\":\"This paper gives an ensemble method called EKNN-RF. Its base classifiers use an enhanced KNN algorithm where an optimal nearest neighbor number and a distance function on a validation set are obtained to make these parameters better reflect the distribution of real data. The feature set of each base classifier is obtained through bootstrap sampling from original feature set, and make the features with higher importance have a better weight. Then the training set of each base classifier is also obtained by bootstrap sampling based original training set and the newly generated feature set. Finally, each base classifier votes to determine the classification result. Experimental results show that compared with Adaboost, Naive Bayes, RandomForest, DCT-KNN [1], LMKNN+DWKNN [2], W-KNN [3], dwh-KNN [4] and LI-KNN [5], the ensemble method EKNN-RF has certain advantages and higher classification accuracy on some datasets.\",\"PeriodicalId\":259201,\"journal\":{\"name\":\"International Conference on Critical Infrastructure Protection\",\"volume\":\"1 1\",\"pages\":\"0\"},\"PeriodicalIF\":0.0000,\"publicationDate\":\"2018-11-02\",\"publicationTypes\":\"Journal Article\",\"fieldsOfStudy\":null,\"isOpenAccess\":false,\"openAccessPdf\":\"\",\"citationCount\":\"10\",\"resultStr\":null,\"platform\":\"Semanticscholar\",\"paperid\":null,\"PeriodicalName\":\"International Conference on Critical Infrastructure Protection\",\"FirstCategoryId\":\"1085\",\"ListUrlMain\":\"https://doi.org/10.1145/3290420.3290454\",\"RegionNum\":0,\"RegionCategory\":null,\"ArticlePicture\":[],\"TitleCN\":null,\"AbstractTextCN\":null,\"PMCID\":null,\"EPubDate\":\"\",\"PubModel\":\"\",\"JCR\":\"\",\"JCRName\":\"\",\"Score\":null,\"Total\":0}","platform":"Semanticscholar","paperid":null,"PeriodicalName":"International Conference on Critical Infrastructure Protection","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/10.1145/3290420.3290454","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
This paper gives an ensemble method called EKNN-RF. Its base classifiers use an enhanced KNN algorithm where an optimal nearest neighbor number and a distance function on a validation set are obtained to make these parameters better reflect the distribution of real data. The feature set of each base classifier is obtained through bootstrap sampling from original feature set, and make the features with higher importance have a better weight. Then the training set of each base classifier is also obtained by bootstrap sampling based original training set and the newly generated feature set. Finally, each base classifier votes to determine the classification result. Experimental results show that compared with Adaboost, Naive Bayes, RandomForest, DCT-KNN [1], LMKNN+DWKNN [2], W-KNN [3], dwh-KNN [4] and LI-KNN [5], the ensemble method EKNN-RF has certain advantages and higher classification accuracy on some datasets.