{"title":"MODIFIED CORRELATION WEIGHT K-NEAREST NEIGHBOR CLASSIFIER USING TRAINING DATASET CLEANING METHOD","authors":"E. K. D. Kette","doi":"10.5614/itb.ijp.2021.32.2.5","DOIUrl":null,"url":null,"abstract":"In pattern recognition, the k-Nearest Neighbor (kNN) algorithm is the simplest non-parametric algorithm. Due to its simplicity, the model cases and the quality of the training data itself usually influence kNN algorithm classification performance. Therefore, this article proposes a sparse correlation weight model, combined with the Training Data Set Cleaning (TDC) method by Classification Ability Ranking (CAR) called the CAR classification method based on Coefficient-Weighted kNN (CAR-CWKNN) to improve kNN classifier performance. Correlation weight in Sparse Representation (SR) has been proven can increase classification accuracy. The SR can show the 'neighborhood' structure of the data, which is why it is very suitable for classification based on the Nearest Neighbor. The Classification Ability (CA) function is applied to classify the best training sample data based on rank in the cleaning stage. The Leave One Out (LV1) concept in the CA works by cleaning data that is considered likely to have the wrong classification results from the original training data, thereby reducing the influence of the training sample data quality on the kNN classification performance. The results of experiments with four public UCI data sets related to classification problems show that the CAR-CWKNN method provides better performance in terms of accuracy.","PeriodicalId":13535,"journal":{"name":"Indonesian Journal of Physics","volume":null,"pages":null},"PeriodicalIF":0.0000,"publicationDate":"2021-12-28","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"Indonesian Journal of Physics","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/10.5614/itb.ijp.2021.32.2.5","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
引用次数: 0
Abstract
In pattern recognition, the k-Nearest Neighbor (kNN) algorithm is the simplest non-parametric algorithm. Due to its simplicity, the model cases and the quality of the training data itself usually influence kNN algorithm classification performance. Therefore, this article proposes a sparse correlation weight model, combined with the Training Data Set Cleaning (TDC) method by Classification Ability Ranking (CAR) called the CAR classification method based on Coefficient-Weighted kNN (CAR-CWKNN) to improve kNN classifier performance. Correlation weight in Sparse Representation (SR) has been proven can increase classification accuracy. The SR can show the 'neighborhood' structure of the data, which is why it is very suitable for classification based on the Nearest Neighbor. The Classification Ability (CA) function is applied to classify the best training sample data based on rank in the cleaning stage. The Leave One Out (LV1) concept in the CA works by cleaning data that is considered likely to have the wrong classification results from the original training data, thereby reducing the influence of the training sample data quality on the kNN classification performance. The results of experiments with four public UCI data sets related to classification problems show that the CAR-CWKNN method provides better performance in terms of accuracy.
在模式识别中,k近邻(kNN)算法是最简单的非参数算法。由于其简单性,模型案例和训练数据本身的质量通常会影响kNN算法的分类性能。因此,本文提出了一种稀疏相关权模型,结合基于分类能力排序(CAR)的训练数据集清洗(TDC)方法,称为基于系数加权kNN的CAR分类方法(CAR- cwknn),以提高kNN分类器的性能。稀疏表示中的相关权重被证明可以提高分类精度。SR可以显示数据的“邻域”结构,这就是为什么它非常适合基于最近邻的分类。在清洗阶段,使用分类能力(CA)函数对基于秩的最佳训练样本数据进行分类。CA中的Leave One Out (LV1)概念通过从原始训练数据中清除被认为可能具有错误分类结果的数据,从而减少训练样本数据质量对kNN分类性能的影响。在4个与分类问题相关的公共UCI数据集上进行的实验结果表明,CAR-CWKNN方法在准确率方面具有更好的性能。