基于多平均值的伪近邻分类器

Dapeng Li, Jing Guo
{"title":"基于多平均值的伪近邻分类器","authors":"Dapeng Li, Jing Guo","doi":"10.3233/aic-230312","DOIUrl":null,"url":null,"abstract":"Conventional k nearest neighbor (KNN) rule is a simple yet effective method for classification, but its classification performance is easily degraded in the case of small size training samples with existing outliers. To address this issue, A multi-average based pseudo nearest neighbor classifier (MAPNN) rule is proposed. In the proposed MAPNN rule, k ( k − 1 ) / 2 ( k > 1) local mean vectors of each class are obtained by taking the average of two points randomly from k nearest neighbors in every category, and then k pseudo nearest neighbors are chosen from k ( k − 1 ) / 2 local mean neighbors of every class to determine the category of a query point. The selected k pseudo nearest neighbors can reduce the negative impact of outliers in some degree. Extensive experiments are carried out on twenty-one numerical real data sets and four artificial data sets by comparing MAPNN to other five KNN-based methods. The experimental results demonstrate that the proposed MAPNN is effective for classification task and achieves better classification results in the small-size samples cases comparing to five relative KNN-based classifiers.","PeriodicalId":505412,"journal":{"name":"AI Communications","volume":"44 5","pages":""},"PeriodicalIF":0.0000,"publicationDate":"2024-03-28","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":"{\"title\":\"A multi-average based pseudo nearest neighbor classifier\",\"authors\":\"Dapeng Li, Jing Guo\",\"doi\":\"10.3233/aic-230312\",\"DOIUrl\":null,\"url\":null,\"abstract\":\"Conventional k nearest neighbor (KNN) rule is a simple yet effective method for classification, but its classification performance is easily degraded in the case of small size training samples with existing outliers. To address this issue, A multi-average based pseudo nearest neighbor classifier (MAPNN) rule is proposed. In the proposed MAPNN rule, k ( k − 1 ) / 2 ( k > 1) local mean vectors of each class are obtained by taking the average of two points randomly from k nearest neighbors in every category, and then k pseudo nearest neighbors are chosen from k ( k − 1 ) / 2 local mean neighbors of every class to determine the category of a query point. The selected k pseudo nearest neighbors can reduce the negative impact of outliers in some degree. Extensive experiments are carried out on twenty-one numerical real data sets and four artificial data sets by comparing MAPNN to other five KNN-based methods. The experimental results demonstrate that the proposed MAPNN is effective for classification task and achieves better classification results in the small-size samples cases comparing to five relative KNN-based classifiers.\",\"PeriodicalId\":505412,\"journal\":{\"name\":\"AI Communications\",\"volume\":\"44 5\",\"pages\":\"\"},\"PeriodicalIF\":0.0000,\"publicationDate\":\"2024-03-28\",\"publicationTypes\":\"Journal Article\",\"fieldsOfStudy\":null,\"isOpenAccess\":false,\"openAccessPdf\":\"\",\"citationCount\":\"0\",\"resultStr\":null,\"platform\":\"Semanticscholar\",\"paperid\":null,\"PeriodicalName\":\"AI Communications\",\"FirstCategoryId\":\"1085\",\"ListUrlMain\":\"https://doi.org/10.3233/aic-230312\",\"RegionNum\":0,\"RegionCategory\":null,\"ArticlePicture\":[],\"TitleCN\":null,\"AbstractTextCN\":null,\"PMCID\":null,\"EPubDate\":\"\",\"PubModel\":\"\",\"JCR\":\"\",\"JCRName\":\"\",\"Score\":null,\"Total\":0}","platform":"Semanticscholar","paperid":null,"PeriodicalName":"AI Communications","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/10.3233/aic-230312","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
引用次数: 0

摘要

传统的 k 近邻(KNN)规则是一种简单而有效的分类方法,但在训练样本较小且存在异常值的情况下,其分类性能很容易下降。为了解决这个问题,我们提出了一种基于多平均值的伪近邻分类器(MAPNN)规则。在所提出的 MAPNN 规则中,每个类别的 k ( k - 1 ) / 2 ( k > 1 ) 个局部均值向量是通过从每个类别的 k 个近邻中随机取两个点的平均值得到的,然后从每个类别的 k ( k - 1 ) / 2 个局部均值近邻中选择 k 个伪近邻来确定查询点的类别。选出的 k 个伪近邻可以在一定程度上减少异常值的负面影响。通过将 MAPNN 与其他五种基于 KNN 的方法进行比较,在 21 个数值真实数据集和 4 个人工数据集上进行了广泛的实验。实验结果表明,与基于 KNN 的五种分类器相比,所提出的 MAPNN 能有效地完成分类任务,并在小样本情况下取得更好的分类结果。
本文章由计算机程序翻译,如有差异,请以英文原文为准。
A multi-average based pseudo nearest neighbor classifier
Conventional k nearest neighbor (KNN) rule is a simple yet effective method for classification, but its classification performance is easily degraded in the case of small size training samples with existing outliers. To address this issue, A multi-average based pseudo nearest neighbor classifier (MAPNN) rule is proposed. In the proposed MAPNN rule, k ( k − 1 ) / 2 ( k > 1) local mean vectors of each class are obtained by taking the average of two points randomly from k nearest neighbors in every category, and then k pseudo nearest neighbors are chosen from k ( k − 1 ) / 2 local mean neighbors of every class to determine the category of a query point. The selected k pseudo nearest neighbors can reduce the negative impact of outliers in some degree. Extensive experiments are carried out on twenty-one numerical real data sets and four artificial data sets by comparing MAPNN to other five KNN-based methods. The experimental results demonstrate that the proposed MAPNN is effective for classification task and achieves better classification results in the small-size samples cases comparing to five relative KNN-based classifiers.
求助全文
通过发布文献求助,成功后即可免费获取论文全文。 去求助
来源期刊
自引率
0.00%
发文量
0
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
确定
请完成安全验证×
copy
已复制链接
快去分享给好友吧!
我知道了
右上角分享
点击右上角分享
0
联系我们:info@booksci.cn Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。 Copyright © 2023 布克学术 All rights reserved.
京ICP备2023020795号-1
ghs 京公网安备 11010802042870号
Book学术文献互助
Book学术文献互助群
群 号:604180095
Book学术官方微信