基于KD-Tree和聚类分析的K近邻非参数回归算法研究

Zheng-Wu Yuan, Yuan-Hui Wang
{"title":"基于KD-Tree和聚类分析的K近邻非参数回归算法研究","authors":"Zheng-Wu Yuan, Yuan-Hui Wang","doi":"10.1109/ICCIS.2012.246","DOIUrl":null,"url":null,"abstract":"Regarding to the limitations of the existing K nearest neighbor non-parametric regression methods, spatial autocorrelation analysis is used to determine the state vector in this paper. In order to improve the speed of searching data, this paper uses the method of clipping samples to reduce data storage and split the sample quickly by KD-Tree. It also reduces the search volume of the nearest neighbor through the pruning principle of KD-Tree, gets the subset by proportional sampling in the KD-Tree subset, and runs K-Means clustering multiple times. Then the optimal K value is selected which can improve the forecast error of the uniform K value on the traditional non-parametric regression. The experimental results show that improved forecasting method is superior to the traditional method.","PeriodicalId":269967,"journal":{"name":"2012 Fourth International Conference on Computational and Information Sciences","volume":"45 1","pages":"0"},"PeriodicalIF":0.0000,"publicationDate":"2012-08-17","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"10","resultStr":"{\"title\":\"Research on K Nearest Neighbor Non-parametric Regression Algorithm Based on KD-Tree and Clustering Analysis\",\"authors\":\"Zheng-Wu Yuan, Yuan-Hui Wang\",\"doi\":\"10.1109/ICCIS.2012.246\",\"DOIUrl\":null,\"url\":null,\"abstract\":\"Regarding to the limitations of the existing K nearest neighbor non-parametric regression methods, spatial autocorrelation analysis is used to determine the state vector in this paper. In order to improve the speed of searching data, this paper uses the method of clipping samples to reduce data storage and split the sample quickly by KD-Tree. It also reduces the search volume of the nearest neighbor through the pruning principle of KD-Tree, gets the subset by proportional sampling in the KD-Tree subset, and runs K-Means clustering multiple times. Then the optimal K value is selected which can improve the forecast error of the uniform K value on the traditional non-parametric regression. The experimental results show that improved forecasting method is superior to the traditional method.\",\"PeriodicalId\":269967,\"journal\":{\"name\":\"2012 Fourth International Conference on Computational and Information Sciences\",\"volume\":\"45 1\",\"pages\":\"0\"},\"PeriodicalIF\":0.0000,\"publicationDate\":\"2012-08-17\",\"publicationTypes\":\"Journal Article\",\"fieldsOfStudy\":null,\"isOpenAccess\":false,\"openAccessPdf\":\"\",\"citationCount\":\"10\",\"resultStr\":null,\"platform\":\"Semanticscholar\",\"paperid\":null,\"PeriodicalName\":\"2012 Fourth International Conference on Computational and Information Sciences\",\"FirstCategoryId\":\"1085\",\"ListUrlMain\":\"https://doi.org/10.1109/ICCIS.2012.246\",\"RegionNum\":0,\"RegionCategory\":null,\"ArticlePicture\":[],\"TitleCN\":null,\"AbstractTextCN\":null,\"PMCID\":null,\"EPubDate\":\"\",\"PubModel\":\"\",\"JCR\":\"\",\"JCRName\":\"\",\"Score\":null,\"Total\":0}","platform":"Semanticscholar","paperid":null,"PeriodicalName":"2012 Fourth International Conference on Computational and Information Sciences","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/10.1109/ICCIS.2012.246","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
引用次数: 10

摘要

针对现有K近邻非参数回归方法的局限性,本文采用空间自相关分析来确定状态向量。为了提高搜索数据的速度,本文采用样本裁剪的方法减少数据存储,并采用KD-Tree快速分割样本。利用KD-Tree的剪枝原理减少最近邻的搜索量,在KD-Tree子集中按比例采样得到子集,并多次运行K-Means聚类。然后选取最优K值,改善了传统非参数回归中均匀K值的预测误差。实验结果表明,改进的预测方法优于传统的预测方法。
本文章由计算机程序翻译,如有差异,请以英文原文为准。
Research on K Nearest Neighbor Non-parametric Regression Algorithm Based on KD-Tree and Clustering Analysis
Regarding to the limitations of the existing K nearest neighbor non-parametric regression methods, spatial autocorrelation analysis is used to determine the state vector in this paper. In order to improve the speed of searching data, this paper uses the method of clipping samples to reduce data storage and split the sample quickly by KD-Tree. It also reduces the search volume of the nearest neighbor through the pruning principle of KD-Tree, gets the subset by proportional sampling in the KD-Tree subset, and runs K-Means clustering multiple times. Then the optimal K value is selected which can improve the forecast error of the uniform K value on the traditional non-parametric regression. The experimental results show that improved forecasting method is superior to the traditional method.
求助全文
通过发布文献求助,成功后即可免费获取论文全文。 去求助
来源期刊
自引率
0.00%
发文量
0
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
确定
请完成安全验证×
copy
已复制链接
快去分享给好友吧!
我知道了
右上角分享
点击右上角分享
0
联系我们:info@booksci.cn Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。 Copyright © 2023 布克学术 All rights reserved.
京ICP备2023020795号-1
ghs 京公网安备 11010802042870号
Book学术文献互助
Book学术文献互助群
群 号:481959085
Book学术官方微信