{"title":"一种新的支持向量分类样本约简方法","authors":"Ling Wang, Meiling Sui, Qin Li, Haijun Xiao","doi":"10.1109/APSCC.2012.57","DOIUrl":null,"url":null,"abstract":"As a powerful tool in machine learning, Support Vector Machine(SVM) also suffers from expensive computational cost in the training phase due to the large number of original training samples. To overcome this problem, this paper presents a new method based on a two steps of sample reduction to reduce training samples. This algorithm includes cluster detection by Fuzzy C-Means Clustering (FCM) Cluster and sample reduction by Multivariate Gaussian Distribution (MGD). In its implementation, the FCM algorithm is used to cluster the original training samples, and then the MGD is used to reduce the training samples by choosing the only boundary samples for the next training. Experiments show that the algorithm accelerates the training speed without the decrease of classification accuracy.","PeriodicalId":256842,"journal":{"name":"2012 IEEE Asia-Pacific Services Computing Conference","volume":"53 1","pages":"0"},"PeriodicalIF":0.0000,"publicationDate":"2012-12-06","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"3","resultStr":"{\"title\":\"A New Method of Sample Reduction for Support Vector Classification\",\"authors\":\"Ling Wang, Meiling Sui, Qin Li, Haijun Xiao\",\"doi\":\"10.1109/APSCC.2012.57\",\"DOIUrl\":null,\"url\":null,\"abstract\":\"As a powerful tool in machine learning, Support Vector Machine(SVM) also suffers from expensive computational cost in the training phase due to the large number of original training samples. To overcome this problem, this paper presents a new method based on a two steps of sample reduction to reduce training samples. This algorithm includes cluster detection by Fuzzy C-Means Clustering (FCM) Cluster and sample reduction by Multivariate Gaussian Distribution (MGD). In its implementation, the FCM algorithm is used to cluster the original training samples, and then the MGD is used to reduce the training samples by choosing the only boundary samples for the next training. Experiments show that the algorithm accelerates the training speed without the decrease of classification accuracy.\",\"PeriodicalId\":256842,\"journal\":{\"name\":\"2012 IEEE Asia-Pacific Services Computing Conference\",\"volume\":\"53 1\",\"pages\":\"0\"},\"PeriodicalIF\":0.0000,\"publicationDate\":\"2012-12-06\",\"publicationTypes\":\"Journal Article\",\"fieldsOfStudy\":null,\"isOpenAccess\":false,\"openAccessPdf\":\"\",\"citationCount\":\"3\",\"resultStr\":null,\"platform\":\"Semanticscholar\",\"paperid\":null,\"PeriodicalName\":\"2012 IEEE Asia-Pacific Services Computing Conference\",\"FirstCategoryId\":\"1085\",\"ListUrlMain\":\"https://doi.org/10.1109/APSCC.2012.57\",\"RegionNum\":0,\"RegionCategory\":null,\"ArticlePicture\":[],\"TitleCN\":null,\"AbstractTextCN\":null,\"PMCID\":null,\"EPubDate\":\"\",\"PubModel\":\"\",\"JCR\":\"\",\"JCRName\":\"\",\"Score\":null,\"Total\":0}","platform":"Semanticscholar","paperid":null,"PeriodicalName":"2012 IEEE Asia-Pacific Services Computing Conference","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/10.1109/APSCC.2012.57","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
A New Method of Sample Reduction for Support Vector Classification
As a powerful tool in machine learning, Support Vector Machine(SVM) also suffers from expensive computational cost in the training phase due to the large number of original training samples. To overcome this problem, this paper presents a new method based on a two steps of sample reduction to reduce training samples. This algorithm includes cluster detection by Fuzzy C-Means Clustering (FCM) Cluster and sample reduction by Multivariate Gaussian Distribution (MGD). In its implementation, the FCM algorithm is used to cluster the original training samples, and then the MGD is used to reduce the training samples by choosing the only boundary samples for the next training. Experiments show that the algorithm accelerates the training speed without the decrease of classification accuracy.