{"title":"基于稀疏的最小二乘支持向量机训练算法","authors":"Jie Yang, Jun Ma","doi":"10.1109/CIDM.2014.7008688","DOIUrl":null,"url":null,"abstract":"We address the training problem of the sparse Least Squares Support Vector Machines (SVM) using compressed sensing. The proposed algorithm regards the support vectors as a dictionary and selects the important ones that minimize the residual output error iteratively. A measurement matrix is also introduced to reduce the computational cost. The main advantage is that the proposed algorithm performs model training and support vector selection simultaneously. The performance of the proposed algorithm is tested with several benchmark classification problems in terms of number of selected support vectors and size of the measurement matrix. Simulation results show that the proposed algorithm performs competitively when compared to existing methods.","PeriodicalId":117542,"journal":{"name":"2014 IEEE Symposium on Computational Intelligence and Data Mining (CIDM)","volume":"22 1","pages":"0"},"PeriodicalIF":0.0000,"publicationDate":"2014-12-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"5","resultStr":"{\"title\":\"A sparsity-based training algorithm for Least Squares SVM\",\"authors\":\"Jie Yang, Jun Ma\",\"doi\":\"10.1109/CIDM.2014.7008688\",\"DOIUrl\":null,\"url\":null,\"abstract\":\"We address the training problem of the sparse Least Squares Support Vector Machines (SVM) using compressed sensing. The proposed algorithm regards the support vectors as a dictionary and selects the important ones that minimize the residual output error iteratively. A measurement matrix is also introduced to reduce the computational cost. The main advantage is that the proposed algorithm performs model training and support vector selection simultaneously. The performance of the proposed algorithm is tested with several benchmark classification problems in terms of number of selected support vectors and size of the measurement matrix. Simulation results show that the proposed algorithm performs competitively when compared to existing methods.\",\"PeriodicalId\":117542,\"journal\":{\"name\":\"2014 IEEE Symposium on Computational Intelligence and Data Mining (CIDM)\",\"volume\":\"22 1\",\"pages\":\"0\"},\"PeriodicalIF\":0.0000,\"publicationDate\":\"2014-12-01\",\"publicationTypes\":\"Journal Article\",\"fieldsOfStudy\":null,\"isOpenAccess\":false,\"openAccessPdf\":\"\",\"citationCount\":\"5\",\"resultStr\":null,\"platform\":\"Semanticscholar\",\"paperid\":null,\"PeriodicalName\":\"2014 IEEE Symposium on Computational Intelligence and Data Mining (CIDM)\",\"FirstCategoryId\":\"1085\",\"ListUrlMain\":\"https://doi.org/10.1109/CIDM.2014.7008688\",\"RegionNum\":0,\"RegionCategory\":null,\"ArticlePicture\":[],\"TitleCN\":null,\"AbstractTextCN\":null,\"PMCID\":null,\"EPubDate\":\"\",\"PubModel\":\"\",\"JCR\":\"\",\"JCRName\":\"\",\"Score\":null,\"Total\":0}","platform":"Semanticscholar","paperid":null,"PeriodicalName":"2014 IEEE Symposium on Computational Intelligence and Data Mining (CIDM)","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/10.1109/CIDM.2014.7008688","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
A sparsity-based training algorithm for Least Squares SVM
We address the training problem of the sparse Least Squares Support Vector Machines (SVM) using compressed sensing. The proposed algorithm regards the support vectors as a dictionary and selects the important ones that minimize the residual output error iteratively. A measurement matrix is also introduced to reduce the computational cost. The main advantage is that the proposed algorithm performs model training and support vector selection simultaneously. The performance of the proposed algorithm is tested with several benchmark classification problems in terms of number of selected support vectors and size of the measurement matrix. Simulation results show that the proposed algorithm performs competitively when compared to existing methods.