利用随机探针和线性支持向量机进行特征选择

Hoi-Ming Chi, O. Ersoy, H. Moskowitz
{"title":"利用随机探针和线性支持向量机进行特征选择","authors":"Hoi-Ming Chi, O. Ersoy, H. Moskowitz","doi":"10.1109/CIMA.2005.1662318","DOIUrl":null,"url":null,"abstract":"A novel feature selection algorithm that combines the ideas of linear support vector machines (SVMs) and random probes is proposed. A random probe is first artificially generated from a Gaussian distribution and appended to the data set as an extra input variable. Next, a standard 2-norm or 1-norm linear support vector machine is trained using this new data set. Each coefficient, or weight, in a linear SVM is compared to that of the random probe feature. Under several statistical assumptions, the probability of each input feature being more relevant than the random probe can be computed easily. The proposed feature selection method is intuitive to use in real-world problems, and it automatically determines the optimal number of features needed. It can also be extended to selecting significant interaction and/or quadratic terms in a 2nd-order polynomial representation","PeriodicalId":306045,"journal":{"name":"2005 ICSC Congress on Computational Intelligence Methods and Applications","volume":"1 1","pages":"0"},"PeriodicalIF":0.0000,"publicationDate":"2005-12-15","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"1","resultStr":"{\"title\":\"Feature selection using random probes and linear support vector machines\",\"authors\":\"Hoi-Ming Chi, O. Ersoy, H. Moskowitz\",\"doi\":\"10.1109/CIMA.2005.1662318\",\"DOIUrl\":null,\"url\":null,\"abstract\":\"A novel feature selection algorithm that combines the ideas of linear support vector machines (SVMs) and random probes is proposed. A random probe is first artificially generated from a Gaussian distribution and appended to the data set as an extra input variable. Next, a standard 2-norm or 1-norm linear support vector machine is trained using this new data set. Each coefficient, or weight, in a linear SVM is compared to that of the random probe feature. Under several statistical assumptions, the probability of each input feature being more relevant than the random probe can be computed easily. The proposed feature selection method is intuitive to use in real-world problems, and it automatically determines the optimal number of features needed. It can also be extended to selecting significant interaction and/or quadratic terms in a 2nd-order polynomial representation\",\"PeriodicalId\":306045,\"journal\":{\"name\":\"2005 ICSC Congress on Computational Intelligence Methods and Applications\",\"volume\":\"1 1\",\"pages\":\"0\"},\"PeriodicalIF\":0.0000,\"publicationDate\":\"2005-12-15\",\"publicationTypes\":\"Journal Article\",\"fieldsOfStudy\":null,\"isOpenAccess\":false,\"openAccessPdf\":\"\",\"citationCount\":\"1\",\"resultStr\":null,\"platform\":\"Semanticscholar\",\"paperid\":null,\"PeriodicalName\":\"2005 ICSC Congress on Computational Intelligence Methods and Applications\",\"FirstCategoryId\":\"1085\",\"ListUrlMain\":\"https://doi.org/10.1109/CIMA.2005.1662318\",\"RegionNum\":0,\"RegionCategory\":null,\"ArticlePicture\":[],\"TitleCN\":null,\"AbstractTextCN\":null,\"PMCID\":null,\"EPubDate\":\"\",\"PubModel\":\"\",\"JCR\":\"\",\"JCRName\":\"\",\"Score\":null,\"Total\":0}","platform":"Semanticscholar","paperid":null,"PeriodicalName":"2005 ICSC Congress on Computational Intelligence Methods and Applications","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/10.1109/CIMA.2005.1662318","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
引用次数: 1

摘要

提出了一种结合线性支持向量机和随机探针思想的特征选择算法。随机探针首先由高斯分布人工生成,并作为额外的输入变量附加到数据集。接下来,使用这个新数据集训练一个标准的2范数或1范数线性支持向量机。线性支持向量机中的每个系数或权重都与随机探测特征的系数或权重进行比较。在几个统计假设下,可以很容易地计算出每个输入特征比随机探测更相关的概率。所提出的特征选择方法可以直观地应用于实际问题,并且可以自动确定所需的最优特征数量。它也可以扩展到在二阶多项式表示中选择重要的相互作用和/或二次项
本文章由计算机程序翻译,如有差异,请以英文原文为准。
Feature selection using random probes and linear support vector machines
A novel feature selection algorithm that combines the ideas of linear support vector machines (SVMs) and random probes is proposed. A random probe is first artificially generated from a Gaussian distribution and appended to the data set as an extra input variable. Next, a standard 2-norm or 1-norm linear support vector machine is trained using this new data set. Each coefficient, or weight, in a linear SVM is compared to that of the random probe feature. Under several statistical assumptions, the probability of each input feature being more relevant than the random probe can be computed easily. The proposed feature selection method is intuitive to use in real-world problems, and it automatically determines the optimal number of features needed. It can also be extended to selecting significant interaction and/or quadratic terms in a 2nd-order polynomial representation
求助全文
通过发布文献求助,成功后即可免费获取论文全文。 去求助
来源期刊
自引率
0.00%
发文量
0
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
确定
请完成安全验证×
copy
已复制链接
快去分享给好友吧!
我知道了
右上角分享
点击右上角分享
0
联系我们:info@booksci.cn Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。 Copyright © 2023 布克学术 All rights reserved.
京ICP备2023020795号-1
ghs 京公网安备 11010802042870号
Book学术文献互助
Book学术文献互助群
群 号:481959085
Book学术官方微信