Design and Development of Hybrid Feature Selection Method for Classification

M. Sreedevi, G. Manasa, Idupulapati Apurva
{"title":"Design and Development of Hybrid Feature Selection Method for Classification","authors":"M. Sreedevi, G. Manasa, Idupulapati Apurva","doi":"10.1109/CONIT51480.2021.9498548","DOIUrl":null,"url":null,"abstract":"To enhance the capability of the learning model in this research paper we have developed a hybrid feature selection method. To defeat the curse of dimensionality, to speed up the classification process, and to get more accurate results a hybrid feature selection model is developed which is a combination of multiple filter methods and a wrapper method. In this model, we employed two sets of Filter Methods-Basic filter methods, a correlation filter method and two Statistical & Ranking filter methods (ANOVA and ROC-AUC) to generate two different subsets of important features, and a Wrapper method (Recursive Feature Elimination with Cross-Validation) is applied on the combined subset to generate a final subset of important features for better prediction results. Five machine learning algorithms-Logistic Regression (LR), Decision Tree, Random Forest (RF), Support Vector Machine (SVM), and K-Nearest Neighbour are used to evaluate the classification accuracy. The proposed hybrid method is applied over four low and four microarray datasets. Outputs are compared to find out which algorithm works best with the proposed model as the results diverge with the machine learning algorithm. Precision, Sensitivity, and Specificity are calculated for each outcome and they demonstrate that the suggested method improved the accuracy of the algorithms.","PeriodicalId":426131,"journal":{"name":"2021 International Conference on Intelligent Technologies (CONIT)","volume":"12 1","pages":"0"},"PeriodicalIF":0.0000,"publicationDate":"2021-06-25","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"2021 International Conference on Intelligent Technologies (CONIT)","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/10.1109/CONIT51480.2021.9498548","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
引用次数: 0

Abstract

To enhance the capability of the learning model in this research paper we have developed a hybrid feature selection method. To defeat the curse of dimensionality, to speed up the classification process, and to get more accurate results a hybrid feature selection model is developed which is a combination of multiple filter methods and a wrapper method. In this model, we employed two sets of Filter Methods-Basic filter methods, a correlation filter method and two Statistical & Ranking filter methods (ANOVA and ROC-AUC) to generate two different subsets of important features, and a Wrapper method (Recursive Feature Elimination with Cross-Validation) is applied on the combined subset to generate a final subset of important features for better prediction results. Five machine learning algorithms-Logistic Regression (LR), Decision Tree, Random Forest (RF), Support Vector Machine (SVM), and K-Nearest Neighbour are used to evaluate the classification accuracy. The proposed hybrid method is applied over four low and four microarray datasets. Outputs are compared to find out which algorithm works best with the proposed model as the results diverge with the machine learning algorithm. Precision, Sensitivity, and Specificity are calculated for each outcome and they demonstrate that the suggested method improved the accuracy of the algorithms.
混合特征选择分类方法的设计与开发
为了提高学习模型的能力,本文提出了一种混合特征选择方法。为了克服维数的困扰,加快分类速度,获得更准确的分类结果,提出了一种混合特征选择模型,该模型将多种滤波方法与一种包装方法相结合。在该模型中,我们使用了两组过滤方法——基本过滤方法、一种相关过滤方法和两种统计和排名过滤方法(ANOVA和ROC-AUC)来生成两个不同的重要特征子集,并在组合子集上应用Wrapper方法(递归特征消除与交叉验证)来生成最终的重要特征子集,以获得更好的预测结果。采用逻辑回归(LR)、决策树(Decision Tree)、随机森林(Random Forest)、支持向量机(SVM)和k近邻(K-Nearest Neighbour)五种机器学习算法来评估分类精度。提出的混合方法应用于四个低和四个微阵列数据集。当结果与机器学习算法不同时,比较输出以找出哪种算法最适合所提出的模型。计算了每个结果的精度、灵敏度和特异性,结果表明所建议的方法提高了算法的准确性。
本文章由计算机程序翻译,如有差异,请以英文原文为准。
求助全文
约1分钟内获得全文 求助全文
来源期刊
自引率
0.00%
发文量
0
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
确定
请完成安全验证×
copy
已复制链接
快去分享给好友吧!
我知道了
右上角分享
点击右上角分享
0
联系我们:info@booksci.cn Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。 Copyright © 2023 布克学术 All rights reserved.
京ICP备2023020795号-1
ghs 京公网安备 11010802042870号
Book学术文献互助
Book学术文献互助群
群 号:481959085
Book学术官方微信