Yanting Guo, Meng Hu, Eric C. C. Tsang, Degang Chen, Weihua Xu
{"title":"基于最小冗余和最大一致性的特征选择","authors":"Yanting Guo, Meng Hu, Eric C. C. Tsang, Degang Chen, Weihua Xu","doi":"10.1007/s43674-021-00021-9","DOIUrl":null,"url":null,"abstract":"<div><p>Feature selection can effectively eliminate irrelevant or redundant features without changing features semantics, so as to improve the performance of learning and reduce the training time. In most of the existing feature selection methods based on rough sets, eliminating the redundant features between features and decisions, and deleting the redundant features between features are performed separately. This will greatly increase the search time of feature subset. To quickly remove redundant features, we define a series of feature evaluation functions that consider both the consistency between features and decisions, and redundancy between features, then propose a novel feature selection method based on min-redundancy and max-consistency. Firstly, we define the consistency of features with respect to decisions and the redundancy between features from neighborhood information granules. Then we propose a combined criterion to measure the importance of features and design a feature selection algorithm based on minimal-redundancy-maximal-consistency (mRMC). Finally, on UCI data sets, mRMC is compared with three other popular feature selection algorithms based on neighborhood idea, from classification accuracy, the number of selected features and running time. The experimental comparison shows that mRMC can quickly delete redundant features and select useful features while ensuring classification accuracy.</p></div>","PeriodicalId":72089,"journal":{"name":"Advances in computational intelligence","volume":"2 1","pages":""},"PeriodicalIF":0.0000,"publicationDate":"2021-12-17","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://link.springer.com/content/pdf/10.1007/s43674-021-00021-9.pdf","citationCount":"1","resultStr":"{\"title\":\"Feature selection based on min-redundancy and max-consistency\",\"authors\":\"Yanting Guo, Meng Hu, Eric C. C. Tsang, Degang Chen, Weihua Xu\",\"doi\":\"10.1007/s43674-021-00021-9\",\"DOIUrl\":null,\"url\":null,\"abstract\":\"<div><p>Feature selection can effectively eliminate irrelevant or redundant features without changing features semantics, so as to improve the performance of learning and reduce the training time. In most of the existing feature selection methods based on rough sets, eliminating the redundant features between features and decisions, and deleting the redundant features between features are performed separately. This will greatly increase the search time of feature subset. To quickly remove redundant features, we define a series of feature evaluation functions that consider both the consistency between features and decisions, and redundancy between features, then propose a novel feature selection method based on min-redundancy and max-consistency. Firstly, we define the consistency of features with respect to decisions and the redundancy between features from neighborhood information granules. Then we propose a combined criterion to measure the importance of features and design a feature selection algorithm based on minimal-redundancy-maximal-consistency (mRMC). Finally, on UCI data sets, mRMC is compared with three other popular feature selection algorithms based on neighborhood idea, from classification accuracy, the number of selected features and running time. The experimental comparison shows that mRMC can quickly delete redundant features and select useful features while ensuring classification accuracy.</p></div>\",\"PeriodicalId\":72089,\"journal\":{\"name\":\"Advances in computational intelligence\",\"volume\":\"2 1\",\"pages\":\"\"},\"PeriodicalIF\":0.0000,\"publicationDate\":\"2021-12-17\",\"publicationTypes\":\"Journal Article\",\"fieldsOfStudy\":null,\"isOpenAccess\":false,\"openAccessPdf\":\"https://link.springer.com/content/pdf/10.1007/s43674-021-00021-9.pdf\",\"citationCount\":\"1\",\"resultStr\":null,\"platform\":\"Semanticscholar\",\"paperid\":null,\"PeriodicalName\":\"Advances in computational intelligence\",\"FirstCategoryId\":\"1085\",\"ListUrlMain\":\"https://link.springer.com/article/10.1007/s43674-021-00021-9\",\"RegionNum\":0,\"RegionCategory\":null,\"ArticlePicture\":[],\"TitleCN\":null,\"AbstractTextCN\":null,\"PMCID\":null,\"EPubDate\":\"\",\"PubModel\":\"\",\"JCR\":\"\",\"JCRName\":\"\",\"Score\":null,\"Total\":0}","platform":"Semanticscholar","paperid":null,"PeriodicalName":"Advances in computational intelligence","FirstCategoryId":"1085","ListUrlMain":"https://link.springer.com/article/10.1007/s43674-021-00021-9","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
Feature selection based on min-redundancy and max-consistency
Feature selection can effectively eliminate irrelevant or redundant features without changing features semantics, so as to improve the performance of learning and reduce the training time. In most of the existing feature selection methods based on rough sets, eliminating the redundant features between features and decisions, and deleting the redundant features between features are performed separately. This will greatly increase the search time of feature subset. To quickly remove redundant features, we define a series of feature evaluation functions that consider both the consistency between features and decisions, and redundancy between features, then propose a novel feature selection method based on min-redundancy and max-consistency. Firstly, we define the consistency of features with respect to decisions and the redundancy between features from neighborhood information granules. Then we propose a combined criterion to measure the importance of features and design a feature selection algorithm based on minimal-redundancy-maximal-consistency (mRMC). Finally, on UCI data sets, mRMC is compared with three other popular feature selection algorithms based on neighborhood idea, from classification accuracy, the number of selected features and running time. The experimental comparison shows that mRMC can quickly delete redundant features and select useful features while ensuring classification accuracy.