{"title":"Regularized sparse feature selection with constraints embedded in graph Laplacian matrix","authors":"Zahir Noorie, F. Afsari","doi":"10.1109/ICSPIS.2017.8311602","DOIUrl":null,"url":null,"abstract":"Feature selection is an important pre-processing stage in many machine learning and pattern recognition tasks, which eliminates irrelevant and redundant features and improves learning performance. Regularized sparse feature selection methods like Lasso and its variants using ℓ1-norm regularization term in their optimization problem have received much attention in recent years. Prior information could be represented as the class labels or pairwise constraints, i.e., must-link (positive) and cannot-link (negative) constraints. In this paper, besides the ℓ1-norm regularization term, a normalized adapted affinity matrix is applied to embed the pairwise constraints in the affinity matrix. In the proposed affinity matrix, the weights are strengthened/weakened according to the positive/negative constraints. The experimental results on several data sets from University of California-Irvine (UCI) machine learning repository and a high dimensional data set, show the effectiveness of the proposed method in the classification tasks compared to some similar feature selection methods.","PeriodicalId":380266,"journal":{"name":"2017 3rd Iranian Conference on Intelligent Systems and Signal Processing (ICSPIS)","volume":"203 2 1","pages":"0"},"PeriodicalIF":0.0000,"publicationDate":"2017-12-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"1","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"2017 3rd Iranian Conference on Intelligent Systems and Signal Processing (ICSPIS)","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/10.1109/ICSPIS.2017.8311602","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
引用次数: 1
Abstract
Feature selection is an important pre-processing stage in many machine learning and pattern recognition tasks, which eliminates irrelevant and redundant features and improves learning performance. Regularized sparse feature selection methods like Lasso and its variants using ℓ1-norm regularization term in their optimization problem have received much attention in recent years. Prior information could be represented as the class labels or pairwise constraints, i.e., must-link (positive) and cannot-link (negative) constraints. In this paper, besides the ℓ1-norm regularization term, a normalized adapted affinity matrix is applied to embed the pairwise constraints in the affinity matrix. In the proposed affinity matrix, the weights are strengthened/weakened according to the positive/negative constraints. The experimental results on several data sets from University of California-Irvine (UCI) machine learning repository and a high dimensional data set, show the effectiveness of the proposed method in the classification tasks compared to some similar feature selection methods.