采用双网络架构的嵌入式特征选择

IF 4.9
Abderrahim Abbassi, Arved Dörpinghaus, Niklas Römgens, Tanja Grießmann, Raimund Rolfes
{"title":"采用双网络架构的嵌入式特征选择","authors":"Abderrahim Abbassi,&nbsp;Arved Dörpinghaus,&nbsp;Niklas Römgens,&nbsp;Tanja Grießmann,&nbsp;Raimund Rolfes","doi":"10.1016/j.mlwa.2025.100672","DOIUrl":null,"url":null,"abstract":"<div><div>Feature selection is essential for eliminating noise, reducing redundancy, simplifying computational complexity, and lowering data collection and processing costs. However, existing methods often face challenges due to the complexity of feature interdependencies, uncertainty regarding the exact number of relevant features, and the need for hyperparameter optimization, which increases methodological complexity.</div><div>This research proposes a novel dual-network architecture for feature selection that addresses these issues. The architecture consists of a task model and a selection model. First, redundant features are fed into the selection model, which generates a binary mask aligned with the input feature dimensions. This mask is applied to a shifted version of the original features, serving as input to the task model. The task model then uses the selected features to perform the target supervised task. Simultaneously, the selection model aims to minimize the cumulative value of the mask, thus selecting the most relevant features with minimal impact on the task model’s performance.</div><div>The method is evaluated using benchmark and synthetic datasets across different supervised tasks. Comparative evaluation with state-of-the-art techniques demonstrates that the proposed approach exhibits superior or competitive feature selection capabilities, achieving a reduction of 90% or more in feature count. This is particularly notable in the presence of non-linear feature interdependencies. The key advantages of the proposed method are its ability to self-determine the number of relevant features needed for the supervised task and its simplicity, requiring the pre-definition of only a single hyperparameter, for which an estimation approach is suggested.</div></div>","PeriodicalId":74093,"journal":{"name":"Machine learning with applications","volume":"21 ","pages":"Article 100672"},"PeriodicalIF":4.9000,"publicationDate":"2025-06-18","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":"{\"title\":\"Embedded feature selection using dual-network architecture\",\"authors\":\"Abderrahim Abbassi,&nbsp;Arved Dörpinghaus,&nbsp;Niklas Römgens,&nbsp;Tanja Grießmann,&nbsp;Raimund Rolfes\",\"doi\":\"10.1016/j.mlwa.2025.100672\",\"DOIUrl\":null,\"url\":null,\"abstract\":\"<div><div>Feature selection is essential for eliminating noise, reducing redundancy, simplifying computational complexity, and lowering data collection and processing costs. However, existing methods often face challenges due to the complexity of feature interdependencies, uncertainty regarding the exact number of relevant features, and the need for hyperparameter optimization, which increases methodological complexity.</div><div>This research proposes a novel dual-network architecture for feature selection that addresses these issues. The architecture consists of a task model and a selection model. First, redundant features are fed into the selection model, which generates a binary mask aligned with the input feature dimensions. This mask is applied to a shifted version of the original features, serving as input to the task model. The task model then uses the selected features to perform the target supervised task. Simultaneously, the selection model aims to minimize the cumulative value of the mask, thus selecting the most relevant features with minimal impact on the task model’s performance.</div><div>The method is evaluated using benchmark and synthetic datasets across different supervised tasks. Comparative evaluation with state-of-the-art techniques demonstrates that the proposed approach exhibits superior or competitive feature selection capabilities, achieving a reduction of 90% or more in feature count. This is particularly notable in the presence of non-linear feature interdependencies. The key advantages of the proposed method are its ability to self-determine the number of relevant features needed for the supervised task and its simplicity, requiring the pre-definition of only a single hyperparameter, for which an estimation approach is suggested.</div></div>\",\"PeriodicalId\":74093,\"journal\":{\"name\":\"Machine learning with applications\",\"volume\":\"21 \",\"pages\":\"Article 100672\"},\"PeriodicalIF\":4.9000,\"publicationDate\":\"2025-06-18\",\"publicationTypes\":\"Journal Article\",\"fieldsOfStudy\":null,\"isOpenAccess\":false,\"openAccessPdf\":\"\",\"citationCount\":\"0\",\"resultStr\":null,\"platform\":\"Semanticscholar\",\"paperid\":null,\"PeriodicalName\":\"Machine learning with applications\",\"FirstCategoryId\":\"1085\",\"ListUrlMain\":\"https://www.sciencedirect.com/science/article/pii/S2666827025000556\",\"RegionNum\":0,\"RegionCategory\":null,\"ArticlePicture\":[],\"TitleCN\":null,\"AbstractTextCN\":null,\"PMCID\":null,\"EPubDate\":\"\",\"PubModel\":\"\",\"JCR\":\"\",\"JCRName\":\"\",\"Score\":null,\"Total\":0}","platform":"Semanticscholar","paperid":null,"PeriodicalName":"Machine learning with applications","FirstCategoryId":"1085","ListUrlMain":"https://www.sciencedirect.com/science/article/pii/S2666827025000556","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
引用次数: 0

摘要

特征选择对于消除噪声、减少冗余、简化计算复杂度、降低数据收集和处理成本至关重要。然而,由于特征相互依赖的复杂性、相关特征的确切数量的不确定性以及对超参数优化的需求,现有方法经常面临挑战,这增加了方法的复杂性。本研究提出了一种用于特征选择的新型双网络架构来解决这些问题。该体系结构由任务模型和选择模型组成。首先,将冗余特征输入到选择模型中,生成与输入特征尺寸对齐的二值掩码。此掩码应用于原始特征的移位版本,作为任务模型的输入。然后,任务模型使用选定的特征来执行目标监督任务。同时,选择模型的目标是最小化掩码的累积值,从而在对任务模型性能影响最小的情况下选择最相关的特征。该方法使用不同监督任务的基准和合成数据集进行评估。与最先进的技术进行比较评估表明,所提出的方法显示出优越或具有竞争力的特征选择能力,实现特征计数减少90%或更多。这在非线性特征相互依赖的情况下尤为明显。该方法的主要优点是能够自确定监督任务所需的相关特征的数量,并且简单,只需要预先定义一个超参数,为此提出了一种估计方法。
本文章由计算机程序翻译,如有差异,请以英文原文为准。
Embedded feature selection using dual-network architecture
Feature selection is essential for eliminating noise, reducing redundancy, simplifying computational complexity, and lowering data collection and processing costs. However, existing methods often face challenges due to the complexity of feature interdependencies, uncertainty regarding the exact number of relevant features, and the need for hyperparameter optimization, which increases methodological complexity.
This research proposes a novel dual-network architecture for feature selection that addresses these issues. The architecture consists of a task model and a selection model. First, redundant features are fed into the selection model, which generates a binary mask aligned with the input feature dimensions. This mask is applied to a shifted version of the original features, serving as input to the task model. The task model then uses the selected features to perform the target supervised task. Simultaneously, the selection model aims to minimize the cumulative value of the mask, thus selecting the most relevant features with minimal impact on the task model’s performance.
The method is evaluated using benchmark and synthetic datasets across different supervised tasks. Comparative evaluation with state-of-the-art techniques demonstrates that the proposed approach exhibits superior or competitive feature selection capabilities, achieving a reduction of 90% or more in feature count. This is particularly notable in the presence of non-linear feature interdependencies. The key advantages of the proposed method are its ability to self-determine the number of relevant features needed for the supervised task and its simplicity, requiring the pre-definition of only a single hyperparameter, for which an estimation approach is suggested.
求助全文
通过发布文献求助,成功后即可免费获取论文全文。 去求助
来源期刊
Machine learning with applications
Machine learning with applications Management Science and Operations Research, Artificial Intelligence, Computer Science Applications
自引率
0.00%
发文量
0
审稿时长
98 days
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
确定
请完成安全验证×
copy
已复制链接
快去分享给好友吧!
我知道了
右上角分享
点击右上角分享
0
联系我们:info@booksci.cn Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。 Copyright © 2023 布克学术 All rights reserved.
京ICP备2023020795号-1
ghs 京公网安备 11010802042870号
Book学术文献互助
Book学术文献互助群
群 号:604180095
Book学术官方微信