{"title":"A self-adjusting representation-based multitask PSO for high-dimensional feature selection","authors":"Li Deng , Xiaohui Su , Bo Wei","doi":"10.1016/j.swevo.2025.102084","DOIUrl":null,"url":null,"abstract":"<div><div>As a critical preprocessing step in machine learning tasks, feature selection (FS) aims to identify informative features from the original datasets. However, FS is commonly formulated as an NP-hard combinatorial optimization problem, particularly when compounded by exponentially expanding search spaces and complex feature interactions. Due to its simplicity of implementation, Particle Swarm Optimization (PSO) has been extensively utilized in FS tasks. Concurrently, the optimization process frequently converges to local optima when handling high-dimensional (<span><math><mo>></mo></math></span>1000D) datasets. To address this issue, a self-adjusting representation-based multitask PSO (SAR-MTPSO) is proposed in this paper. Firstly, the knee point strategy and an elite feature-preserving strategy are employed to obtain promising particles with key features. Based on these particles, a new multitask framework is introduced, where two tasks are constructed by using a self-adjusting representation strategy. Secondly, a two-layer knowledge transfer strategy is proposed to promote the useful information sharing and exchanging between two tasks dynamically. Finally, an adaptive re-initialization strategy is proposed to enhance the exploitation and exploration capabilities of the two tasks respectively. SAR-MTPSO was compared with 10 representative FS algorithms on 21 high-dimensional datasets. Experimental results show that SAR-MTPSO can achieve the highest classification accuracies with smaller sizes of feature subsets on 17 out of 21 datasets.</div></div>","PeriodicalId":48682,"journal":{"name":"Swarm and Evolutionary Computation","volume":"98 ","pages":"Article 102084"},"PeriodicalIF":8.2000,"publicationDate":"2025-07-23","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"Swarm and Evolutionary Computation","FirstCategoryId":"94","ListUrlMain":"https://www.sciencedirect.com/science/article/pii/S2210650225002421","RegionNum":1,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"Q1","JCRName":"COMPUTER SCIENCE, ARTIFICIAL INTELLIGENCE","Score":null,"Total":0}
引用次数: 0
Abstract
As a critical preprocessing step in machine learning tasks, feature selection (FS) aims to identify informative features from the original datasets. However, FS is commonly formulated as an NP-hard combinatorial optimization problem, particularly when compounded by exponentially expanding search spaces and complex feature interactions. Due to its simplicity of implementation, Particle Swarm Optimization (PSO) has been extensively utilized in FS tasks. Concurrently, the optimization process frequently converges to local optima when handling high-dimensional (1000D) datasets. To address this issue, a self-adjusting representation-based multitask PSO (SAR-MTPSO) is proposed in this paper. Firstly, the knee point strategy and an elite feature-preserving strategy are employed to obtain promising particles with key features. Based on these particles, a new multitask framework is introduced, where two tasks are constructed by using a self-adjusting representation strategy. Secondly, a two-layer knowledge transfer strategy is proposed to promote the useful information sharing and exchanging between two tasks dynamically. Finally, an adaptive re-initialization strategy is proposed to enhance the exploitation and exploration capabilities of the two tasks respectively. SAR-MTPSO was compared with 10 representative FS algorithms on 21 high-dimensional datasets. Experimental results show that SAR-MTPSO can achieve the highest classification accuracies with smaller sizes of feature subsets on 17 out of 21 datasets.
期刊介绍:
Swarm and Evolutionary Computation is a pioneering peer-reviewed journal focused on the latest research and advancements in nature-inspired intelligent computation using swarm and evolutionary algorithms. It covers theoretical, experimental, and practical aspects of these paradigms and their hybrids, promoting interdisciplinary research. The journal prioritizes the publication of high-quality, original articles that push the boundaries of evolutionary computation and swarm intelligence. Additionally, it welcomes survey papers on current topics and novel applications. Topics of interest include but are not limited to: Genetic Algorithms, and Genetic Programming, Evolution Strategies, and Evolutionary Programming, Differential Evolution, Artificial Immune Systems, Particle Swarms, Ant Colony, Bacterial Foraging, Artificial Bees, Fireflies Algorithm, Harmony Search, Artificial Life, Digital Organisms, Estimation of Distribution Algorithms, Stochastic Diffusion Search, Quantum Computing, Nano Computing, Membrane Computing, Human-centric Computing, Hybridization of Algorithms, Memetic Computing, Autonomic Computing, Self-organizing systems, Combinatorial, Discrete, Binary, Constrained, Multi-objective, Multi-modal, Dynamic, and Large-scale Optimization.