{"title":"ESARSA-MRFO-FS:使用期望sarsa强化学习进行特征选择的优化蝠鲼觅食优化器","authors":"Yousry AbdulAzeem , Hossam Magdy Balaha , Amna Bamaqa , Mahmoud Badawy , Mostafa A. Elhosseini","doi":"10.1016/j.knosys.2025.113695","DOIUrl":null,"url":null,"abstract":"<div><div>Disease prediction with the help of computers has achieved significant progress in this area; however, it still requires a more accurate identification of each data feature. In the past few years, ML-based medical diagnoses have become increasingly dependent on the datasets rather than the algorithms. Consequently, the more features in a dataset, the more difficult the model training process will be. A feature selection (FS) strategy can be used to eliminate unwanted features before the training in order to improve the model performance. Due to the great increase in the number of features, FS causes a complex combinatorial problem, which is NP-hard. The primary goal of this research is to create a novel framework for improving the Manta-ray Foraging Optimizer (MRFO) algorithm’s performance. By the method of exploration–exploitation toggle adjustments, this process becomes the most efficient one by using the Expected State–action-reward-state–action (ESARSA) reinforcement learning strategy. The study employed 12 datasets as well as two classifiers (K-Nearest-Neighbor and Decision-Tree), both with and without hyperparameter adjustment. The study compared the performances of three feature selection methods: no feature selection, the original MRFO algorithm, and the proposed ESARSAMRFO-FS algorithm. The results revealed that ESARSA-MRFO-FS is the most effective of the three algorithms with classifiers and optimization scenarios. Besides, experiments comparing MRFO and ESARSA-MRFO performance on 44 benchmarks are presented in the article. In five functions, ESARSA-MRFO outperforms MRFO. These results confirm that the proposed method is efficient for feature selection in medical diagnosis because of its accuracy and low processing costs.</div></div>","PeriodicalId":49939,"journal":{"name":"Knowledge-Based Systems","volume":"321 ","pages":"Article 113695"},"PeriodicalIF":7.6000,"publicationDate":"2025-05-17","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":"{\"title\":\"ESARSA-MRFO-FS: Optimizing Manta-ray Foraging Optimizer using Expected-SARSA reinforcement learning for features selection\",\"authors\":\"Yousry AbdulAzeem , Hossam Magdy Balaha , Amna Bamaqa , Mahmoud Badawy , Mostafa A. Elhosseini\",\"doi\":\"10.1016/j.knosys.2025.113695\",\"DOIUrl\":null,\"url\":null,\"abstract\":\"<div><div>Disease prediction with the help of computers has achieved significant progress in this area; however, it still requires a more accurate identification of each data feature. In the past few years, ML-based medical diagnoses have become increasingly dependent on the datasets rather than the algorithms. Consequently, the more features in a dataset, the more difficult the model training process will be. A feature selection (FS) strategy can be used to eliminate unwanted features before the training in order to improve the model performance. Due to the great increase in the number of features, FS causes a complex combinatorial problem, which is NP-hard. The primary goal of this research is to create a novel framework for improving the Manta-ray Foraging Optimizer (MRFO) algorithm’s performance. By the method of exploration–exploitation toggle adjustments, this process becomes the most efficient one by using the Expected State–action-reward-state–action (ESARSA) reinforcement learning strategy. The study employed 12 datasets as well as two classifiers (K-Nearest-Neighbor and Decision-Tree), both with and without hyperparameter adjustment. The study compared the performances of three feature selection methods: no feature selection, the original MRFO algorithm, and the proposed ESARSAMRFO-FS algorithm. The results revealed that ESARSA-MRFO-FS is the most effective of the three algorithms with classifiers and optimization scenarios. Besides, experiments comparing MRFO and ESARSA-MRFO performance on 44 benchmarks are presented in the article. In five functions, ESARSA-MRFO outperforms MRFO. These results confirm that the proposed method is efficient for feature selection in medical diagnosis because of its accuracy and low processing costs.</div></div>\",\"PeriodicalId\":49939,\"journal\":{\"name\":\"Knowledge-Based Systems\",\"volume\":\"321 \",\"pages\":\"Article 113695\"},\"PeriodicalIF\":7.6000,\"publicationDate\":\"2025-05-17\",\"publicationTypes\":\"Journal Article\",\"fieldsOfStudy\":null,\"isOpenAccess\":false,\"openAccessPdf\":\"\",\"citationCount\":\"0\",\"resultStr\":null,\"platform\":\"Semanticscholar\",\"paperid\":null,\"PeriodicalName\":\"Knowledge-Based Systems\",\"FirstCategoryId\":\"94\",\"ListUrlMain\":\"https://www.sciencedirect.com/science/article/pii/S0950705125007415\",\"RegionNum\":1,\"RegionCategory\":\"计算机科学\",\"ArticlePicture\":[],\"TitleCN\":null,\"AbstractTextCN\":null,\"PMCID\":null,\"EPubDate\":\"\",\"PubModel\":\"\",\"JCR\":\"Q1\",\"JCRName\":\"COMPUTER SCIENCE, ARTIFICIAL INTELLIGENCE\",\"Score\":null,\"Total\":0}","platform":"Semanticscholar","paperid":null,"PeriodicalName":"Knowledge-Based Systems","FirstCategoryId":"94","ListUrlMain":"https://www.sciencedirect.com/science/article/pii/S0950705125007415","RegionNum":1,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"Q1","JCRName":"COMPUTER SCIENCE, ARTIFICIAL INTELLIGENCE","Score":null,"Total":0}
ESARSA-MRFO-FS: Optimizing Manta-ray Foraging Optimizer using Expected-SARSA reinforcement learning for features selection
Disease prediction with the help of computers has achieved significant progress in this area; however, it still requires a more accurate identification of each data feature. In the past few years, ML-based medical diagnoses have become increasingly dependent on the datasets rather than the algorithms. Consequently, the more features in a dataset, the more difficult the model training process will be. A feature selection (FS) strategy can be used to eliminate unwanted features before the training in order to improve the model performance. Due to the great increase in the number of features, FS causes a complex combinatorial problem, which is NP-hard. The primary goal of this research is to create a novel framework for improving the Manta-ray Foraging Optimizer (MRFO) algorithm’s performance. By the method of exploration–exploitation toggle adjustments, this process becomes the most efficient one by using the Expected State–action-reward-state–action (ESARSA) reinforcement learning strategy. The study employed 12 datasets as well as two classifiers (K-Nearest-Neighbor and Decision-Tree), both with and without hyperparameter adjustment. The study compared the performances of three feature selection methods: no feature selection, the original MRFO algorithm, and the proposed ESARSAMRFO-FS algorithm. The results revealed that ESARSA-MRFO-FS is the most effective of the three algorithms with classifiers and optimization scenarios. Besides, experiments comparing MRFO and ESARSA-MRFO performance on 44 benchmarks are presented in the article. In five functions, ESARSA-MRFO outperforms MRFO. These results confirm that the proposed method is efficient for feature selection in medical diagnosis because of its accuracy and low processing costs.
期刊介绍:
Knowledge-Based Systems, an international and interdisciplinary journal in artificial intelligence, publishes original, innovative, and creative research results in the field. It focuses on knowledge-based and other artificial intelligence techniques-based systems. The journal aims to support human prediction and decision-making through data science and computation techniques, provide a balanced coverage of theory and practical study, and encourage the development and implementation of knowledge-based intelligence models, methods, systems, and software tools. Applications in business, government, education, engineering, and healthcare are emphasized.