{"title":"特征选择技术在学生成绩分类中的比较研究","authors":"Wattana Punlumjeak, Nachirat Rachburee","doi":"10.1109/ICITEED.2015.7408984","DOIUrl":null,"url":null,"abstract":"Student performance classification is a challenging task for teacher and stakeholder for better academic planning and management. Data mining can be used to find knowledge from student data to improve the performance of classifying model. Before applying a classification model, feature selection method is proposed in data preprocessing process to find out the most significant and intrinsic features. In this research, we propose a comparison of four feature selection methods: genetic algorithms, support vector machine, information gain, and minimum redundancy and maximum relevance with four supervised classifiers: naive bays, decision tree, k-nearest neighbor, and neural network. The experimental results show that the minimum redundancy and maximum relevance feature selection method with 10 feature selected give the best result on 91.12% accuracy with a k-nearest neighbor classifier. The result of the present study shows that the advantage of future selection to find a minimum and significant of feature is more effective to classify the student performance.","PeriodicalId":207985,"journal":{"name":"2015 7th International Conference on Information Technology and Electrical Engineering (ICITEE)","volume":"3 1","pages":"0"},"PeriodicalIF":0.0000,"publicationDate":"2015-10-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"32","resultStr":"{\"title\":\"A comparative study of feature selection techniques for classify student performance\",\"authors\":\"Wattana Punlumjeak, Nachirat Rachburee\",\"doi\":\"10.1109/ICITEED.2015.7408984\",\"DOIUrl\":null,\"url\":null,\"abstract\":\"Student performance classification is a challenging task for teacher and stakeholder for better academic planning and management. Data mining can be used to find knowledge from student data to improve the performance of classifying model. Before applying a classification model, feature selection method is proposed in data preprocessing process to find out the most significant and intrinsic features. In this research, we propose a comparison of four feature selection methods: genetic algorithms, support vector machine, information gain, and minimum redundancy and maximum relevance with four supervised classifiers: naive bays, decision tree, k-nearest neighbor, and neural network. The experimental results show that the minimum redundancy and maximum relevance feature selection method with 10 feature selected give the best result on 91.12% accuracy with a k-nearest neighbor classifier. The result of the present study shows that the advantage of future selection to find a minimum and significant of feature is more effective to classify the student performance.\",\"PeriodicalId\":207985,\"journal\":{\"name\":\"2015 7th International Conference on Information Technology and Electrical Engineering (ICITEE)\",\"volume\":\"3 1\",\"pages\":\"0\"},\"PeriodicalIF\":0.0000,\"publicationDate\":\"2015-10-01\",\"publicationTypes\":\"Journal Article\",\"fieldsOfStudy\":null,\"isOpenAccess\":false,\"openAccessPdf\":\"\",\"citationCount\":\"32\",\"resultStr\":null,\"platform\":\"Semanticscholar\",\"paperid\":null,\"PeriodicalName\":\"2015 7th International Conference on Information Technology and Electrical Engineering (ICITEE)\",\"FirstCategoryId\":\"1085\",\"ListUrlMain\":\"https://doi.org/10.1109/ICITEED.2015.7408984\",\"RegionNum\":0,\"RegionCategory\":null,\"ArticlePicture\":[],\"TitleCN\":null,\"AbstractTextCN\":null,\"PMCID\":null,\"EPubDate\":\"\",\"PubModel\":\"\",\"JCR\":\"\",\"JCRName\":\"\",\"Score\":null,\"Total\":0}","platform":"Semanticscholar","paperid":null,"PeriodicalName":"2015 7th International Conference on Information Technology and Electrical Engineering (ICITEE)","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/10.1109/ICITEED.2015.7408984","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
A comparative study of feature selection techniques for classify student performance
Student performance classification is a challenging task for teacher and stakeholder for better academic planning and management. Data mining can be used to find knowledge from student data to improve the performance of classifying model. Before applying a classification model, feature selection method is proposed in data preprocessing process to find out the most significant and intrinsic features. In this research, we propose a comparison of four feature selection methods: genetic algorithms, support vector machine, information gain, and minimum redundancy and maximum relevance with four supervised classifiers: naive bays, decision tree, k-nearest neighbor, and neural network. The experimental results show that the minimum redundancy and maximum relevance feature selection method with 10 feature selected give the best result on 91.12% accuracy with a k-nearest neighbor classifier. The result of the present study shows that the advantage of future selection to find a minimum and significant of feature is more effective to classify the student performance.