J. Rabcan, Patrik Rusnak, J. Kostolny, R. Stankovic
{"title":"Comparison of Algorithms for Fuzzy Decision Tree Induction","authors":"J. Rabcan, Patrik Rusnak, J. Kostolny, R. Stankovic","doi":"10.1109/ICETA51985.2020.9379189","DOIUrl":null,"url":null,"abstract":"The objective of the classification is to assign a class label to a data sample based on previous, learned experience. Despite the long tradition of classification algorithms research, there is not a unique technique that yields the best classification performance in all scenarios. Many real-world problems are uncertain. In this case, the crisp classification can be difficult to perform. The usage of fuzzy logic can be useful to describe real-world problems with higher accuracy. In this paper, we describe and compare algorithms of Fuzzy Decision Trees (FDTs), which are an extension of traditional decision trees. These algorithms are popular for their easy understandability and interpretability. But today there are many algorithms for FDT induction. These algorithms differ in many aspects. The most common difference is in information measures for selecting splitting attributes. Generally, the goal of decision tree induction is to create the smallest tree which is the most accurate as possible. In this paper, we use various information measures to induct FDTs and compare the accuracy and size of obtained FDTs. In comparison, information measures that are based on a generalization of the Shannon entropy and cumulative mutual information are included. The comparison shows that FDTs based on the cumulative mutual information acquired the best results. These results will also be included in a course on data mining that is taught at Faculty of Management Science and Informatics of University of Zilina.","PeriodicalId":149716,"journal":{"name":"2020 18th International Conference on Emerging eLearning Technologies and Applications (ICETA)","volume":null,"pages":null},"PeriodicalIF":0.0000,"publicationDate":"2020-11-12","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"1","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"2020 18th International Conference on Emerging eLearning Technologies and Applications (ICETA)","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/10.1109/ICETA51985.2020.9379189","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
引用次数: 1
Abstract
The objective of the classification is to assign a class label to a data sample based on previous, learned experience. Despite the long tradition of classification algorithms research, there is not a unique technique that yields the best classification performance in all scenarios. Many real-world problems are uncertain. In this case, the crisp classification can be difficult to perform. The usage of fuzzy logic can be useful to describe real-world problems with higher accuracy. In this paper, we describe and compare algorithms of Fuzzy Decision Trees (FDTs), which are an extension of traditional decision trees. These algorithms are popular for their easy understandability and interpretability. But today there are many algorithms for FDT induction. These algorithms differ in many aspects. The most common difference is in information measures for selecting splitting attributes. Generally, the goal of decision tree induction is to create the smallest tree which is the most accurate as possible. In this paper, we use various information measures to induct FDTs and compare the accuracy and size of obtained FDTs. In comparison, information measures that are based on a generalization of the Shannon entropy and cumulative mutual information are included. The comparison shows that FDTs based on the cumulative mutual information acquired the best results. These results will also be included in a course on data mining that is taught at Faculty of Management Science and Informatics of University of Zilina.