{"title":"Genetic algorithm based multiple decision tree induction","authors":"Z. Bandar, H. Al-Attar, D. Mclean","doi":"10.1109/ICONIP.1999.845633","DOIUrl":null,"url":null,"abstract":"There are two fundamental weaknesses which may have a great impact on the performance of decision tree (DT) induction. These are the limitations in the ability of the DT language to represent some of the underlying patterns of the domain and the degradation in the quality of evidence available to the induction process caused by its recursive partitioning of the training data. The impact of these two weaknesses is greatest when the induction process attempts to overcome the first weakness by resorting to more partitioning of the training data, thus increasing its vulnerability to the second weakness. The authors investigate the use of multiple DT models as a method of overcoming the limitations of the DT modeling language and describe a new and novel algorithm to automatically generate multiple DT models from the same training data. The algorithm is compared to a single-tree classifier by experiments on two well known data sets. Results clearly demonstrate the superiority of our algorithm.","PeriodicalId":237855,"journal":{"name":"ICONIP'99. ANZIIS'99 & ANNES'99 & ACNN'99. 6th International Conference on Neural Information Processing. Proceedings (Cat. No.99EX378)","volume":"33 1","pages":"0"},"PeriodicalIF":0.0000,"publicationDate":"1999-11-16","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"14","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"ICONIP'99. ANZIIS'99 & ANNES'99 & ACNN'99. 6th International Conference on Neural Information Processing. Proceedings (Cat. No.99EX378)","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/10.1109/ICONIP.1999.845633","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
引用次数: 14
Abstract
There are two fundamental weaknesses which may have a great impact on the performance of decision tree (DT) induction. These are the limitations in the ability of the DT language to represent some of the underlying patterns of the domain and the degradation in the quality of evidence available to the induction process caused by its recursive partitioning of the training data. The impact of these two weaknesses is greatest when the induction process attempts to overcome the first weakness by resorting to more partitioning of the training data, thus increasing its vulnerability to the second weakness. The authors investigate the use of multiple DT models as a method of overcoming the limitations of the DT modeling language and describe a new and novel algorithm to automatically generate multiple DT models from the same training data. The algorithm is compared to a single-tree classifier by experiments on two well known data sets. Results clearly demonstrate the superiority of our algorithm.