Narayana Darapaneni, Pramod Srinivas, K. Reddy, A. Paduri, Lakshmikanth Kanugovi, Pavithra J, S. B. G., B. S
{"title":"Tree Based Models: A Comparative And Explainable Study For Credit Default Classification","authors":"Narayana Darapaneni, Pramod Srinivas, K. Reddy, A. Paduri, Lakshmikanth Kanugovi, Pavithra J, S. B. G., B. S","doi":"10.1109/UPCON56432.2022.9986411","DOIUrl":null,"url":null,"abstract":"Most of the real-world problems can be solved with the help of classification. Some problems that might feel to be regression can also be solved using classification by binning the values. But due to the presence of a large number of classification algorithms, it becomes difficult to choose one particular model. SVC, KNN, and Naive Bayes are some of the traditional models available which do the task quite efficiently, but when it comes to the explainability of the outcome these models fail to address them. This is when the real use of tree-based models comes in handy as we can visualize the entire flow of conditions that leads to a particular outcome. In this paper, we have tried to compare the performance and explainability of the Decision tree, Random Forest, AdaBoost, GBoost, and XGBoost for this we have considered the credit default dataset from Kaggle.","PeriodicalId":185782,"journal":{"name":"2022 IEEE 9th Uttar Pradesh Section International Conference on Electrical, Electronics and Computer Engineering (UPCON)","volume":"25 1","pages":"0"},"PeriodicalIF":0.0000,"publicationDate":"2022-12-02","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"1","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"2022 IEEE 9th Uttar Pradesh Section International Conference on Electrical, Electronics and Computer Engineering (UPCON)","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/10.1109/UPCON56432.2022.9986411","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
引用次数: 1
Abstract
Most of the real-world problems can be solved with the help of classification. Some problems that might feel to be regression can also be solved using classification by binning the values. But due to the presence of a large number of classification algorithms, it becomes difficult to choose one particular model. SVC, KNN, and Naive Bayes are some of the traditional models available which do the task quite efficiently, but when it comes to the explainability of the outcome these models fail to address them. This is when the real use of tree-based models comes in handy as we can visualize the entire flow of conditions that leads to a particular outcome. In this paper, we have tried to compare the performance and explainability of the Decision tree, Random Forest, AdaBoost, GBoost, and XGBoost for this we have considered the credit default dataset from Kaggle.