Lu Wang, Zecheng Yu, Jingling Ma, Xiaofang Chen, Chong Wu
{"title":"A Two-Stage Interpretable Model to Explain Classifier in Credit Risk Prediction","authors":"Lu Wang, Zecheng Yu, Jingling Ma, Xiaofang Chen, Chong Wu","doi":"10.1002/for.3288","DOIUrl":null,"url":null,"abstract":"<div>\n \n <p>In the financial sector, credit risk represents a critical issue, and accurate prediction is essential for mitigating financial risk and ensuring economic stability. Although artificial intelligence methods can achieve satisfactory accuracy, explaining their predictive results poses a significant challenge, thereby prompting research on interpretability. Current research primarily focuses on individual interpretability methods and seldom investigates the combined application of multiple approaches. To address the limitations of existing research, this study proposes a two-stage interpretability model that integrates SHAP and counterfactual explanations. In the first stage, SHAP is employed to analyze feature importance, categorizing features into subsets according to their positive or negative impact on predicted outcomes. In the second stage, a genetic algorithm generates counterfactual explanations by considering feature importance and applying perturbations in various directions based on predefined subsets, thereby accurately identifying counterfactual samples that can modify predicted outcomes. We conducted experiments on the German credit datasets, HMEQ datasets, and the Taiwan Default of Credit Card Clients dataset using SVM, XGB, MLP, and LSTM as base classifiers, respectively. The experimental results indicate that the frequency of feature changes in the counterfactual explanations generated closely aligns with the feature importance derived from the SHAP method. Under the evaluation metrics of effectiveness and sparsity, the performance demonstrates improvements over both basic counterfactual explanation methods and prototype-based counterfactuals. Furthermore, this study offers recommendations based on features derived from SHAP analysis results and counterfactual explanations to reduce the risk of classification as a default.</p>\n </div>","PeriodicalId":47835,"journal":{"name":"Journal of Forecasting","volume":"44 7","pages":"2132-2150"},"PeriodicalIF":2.7000,"publicationDate":"2025-06-02","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"Journal of Forecasting","FirstCategoryId":"96","ListUrlMain":"https://onlinelibrary.wiley.com/doi/10.1002/for.3288","RegionNum":3,"RegionCategory":"经济学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"Q1","JCRName":"ECONOMICS","Score":null,"Total":0}
引用次数: 0
Abstract
In the financial sector, credit risk represents a critical issue, and accurate prediction is essential for mitigating financial risk and ensuring economic stability. Although artificial intelligence methods can achieve satisfactory accuracy, explaining their predictive results poses a significant challenge, thereby prompting research on interpretability. Current research primarily focuses on individual interpretability methods and seldom investigates the combined application of multiple approaches. To address the limitations of existing research, this study proposes a two-stage interpretability model that integrates SHAP and counterfactual explanations. In the first stage, SHAP is employed to analyze feature importance, categorizing features into subsets according to their positive or negative impact on predicted outcomes. In the second stage, a genetic algorithm generates counterfactual explanations by considering feature importance and applying perturbations in various directions based on predefined subsets, thereby accurately identifying counterfactual samples that can modify predicted outcomes. We conducted experiments on the German credit datasets, HMEQ datasets, and the Taiwan Default of Credit Card Clients dataset using SVM, XGB, MLP, and LSTM as base classifiers, respectively. The experimental results indicate that the frequency of feature changes in the counterfactual explanations generated closely aligns with the feature importance derived from the SHAP method. Under the evaluation metrics of effectiveness and sparsity, the performance demonstrates improvements over both basic counterfactual explanation methods and prototype-based counterfactuals. Furthermore, this study offers recommendations based on features derived from SHAP analysis results and counterfactual explanations to reduce the risk of classification as a default.
期刊介绍:
The Journal of Forecasting is an international journal that publishes refereed papers on forecasting. It is multidisciplinary, welcoming papers dealing with any aspect of forecasting: theoretical, practical, computational and methodological. A broad interpretation of the topic is taken with approaches from various subject areas, such as statistics, economics, psychology, systems engineering and social sciences, all encouraged. Furthermore, the Journal welcomes a wide diversity of applications in such fields as business, government, technology and the environment. Of particular interest are papers dealing with modelling issues and the relationship of forecasting systems to decision-making processes.