Author
Listed:
- Lu Wang
- Zecheng Yu
- Jingling Ma
- Xiaofang Chen
- Chong Wu
Abstract
In the financial sector, credit risk represents a critical issue, and accurate prediction is essential for mitigating financial risk and ensuring economic stability. Although artificial intelligence methods can achieve satisfactory accuracy, explaining their predictive results poses a significant challenge, thereby prompting research on interpretability. Current research primarily focuses on individual interpretability methods and seldom investigates the combined application of multiple approaches. To address the limitations of existing research, this study proposes a two‐stage interpretability model that integrates SHAP and counterfactual explanations. In the first stage, SHAP is employed to analyze feature importance, categorizing features into subsets according to their positive or negative impact on predicted outcomes. In the second stage, a genetic algorithm generates counterfactual explanations by considering feature importance and applying perturbations in various directions based on predefined subsets, thereby accurately identifying counterfactual samples that can modify predicted outcomes. We conducted experiments on the German credit datasets, HMEQ datasets, and the Taiwan Default of Credit Card Clients dataset using SVM, XGB, MLP, and LSTM as base classifiers, respectively. The experimental results indicate that the frequency of feature changes in the counterfactual explanations generated closely aligns with the feature importance derived from the SHAP method. Under the evaluation metrics of effectiveness and sparsity, the performance demonstrates improvements over both basic counterfactual explanation methods and prototype‐based counterfactuals. Furthermore, this study offers recommendations based on features derived from SHAP analysis results and counterfactual explanations to reduce the risk of classification as a default.
Suggested Citation
Lu Wang & Zecheng Yu & Jingling Ma & Xiaofang Chen & Chong Wu, 2025.
"A Two‐Stage Interpretable Model to Explain Classifier in Credit Risk Prediction,"
Journal of Forecasting, John Wiley & Sons, Ltd., vol. 44(7), pages 2132-2150, November.
Handle:
RePEc:wly:jforec:v:44:y:2025:i:7:p:2132-2150
DOI: 10.1002/for.3288
Download full text from publisher
Corrections
All material on this site has been provided by the respective publishers and authors. You can help correct errors and omissions. When requesting a correction, please mention this item's handle: RePEc:wly:jforec:v:44:y:2025:i:7:p:2132-2150. See general information about how to correct material in RePEc.
If you have authored this item and are not yet registered with RePEc, we encourage you to do it here. This allows to link your profile to this item. It also allows you to accept potential citations to this item that we are uncertain about.
We have no bibliographic references for this item. You can help adding them by using this form .
If you know of missing items citing this one, you can help us creating those links by adding the relevant references in the same way as above, for each refering item. If you are a registered author of this item, you may also want to check the "citations" tab in your RePEc Author Service profile, as there may be some citations waiting for confirmation.
For technical questions regarding this item, or to correct its authors, title, abstract, bibliographic or download information, contact: Wiley Content Delivery (email available below). General contact details of provider: http://www3.interscience.wiley.com/cgi-bin/jhome/2966 .
Please note that corrections may take a couple of weeks to filter through
the various RePEc services.