4.6 Article

Explainability of Machine Learning Models for Bankruptcy Prediction

Journal

IEEE ACCESS
Volume 9, Issue -, Pages 124887-124899

Publisher

IEEE-INST ELECTRICAL ELECTRONICS ENGINEERS INC
DOI: 10.1109/ACCESS.2021.3110270

Keywords

Predictive models; Bankruptcy; Data models; Machine learning; Companies; Feature extraction; Analytical models; Bankruptcy prediction; machine learning; explainable AI; feature importance

Funding

  1. National Research Foundation of Korea (NRF) Grants through Korean Government [Ministry of Science and ICT (MSIT)] [NRF-2017R1E1A1A03070105, NRF-2019R1A5A1028324]
  2. Institute for the Information and Communications Technology Promotion (IITP) Grant through Korean Government [Ministry of Science, ICT and Future Planning (MSIP)] [2019-0-01906]
  3. Information Technology Research Center (ITRC) [IITP-2020-2018-0-01441]

Ask authors/readers for more resources

With the increasing amount of data, there is a growing demand for utilizing machine learning methodologies in bankruptcy prediction. This study focuses on the interpretability of machine learning models, particularly measuring feature importance using the LIME algorithm, to address the black-box issue.
As the amount of data increases, it is more likely that the assumptions in the existing economic analysis model are unsatisfied or make it difficult to establish a new analysis model. Therefore, there has been increased demand for applying the machine learning methodology to bankruptcy prediction due to its high performance. By contrast, machine learning models usually operate as black-boxes but credit rating regulatory systems require the provisioning of appropriate information regarding credit rating standards. If machine learning models have sufficient interpretablility, they would have the potential to be used as effective analytical models in bankruptcy prediction. From this aspect, we study the explainability of machine learning models for bankruptcy prediction by applying the Local Interpretable Model-Agnostic Explanations (LIME) algorithm, which measures the feature importance for each data point. To compare how the feature importance measured through LIME differs from that of models themselves, we first applied this algorithm to typical tree-based models that have ability to measure the feature importance of the models themselves. We showed that the feature importance measured through LIME could be a consistent generalization of the feature importance measured by tree-based models themselves. Moreover, we study the consistency of the feature importance through the model's predicted bankruptcy probability, which suggests the possibility that observations of important features can be used as a basis for the fair treatment of loan eligibility requirements.

Authors

I am an author on this paper
Click your name to claim this paper and add it to your profile.

Reviews

Primary Rating

4.6
Not enough ratings

Secondary Ratings

Novelty
-
Significance
-
Scientific rigor
-
Rate this paper

Recommended

No Data Available
No Data Available