4.7 Article

Explainable machine learning can outperform Cox regression predictions and provide insights in breast cancer survival

期刊

SCIENTIFIC REPORTS
卷 11, 期 1, 页码 -

出版社

NATURE PORTFOLIO
DOI: 10.1038/s41598-021-86327-7

关键词

-

向作者/读者索取更多资源

This study compares the performance of Cox Proportional Hazards (CPH) analysis and machine learning techniques in predicting survival rates for non-metastatic breast cancer patients. The results show that machine learning models can outperform classical methods, with Extreme Gradient Boosting (XGB) performing even better. By utilizing Shapley Additive Explanation (SHAP) values, the study explains the predictions of the models, demonstrating the importance of explainable machine learning in healthcare.
Cox Proportional Hazards (CPH) analysis is the standard for survival analysis in oncology. Recently, several machine learning (ML) techniques have been adapted for this task. Although they have shown to yield results at least as good as classical methods, they are often disregarded because of their lack of transparency and little to no explainability, which are key for their adoption in clinical settings. In this paper, we used data from the Netherlands Cancer Registry of 36,658 non-metastatic breast cancer patients to compare the performance of CPH with ML techniques (Random Survival Forests, Survival Support Vector Machines, and Extreme Gradient Boosting [XGB]) in predicting survival using the c -index. We demonstrated that in our dataset, ML-based models can perform at least as good as the classical CPH regression ( c-index similar to 0.63), and in the case of XGB even better ( c-index similar to 0.73). Furthermore, we used Shapley Additive Explanation (SHAP) values to explain the models' predictions. We concluded that the difference in performance can be attributed to XGB's ability to model nonlinearities and complex interactions. We also investigated the impact of specific features on the models' predictions as well as their corresponding insights. Lastly, we showed that explainable ML can generate explicit knowledge of how models make their predictions, which is crucial in increasing the trust and adoption of innovative ML techniques in oncology and healthcare overall.

作者

我是这篇论文的作者
点击您的名字以认领此论文并将其添加到您的个人资料中。

评论

主要评分

4.7
评分不足

次要评分

新颖性
-
重要性
-
科学严谨性
-
评价这篇论文

推荐

暂无数据
暂无数据