4.4 Article

Interpretation of machine learning models using shapley values: application to compound potency and multi-target activity predictions

Journal

JOURNAL OF COMPUTER-AIDED MOLECULAR DESIGN
Volume 34, Issue 10, Pages 1013-1026

Publisher

SPRINGER
DOI: 10.1007/s10822-020-00314-0

Keywords

Machine learning; Black box character; Structure-activity relationships; Compound activity; Compound potency prediction; Multi-target modeling; Model interpretation; Feature importance; Shapley values

Funding

  1. Projekt DEAL

Ask authors/readers for more resources

Difficulties in interpreting machine learning (ML) models and their predictions limit the practical applicability of and confidence in ML in pharmaceutical research. There is a need for agnostic approaches aiding in the interpretation of ML models regardless of their complexity that is also applicable to deep neural network (DNN) architectures and model ensembles. To these ends, the SHapley Additive exPlanations (SHAP) methodology has recently been introduced. The SHAP approach enables the identification and prioritization of features that determine compound classification and activity prediction using any ML model. Herein, we further extend the evaluation of the SHAP methodology by investigating a variant for exact calculation of Shapley values for decision tree methods and systematically compare this variant in compound activity and potency value predictions with the model-independent SHAP method. Moreover, new applications of the SHAP analysis approach are presented including interpretation of DNN models for the generation of multi-target activity profiles and ensemble regression models for potency prediction.

Authors

I am an author on this paper
Click your name to claim this paper and add it to your profile.

Reviews

Primary Rating

4.4
Not enough ratings

Secondary Ratings

Novelty
-
Significance
-
Scientific rigor
-
Rate this paper

Recommended

No Data Available
No Data Available