4.7 Article

An evidential reasoning rule based feature selection for improving trauma outcome prediction

期刊

APPLIED SOFT COMPUTING
卷 103, 期 -, 页码 -

出版社

ELSEVIER
DOI: 10.1016/j.asoc.2021.107112

关键词

Feature selection; Trauma; Evidential reasoning rule; Random forest; ReliefF; Imbalance classes

资金

  1. Saudi Arabian Government
  2. EU project, United Kingdom [823759]
  3. NSFC-Zhejiang Joint Fund for the Integration of Industrialization and Informatization, United Kingdom [U1709215]

向作者/读者索取更多资源

Key features for accurately predicting patient outcomes can be selected through random forest, ReliefF, and evidential reasoning (ER) rule. The impact of outcome class imbalance on feature selection is discussed, with synthetic minority over-sampling technique (SMOTE) showing differences in selected features. The highest prediction performance is achieved by the ER rule for selecting features, followed by ReliefF and random forest.
Various demographic and medical factors can be linked to severe deterioration of patients suffering from traumatic injuries. Accurate identification of the most relevant variables is essential for building more accurate prediction models and making more rapid life-saving medical decision. The intention of this paper is to select a number of features that can be used to accurately predict patients' outcomes through three feature selection methods: random forest, ReliefF and the evidential reasoning (ER) rule. The impact of an outcome's class imbalance on feature selection is discussed, and synthetic minority over-sampling technique (SMOTE) is performed to show the differences in the selected features. The results show that length of stay in hospital, length of stay in intensive care unit, age, and Glasgow Coma Scale (GCS) are the most selected features across different techniques. The prediction models based on the features selected by the ER rule show the highest prediction performance represented by the area under the receiver operating characteristic curve (AUC) values, which has a median of 0.895 for the model employed by the ten highest-weighted variables, while the median AUC values are 0.827 and 0.885 if the ten highest-weighted variables are selected by ReliefF and random forest respectively. The results also show that after the ten most important features, increasing the number of the less important features has only a slight increase in prediction accuracy. (C) 2021 Published by Elsevier B.V.

作者

我是这篇论文的作者
点击您的名字以认领此论文并将其添加到您的个人资料中。

评论

主要评分

4.7
评分不足

次要评分

新颖性
-
重要性
-
科学严谨性
-
评价这篇论文

推荐

暂无数据
暂无数据