4.7 Article

Tabular data: Deep learning is not all you need

期刊

INFORMATION FUSION
卷 81, 期 -, 页码 84-90

出版社

ELSEVIER
DOI: 10.1016/j.inffus.2021.11.011

关键词

Tabular data; Deep neural networks; Tree-based models; Hyperparameter optimization

向作者/读者索取更多资源

The study compares the performance of deep learning models and XGBoost on various datasets, finding that XGBoost outperforms deep models and requires less tuning. However, an ensemble of deep models and XGBoost performs better on these datasets than XGBoost alone.
A key element in solving real-life data science problems is selecting the types of models to use. Tree ensemble models (such as XGBoost) are usually recommended for classification and regression problems with tabular data. However, several deep learning models for tabular data have recently been proposed, claiming to outperform XGBoost for some use cases. This paper explores whether these deep models should be a recommended option for tabular data by rigorously comparing the new deep models to XGBoost on various datasets. In addition to systematically comparing their performance, we consider the tuning and computation they require. Our study shows that XGBoost outperforms these deep models across the datasets, including the datasets used in the papers that proposed the deep models. We also demonstrate that XGBoost requires much less tuning. On the positive side, we show that an ensemble of deep models and XGBoost performs better on these datasets than XGBoost alone.

作者

我是这篇论文的作者
点击您的名字以认领此论文并将其添加到您的个人资料中。

评论

主要评分

4.7
评分不足

次要评分

新颖性
-
重要性
-
科学严谨性
-
评价这篇论文

推荐

暂无数据
暂无数据