4.6 Article

Gradient boosting Bayesian neural networks via Langevin MCMC

期刊

NEUROCOMPUTING
卷 558, 期 -, 页码 -

出版社

ELSEVIER
DOI: 10.1016/j.neucom.2023.126726

关键词

MCMC; Gradient boosting; Ensemble learning; Bayesian neural networks

向作者/读者索取更多资源

This paper presents a Bayesian ensemble learning framework that utilizes Bayesian inference to improve prediction accuracy and quantify uncertainty. By combining multiple base learners and training them using MCMC sampling, the framework has good scalability on large-scale models.
Bayesian neural networks harness the power of Bayesian inference which provides an approach to neural learning that not only focuses on accuracy but also uncertainty quantification. Markov Chain Monte Carlo (MCMC) methods implement Bayesian inference by sampling from the posterior distribution of the model parameters. In the case of Bayesian neural networks, the model parameters refer to weights and biases. MCMC methods suffer from scalability issues in large models, such as deep neural networks with thousands to millions of parameters. In this paper, we present a Bayesian ensemble learning framework that utilizes gradient boosting by combining multiple shallow neural networks (base learners) that are trained by MCMC sampling. We present two Bayesian gradient boosting strategies that employ simple neural networks as base learners with Langevin MCMC sampling. We evaluate the performance of these methods on various classification and time -series prediction problems. We demonstrate that the proposed framework improves the prediction accuracy of canonical gradient boosting while providing uncertainty quantification via Bayesian inference. Furthermore, we demonstrate that the respective methods scale well when the size of the dataset and model increases.

作者

我是这篇论文的作者
点击您的名字以认领此论文并将其添加到您的个人资料中。

评论

主要评分

4.6
评分不足

次要评分

新颖性
-
重要性
-
科学严谨性
-
评价这篇论文

推荐

暂无数据
暂无数据