4.7 Article

Sparse Bayesian Learning With Weakly Informative Hyperprior and Extended Predictive Information Criterion

出版社

IEEE-INST ELECTRICAL ELECTRONICS ENGINEERS INC
DOI: 10.1109/TNNLS.2021.3131357

关键词

Task analysis; Kernel; Bayes methods; Shape; Predictive models; Support vector machines; Mathematical models; Predictive information criterion (PIC); relevance vector machine (RVM); weakly informative prior

向作者/读者索取更多资源

This article discusses the regression problem with sparse Bayesian learning when the number of weights is larger than the data size. The author proposes a strategy consisting of applying an inverse gamma hyperprior to control sparsity and selecting an optimal scale parameter. Empirical evaluation shows that this strategy prevents overfitting and maintains good sparsity.
This article considers the regression problem with sparse Bayesian learning (SBL) when the number of weights P is larger than the data size N, i.e., P(sic) N. The situation induces overfitting and makes regression tasks, such as prediction and basis selection, challenging. We show a strategy to address this problem. Our strategy consists of two steps. The first is to apply an inverse gamma hyperprior with a shape parameter close to zero over the noise precision of automatic relevance determination (ARD) prior. This hyperprior is associated with the concept of a weakly informative prior in terms of enhancing sparsity. The model sparsity can be controlled by adjusting a scale parameter of inverse gamma hyperprior, leading to the prevention of overfitting. The second is to select an optimal scale parameter. We develop an extended predictive information criterion (EPIC) for optimal selection. We investigate the strategy through relevance vector machine (RVM) with a multiple-kernel scheme dealing with highly nonlinear data, including smooth and less smooth regions. This setting is one form of the regression task with SBL in the P(sic) N situation. As an empirical evaluation, regression analyses on four artificial datasets and eight real datasets are performed. We see that the overfitting is prevented, while predictive performance may be not drastically superior to comparative methods. Our methods allow us to select a small number of nonzero weights while keeping the model sparse. Thus, the methods are expected to be useful for basis and variable selection.

作者

我是这篇论文的作者
点击您的名字以认领此论文并将其添加到您的个人资料中。

评论

主要评分

4.7
评分不足

次要评分

新颖性
-
重要性
-
科学严谨性
-
评价这篇论文

推荐

暂无数据
暂无数据