4.4 Article

An empirical bias-variance analysis of DECORATE ensemble method at different training sample sizes

期刊

JOURNAL OF APPLIED STATISTICS
卷 39, 期 4, 页码 829-850

出版社

TAYLOR & FRANCIS LTD
DOI: 10.1080/02664763.2011.620949

关键词

classifier combination method; AdaBoost; random forest; bias-variance decomposition; training sample size

资金

  1. National Natural Science Foundation of China [61075006]
  2. Research Fund for the Doctoral Program of Higher Education of China [20100201120048]
  3. Fundamental Research Funds for the Central Universities of China

向作者/读者索取更多资源

DECORATE (Diverse Ensemble Creation by Oppositional Relabeling of Artificial Training Examples) is a classifier combination technique to construct a set of diverse base classifiers using additional artificially generated training instances. The predictions from the base classifiers are then integrated into one by the mean combination rule. In order to gain more insight about its effectiveness and advantages, this paper utilizes a large experiment to study the bias-variance analysis of DECORATE as well as some other widely used ensemble methods (such as bagging, AdaBoost, random forest) at different training sample sizes. The experimental results yield the following conclusions. For small training sets, DECORATE has a dominant advantage over its rivals and its success is attributed to the larger bias reduction achieved by it than the other algorithms. With increase in training data, AdaBoost benefits most and the bias reduced by it gradually turns to be significant while its variance reduction is also medium. Thus, AdaBoost performs best with large training samples. Moreover, random forest behaves always second best regardless of small or large training sets and it is seen to mainly decrease variance while maintaining low bias. Bagging seems to be an intermediate one since it reduces variance primarily.

作者

我是这篇论文的作者
点击您的名字以认领此论文并将其添加到您的个人资料中。

评论

主要评分

4.4
评分不足

次要评分

新颖性
-
重要性
-
科学严谨性
-
评价这篇论文

推荐

暂无数据
暂无数据