4.7 Article

An experimental comparison of three methods for constructing ensembles of decision trees: Bagging, boosting, and randomization

期刊

MACHINE LEARNING
卷 40, 期 2, 页码 139-157

出版社

KLUWER ACADEMIC PUBL
DOI: 10.1023/A:1007607513941

关键词

decision trees; ensemble learning; bagging; boosting; C4.5; Monte Carlo methods

向作者/读者索取更多资源

Bagging and boosting are methods that generate a diverse ensemble of classifiers by manipulating the training data given to a base learning algorithm. Breiman has pointed out that they rely for their effectiveness on the instability of the base learning algorithm. An alternative approach to generating an ensemble is to randomize the internal decisions made by the base algorithm. This general approach has been studied previously by Ali and Pazzani and by Dietterich and Kong. This paper compares the effectiveness of randomization, bagging, and boosting for improving the performance of the decision-tree algorithm C4.5. The experiments show that in situations with little or no classification noise, randomization is competitive with (and perhaps slightly superior to) bagging but not as accurate as boosting. In situations with substantial classification noise, bagging is much better than boosting, and sometimes better than randomization.

作者

我是这篇论文的作者
点击您的名字以认领此论文并将其添加到您的个人资料中。

评论

主要评分

4.7
评分不足

次要评分

新颖性
-
重要性
-
科学严谨性
-
评价这篇论文

推荐

暂无数据
暂无数据