4.2 Article

Bootstrap analysis of multiple repetitions of experiments using an interval-valued multiple comparison procedure

Journal

JOURNAL OF COMPUTER AND SYSTEM SCIENCES
Volume 80, Issue 1, Pages 88-100

Publisher

ACADEMIC PRESS INC ELSEVIER SCIENCE
DOI: 10.1016/j.jcss.2013.03.009

Keywords

Cross validation; Statistical comparisons of algorithms; Tests for interval-valued data

Funding

  1. Spanish Ministry of Economy and Competitiveness [TIN2011-24302]

Ask authors/readers for more resources

A new bootstrap test is introduced that allows for assessing the significance of the differences between stochastic algorithms in a cross-validation with repeated folds experimental setup. Intervals are used for modeling the variability of the data that can be attributed to the repetition of learning and testing stages over the same folds in cross validation. Numerical experiments are provided that support the following three claims: (1) Bootstrap tests can be more powerful than ANOVA or Friedman test for comparing multiple classifiers. (2) In the presence of outliers, interval-valued bootstrap tests achieve a better discrimination between stochastic algorithms than nonparametric tests. (3) Choosing ANOVA, Friedman or Bootstrap can produce different conclusions in experiments involving actual data from machine learning tasks. (C) 2013 Elsevier Inc. All rights reserved.

Authors

I am an author on this paper
Click your name to claim this paper and add it to your profile.

Reviews

Primary Rating

4.2
Not enough ratings

Secondary Ratings

Novelty
-
Significance
-
Scientific rigor
-
Rate this paper

Recommended

No Data Available
No Data Available