4.6 Article

Prediction, Estimation, and Attribution

Journal

JOURNAL OF THE AMERICAN STATISTICAL ASSOCIATION
Volume 115, Issue 530, Pages 636-655

Publisher

TAYLOR & FRANCIS INC
DOI: 10.1080/01621459.2020.1762613

Keywords

Black box; Ephemeral predictors; Random forests; Surface plus noise

Ask authors/readers for more resources

The scientific needs and computational limitations of the twentieth century fashioned classical statistical methodology. Both the needs and limitations have changed in the twenty-first, and so has the methodology. Large-scale prediction algorithms-neural nets, deep learning, boosting, support vector machines, random forests-have achieved star status in the popular press. They are recognizable as heirs to the regression tradition, but ones carried out at enormous scale and on titanic datasets. How do these algorithms compare with standard regression techniques such as ordinary least squares or logistic regression? Several key discrepancies will be examined, centering on the differences between prediction and estimation or prediction and attribution (significance testing). Most of the discussion is carried out through small numerical examples.

Authors

I am an author on this paper
Click your name to claim this paper and add it to your profile.

Reviews

Primary Rating

4.6
Not enough ratings

Secondary Ratings

Novelty
-
Significance
-
Scientific rigor
-
Rate this paper

Recommended

No Data Available
No Data Available