4.2 Article

Model selection bias and Freedman's paradox

Journal

Publisher

SPRINGER HEIDELBERG
DOI: 10.1007/s10463-009-0234-4

Keywords

Akaike's information criterion; Confidence interval coverage; Freedman's paradox; Model averaging; Model selection bias; Model selection uncertainty; Multimodel inference; Stepwise selection

Funding

  1. U.S. Geological Survey/ Biological Resources Division
  2. Colorado Division of Wildlife

Ask authors/readers for more resources

In situations where limited knowledge of a system exists and the ratio of data points to variables is small, variable selection methods can often be misleading. Freedman (Am Stat 37:152-155, 1983) demonstrated how common it is to select completely unrelated variables as highly significant when the number of data points is similar in magnitude to the number of variables. A new type of model averaging estimator based on model selection with Akaike's AIC is used with linear regression to investigate the problems of likely inclusion of spurious effects and model selection bias, the bias introduced while using the data to select a single seemingly best model from a (often large) set of models employing many predictor variables. The new model averaging estimator helps reduce these problems and provides confidence interval coverage at the nominal level while traditional stepwise selection has poor inferential properties.

Authors

I am an author on this paper
Click your name to claim this paper and add it to your profile.

Reviews

Primary Rating

4.2
Not enough ratings

Secondary Ratings

Novelty
-
Significance
-
Scientific rigor
-
Rate this paper

Recommended

No Data Available
No Data Available