4.5 Article

The prevalence of statistical reporting errors in psychology (1985-2013)

Journal

BEHAVIOR RESEARCH METHODS
Volume 48, Issue 4, Pages 1205-1226

Publisher

SPRINGER
DOI: 10.3758/s13428-015-0664-2

Keywords

Reporting errors; p-values; Significance; False positives; NHST; Questionable research practices; Publication bias

Funding

  1. Innovational Research Incentives Scheme Vidi from the Netherlands Organization for Scientific Research [452-11-004]

Ask authors/readers for more resources

This study documents reporting errors in a sample of over 250,000 p-values reported in eight major psychology journals from 1985 until 2013, using the new R package statcheck. statcheck retrieved null-hypothesis significance testing (NHST) results from over half of the articles from this period. In line with earlier research, we found that half of all published psychology papers that use NHST contained at least one p-value that was inconsistent with its test statistic and degrees of freedom. One in eight papers contained a grossly inconsistent p-value that may have affected the statistical conclusion. In contrast to earlier findings, we found that the average prevalence of inconsistent p-values has been stable over the years or has declined. The prevalence of gross inconsistencies was higher in p-values reported as significant than in p-values reported as nonsignificant. This could indicate a systematic bias in favor of significant results. Possible solutions for the high prevalence of reporting inconsistencies could be to encourage sharing data, to let co-authors check results in a so-called co-pilot model, and to use statcheck to flag possible inconsistencies in one's own manuscript or during the review process.

Authors

I am an author on this paper
Click your name to claim this paper and add it to your profile.

Reviews

Primary Rating

4.5
Not enough ratings

Secondary Ratings

Novelty
-
Significance
-
Scientific rigor
-
Rate this paper

Recommended

No Data Available
No Data Available