Journal
LEARNED PUBLISHING
Volume 23, Issue 2, Pages 124-131Publisher
WILEY
DOI: 10.1087/20100207
Keywords
-
Categories
Ask authors/readers for more resources
If a manuscript meets scientific standards and contributes to the advancement of science, it can be expected that two or more reviewers will agree on its value. Manuscripts are rated reliably when there is a high level of agreement between independent reviewers. This study investigates for the first time whether inter-rater reliability, which is low with the traditional model of closed peer review, is also low with the new system of public peer review or whether higher coefficients can be found for public peer review. To investigate this question we examined the peer-review process practiced by the interactive open access journal Atmospheric Chemistry and Physics (based on 465 manuscripts submitted between 2004 and 2006 receiving 1,058 reviews in total). The results of the study show that inter-rater reliability is low (kappa coefficient) or reasonable (Intraclass Correlation Coefficient) in public peer review.
Authors
I am an author on this paper
Click your name to claim this paper and add it to your profile.
Reviews
Recommended
No Data Available