4.5 Article

A validation of Amazon Mechanical Turk for the collection of acceptability judgments in linguistic theory

期刊

BEHAVIOR RESEARCH METHODS
卷 43, 期 1, 页码 155-167

出版社

PSYCHONOMIC SOC INC
DOI: 10.3758/s13428-010-0039-7

关键词

Amazon Mechanical Turk; Acceptability judgments; Grammaticality judgments; Experimental syntax; Linguistic theory

资金

  1. Division Of Behavioral and Cognitive Sci
  2. Direct For Social, Behav & Economic Scie [0843896] Funding Source: National Science Foundation

向作者/读者索取更多资源

Amazon's Mechanical Turk (AMT) is a Web application that provides instant access to thousands of potential participants for survey-based psychology experiments, such as the acceptability judgment task used extensively in syntactic theory. Because AMT is a Web-based system, syntacticians may worry that the move out of the experimenter-controlled environment of the laboratory and onto the user-controlled environment of AMT could adversely affect the quality of the judgment data collected. This article reports a quantitative comparison of two identical acceptability judgment experiments, each with 176 participants (352 total): one conducted in the laboratory, and one conducted on AMT. Crucial indicators of data quality-such as participant rejection rates, statistical power, and the shape of the distributions of the judgments for each sentence type-are compared between the two samples. The results suggest that aside from slightly higher participant rejection rates, AMT data are almost indistinguishable from laboratory data.

作者

我是这篇论文的作者
点击您的名字以认领此论文并将其添加到您的个人资料中。

评论

主要评分

4.5
评分不足

次要评分

新颖性
-
重要性
-
科学严谨性
-
评价这篇论文

推荐

暂无数据
暂无数据