4.1 Article

Automated Scoring of Constructed-Response Science Items: Prospects and Obstacles

期刊

出版社

WILEY
DOI: 10.1111/emip.12028

关键词

automated scoring; constructed-response items; c-rater (TM); science assessment

资金

  1. Direct For Education and Human Resources [0822388] Funding Source: National Science Foundation
  2. Division Of Research On Learning [0822388] Funding Source: National Science Foundation
  3. Division Of Research On Learning
  4. Direct For Education and Human Resources [1119670] Funding Source: National Science Foundation

向作者/读者索取更多资源

Content-based automated scoring has been applied in a variety of science domains. However, many prior applications involved simplified scoring rubrics without considering rubrics representing multiple levels of understanding. This study tested a concept-based scoring tool for content-based scoring, c-rater (TM), for four science items with rubrics aiming to differentiate among multiple levels of understanding. The items showed moderate to good agreement with human scores. The findings suggest that automated scoring has the potential to score constructed-response items with complex scoring rubrics, but in its current design cannot replace human raters. This article discusses sources of disagreement and factors that could potentially improve the accuracy of concept-based automated scoring.

作者

我是这篇论文的作者
点击您的名字以认领此论文并将其添加到您的个人资料中。

评论

主要评分

4.1
评分不足

次要评分

新颖性
-
重要性
-
科学严谨性
-
评价这篇论文

推荐

暂无数据
暂无数据