4.2 Article

Assessing L2 English speaking using automated scoring technology: examining automarker reliability

期刊

出版社

ROUTLEDGE JOURNALS, TAYLOR & FRANCIS LTD
DOI: 10.1080/0969594X.2021.1979467

关键词

Automated scoring; L2 speaking assessment; limits of agreement

向作者/读者索取更多资源

This study found that the reliability of the automarker in online oral English test is good, but it tends to be more lenient towards low-proficiency speakers. The uncertainty measure named Language Quality, which indicates the confidence of speech recognition, was found to be useful in predicting reliability and identifying abnormal speech.
Recent advances in machine learning have made automated scoring of learner speech widespread, and yet validation research that provides support for applying automated scoring technology to assessment is still in its infancy. Both the educational measurement and language assessment communities have called for greater transparency in describing scoring algorithms and research evidence about the reliability of automated scoring. This paper reports on a study that investigated the reliability of an automarker using candidate responses produced in an online oral English test. Based on 'limits of agreement' and multi-faceted Rasch analyses on automarker scores and individual examiner scores, the study found that the automarker, while exhibiting excellent internal consistency, was slightly more lenient than examiner fair average scores, particularly for low-proficiency speakers. Additionally, it was found that an automarker uncertainty measure termed Language Quality, which indicates the confidence of speech recognition, was useful for predicting automarker reliability and flagging abnormal speech.

作者

我是这篇论文的作者
点击您的名字以认领此论文并将其添加到您的个人资料中。

评论

主要评分

4.2
评分不足

次要评分

新颖性
-
重要性
-
科学严谨性
-
评价这篇论文

推荐

暂无数据
暂无数据