...
首页> 外文期刊>Applied Measurement in Education >Validating Automated Essay Scoring: A (Modest) Refinement of the 'Gold Standard'
【24h】

Validating Automated Essay Scoring: A (Modest) Refinement of the 'Gold Standard'

机译:验证自动论文评分:“适度”完善“黄金标准”

获取原文
获取原文并翻译 | 示例
           

摘要

By far, the most frequently used method of validating (the interpretation and use of) automated essay scores has been to compare them with scores awarded by human raters. Although this practice is questionable, human-machine agreement is still often regarded as the "gold standard." Our objective was to refine this model and apply it to data from a major testing program and one system of automated essay scoring. The refinement capitalizes on the fact that essay raters differ in numerous ways (e.g., training and experience), any of which may affect the quality of ratings. We found that automated scores exhibited different correlations with scores awarded by experienced raters (a more compelling criterion) than with those awarded by untrained raters (a less compelling criterion). The results suggest potential for a refined machine-human agreement model that differentiates raters with respect to experience, expertise, and possibly even more salient characteristics.
机译:到目前为止,验证(解释和使用)自动作文成绩的最常用方法是将其与人类评分者授予的分数进行比较。尽管这种做法值得怀疑,但是人机协议仍然经常被视为“黄金标准”。我们的目标是完善该模型,并将其应用于主要测试程序和自动作文评分系统中的数据。这种改进利用了散文评分者在许多方面(例如培训和经验)不同的事实,其中任何一种都可能影响评分的质量。我们发现,自动评分与有经验的评分员(较引人注目的标准)所授予的评分相比,与未经培训的评分员(较不引人注目的标准)所授予的评分具有不同的相关性。结果表明,有可能建立一种完善的机器人为协议模型,从而使评估者在经验,专业知识以及可能甚至更显着的特征方面与众不同。

著录项

相似文献

  • 外文文献
  • 中文文献
  • 专利
获取原文

客服邮箱:kefu@zhangqiaokeyan.com

京公网安备:11010802029741号 ICP备案号:京ICP备15016152号-6 六维联合信息科技 (北京) 有限公司©版权所有
  • 客服微信

  • 服务号