首页> 外文会议>Workshop on Innovative use of NLP for Building Educational Applications >Using PRMSE to evaluate automated scoring systems in the presence of label noise
【24h】

Using PRMSE to evaluate automated scoring systems in the presence of label noise

机译:使用PRMSE在存在标签噪声时评估自动评分系统

获取原文

摘要

The effect of noisy labels on the performance of NLP systems has been studied extensively for system training. In this paper, we focus on the effect that noisy labels have on system evaluation. Using automated scoring as an example, we demonstrate that the quality of human ratings used for system evaluation have a substantial impact on traditional performance metrics, making it impossible to compare system evaluations on labels with different quality. We propose that a new metric, proportional reduction in mean squared error (PRMSE), developed within the educational measurement community, can help address this issue, and provide practical guidelines on using PRMSE,
机译:对系统培训进行了广泛研究了噪声标签对NLP系统性能的影响。在本文中,我们专注于嘈杂标签对系统评估的影响。使用自动评分作为示例,我们证明了用于系统评估的人类评级质量对传统绩效指标具有大量影响,从而无法对具有不同质量的标签进行比较系统评估。我们建议在教育测量社区中开发的新的公制,比例减少(PRMSE),可以帮助解决这个问题,并提供使用PRMSE的实用指南,

著录项

相似文献

  • 外文文献
  • 中文文献
  • 专利
获取原文

客服邮箱:kefu@zhangqiaokeyan.com

京公网安备:11010802029741号 ICP备案号:京ICP备15016152号-6 六维联合信息科技 (北京) 有限公司©版权所有
  • 客服微信

  • 服务号