【24h】

Summarization Evaluation meets Short-Answer Grading

机译:摘要评估符合短答案评分

获取原文

摘要

Summarization Evaluation and Short-Answer Grading share the challenge of automatically evaluating content quality. Therefore, we explore the use of ROUGE, a well-known Summarization Evaluation method, for Short-Answer Grading. We find a reliable ROUGE parametrization that is robust across corpora and languages and produces scores that are significantly cor-related with human short-answer grades. ROUGE adds no information to Short-Answer Grading NLP-based machine learn-ing features in a by-corpus evaluation. However, on a question-by-question basis, we find that the ROUGE Recall score may outperform standard NLP features. We therefore suggest to use ROUGE within a framework for per-question feature se-lection or as a reliable and reproducible baseline for SAG.
机译:摘要评估和缩写分级份额分享自动评估内容质量的挑战。因此,我们探索胭脂,众所周知的摘要评估方法,用于短答案分级。我们发现一个可靠的胭脂参数化参数化,跨越语言和语言强大,并产生与人类短答复等级有关的分数。 Rouge在逐个语料库评估中增加了基于基于NLP的基于NLP的机器学习功能的信息。但是,在一个问题的基础上,我们发现Rouge召回得分可能优于标准的NLP功能。因此,我们建议在框架内使用Rouge进行架构特征SE-enction或作为下垂的可靠和可重复的基线。

著录项

相似文献

  • 外文文献
  • 中文文献
  • 专利
获取原文

客服邮箱:kefu@zhangqiaokeyan.com

京公网安备:11010802029741号 ICP备案号:京ICP备15016152号-6 六维联合信息科技 (北京) 有限公司©版权所有
  • 客服微信

  • 服务号