【24h】

Pre-Training BERT on Domain Resources for Short Answer Grading

机译:为简短答案分级对域资源进行预训练BERT

获取原文

摘要

Pre-trained BERT contextualized representations have achieved state-of-the-art results on multiple downstream NLP tasks by fine-tuning with task-specific data. While there has been a lot of focus on task-specific fine-tuning, there has been limited work on improving the pre-trained representations. In this paper, we explore ways of improving the pre-trained contextual representations for the task of automatic short answer grading, a critical component of intelligent tutoring systems. We show that the pre-trained BERT model can be improved by augmenting data from the domain-specific resources like textbooks. We also present a new approach to use labeled short answering grading data for further enhancement of the language model. Empirical evaluation on multi-domain datasets shows that task-specific fine-tuning on the enhanced pre-trained language model achieves superior performance for short answer grading.
机译:通过对特定于任务的数据进行微调,预训练的BERT上下文化表示已在多个下游NLP任务上获得了最新的结果。尽管人们一直将重点放在针对特定任务的微调上,但是在改进预训练表示方面的工作却很少。在本文中,我们探索了为自动简短答案评分(智能补习系统的关键组成部分)的任务改进预训练的上下文表示的方法。我们表明,可以通过增加特定领域资源(如教科书)中的数据来改进预训练的BERT模型。我们还提出了一种使用标记的简短回答评分数据来进一步增强语言模型的新方法。对多域数据集的经验评估表明,增强型预训练语言模型上的特定于任务的微调可为短答案评分提供出色的性能。

著录项

相似文献

  • 外文文献
  • 中文文献
  • 专利
获取原文

客服邮箱:kefu@zhangqiaokeyan.com

京公网安备:11010802029741号 ICP备案号:京ICP备15016152号-6 六维联合信息科技 (北京) 有限公司©版权所有
  • 客服微信

  • 服务号