首页> 外文会议>Conference on Computational Natural Language Learning >Cloze Distillation: Improving Neural Language Models with Human Next-Word Predictions
【24h】

Cloze Distillation: Improving Neural Language Models with Human Next-Word Predictions

机译:强化蒸馏:改善人类下一词预测的神经语言模型

获取原文

摘要

Contemporary autoregressive language models (LMs) trained purely on corpus data have been shown to capture numerous features of human incremental processing. However, past work has also suggested dissociations between corpus probabilities and human next-word predictions. Here we evaluate several state-of-the-art language models for their match to human next-word predictions and to reading time behavior from eye movements. We then propose a novel method for distilling the linguistic information implicit in human linguistic predictions into pre-trained LMs: Cloze Distillation. We apply this method to a baseline neural LM and show potential improvement in reading time prediction and generalization to held-out human cloze data.
机译:已经显示了当代自回归语言模型(LMS)训练的语料库数据,以捕获人类增量处理的许多特征。然而,过去的工作也建议在语料库概率和人类下一词预测之间进行解散。在这里,我们评估了几种最先进的语言模型,以便与人类的下一词预测匹配并从眼睛运动中读取时间行为。然后,我们提出了一种新的方法,用于将人类语言预测中隐含的语言信息蒸馏到预先训练的LMS:强缩蒸馏。我们将该方法应用于基线神经LM,并显示读取时间预测和概括的潜在改进,以阻止人的渗出数据。

著录项

相似文献

  • 外文文献
  • 中文文献
  • 专利
获取原文

客服邮箱:kefu@zhangqiaokeyan.com

京公网安备:11010802029741号 ICP备案号:京ICP备15016152号-6 六维联合信息科技 (北京) 有限公司©版权所有
  • 客服微信

  • 服务号