首页> 外文会议>International Conference on Intelligent Tutoring Systems >Automated Paraphrase Quality Assessment Using Recurrent Neural Networks and Language Models
【24h】

Automated Paraphrase Quality Assessment Using Recurrent Neural Networks and Language Models

机译:使用递归神经网络和语言模型的自动复述质量评估

获取原文

摘要

The ability to automatically assess the quality of paraphrases can be very useful for facilitating literacy skills and providing timely feedback to learners. Our aim is twofold: a) to automatically evaluate the quality of paraphrases across four dimensions: lexical similarity, syntactic similarity, semantic similarity and paraphrase quality, and b) to assess how well models trained for this task generalize. The task is modeled as a classification problem and three different methods are explored: a) manual feature extraction combined with an Extra Trees model, b) GloVe embeddings and a Siamese neural network, and c) using a pretrained BERT model fine-tuned on our task. Starting from a dataset of 1998 paraphrases from the User Language Paraphrase Corpus (ULPC), we explore how the three models trained on the ULPC dataset generalize when applied on a separate, small paraphrase corpus based on children inputs. The best out-of-the-box generalization performance is obtained by the Extra Trees model with at least 75% average F1-scores for the three similarity dimensions. We also show that the Siamese neural network and BERT models can obtain an improvement of at least 5% after fine-tuning across all dimensions.
机译:自动评估释义质量的能力对于提高识字技能和向学习者提供及时反馈非常有用。我们的目标有两个:a)自动评估四个维度的释义质量:词汇相似性、句法相似性、语义相似性和释义质量,以及b)评估为这项任务训练的模型的通用性。该任务被建模为一个分类问题,并探索了三种不同的方法:a)结合额外树模型的手动特征提取,b)手套嵌入和连体神经网络,以及c)使用经过微调的预训练伯特模型。从用户语言释义语料库(ULPC)1998年的释义数据集开始,我们探讨了在ULPC数据集上训练的三个模型在基于儿童输入的单独小型释义语料库上的应用时是如何概括的。通过额外树模型获得最佳的开箱即用泛化性能,三个相似维度的F1平均分数至少为75%。我们还表明,在对所有维度进行微调后,暹罗神经网络和伯特模型可以获得至少5%的改善。

著录项

相似文献

  • 外文文献
  • 中文文献
  • 专利
获取原文

客服邮箱:kefu@zhangqiaokeyan.com

京公网安备:11010802029741号 ICP备案号:京ICP备15016152号-6 六维联合信息科技 (北京) 有限公司©版权所有
  • 客服微信

  • 服务号