首页> 外文会议>SIGBioMed workshop on biomedical natural language processing;Annual meeting of the Association for Computational Linguistics >PANLP at MEDIQA 2019: Pre-trained Language Models, Transfer Learning and Knowledge Distillation
【24h】

PANLP at MEDIQA 2019: Pre-trained Language Models, Transfer Learning and Knowledge Distillation

机译:PANLP在MEDIQA 2019上:预训练的语言模型,迁移学习和知识提炼

获取原文

摘要

This paper describes the models designated for the MEDIQA 2019 shared tasks by the team PANLP. We take advantages of the recent advances in pre-trained bidirectional transformer language models such as BERT (Devlin et al., 2018) and MT-DNN (Liu et al., 2019b). We find that pre-trained language models can significantly outperform traditional deep learning models. Transfer learning from the NLI task to the RQE task is also experimented, which proves to be useful in improving the results of fine-tuning MT-DNN large. A knowledge distillation process is implemented, to distill the knowledge contained in a set of models and transfer it into an single model, whose performance turns out to be comparable with that obtained by the ensemble of that set of models. Finally, for test submissions, model ensemble and a re-ranking process are implemented to boost the performances. Our models participated in all three tasks and ranked the 1st place for the RQE task, and the 2nd place for the NLI task, and also the 2nd place for the QA task.
机译:本文介绍了由PANLP团队指定用于MEDIQA 2019共享任务的模型。我们利用BERT(Devlin等人,2018)和MT-DNN(Liu等人,2019b)等预训练双向转换器语言模型的最新进展。我们发现,经过预训练的语言模型可以大大优于传统的深度学习模型。还对从NLI任务到RQE任务的转移学习进行了实验,这被证明有助于改善MT-DNN的微调结果。实施知识提炼过程,以提炼一组模型中包含的知识并将其转移到单个模型中,该模型的性能与该组模型的集成所获得的性能相当。最后,对于测试提交,实施了模型集成和重新排序过程以提高性能。我们的模型参与了所有三个任务,在RQE任务中排名第一,在NLI任务中排名第二,在QA任务中排名第二。

著录项

相似文献

  • 外文文献
  • 中文文献
  • 专利
获取原文

客服邮箱:kefu@zhangqiaokeyan.com

京公网安备:11010802029741号 ICP备案号:京ICP备15016152号-6 六维联合信息科技 (北京) 有限公司©版权所有
  • 客服微信

  • 服务号