首页> 外文会议>SIGBioMed workshop on biomedical natural language processing;Annual meeting of the Association for Computational Linguistics >KU ai at MEDIQA 2019: Domain-specific Pre-training and Transfer Learning for Medical NLI
【24h】

KU ai at MEDIQA 2019: Domain-specific Pre-training and Transfer Learning for Medical NLI

机译:KU Ai在MEDIQA 2019:针对医学NLI的领域特定的预培训和转移学习

获取原文
获取外文期刊封面目录资料

摘要

In this paper, we describe our system and results submitted for the Natural Language Inference (NLI) track of the MEDIQA 2019 Shared Task (Ben Abacha et al., 2019). As KLLai team, we used BERT (Devlin et al., 2018) as our baseline model and pre-processed the MedNLI dataset to mitigate the negative impact of de-identification artifacts. Moreover, we investigated different pre-training and transfer learning approaches to improve the performance. We show that pre-training the language model on rich biomedical corpora has a significant effect in teaching the model domain-specific language. In addition, training the model on large NLI datasets such as MultiNLI and SNLI helps in learning task-specific reasoning. Finally, we ensembled our highest-performing models, and achieved 84.7% accuracy on the unseen test dataset and ranked 10th out of 17 teams in the official results.
机译:在本文中,我们描述了针对MEDIQA 2019共享任务的自然语言推理(NLI)轨道提交的系统和结果(Ben Abacha等人,2019)。作为KLLai团队,我们使用BERT(Devlin等人,2018)作为我们的基线模型,并对MedNLI数据集进行了预处理,以减轻去识别工件的负面影响。此外,我们研究了不同的预培训和转移学习方法以提高性能。我们表明,在丰富的生物医学语料库上对语言模型进行预训练在教授模型领域特定的语言方面具有重要作用。此外,在大型NLI数据集(例如MultiNLI和SNLI)上训练模型有助于学习特定于任务的推理。最后,我们整合了性能最高的模型,在看不见的测试数据集上实现了84.7%的准确度,在官方结果中的17个团队中排名第10。

著录项

相似文献

  • 外文文献
  • 中文文献
  • 专利
获取原文

客服邮箱:kefu@zhangqiaokeyan.com

京公网安备:11010802029741号 ICP备案号:京ICP备15016152号-6 六维联合信息科技 (北京) 有限公司©版权所有
  • 客服微信

  • 服务号