首页> 外文会议>Conference on Empirical Methods in Natural Language Processing >PatchBERT: Just-in-Time, Out-of-Vocabulary Patching
【24h】

PatchBERT: Just-in-Time, Out-of-Vocabulary Patching

机译:Patchbert:正常,失败的upbulary修补

获取原文

摘要

Large scale pre-trained language models have shown groundbreaking performance improve­ments for transfer learning in the domain of natural language processing. In our paper, we study a pre-trained multilingual BERT model and analyze the OOV rate on downstream tasks, how it introduces information loss, and as a side-effect, obstructs the potential of the underlying model. We then propose multi­ple approaches for mitigation and demonstrate that it improves performance with the same parameter count when combined with fine-tuning.
机译:大规模预训练的语言模型显示了在自然语言处理领域的转移学习的突破性性能改进。在我们的论文中,我们研究了预先训练的多语言BERT模型,并分析了下游任务的OOV率,如何引入信息丢失,以及作为副作用,妨碍潜在模型的潜力。然后,我们提出了多种方法来缓解,并证明它在与微调时,它可以提高相同的参数计数。

著录项

相似文献

  • 外文文献
  • 中文文献
  • 专利
获取原文

客服邮箱:kefu@zhangqiaokeyan.com

京公网安备:11010802029741号 ICP备案号:京ICP备15016152号-6 六维联合信息科技 (北京) 有限公司©版权所有
  • 客服微信

  • 服务号