首页> 外文会议>International conference on text, speech, and dialogue >FinEst BERT and CroSloEngual BERT Less Is More in Multilingual Models
【24h】

FinEst BERT and CroSloEngual BERT Less Is More in Multilingual Models

机译:多语言模型中的FinEst BERT和CroSloEngual BERT少即是多

获取原文

摘要

Large pretrained masked language models have become state-of-the-art solutions for many NLP problems. The research has been mostly focused on English language, though. While massively multilingual models exist, studies have shown that monolingual models produce much better results. We train two trilingual BERT-like models, one for Finnish, Estonian, and English, the other for Croatian, Slovenian, and English. We evaluate their performance on several downstream tasks, NER, POS-tagging, and dependency parsing, using the multilingual BERT and XLM-R as baselines. The newly created FinEst BERT and CroSloEngual BERT improve the results on all tasks in most monolingual and cross-lingual situations.
机译:大型的预训练掩蔽语言模型已成为解决许多NLP问题的最新解决方案。不过,这项研究主要集中在英语方面。尽管存在大量的多语言模型,但研究表明,单语言模型会产生更好的结果。我们训练了两种类似BERT的三种语言,一种用于芬兰语,爱沙尼亚语和英语,另一种用于克罗地亚语,斯洛文尼亚语和英语。我们使用多语言BERT和XLM-R作为基准,评估它们在多个下游任务,NER,POS标记和依赖项解析上的性能。新创建的FinEst BERT和CroSloEngual BERT可改善大多数单语和跨语种情况下所有任务的结果。

著录项

相似文献

  • 外文文献
  • 中文文献
  • 专利
获取原文

客服邮箱:kefu@zhangqiaokeyan.com

京公网安备:11010802029741号 ICP备案号:京ICP备15016152号-6 六维联合信息科技 (北京) 有限公司©版权所有
  • 客服微信

  • 服务号