首页> 外文会议>Conference of the European Chapter of the Association for Computational Linguistics >Multilingual LAMA: Investigating Knowledge in Multilingual Pretrained Language Models
【24h】

Multilingual LAMA: Investigating Knowledge in Multilingual Pretrained Language Models

机译:多语种喇嘛:调查多语种预留语言模型的知识

获取原文

摘要

Recently, it has been found that monolingual English language models can be used as knowledge bases. Instead of structural knowledge base queries, masked sentences such as "Paris is the capital of [MASK]" are used as probes. We translate the established benchmarks TREx and GoogleRE into 53 languages. Working with mBERT, we investigate three questions, (ⅰ) Can mBERT be used as a multilingual knowledge base? Most prior work only considers English. Extending research to multiple languages is important for diversity and accessibility, (ⅱ) Is mBERT's performance as knowledge base language-independent or does it vary from language to language? (ⅲ) A multilingual model is trained on more text, e.g., mBERT is trained on 104 Wikipedias. Can mBERT leverage this for better performance? We find that using mBERT as a knowledge base yields varying performance across languages and pooling predictions across languages improves performance. Conversely, mBERT exhibits a language bias; e.g.. when queried in Italian, it tends to predict Italy as the country of origin.
机译:最近,已经发现单格式语言模型可以用作知识库。代替结构知识库查询,诸如“巴黎是[掩码]的首都之类的屏蔽句子用作探针。我们将建立的基准Trex和GoogleRe翻译成53种语言。我们使用MBERT,我们调查了三个问题,(Ⅰ)MBERT可以用作多语言知识库?大多数事先工作只考虑英语。扩展对多种语言的研究对于多样性和可访问性很重要,(Ⅱ)是MBERT作为知识库语言无关的表现,或者它与语言语言有所不同吗? (三)多语言模型培训更多文本,例如,MBENT在104维基百科培训。 MBERT可以利用这一点以实现更好的性能吗?我们发现,使用MBERT作为知识库产生不同的性能,跨语言的汇集预测可以提高性能。相反,Mbert表现出语言偏见;例如..当在意大利语中查询时,它往往将意大利预测为原籍国。

著录项

相似文献

  • 外文文献
  • 中文文献
  • 专利
获取原文

客服邮箱:kefu@zhangqiaokeyan.com

京公网安备:11010802029741号 ICP备案号:京ICP备15016152号-6 六维联合信息科技 (北京) 有限公司©版权所有
  • 客服微信

  • 服务号