首页> 外文会议>BlackboxNLP Workshop on Analyzing and Interpreting Neural Networks for NLP >This is a BERT. Now there are several of them. Can they generalize to novel words?
【24h】

This is a BERT. Now there are several of them. Can they generalize to novel words?

机译:这是伯特。现在有几个。他们可以推广到新颖的话语吗?

获取原文

摘要

Recently, large-scale pre-trained neural network models such as BERT have achieved many state-of-the-art results in natural language processing. Recent work has explored the linguistic capacities of these models. However, no work has focused on the ability of these models to generalize these capacities to novel words. This type of generalization is exhibited by humans (Berko, 1958), and is intimately related to morphology-humans are in many cases able to identify inflections of novel words in the appropriate context. This type of morphological capacity has not been previously tested in BERT models, and is important for morphologically-rich languages, which are under-studied in the literature regarding BERT's linguistic capacities. In this work, we investigate this by considering monolingual and multilingual BERT models' abilities to agree in number with novel plural words in English, French, German, Spanish, and Dutch. We find that many models are not able to reliably determine plurality of novel words, suggesting potential deficiencies in the morphological capacities of BERT models.
机译:最近,诸如BERT的大型预训练的神经网络模型已经实现了许多最先进的自然语言处理。最近的工作探索了这些模型的语言能力。然而,没有任何工作侧重于这些模型将这些能力概括为新型词语的能力。这种类型的概括由人类(Berko,1958)展示,与形态学 - 人类密切相关,在许多情况下,能够在适当的背景下识别新颖单词的拐点。这种类型的形态容量以前尚未在伯特模型中进行测试,对形态学的语言很重要,这在文献中研究了关于BERT的语言能力。在这项工作中,我们通过考虑单语和多语言BERT模型的能力来调查这一点,以便用英语,法语,德语,西班牙语和荷兰语的新型复数单词同意。我们发现许多模型无法可靠地确定多个新颖的单词,这表明伯特模型的形态能力的潜在不足。

著录项

相似文献

  • 外文文献
  • 中文文献
  • 专利
获取原文

客服邮箱:kefu@zhangqiaokeyan.com

京公网安备:11010802029741号 ICP备案号:京ICP备15016152号-6 六维联合信息科技 (北京) 有限公司©版权所有
  • 客服微信

  • 服务号