首页> 外文会议>International joint conference on natural language processing;Conference on empirical methods in natural language processing >How Contextual are Contextualized Word Representations? Comparing the Geometry of BERT, ELMo, and GPT-2 Embeddings
【24h】

How Contextual are Contextualized Word Representations? Comparing the Geometry of BERT, ELMo, and GPT-2 Embeddings

机译:上下文化的词表示形式如何上下文化?比较BERT,ELMo和GPT-2嵌入的几何

获取原文

摘要

Replacing static word embeddings with contextualized word representations has yielded significant improvements on many NLP tasks. However, just how contextual are the contextualized representations produced by models such as ELMo and BERT? Are there infinitely many context-specific representations for each word, or are words essentially assigned one of a finite number of word-sense representations? For one, we find that the contextualized representations of all words are not isotropic in any layer of the contextualizing model. While representations of the same word in different contexts still have a greater cosine similarity than those of two different words, this self-similarity is much lower in upper layers. This suggests that upper layers of contextualizing models produce more context-specific representations, much like how upper layers of LSTMs produce more task-specific representations. In all layers of ELMo, BERT, and GPT-2, on average, less than 5% of the variance in a word's contextualized representations can be explained by a static embedding for that word, providing some justification for the success of contextualized representations.
机译:用上下文化的单词表示代替静态单词嵌入已在许多NLP任务上产生了重大改进。但是,像ELMo和BERT这样的模型产生的上下文表示形式又是如何上下文的呢?每个单词有无数种特定于上下文的表示形式,还是本质上为单词分配了有限数量的单词感测表示形式之一?首先,我们发现在上下文模型的任何层中,所有单词的上下文表示都不是各向同性的。尽管同一词在不同上下文中的表示仍然比两个不同词的表示具有更高的余弦相似度,但这种自相似性在上层中要低得多。这表明上下文化模型的上层会产生更多特定于上下文的表示,就像LSTM的上层如何产生更多特定于任务的表示一样。在ELMo,BERT和GPT-2的所有各层中,平均而言,单词的上下文表示中方差的不到5%可以通过对该单词的静态嵌入来解释,从而为上下文表示的成功提供了一些依据。

著录项

相似文献

  • 外文文献
  • 中文文献
  • 专利
获取原文

客服邮箱:kefu@zhangqiaokeyan.com

京公网安备:11010802029741号 ICP备案号:京ICP备15016152号-6 六维联合信息科技 (北京) 有限公司©版权所有
  • 客服微信

  • 服务号