首页> 外文会议>Conference on Empirical Methods in Natural Language Processing >LUKE: Deep Contextualized Entity Representations with Entity-aware Self-attention
【24h】

LUKE: Deep Contextualized Entity Representations with Entity-aware Self-attention

机译:卢克:具有实体意识的自我关注的深层情境化实体表示

获取原文

摘要

Entity representations are useful in natural language tasks involving entities. In this paper, we propose new pretrained contextualized representations of words and entities based on the bidirectional transformer (Vaswani et al., 2017). The proposed model treats words and entities in a given text as independent tokens, and outputs contextualized representations of them. Our model is trained using a new pretraining task based on the masked language model of BERT (Devlin et al., 2019). The task involves predicting randomly masked words and entities in a large entity-annotated corpus retrieved from Wikipedia. We also propose an entity-aware self-attention mechanism that is an extension of the self-attention mechanism of the transformer, and considers the types of tokens (words or entities) when computing attention scores. The proposed model achieves impressive empirical performance on a wide range of entity-related tasks. In particular, it obtains state-of-the-art results on five well-known datasets: Open Entity (entity typing), TACRED (relation classification), CoNLL-2003 (named entity recognition), ReCoRD (cloze-style question answering), and SQuAD 1.1 (extractive question answering).
机译:实体表示在涉及实体的自然语言任务中是有用的。在本文中,我们提出了基于双向变压器的新预借出的语境化表示(Vaswani等,2017)。所提出的模型将给定文本中的单词和实体视为独立的令牌,并输出它们的上下文化表示。我们的模型采用基于BERT屏蔽语言模型的新预制任务培训(Devlin等,2019)。该任务涉及预测从维基百科检索的大实体注释语料库中的随机屏蔽的单词和实体。我们还提出了一个实体感知的自我关注机制,该机制是变压器自我关注机制的延伸,并在计算注意力分数时考虑令牌(单词或实体)的类型。拟议的模型在广泛的实体相关任务中实现了令人印象深刻的经验性能。特别地,它获得了五个众所周知的数据集的最先进结果:开放实体(实体键入),TACRED(关系分类),CONLL-2003(命名实体识别),记录(CLOZE样式问题应答)和小队1.1(提取问题回答)。

著录项

相似文献

  • 外文文献
  • 中文文献
  • 专利
获取原文

客服邮箱:kefu@zhangqiaokeyan.com

京公网安备:11010802029741号 ICP备案号:京ICP备15016152号-6 六维联合信息科技 (北京) 有限公司©版权所有
  • 客服微信

  • 服务号