首页> 外文会议>Conference on computational natural language learning >Joint Learning of the Embedding of Words and Entities for Named Entity Disambiguation
【24h】

Joint Learning of the Embedding of Words and Entities for Named Entity Disambiguation

机译:联合学习为命名实体歧义的单词和实体嵌入

获取原文

摘要

Named Entity Disambiguation (NED) refers to the task of resolving multiple named entity mentions in a document to their correct references in a knowledge base (KB) (e.g., Wikipedia). In this paper, we propose a novel embedding method specifically designed for NED. The proposed method jointly maps words and entities into the same continuous vector space. We extend the skip-gram model by using two models. The KB graph model learns the relatedness of entities using the link structure of the KB, whereas the anchor context model aims to align vectors such that similar words and entities occur close to one another in the vector space by leveraging KB anchors and their context words. By combining contexts based on the proposed embedding with standard NED features, we achieved state-of-the-art accuracy of 93.1% on the standard CoNLL dataset and 85.2% on the TAC 2010 dataset.
机译:命名实体消除歧义(NED)是指在文档中解析到文档中的多个命名实体提到的任务,以在知识库(KB)(例如,维基百科)中的正确引用。在本文中,我们提出了一种专门为NED设计的新型嵌入方法。所提出的方法将单词和实体联合地将单词和实体映射到同一个连续的矢量空间中。我们使用两种型号扩展了Skip-Gram模型。 KB图模型使用KB的链路结构来学习实体的相关性,而锚上上下文模型旨在通过利用KB锚和其上下文词来对准矢量使得类似的单词和实体在矢量空间中彼此接近。通过将基于建议的嵌入与标准NED特征的嵌入的组合,我们在标准Conll数据集中实现了最先进的准确性,而TAC 2010数据集上的85.2%。

著录项

相似文献

  • 外文文献
  • 中文文献
  • 专利
获取原文

客服邮箱:kefu@zhangqiaokeyan.com

京公网安备:11010802029741号 ICP备案号:京ICP备15016152号-6 六维联合信息科技 (北京) 有限公司©版权所有
  • 客服微信

  • 服务号