首页> 外文会议>Annual meeting of the Association for Computational Linguistics >Barack's Wife Hillary: Using Knowledge Graphs for Fact-Aware Language Modeling
【24h】

Barack's Wife Hillary: Using Knowledge Graphs for Fact-Aware Language Modeling

机译:巴拉克的妻子希拉里:使用知识图进行事实感知的语言建模

获取原文

摘要

Modeling human language requires the ability to not only generate fluent text but also encode factual knowledge. However, traditional language models are only capable of remembering facts seen at training time, and often have difficulty recalling them. To address this, we introduce the knowledge graph language model (KGLM), a neural language model with mechanisms for selecting and copying facts from a knowledge graph that are relevant to the context. These mechanisms enable the model to render information it has never seen before, as well as generate out-of-vocabulary tokens. We also introduce the Linked WikiText-2 dataset,~(1) a corpus of annotated text aligned to the Wikidata knowledge graph whose contents (roughly) match the popular WikiText-2 benchmark (Merity et al., 2017). In experiments, we demonstrate that the KGLM achieves significantly better performance than a strong baseline language model. We additionally compare different language models' ability to complete sentences requiring factual knowledge, and show that the KGLM outperforms even very large language models in generating facts.
机译:对人类语言进行建模不仅需要能够生成流利的文本,而且还可以对事实知识进行编码。但是,传统语言模型只能记住训练时看到的事实,并且通常很难回忆起它们。为了解决这个问题,我们介绍了知识图语言模型(KGLM),这是一种神经语言模型,具有从知识图中选择和复制与上下文相关的事实的机制。这些机制使模型能够呈现以前从未见过的信息,并生成语音提示。我们还介绍了链接的WikiText-2数据集,〜(1)与Wikidata知识图对齐的带注释文本的语料库,其内容(大致)与流行的WikiText-2基准匹配(Merity等,2017)。在实验中,我们证明了KGLM比强大的基准语言模型具有明显更好的性能。我们还比较了不同语言模型完成需要事实知识的句子的能力,并表明KGLM在生成事实方面甚至胜过非常大的语言模型。

著录项

相似文献

  • 外文文献
  • 中文文献
  • 专利
获取原文

客服邮箱:kefu@zhangqiaokeyan.com

京公网安备:11010802029741号 ICP备案号:京ICP备15016152号-6 六维联合信息科技 (北京) 有限公司©版权所有
  • 客服微信

  • 服务号