首页> 外文期刊>Neurocomputing >Syntactically-informed word representations from graph neural network
【24h】

Syntactically-informed word representations from graph neural network

机译:图形神经网络的语法上通知的字表示

获取原文
获取原文并翻译 | 示例
       

摘要

Most deep language understanding models depend only on word representations, which are mainly based on language modelling derived from a large amount of raw text. These models encode distributional knowledge without considering syntactic structural information, although several studies have shown benefits of including such information. Therefore, we propose new syntactically-informed word representations (SIWRs), which allow us to enrich the pre-trained word representations with syntactic information without training language models from scratch. To obtain SIWRs, a graph-based neural model is built on top of either static or contextualised word representations such as GloVe, ELMo and BERT. The model is first pre-trained with only a relatively modest amount of task-independent data that are automatically annotated using existing syntactic tools. SIWRs are then obtained by applying the model to downstream task data and extracting the intermediate word representations. We finally replace word representations in downstream models with SIWRs for applications. We evaluate SIWRs on three information extraction tasks, namely nested named entity recognition (NER), binary and n-ary relation extractions (REs). The results demonstrate that our SIWRs yield performance gains over the base representations in these NLP tasks with 3-9% relative error reduction. Our SIWRs also perform better than fine-tuning BERT in binary RE. We also conduct extensive experiments to analyse the proposed method. (C) 2020 The Authors. Published by Elsevier B.V.
机译:最深刻的语言理解模型只取决于Word表示,它们主要基于从大量原始文本派生的语言建模。这些模型在不考虑句法结构信息的情况下编码分配知识,尽管有几项研究表明包括包括此类信息的好处。因此,我们提出了新的句法上通知的单词表示(SIWRS),这允许我们在不培训语言模型的情况下以句法信息来丰富训练有素的单词表示。为了获得SIWRS,基于图形的神经模型构建在静态或上下文化的单词表示之上,例如手套,ELMO和BERT。首先是使用现有语法工具自动注释的相对适度的任务无关数据进行预先培训。然后通过将模型应用于下游任务数据并提取中间字表示来获得SIWRS。我们终于将下游模型中的Word表示替换为SIWRS供申请。我们在三个信息提取任务中评估SIWRS,即嵌套命名实体识别(NER),二进制和N-ARY关系提取(RES)。结果表明,我们的SIWRS在这些NLP任务中的基础表示中产生性能提升,相对误差减少3-9%。我们的SIWRS在二进制RE中的微调伯爵也表现优于微调。我们还开展了广泛的实验来分析所提出的方法。 (c)2020作者。由elsevier b.v出版。

著录项

  • 来源
    《Neurocomputing》 |2020年第6期|431-443|共13页
  • 作者单位

    Univ Manchester Natl Ctr Text Min Dept Comp Sci Manchester Lancs England|Natl Inst Adv Ind Sci & Technol Artificial Intelligence Res Ctr Tokyo Japan;

    Toyota Technol Inst Nagoya Aichi Japan|Natl Inst Adv Ind Sci & Technol Artificial Intelligence Res Ctr Tokyo Japan;

    Univ Manchester Natl Ctr Text Min Dept Comp Sci Manchester Lancs England|Natl Inst Adv Ind Sci & Technol Artificial Intelligence Res Ctr Tokyo Japan|Alan Turing Inst London England;

  • 收录信息 美国《科学引文索引》(SCI);美国《工程索引》(EI);
  • 原文格式 PDF
  • 正文语种 eng
  • 中图分类
  • 关键词

    Natural language processing; Contextual word representation; Word representation; Word embedding; Syntactic word representation;

    机译:自然语言处理;语境词表示;词表示;词嵌入;句子词表示;
  • 入库时间 2022-08-18 22:26:49

相似文献

  • 外文文献
  • 中文文献
  • 专利
获取原文

客服邮箱:kefu@zhangqiaokeyan.com

京公网安备:11010802029741号 ICP备案号:京ICP备15016152号-6 六维联合信息科技 (北京) 有限公司©版权所有
  • 客服微信

  • 服务号