首页> 外文会议>Conference of the European Chapter of the Association for Computational Linguistics >Do Syntax Trees Help Pre-trained Transformers Extract Information?
【24h】

Do Syntax Trees Help Pre-trained Transformers Extract Information?

机译:语法树木有助于预先训练的变压器提取信息吗?

获取原文

摘要

Much recent work suggests that incorporating syntax information from dependency trees can improve task-specific transformer models. However, the effect of incorporating dependency tree information into pre-trained transformer models (e.g., BERT) remains unclear, especially given recent studies highlighting how these models implicitly encode syntax. In this work, we systematically study the utility of incorporating dependency trees into pre-trained transformers on three representative information extraction tasks: semantic role labeling (SRL), named entity recognition, and relation extraction. We propose and investigate two distinct strategies for incorporating dependency structure: a late fusion approach, which applies a graph neural network on the output of a transformer, and a joint fusion approach, which infuses syntax structure into the transformer attention layers. These strategies are representative of prior work, but we introduce additional model design elements that are necessary for obtaining improved performance. Our empirical analysis demonstrates that these syntax-infused transformers obtain state-of-the-art results on SRL and relation extraction tasks. However, our analysis also reveals a critical shortcoming of these models: we find that their performance gains are highly contingent on the availability of human-annotated dependency parses, which raises important questions regarding the viability of syntax-augmented transformers in real-world applications.
机译:最近的最近的工作表明,从依赖树中融合了语法信息可以改善特定于任务的变压器模型。然而,将依赖性树信息纳入预先接受的变压器模型(例如,BERT)的效果仍不清楚,特别是初步研究突出了这些模型如何隐含地编码语法。在这项工作中,我们系统地研究了在三个代表信息提取任务上将依赖树纳入预训练的变压器的实用程序:语义角色标记(SRL),命名实体识别和关系提取。我们提出并调查了一种结合依赖结构的两种不同的策略:一种晚期融合方法,它在变压器的输出上应用图形神经网络,以及带有联合融合方法,将语法结构注入变压器注意层。这些策略代表了现有工作,但我们引入了获得提高性能所需的额外模型设计元素。我们的实证分析表明,这些语法注入的变压器在SRL和关系提取任务上获得最先进的结果。然而,我们的分析还揭示了这些模型的关键缺点:我们发现他们的绩效收益对人类注释的依赖解析的可用性高度抵销,这提出了关于现实世界应用中的语法增强变压器的可行性的重要问题。

著录项

相似文献

  • 外文文献
  • 中文文献
  • 专利
获取原文

客服邮箱:kefu@zhangqiaokeyan.com

京公网安备:11010802029741号 ICP备案号:京ICP备15016152号-6 六维联合信息科技 (北京) 有限公司©版权所有
  • 客服微信

  • 服务号