首页> 外文会议>International Conference on Computational Linguistics >The Devil is in the Details: Evaluating Limitations of Transformer-based Methods for Granular Tasks
【24h】

The Devil is in the Details: Evaluating Limitations of Transformer-based Methods for Granular Tasks

机译:魔鬼在细节:评估基于变压器的粒度任务方法的限制

获取原文

摘要

Contextual embeddings derived from transformer-based neural language models have shown state-of-the-art performance for various tasks such as question answering, sentiment analysis, and textual similarity in recent years. Extensive work shows how accurately such models can represent abstract, semantic information present in text. In this expository work, we explore a tangent direction and analyze such models' performance on tasks that require a more granular level of representation. We focus on the problem of textual similarity from two perspectives: matching documents on a granular level (requiring embeddings to capture fine-grained attributes in the text), and an abstract level (requiring embeddings to capture overall textual semantics). We empirically demonstrate, across two datasets from different domains, that despite high performance in abstract document matching as expected, contextual embeddings are consistently (and at times, vastly) outperformed by simple baselines like TF-IDF for more granular tasks. We then propose a simple but effective method to incorporate TF-IDF into models that use contextual embeddings, achieving relative improvements of up to 36% on granular tasks.
机译:从基于变压器的神经语言模型导出的上下文嵌入物显示了近年来的问题应答,情绪分析和文本相似性等各种任务的最先进的性能。广泛的工作表明,此类模型如何代表文本中存在的抽象,语义信息。在本次要工作中,我们探索了切线方向并分析了需要更粒度的代表水平的任务的这种模型的性能。我们专注于从两个角度来看文本相似性的问题:匹配文档的粒度水平(要求嵌入的嵌入来捕获文本中的细粒度属性)和抽象水平(要求嵌入捕获整体文本语义)。我们经验展示了来自不同域的两个数据集,尽管随着预期的抽象文件匹配中的高性能,但上下文嵌入始终如一(有时,最大限度地)通过简单的基线,如TF-IDF的简单基线,对于更多粒度任务。然后,我们提出了一种简单但有效的方法,将TF-IDF合并到使用上下文嵌入的模型中,在粒度任务中实现高达36%的相对提高。

著录项

相似文献

  • 外文文献
  • 中文文献
  • 专利
获取原文

客服邮箱:kefu@zhangqiaokeyan.com

京公网安备:11010802029741号 ICP备案号:京ICP备15016152号-6 六维联合信息科技 (北京) 有限公司©版权所有
  • 客服微信

  • 服务号