首页> 外文会议>International Conference on Computational Linguistics >A Joint Learning Approach based on Self-Distillation for Keyphrase Extraction from Scientific Documents
【24h】

A Joint Learning Approach based on Self-Distillation for Keyphrase Extraction from Scientific Documents

机译:基于自蒸馏对科学文献的关键蒸馏的联合学习方法

获取原文

摘要

Keyphrase extraction is the task of extracting a small set of phrases that best describe a document. Most existing benchmark datasets for the task typically have limited numbers of annotated documents, making it challenging to train increasingly complex neural networks. In contrast, digital libraries store millions of scientific articles online, covering a wide range of topics. While a significant portion of these articles contain keyphrases provided by their authors, most other articles lack such kind of annotations. Therefore, to effectively utilize these large amounts of unlabeled articles, we propose a simple and efficient joint learning approach based on the idea of self-distillation. Experimental results show that our approach consistently improves the performance of baseline models for keyphrase extraction. Furthermore, our best models outperform previous methods for the task, achieving new state-of-the-art results on two public benchmarks: Inspec and SemEval-2017.
机译:关键词提取是提取最能描述文件的一小组短语的任务。 任务的大多数现有的基准数据集通常具有有限数量的注释文件,使得培训越来越复杂的神经网络的挑战性。 相比之下,数字图书馆在线存储数百万科学文章,涵盖了各种主题。 虽然这些文章的重要部分包含他们的作者提供的关键时代,但大多数其他文章都缺乏这种注释。 因此,为了有效利用这些大量的未标记物品,我们提出了一种基于自蒸馏的想法的简单有效的联合学习方法。 实验结果表明,我们的方法一致地提高了基线模型的关键酶提取。 此外,我们最好的模型优于前面的任务方法,实现了两种公共基准的新的最先进结果:Inspec和Semeval-2017。

著录项

相似文献

  • 外文文献
  • 中文文献
  • 专利
获取原文

客服邮箱:kefu@zhangqiaokeyan.com

京公网安备:11010802029741号 ICP备案号:京ICP备15016152号-6 六维联合信息科技 (北京) 有限公司©版权所有
  • 客服微信

  • 服务号