首页> 外文会议>International natural language generation conference >Generating Abstractive Summaries with Finetuned Language Models
【24h】

Generating Abstractive Summaries with Finetuned Language Models

机译:使用微调的语言模型生成抽象摘要

获取原文

摘要

Neural abstractive document summarization is commonly approached by models that exhibit a mostly extractive behavior. This behavior is facilitated by a copy-attention which allows models to copy words from a source document. While models in the mostly extractive news summarization domain benefit from this inductive bias, they commonly fail to paraphrase or compress information from the source document. Recent advances in transfer-learning from large pretrained language models give rise to alternative approaches that do not rely on copy-attention and instead learn to generate concise and abstractive summaries. In this paper, as part of the TL;DR challenge, we compare the abstractiveness of summaries from different summarization approaches and show that transfer-learning can be efficiendy utilized without any changes to the model architecture. We demonstrate that the approach leads to a higher level of abstraction for a similar performance on the TL;DR challenge tasks, enabling true natural language compression.
机译:神经抽象文档摘要通常通过表现出大部分抽取行为的模型来实现。复制注意促进了此行为,复制注意允许模型从源文档复制单词。尽管主要是提取性新闻摘要领域中的模型受益于这种归纳性偏见,但它们通常无法解释或压缩来自源文档的信息。大型预训练语言模型在转移学习中的最新进展产生了替代方法,这些方法不依赖复制注意,而是学会生成简洁和抽象的摘要。在本文中,作为TL; DR挑战的一部分,我们比较了来自不同汇总方法的汇总的抽象性,并表明可以有效利用转移学习,而无需对模型体系结构进行任何更改。我们证明,该方法可为TL; DR挑战任务提供更高的抽象水平,以实现类似的性能,从而实现真正的自然语言压缩。

著录项

相似文献

  • 外文文献
  • 中文文献
  • 专利
获取原文

客服邮箱:kefu@zhangqiaokeyan.com

京公网安备:11010802029741号 ICP备案号:京ICP备15016152号-6 六维联合信息科技 (北京) 有限公司©版权所有
  • 客服微信

  • 服务号