首页> 外文会议>Annual meeting of the Association for Computational Linguistics >Sentence Centrality Revisited for Unsupervised Summarization
【24h】

Sentence Centrality Revisited for Unsupervised Summarization

机译:重新审视句子中心以实现无监督汇总

获取原文

摘要

Single document summarization has enjoyed renewed interest in recent years thanks to the popularity of neural network models and the availability of large-scale datasets. In this paper we develop an unsupervised approach arguing that it is unrealistic to expect large-scale and high-quality training data to be available or created for different types of summaries, domains, or languages. We revisit a popular graph-based ranking algorithm and modify how node (aka sentence) centrality is computed in two ways: (a) we employ BERT, a state-of-the-art neural representation learning model to better capture sentential meaning and (b) we build graphs with directed edges arguing that the contribution of any two nodes to their respective centrality is influenced by their relative position in a document. Experimental results on three news summarization datasets representative of different languages and writing styles show that our approach outperforms strong baselines by a wide margin.
机译:近年来,由于神经网络模型的普及和大规模数据集的可用性,单一文档摘要引起了新的兴趣。在本文中,我们开发了一种无监督的方法,认为期望为各种类型的摘要,域或语言提供或创建大规模且高质量的培训数据是不现实的。我们重新审视了一种流行的基于图的排名算法,并修改了以两种方式计算节点(aka句子)中心度的方式:(a)我们使用BERT(一种最新的神经表示学习模型)来更好地捕获句子的含义,并且( b)我们建立有向边的图,认为任何两个节点对其各自中心的贡献受它们在文档中的相对位置的影响。在代表不同语言和写作风格的三个新闻摘要数据集上的实验结果表明,我们的方法在很大程度上优于强基准。

著录项

相似文献

  • 外文文献
  • 中文文献
  • 专利
获取原文

客服邮箱:kefu@zhangqiaokeyan.com

京公网安备:11010802029741号 ICP备案号:京ICP备15016152号-6 六维联合信息科技 (北京) 有限公司©版权所有
  • 客服微信

  • 服务号