首页> 外文期刊>ACM Transactions on Information Systems >Sentence Relations for Extractive Summarization with Deep Neural Networks
【24h】

Sentence Relations for Extractive Summarization with Deep Neural Networks

机译:深度神经网络提取摘要的句子关系

获取原文
获取原文并翻译 | 示例
           

摘要

Sentence regression is a type of extractive summarization that achieves state-of-the-art performance and is commonly used in practical systems. The most challenging task within the sentence regression framework is to identify discriminative features to represent each sentence. In this article, we study the use of sentence relations, e.g., Contextual Sentence Relations (CSR), Title Sentence Relations (TSR), and Query Sentence Relations (QSR), so as to improve the performance of sentence regression. CSR, TSR, and QSR refer to the relations between a main body sentence and its local context, its document title, and a given query, respectively.We propose a deep neural network model, Sentence Relation-based Summarization (SRSum), that consists of five sub-models, PriorSum, CSRSum, TSRSum, QSRSum, and SFSum. PriorSum encodes the latent semantic meaning of a sentence using a bi-gram convolutional neural network. SFSum encodes the surface information of a sentence, e.g., sentence length, sentence position, and so on. CSRSum TSRSum, and QSRSum are three sentence relation sub-models corresponding to CSR, TSR, and QSR, respectively. CSRSum evaluates the ability of each sentence to summarize its local contexts. Specifically, CSRSum applies a CSR-based word-level and sentence-level attention mechanism to simulate the context-aware reading of a human reader, where words and sentences that have anaphoric relations or local summarization abilities are easily remembered and paid attention to. TSRSum evaluates the semantic closeness of each sentence with respect to its title, which usually reflects the main ideas of a document. TSRSum applies a TSR-based attention mechanism to simulate people's reading ability with the main idea (title) in mind. QSRSum evaluates the relevance of each sentence with given queries for the query-focused summarization. QSRSum applies a QSR-based attention mechanism to simulate the attentive reading of a human reader with some queries in mind. The mechanism can recognize which parts of the given queries are more likely answered by a sentence under consideration. Finally as a whole, SRSum automatically learns useful latent features by jointly learning representations of query sentences, content sentences, and title sentences as well as their relations.We conduct extensive experiments on six benchmark datasets, including generic multi-document summarization and query-focused multi-document summarization. On both tasks, SRSum achieves comparable or superior performance compared with state-of-the-art approaches in terms of multiple ROUGE metrics.
机译:句子回归是一种提取性摘要,可实现最先进的性能,通常在实际系统中使用。句子回归框架中最具挑战性的任务是识别区分特征以代表每个句子。在本文中,我们研究了句子关系的使用,例如上下文句子关系(CSR),标题句子关系(TSR)和查询句子关系(QSR),以提高句子回归的性能。 CSR,TSR和QSR分别指代主体句子与其局部上下文,其文档标题和给定查询之间的关系。我们提出了一种深度神经网络模型,即基于句子关系的摘要(SRSum),该模型由五个子模型,即PriorSum,CSRSum,TSRSum,QSRSum和SFSum。 PriorSum使用Bi-gram卷积神经网络对句子的潜在语义进行编码。 SFSum对句子的表面信息进行编码,例如,句子的长度,句子的位置等。 CSRSum TSRSum和QSRSum是分别对应于CSR,TSR和QSR的三个句子关系子模型。 CSRSum评估每个句子总结其本地上下文的能力。具体而言,CSRSum应用基于CSR的单词级和句子级注意力机制来模拟人类读者的上下文感知阅读,在其中容易理解和注意具有隐喻关系或局部摘要能力的单词和句子。 TSRSum评估每个句子相对于其标题的语义接近度,这通常反映出文档的主要思想。 TSRSum应用基于TSR的注意力机制来模拟人们的阅读能力,并牢记主要思想(标题)。 QSRSum评估每个句子与给定查询的相关性,以针对查询进行汇总。 QSRSum应用基于QSR的注意力机制来模拟人类读者的细心阅读,并牢记一些疑问。该机制可以识别正在考虑的句子更可能回答给定查询的哪些部分。最终,SRSum通过联合学习查询语句,内容语句和标题语句的表示及其关系来自动学习有用的潜在特征。我们对六个基准数据集进行了广泛的实验,包括通用的多文档摘要和针对查询的数据多文档摘要。在两个任务上,SRSum在多个ROUGE指标方面都可以与最新技术相媲美或出色。

著录项

相似文献

  • 外文文献
  • 中文文献
  • 专利
获取原文

客服邮箱:kefu@zhangqiaokeyan.com

京公网安备:11010802029741号 ICP备案号:京ICP备15016152号-6 六维联合信息科技 (北京) 有限公司©版权所有
  • 客服微信

  • 服务号