首页> 外文会议>International Joint Conference on Neural Networks >Unleashing the Potential of Attention Model for News Headline Generation
【24h】

Unleashing the Potential of Attention Model for News Headline Generation

机译:释放注意力模型在新闻标题生成中的潜力

获取原文

摘要

Headline generation is a special summarization generation task and the difficulty lies in requiring the generated headline to be concise, fluent and informative. Limited by the ability of commonly used encoder and decoder modules to capture long-term dependencies in seq2seq tasks, previous work rarely researched headline generation by end-to-end methods. However, the recent success of Transformer model and its subsequent improved versions demonstrate their remarkable performance on seq2seq tasks, which provide us with a feasible solution. In this paper, we propose a novel model Transformer(XL)-CC to generate headline from the perspective of understanding the whole text, the segment-level recurrence mechanism and relative positional encoding make our model learn ultra-long-term dependencies. In addition, we combine the copy and coverage mechanisms to generate more readable titles. Experimental results on the NYT and Chinese LSCC news datasets also confirm that our method significantly achieves better performance on the headline generation task.
机译:标题生成是一个特别的摘要生成任务,难度在于需要生成的标题简明扼要,流畅和信息。通过常用编码器和解码器模块在SEQ2SEQ任务中捕获长期依赖性的能力的限制,之前的工作很少通过端到端方法研究标题生成。然而,最近变压器模型的成功及其随后的改进版本在SEQ2SEQ任务上展示了它们的显着性能,为我们提供了可行的解决方案。在本文中,我们提出了一种新颖的模型变换器(XL)-CC,从理解整个文本的角度来生成标题,分段级复发机制和相对位置编码使我们的模型学习超长期依赖性。此外,我们结合了副本和覆盖机制来生成更可读的标题。 NYT和中国LSCC新闻数据集的实验结果还证实,我们的方法在标题生成任务上显着实现了更好的性能。

著录项

相似文献

  • 外文文献
  • 中文文献
  • 专利
获取原文

客服邮箱:kefu@zhangqiaokeyan.com

京公网安备:11010802029741号 ICP备案号:京ICP备15016152号-6 六维联合信息科技 (北京) 有限公司©版权所有
  • 客服微信

  • 服务号