【24h】

Noisy Self-Knowledge Distillation for Text Summarization

机译:文本摘要嘈杂的自我知识蒸馏

获取原文
获取外文期刊封面目录资料

摘要

In this paper we apply self-knowledge distillation to text summarization which we argue can alleviate problems with maximum-likelihood training on single reference and noisy datasets. Instead of relying on one-hot annotation labels, our student summarization model is trained with guidance from a teacher which generates smoothed labels to help regularize training. Furthermore, to better model uncertainty during training, we introduce multiple noise signals for both teacher and student models. We demonstrate experimentally on three benchmarks that our framework boosts the performance of both pretrained and non-pretrained summarizers achieving state-of-the-art results.
机译:在本文中,我们将自我知识蒸馏应用于文本摘要,我们认为可以缓解单个参考和嘈杂数据集的最大可能性培训问题。 我们的学生摘要模型而不是依赖单热的注释标签,而是通过来自教师的指导培训,而不是从教师产生平滑标签,以帮助正规化培训。 此外,在培训期间更好地模范不确定性,我们为教师和学生模型引入多种噪声信号。 我们在实验上展示了三个基准,我们的框架推动了普里预先磨普和非净化摘要的性能,实现了最先进的结果。

著录项

相似文献

  • 外文文献
  • 中文文献
  • 专利
获取原文

客服邮箱:kefu@zhangqiaokeyan.com

京公网安备:11010802029741号 ICP备案号:京ICP备15016152号-6 六维联合信息科技 (北京) 有限公司©版权所有
  • 客服微信

  • 服务号