首页> 外文会议>Workshop on Domain Adaptation for NLP >Few-Shot Learning of an Interleaved Text Summarization Model by Pretraining with Synthetic Data
【24h】

Few-Shot Learning of an Interleaved Text Summarization Model by Pretraining with Synthetic Data

机译:用合成数据预先预先预测的近摄文本摘要模型的几次拍摄学习

获取原文

摘要

Interleaved texts, where posts belonging to different threads occur in a sequence, commonly occur in online chat posts, so that it can be time-consuming to quickly obtain an overview of the discussions. Existing systems first disentangle the posts by threads and then extract summaries from those threads. A major issue with such systems is error propagation from the disentanglement component. While end-to-end trainable summarization system could obviate explicit disentanglement, such systems require a large amount of labeled data. To address this, we propose to pretrain an end-to-end trainable hierarchical encoder-decoder system using synthetic interleaved texts. We show that by fine-tuning on a real-world meeting dataset (AMI), such a system out-performs a traditional two-step system by 22%. We also compare against transformer models and observed that pretraining with synthetic data both the encoder and decoder outperforms the BertSumExtAbs transformer model which pre-trains only the encoder on a large dataset.
机译:交错的文本,其中属于不同线程的帖子发生在序列中,通常在在线聊天帖子中出现,以便快速获取讨论的概述耗时。现有系统首先通过线程解除帖子,然后从这些线程中提取摘要。此类系统的主要问题是来自DisonDandlement组件的错误传播。虽然端到端培训摘要系统可以避免显式解剖,但这些系统需要大量标记数据。为了解决这个问题,我们建议使用合成交错文本来预先绘制端到端的培训分层编码器-解码器系统。我们表明,通过对真实世界会议数据集(AMI)进行微调,这样的系统会将传统的两步系统推出22%。我们还与变压器模型进行比较,并观察到使用合成数据的预先介绍编码器和解码器均优于BertsumExtabs变压器模型,该模型仅在大型数据集上仅预先列出编码器。

著录项

相似文献

  • 外文文献
  • 中文文献
  • 专利
获取原文

客服邮箱:kefu@zhangqiaokeyan.com

京公网安备:11010802029741号 ICP备案号:京ICP备15016152号-6 六维联合信息科技 (北京) 有限公司©版权所有
  • 客服微信

  • 服务号