首页> 外文会议>Conference on Empirical Methods in Natural Language Processing >PALM: Pre-training an AutoencodingAutoregressive Language Model for Context-conditioned Generation
【24h】

PALM: Pre-training an AutoencodingAutoregressive Language Model for Context-conditioned Generation

机译:Palm:预先培训用于上下文化生成的自动编码和自动评级语言模型

获取原文

摘要

Self-supervised pre-training, such as BERT (Devlin et al., 2018), MASS (Song et al., 2019) and BART (Lewis et al., 2019), has emerged as a powerful technique for natural language understanding and generation. Existing pre-training techniques employ au-toencoding and/or autoregressive objectives to train Transformer-based models by recovering original word tokens from corrupted text with some masked tokens. The training goals of existing techniques are often inconsistent with the goals of many language generation tasks, such as generative question answering and conversational response generation, for producing new text given context. This work presents PALM with a novel scheme that jointly pre-trains an autoencod-ing and autoregressive language model on a large unlabclcd corpus, specifically designed for generating new text conditioned on context. The new scheme alleviates the mismatch introduced by the existing denoising scheme between pre-training and fine-tuning where generation is more than reconstructing original text. An extensive set of experiments show that PALM achieves new state-of-the-art results on a variety of language generation benchmarks covering generative question answering (Rank 1 on the official MARCO leaderboard), abstractive summarization on CNN/DailyMail as well as Gigaword, question generation on SQuAD, and conversational response generation on Cornell Movie Dialogues.
机译:自我监督的预训练,如伯特(Devlin等,2018),Mass(Song et al。,2019)和Bart(Lewis等,2019),已成为自然语言理解的强大技术一代。现有的预训练技术采用AU-Toescoding和/或自动增加目标来通过使用一些屏蔽令牌从损坏的文本中恢复原始字令牌来培训基于变换器的模型。现有技术的培训目标通常与许多语言生成任务的目标不一致,例如生成问题应答和会话响应生成,用于产生上下文的新文本。这项工作介绍了一种新颖的计划,共同预先在大型unlabclcd语料库上共同预先预先训练一个自动增加的语言模型,专门用于在上下文上生成新的文本。新方案减轻了在预训练和微调之间的现有去噪方案引入的不匹配,在那里的那些时不仅仅是重建原文。一组广泛的实验表明,Palm在涵盖生成问题的各种语言生成基准上实现了新的最先进的结果(在官方Marco排行榜上排名第一),CNN / Dailymail以及Gigaword的抽象摘要,康奈尔电影对话中队的问题生成和对话响应生成。

著录项

相似文献

  • 外文文献
  • 中文文献
  • 专利
获取原文

客服邮箱:kefu@zhangqiaokeyan.com

京公网安备:11010802029741号 ICP备案号:京ICP备15016152号-6 六维联合信息科技 (北京) 有限公司©版权所有
  • 客服微信

  • 服务号