首页> 外文期刊>Journal of machine learning research >Exploring the Limits of Transfer Learning with a Unified Text-to-Text Transformer
【24h】

Exploring the Limits of Transfer Learning with a Unified Text-to-Text Transformer

机译:用统一的文本到文本变压器探索转移学习的限制

获取原文
           

摘要

Transfer learning, where a model is first pre-trained on a data-rich task before being fine-tuned on a downstream task, has emerged as a powerful technique in natural language processing (NLP). The effectiveness of transfer learning has given rise to a diversity of approaches, methodology, and practice. In this paper, we explore the landscape of transfer learning techniques for NLP by introducing a unified framework that converts all text-based language problems into a text-to-text format. Our systematic study compares pre-training objectives, architectures, unlabeled data sets, transfer approaches, and other factors on dozens of language understanding tasks. By combining the insights from our exploration with scale and our new “Colossal Clean Crawled Corpus”, we achieve state-of-the-art results on many benchmarks covering summarization, question answering, text classification, and more. To facilitate future work on transfer learning for NLP, we release our data set, pre-trained models, and code.
机译:转移学习,其中模型首先在富有的数据上进行预先训练,然后在下游任务上进行微调,因此是一种在自然语言处理(NLP)中的强大技术。转移学习的有效性引起了多样性的方法,方法和实践。在本文中,我们通过引入统一的框架来探讨NLP的传输学习技术景观,该框架将所有基于文本的语言问题转换为文本到文本格式。我们的系统研究比较了预训练目标,架构,未标记的数据集,转移方法以及数十个语言理解任务的其他因素。通过将我们的探索与规模和新的“巨大清洁爬行语料库”的洞察结合,我们实现了最先进的结果,涵盖了许多基准,涵盖了总结,问题应答,文本分类等等。为了促进未来的转移学习工作,我们释放了我们的数据集,预先训练的模型和代码。

著录项

相似文献

  • 外文文献
  • 中文文献
  • 专利
获取原文

客服邮箱:kefu@zhangqiaokeyan.com

京公网安备:11010802029741号 ICP备案号:京ICP备15016152号-6 六维联合信息科技 (北京) 有限公司©版权所有
  • 客服微信

  • 服务号