首页> 外文会议>Conference on Empirical Methods in Natural Language Processing >Pretrained Language Model Embryology: The Birth of ALBERT
【24h】

Pretrained Language Model Embryology: The Birth of ALBERT

机译:预付费语言模型胚胎学:Albert的诞生

获取原文

摘要

While behaviors of pretrained language models (LMs) have been thoroughly examined, what happened during pretraining is rarely studied. We thus investigate the developmental process from a set of randomly initialized parameters to a totipotent language model, which we refer to as the embryology of a pretrained language model. Our results show that ALBERT learns to reconstruct and predict tokens of different parts of speech (POS) in different learning speeds during pretraining. We also find that linguistic knowledge and world knowledge do not generally improve as pretraining proceeds, nor do downstream tasks' performance. These findings suggest that knowledge of a pretrained model varies during pretraining, and having more pretrain steps does not necessarily provide a model with more comprehensive knowledge.
机译:虽然已经彻底检查了预先训练的语言模型(LMS)的行为,但很少研究在预磨练期间发生的事情。因此,我们将从一组随机初始化参数调查到Totipotent语言模型的发展过程,我们将其称为预用语言模型的胚胎学。我们的研究结果表明,艾伯特在预先预防期间学会在不同的学习速度下重建和预测不同部分语音(POS)的令牌。我们还发现语言知识和世界知识通常不会随着预押收益而改善,也不会改善下游任务的表现。这些研究结果表明,预先磨普模型的知识在预测期间变化,并且具有更多的预留下步骤并不一定提供具有更全面的知识的模型。

著录项

相似文献

  • 外文文献
  • 中文文献
  • 专利
获取原文

客服邮箱:kefu@zhangqiaokeyan.com

京公网安备:11010802029741号 ICP备案号:京ICP备15016152号-6 六维联合信息科技 (北京) 有限公司©版权所有
  • 客服微信

  • 服务号