首页> 外文会议>IEEE International Conference on Power Electronics, Computer Applications >A fine-tuning approach research of pre-trained model with two stage
【24h】

A fine-tuning approach research of pre-trained model with two stage

机译:两阶段预训练模型的微调方法研究

获取原文

摘要

A Fine-tuning method has been mention in BERT, which is a pre-trained model use widely in NLP. In BERT and GPT, they hold that a standard fine-tuning model should there have a minimal difference between pre-trained architecture and the final downsteam architecture, and the task-special model will harm the result. In this paper, we mention two stream model which use hidden state pre-trained in BERT. In order to facilitate the validity of the verification method, We use sentiment analysis tasks to verify the results, which is a very simple text classification task in natural language process. Experiments on Yelp-review-poliarty show that using the same training data and other fine-tuning method, we can reduce ERROR by 0.21%. With the same setup, we can reduce ERROR of Amazon-review-poliarty by 0.13 %.
机译:BERT中提到了一种微调方法,这是一个在NLP中广泛使用的预先训练的模型。在BERT和GPT中,他们认为,标准的微调模型应该在训练有素的架构和最终下游架构之间存在最小的差异,并且任务 - 特殊模型将损害结果。在本文中,我们提到了两种流模型,该模型使用伯特预先培训的隐藏状态。为了促进验证方法的有效性,我们使用情感分析任务来验证结果,这是自然语言过程中非常简单的文本分类任务。 yelp审查 - 加强的实验表明,使用相同的训练数据和其他微调方法,我们可以将误差减少0.21%。通过相同的设置,我们可以将亚马逊审查的错误减少0.13%。

著录项

相似文献

  • 外文文献
  • 中文文献
  • 专利
获取原文

客服邮箱:kefu@zhangqiaokeyan.com

京公网安备:11010802029741号 ICP备案号:京ICP备15016152号-6 六维联合信息科技 (北京) 有限公司©版权所有
  • 客服微信

  • 服务号