首页> 外文会议>International Conference on Computational Linguistics >Fine-tuning BERT for Low-Resource Natural Language Understanding via Active Learning
【24h】

Fine-tuning BERT for Low-Resource Natural Language Understanding via Active Learning

机译:通过主动学习的低资源自然语言理解微调伯特

获取原文

摘要

Recently, leveraging pre-trained Transformer based language models in down stream, task specific models has advanced state of the art results in natural language understanding tasks. However, only a little research has explored the suitability of this approach in low resource settings with less than 1,000 training data points. In this work, we explore fine-tuning methods of BERT - a pre-trained Transformer based language model - by utilizing pool-based active learning to speed up training while keeping the cost of labeling new data constant. Our experimental results on the GLUE data set show an advantage in model performance by maximizing the approximate knowledge gain of the model when querying from the pool of unlabeled data. Finally, we demonstrate and analyze the benefits of freezing layers of the language model during fine-tuning to reduce the number of trainable parameters, making it more suitable for low-resource settings.
机译:最近,利用预训练的基于变压器的语言模型在下游,任务特定模型具有先进的技术,导致自然语言理解任务。 然而,只有一点研究探索了这种方法在低资源设置中的适用性,培训数据点小于1,000。 在这项工作中,我们探讨了BERT的微调方法 - 通过利用基于池的主动学习来加速训练的频率 - 一种基于训练的变压器的语言模型,同时保持标记新数据常数的成本。 我们对胶水数据集的实验结果通过在从未标记数据池查询时最大化模型的近似知识增益,显示了模型性能的优势。 最后,我们在微调过程中展示和分析了语言模型的冷冻层的好处,以减少培训参数的数量,使其更适合低资源设置。

著录项

相似文献

  • 外文文献
  • 中文文献
  • 专利
获取原文

客服邮箱:kefu@zhangqiaokeyan.com

京公网安备:11010802029741号 ICP备案号:京ICP备15016152号-6 六维联合信息科技 (北京) 有限公司©版权所有
  • 客服微信

  • 服务号