首页> 外文会议>Workshop on Domain Adaptation for NLP >Pseudo-Label Guided Unsupervised Domain Adaptation of Contextual Embeddings
【24h】

Pseudo-Label Guided Unsupervised Domain Adaptation of Contextual Embeddings

机译:伪标签引导无监督域适应上下文嵌入

获取原文

摘要

Contextual embedding models such as BERT can be easily fine-tuned on labeled samples to create a state-of-the-art model for many downstream tasks. However, the fine-tuned BERT model suffers considerably from unla-beled data when applied to a different domain. In unsupervised domain adaptation, we aim to train a model that works well on a target domain when provided with labeled source samples and unlabeled target samples. In this paper, we propose a pseudo-label guided method for unsupervised domain adaptation. Two models are fine-tuned on labeled source samples as pseudo labeling models. To leam representations for the target domain, one of those models is adapted by masked language modeling from the target domain. Then those models are used to assign pseudo-labels to target samples. We train the final model with those samples. We evaluate our method on named entity segmentation and sentiment analysis tasks. These experiments show that our approach outperforms baseline methods.
机译:诸如BERT等上下文嵌入模型可以轻松地微调标记的样本,以为许多下游任务创建最先进的模型。 然而,当应用于不同的域时,微调BERT模型从ULLED数据遭受了很大的影响。 在无监督的域适应中,我们的目标是在提供标记为源样本和未标记的目标样本时培训在目标域上运行良好的模型。 在本文中,我们提出了一种用于无监督域适应的伪标签引导方法。 在标记的源样本上,两种模型是伪标签模型的微调。 对于目标域的LeaM表示,其中一个模型由来自目标域的屏蔽语言建模调整。 然后,这些模型用于将伪标签分配给目标样本。 我们用这些样品训练最终模型。 我们评估我们在命名实体分割和情感分析任务的方法。 这些实验表明,我们的方法优于基线方法。

著录项

相似文献

  • 外文文献
  • 中文文献
  • 专利
获取原文

客服邮箱:kefu@zhangqiaokeyan.com

京公网安备:11010802029741号 ICP备案号:京ICP备15016152号-6 六维联合信息科技 (北京) 有限公司©版权所有
  • 客服微信

  • 服务号