首页> 外文会议>Workshop on Domain Adaptation for NLP >BERTologiCoMix: How does Code-Mixing interact with Multilingual BERT?
【24h】

BERTologiCoMix: How does Code-Mixing interact with Multilingual BERT?

机译:Bertologicomix:如何与多语种伯爵相互作用?

获取原文

摘要

Models such as mBERT and XLMR have shown success in solving Code-Mixed NLP tasks even though they were not exposed to such text during pretraining. Code-Mixed NLP models have relied on using synthetically generated data along with naturally occurring data to improve their performance. Finetun-ing mBERT on such data improves it's code-mixed performance, but the benefits of using the different types of Code-Mixed data aren't clear. In this paper, we study the impact of fine-tuning with different types of code-mixed data and outline the changes that occur to the model during such finetuning. Our findings suggest that using naturally occurring code-mixed data brings in the best performance improvement after finetuning and that finetuning with any type of code-mixed text improves the respon-sivity of it's attention heads to code-mixed text inputs.
机译:即使在预先训练期间没有暴露在这些文本中,诸如Mbert和XLMR之类的模型也在解决代码混合的NLP任务方面取得了成功。 代码混合的NLP模型依赖于使用综合生成的数据以及天然存在的数据来提高其性能。 FineTun-ing Mbert在这些数据上提高了代码混合性能,但使用不同类型的代码混合数据的好处尚不清楚。 在本文中,我们研究了微调与不同类型的代码混合数据的影响,并概述了在这种FineTuning期间模型所发生的变化。 我们的研究结果表明,使用自然发生的代码混合数据在FineTuning之后带来了最佳性能改进,并且具有任何类型的码混合文本的FineTuning将其注意力头的响应性提高到Code-Mixed文本输入。

著录项

相似文献

  • 外文文献
  • 中文文献
  • 专利
获取原文

客服邮箱:kefu@zhangqiaokeyan.com

京公网安备:11010802029741号 ICP备案号:京ICP备15016152号-6 六维联合信息科技 (北京) 有限公司©版权所有
  • 客服微信

  • 服务号