首页> 外文会议>Workshop on Arabic Natural Language Processing >The Interplay of Variant, Size, and Task Type in Arabic Pre-trained Language Models
【24h】

The Interplay of Variant, Size, and Task Type in Arabic Pre-trained Language Models

机译:在阿拉伯语预先接受的语言模型中的变体,大小和任务类型的相互作用

获取原文

摘要

In this paper, we explore the effects of language variants, data sizes, and fine-tuning task types in Arabic pre-trained language models. To do so, we build three pre-trained language models across three variants of Arabic: Modern Standard Arabic (MSA), dialectal Arabic, and classical Arabic, in addition to a fourth language model which is pre-trained on a mix of the three. We also examine the importance of pre-training data size by building additional models that are pre-trained on a scaled-down set of the MSA variant. We compare our different models to each other, as well as to eight publicly available models by fine-tuning them on five NLP tasks spanning 12 datasets. Our results suggest that the variant proximity of pre-training data to fine-tuning data is more important than the pre-training data size. We exploit this insight in defining an optimized system selection model for the studied tasks.
机译:在本文中,我们探讨了语言变体,数据大小和微调任务类型在阿拉伯语预先培训的语言模型中的影响。 为此,我们在阿拉伯语的三种变体中建立了三种预先接受的语言模型:现代标准阿拉伯语(MSA),辩证阿拉伯语和古典阿拉伯语,除了第四种语言模型,该模型还在三个中预先培训 。 我们还通过构建在MSA变体的缩小组上预先培训的其他模型来检查预培训数据大小的重要性。 我们将我们的不同型号互相进行比较,以及通过在跨越12个数据集的五个NLP任务上进行微调,以及八个公开的型号。 我们的结果表明,预训练数据到微调数据的变体附近比预训练数据大小更重要。 我们利用此识别在定义所研究的任务的优化系统选择模型方面。

著录项

相似文献

  • 外文文献
  • 中文文献
  • 专利
获取原文

客服邮箱:kefu@zhangqiaokeyan.com

京公网安备:11010802029741号 ICP备案号:京ICP备15016152号-6 六维联合信息科技 (北京) 有限公司©版权所有
  • 客服微信

  • 服务号