首页> 外文会议>Workshop on Arabic Natural Language Processing >BERT-based Multi-Task Model for Country and Province Level Modern Standard Arabic and Dialectal Arabic Identification
【24h】

BERT-based Multi-Task Model for Country and Province Level Modern Standard Arabic and Dialectal Arabic Identification

机译:基于BERT的国家和省级多任务模型现代标准阿拉伯语和方言阿拉伯语鉴定

获取原文

摘要

Dialect and standard language identification are crucial tasks for many Arabic natural language processing applications. In this paper, we present our deep learning-based system, submitted to the second NADI shared task for country-level and province-level identification of Modern Standard Arabic (MSA) and Dialectal Arabic (DA). The system is based on an end-to-end deep Multi-Task Learning (MTL) model to tackle both country-level and province-level MSA/DA identification. The latter MTL model consists of a shared Bidirectional Encoder Representation Transformers (BERT) encoder, two task-specific attention layers, and two classifiers. Our key idea is to leverage both the task-discriminative and the inter-task shared features for country and province MSA/DA identification. The obtained results show that our MTL model outperforms single-task models on most subtasks.
机译:方言和标准语言识别是许多阿拉伯语自然语言处理应用的重要任务。 在本文中,我们介绍了我们的深度学习的系统,提交了第二个NADI共享任务,以获得现代标准阿拉伯语(MSA)和辩证阿拉伯语(DA)的国家级和省级识别。 该系统基于端到端的深度多任务学习(MTL)模型来解决国家级和省级MSA / DA标识。 后一MTL模型由共享双向编码器表示变压器(BERT)编码器,两个特定于任务的注意层和两个分类器组成。 我们的主要思想是利用国家和省MSA / DA识别的任务鉴别和任务间共享特征。 所获得的结果表明,我们的MTL模型在大多数子任务上优于单任务模型。

著录项

相似文献

  • 外文文献
  • 中文文献
  • 专利
获取原文

客服邮箱:kefu@zhangqiaokeyan.com

京公网安备:11010802029741号 ICP备案号:京ICP备15016152号-6 六维联合信息科技 (北京) 有限公司©版权所有
  • 客服微信

  • 服务号