...
首页> 外文期刊>Computational intelligence and neuroscience >Arabic Syntactic Diacritics Restoration Using BERT Models
【24h】

Arabic Syntactic Diacritics Restoration Using BERT Models

机译:Arabic Syntactic Diacritics Restoration Using BERT Models

获取原文
获取原文并翻译 | 示例

摘要

The Arabic syntactic diacritics restoration problem is often solved using long short-term memory (LSTM) networks. Handcrafted features are used to augment these LSTM networks or taggers to improve performance. A transformer-based machine learning technique known as bidirectional encoder representations from transformers (BERT) has become the state-of-the-art method for natural language understanding in recent years. In this paper, we present a novel tagger based on BERT models to restore Arabic syntactic diacritics. We formulated the syntactic diacritics restoration as a token sequence classification task similar to named-entity recognition (NER). Using the Arabic TreeBank (ATB) corpus, the developed BERT tagger achieves a 1.36 absolute caseending error rate (CEER) over other systems.

著录项

获取原文

客服邮箱:kefu@zhangqiaokeyan.com

京公网安备:11010802029741号 ICP备案号:京ICP备15016152号-6 六维联合信息科技 (北京) 有限公司©版权所有
  • 客服微信

  • 服务号