首页> 外文期刊>ACM transactions on Asian language information processing >A Survey of the Model Transfer Approaches to Cross-Lingual Dependency Parsing
【24h】

A Survey of the Model Transfer Approaches to Cross-Lingual Dependency Parsing

机译:跨语言依赖解析模型转移方法的调查

获取原文
       

摘要

Cross-lingual dependency parsing approaches have been employed to develop dependency parsers for the languages for which little or no treebanks are available using the treebanks of other languages. A language for which the cross-lingual parser is developed is usually referred to as the target language and the language whose treebank is used to train the cross-lingual parser model is referred to as the source language. The cross-lingual parsing approaches for dependency parsing may be broadly classified into three categories: model transfer, annotation projection, and treebank translation. This survey provides an overview of the various aspects of the model transfer approach of cross-lingual dependency parsing. In this survey, we present a classification of the model transfer approaches based on the different aspects of the method. We discuss some of the challenges associated with cross-lingual parsing and the techniques used to address these challenges. In order to address the difference in vocabulary between two languages, some approaches use only non-lexical features of the words to train the models while others use shared representations of the words. Some approaches address the morphological differences by chunk-level transfer rather than word-level transfer. The syntactic differences between the source and target languages are sometimes addressed by transforming the source language treebanks or by combining the resources of multiple source languages. Besides cross-lingual transfer parser models may be developed for a specific target language or it may be trained to parse sentences of multiple languages. With respect to the above-mentioned aspects, we look at the different ways in which the methods can be classified. We further classify and discuss the different approaches from the perspective of the corresponding aspects. We also demonstrate the performance of the transferred models under different settings corresponding to the classification aspects on a common dataset.
机译:已经采用交叉语言依赖解析方法为使用其他语言的TreeBanks开发几乎没有树木班的语言的依赖解析器。开发交叉语言解析器的语言通常被称为目标语言,并且使用TreeBank用于训练交叉语言解析器模型的语言被称为源语言。依赖解析的交叉解析方法可以广泛分为三类:模型传输,注释投影和树班平移。本调查概述了交叉依赖解析的模型传递方法的各个方面。在本调查中,我们基于该方法的不同方面提出了模型传输方法的分类。我们讨论了与交叉解析相关的一些挑战以及用于解决这些挑战的技术。为了解决两种语言之间词汇的差异,一些方法仅使用单词的非词汇特征来训练模型,而其他方法则使用单词的共享表示。有些方法通过块级转移而不是字级转移来解决形态差异。源语言与目标语言之间的语法差异是通过转换源语言树班签,或者通过组合多种源语言的资源来解决。除了交叉语言传输解析器外,可以为特定的目标语言开发,或者可以训练以解析多种语言的句子。关于上述方面,我们看看可以对方法进行分类的不同方式。从相应方面的角度,我们进一步分类和讨论了不同的方法。我们还在与公共数据集上的分类方面对应的不同设置下演示了转移模型的性能。

著录项

相似文献

  • 外文文献
  • 中文文献
  • 专利
获取原文

客服邮箱:kefu@zhangqiaokeyan.com

京公网安备:11010802029741号 ICP备案号:京ICP备15016152号-6 六维联合信息科技 (北京) 有限公司©版权所有
  • 客服微信

  • 服务号