首页> 外文会议>Conference on the North American Chapter of the Association for Computational Linguistics: Human Language Technologies >Addressing Word-order Divergence in Multilingual Neural Machine Translation for Extremely Low Resource Languages
【24h】

Addressing Word-order Divergence in Multilingual Neural Machine Translation for Extremely Low Resource Languages

机译:解决极少资源语言的多语言神经机器翻译中的字序差异

获取原文

摘要

Transfer learning approaches for Neural Machine Translation (NMT) trains a NMT model on an assisting language-target language pair (parent model) which is later fine-tuned for the source language-target language pair of interest (child model), with the target language being the same. In many cases, the assisting language has a different word order from the source language. We show that divergent word order adversely limits the benefits from transfer learning when little to no parallel corpus between the source and target language is available. To bridge this divergence, we propose to pre-order the assisting language sentences to match the word order of the source language and train the parent model. Our experiments on many language pairs show that bridging the word order gap leads to major improvements in the translation quality in extremely low-resource scenarios.
机译:神经机器翻译(NMT)的转移学习方法在辅助语言-目标语言对(父模型)上训练NMT模型,随后针对目标语言对(目标模型)对源语言-目标语言对进行了微调。语言是一样的。在许多情况下,辅助语言的词序与源语言的词序不同。我们表明,当源语言和目标语言之间几乎没有平行语料库时,发散的单词顺序不利地限制了迁移学习的好处。为了弥合这种差异,我们建议对辅助语言句子进行预排序以匹配源语言的词序并训练父模型。我们在许多语言对上的实验表明,在极少资源的情况下,弥合单词顺序差距可以大大改善翻译质量。

著录项

相似文献

  • 外文文献
  • 中文文献
  • 专利
获取原文

客服邮箱:kefu@zhangqiaokeyan.com

京公网安备:11010802029741号 ICP备案号:京ICP备15016152号-6 六维联合信息科技 (北京) 有限公司©版权所有
  • 客服微信

  • 服务号