首页> 外文会议>Conference of the Association for Machine Translation in the Americas >Context Models for OOV Word Translation in Low-Resource Languages
【24h】

Context Models for OOV Word Translation in Low-Resource Languages

机译:低资源语言中OOV字转换的上下文模型

获取原文

摘要

Out-of-vocabulary word translation is a major problem for the translation of low-resource languages that suffer from a lack of parallel training data. This paper evaluates the contributions of target-language context models towards the translation of OOV words, specifically in those cases where OOV translations are derived from external knowledge sources, such as dictionaries. We develop both neural and non-neural context models and evaluate them within both phrase-based and self-attention based neural machine translation systems. Our results show that neural language models that integrate additional context beyond the current sentence are the most effective in disambiguating possible OOV word translations. We present an efficient second-pass lattice-rescoring method for wide-context neural language models and demonstrate performance improvements over state-of-the-art self-attention based neural MT systems in five out of six low-resource language pairs.
机译:失败的单词翻译是缺乏平行训练数据的低资源语言翻译的主要问题。本文评估了目标语言上下文模型对OOV字的翻译的贡献,特别是在从外部知识源导出的那些oov翻译的情况下,例如词典。我们开发了神经和非神经背景模型,并在基于短语和基于自我关注的神经机翻译系统中进行评估。我们的结果表明,整合其他上下文超出当前句子的神经语言模型是歧义可能的OOV字翻译最有效的。我们为广泛的神经语言模型提出了一种有效的二手格式救援方法,并在六个低资源对中的五个中,展示了基于最先进的自我关注的神经MT系统的性能改进。

著录项

相似文献

  • 外文文献
  • 中文文献
  • 专利
获取原文

客服邮箱:kefu@zhangqiaokeyan.com

京公网安备:11010802029741号 ICP备案号:京ICP备15016152号-6 六维联合信息科技 (北京) 有限公司©版权所有
  • 客服微信

  • 服务号