首页> 外文会议>Annual meeting of the Association for Computational Linguistics >Enhancing Language Models in Statistical Machine Translation with Backward N-grams and Mutual Information Triggers
【24h】

Enhancing Language Models in Statistical Machine Translation with Backward N-grams and Mutual Information Triggers

机译:用落后n-grams增强统计机器翻译中的语言模型和互信息触发器

获取原文

摘要

In this paper, with a belief that a language model that embraces a larger context provides better prediction ability, we present two extensions to standard n-gram language models in statistical machine translation: a backward language model that augments the conventional forward language model, and a mutual information trigger model which captures long-distance dependencies that go beyond the scope of standard n-gram language models. We integrate the two proposed models into phrase-based statistical machine translation and conduct experiments on large-scale training data to investigate their effectiveness. Our experimental results show that both models are able to significantly improve translation quality and collectively achieve up to 1 BLEU point over a competitive baseline.
机译:在本文中,通过相信包含较大上下文的语言模型提供更好的预测能力,我们向统计机器翻译中的标准N-Gram语言模型提供了两个扩展:向后语言模型增强传统的前向语言模型,以及一个互信息触发模型,捕获超出标准N-GRAM语言模型范围的远程距离。我们将两种提出的模型集成到基于短语的统计机器翻译中,并对大规模培训数据进行实验,以调查其有效性。我们的实验结果表明,两种模型都能够显着提高翻译质量,并在竞争性基线上集体达到最多1个BLEU点。

著录项

相似文献

  • 外文文献
  • 中文文献
  • 专利
获取原文

客服邮箱:kefu@zhangqiaokeyan.com

京公网安备:11010802029741号 ICP备案号:京ICP备15016152号-6 六维联合信息科技 (北京) 有限公司©版权所有
  • 客服微信

  • 服务号