【24h】

Locally Training the Log-Linear Model for SMT

机译:本地培训SMT的对数线性模型

获取原文

摘要

In statistical machine translation, minimum error rate training (MERT) is a standard method for tuning a single weight with regard to a given development data. However, due to the diversity and uneven distribution of source sentences, there are two problems suffered by this method. First, its performance is highly dependent on the choice of a development set, which may lead to an unstable performance for testing. Second, translations become inconsistent at the sentence level since tuning is performed globally on a document level. In this paper, we propose a novel local training method to address these two problems. Unlike a global training method, such as MERT, in which a single weight is learned and used for all the input sentences, we perform training and testing in one step by learning a sentence-wise weight for each input sentence. We propose efficient incremental training methods to put the local training into practice. In NIST Chinese-to-English translation tasks, our local training method significantly outperforms MERT with the maximal improvements up to 2.0 BLEU points, meanwhile its efficiency is comparable to that of the global method.
机译:在统计机器翻译中,最小错误率训练(MERT)是一个标准方法,用于在给定的开发数据方面调整单一重量。然而,由于源句子的多样性和不均匀分布,这种方法遭受了两个问题。首先,其性能高度依赖于开发集的选择,这可能导致对测试的不稳定性能。其次,由于调整在文档级别上进行调整,翻译在句子级别变得不一致。在本文中,我们提出了一种新的本地培训方法来解决这两个问题。与默认的全局训练方法不同,其中学习了单一重量并用于所有输入句子,我们通过学习每个输入句的句子重量来在一步中执行培训和测试。我们提出了有效的增量培训方法,将本地培训付诸实践。在NIST中文 - 英文翻译任务中,我们的本地培训方法显着优于最大改善的磁力效果,最高可达2.0 BLEU积分,同时其效率与全局方法的效率相当。

著录项

相似文献

  • 外文文献
  • 中文文献
  • 专利
获取原文

客服邮箱:kefu@zhangqiaokeyan.com

京公网安备:11010802029741号 ICP备案号:京ICP备15016152号-6 六维联合信息科技 (北京) 有限公司©版权所有
  • 客服微信

  • 服务号