【24h】

Efficient Minimum Error Rate Training and Minimum Bayes-Risk Decoding for Translation Hypergraphs and Lattices

机译:翻译超图和格的有效最小错误率训练和最小贝叶斯风险解码

获取原文

摘要

Minimum Error Rate Training (MERT) and Minimum Bayes-Risk (MBR) decoding are used in most current state-of-the-art Statistical Machine Translation (SMT) systems. The algorithms were originally developed to work with N-best lists of translations, and recently extended to lattices that encode many more hypotheses than typical N-best lists. We here extend lattice-based MERT and MBR algorithms to work with hypergraphs that encode a vast number of translations produced by MT systems based on Synchronous Context Free Grammars. These algorithms are more efficient than the lattice-based versions presented earlier. We show how MERT can be employed to optimize parameters for MBR decoding. Our experiments show speedups from MERT and MBR as well as performance improvements from MBR decoding on several language pairs.
机译:最小错误率训练(MERT)和最小贝叶斯风险(MBR)解码用于大多数最新的统计机器翻译(SMT)系统中。该算法最初是为处理N个最佳翻译列表而开发的,最近扩展到了比典型N个最佳列表编码更多假设的晶格。我们在这里扩展了基于格的MERT和MBR算法,以与超图一起使用,这些超图对MT系统基于同步上下文无关文法的大量翻译进行编码。这些算法比前面介绍的基于网格的版本更有效。我们展示了如何使用MERT来优化MBR解码的参数。我们的实验显示了MERT和MBR的提速,以及几种语言对的MBR解码的性能提升。

相似文献

  • 外文文献
  • 中文文献
  • 专利
获取原文

客服邮箱:kefu@zhangqiaokeyan.com

京公网安备:11010802029741号 ICP备案号:京ICP备15016152号-6 六维联合信息科技 (北京) 有限公司©版权所有
  • 客服微信

  • 服务号