首页> 外文期刊>Machine translation >Significance tests of automatic machine translation evaluation metrics
【24h】

Significance tests of automatic machine translation evaluation metrics

机译:自动机器翻译评估指标的意义测试

获取原文
获取原文并翻译 | 示例
       

摘要

Automatic evaluation metrics for Machine Translation (MT) systems, such as BLEU, METEOR and the related NIST metric, are becoming increasingly important in MT research and development. This paper presents a significance test-driven comparison of n-gram-based automatic MT evaluation metrics. Statistical significance tests use bootstrapping methods to estimate the reliability of automatic machine translation evaluations. Based on this reliability estimation, we study the characteristics of different MT evaluation metrics and how to construct reliable and efficient evaluation suites.
机译:机器翻译(MT)系统的自动评估指标,例如BLEU,METEOR和相关的NIST指标,在MT研发中变得越来越重要。本文介绍了基于n-gram的自动MT评估指标的显着性测试驱动比较。统计显着性测试使用自举方法来估计自动机器翻译评估的可靠性。基于此可靠性评估,我们研究了不同MT评估指标的特征以及如何构建可靠而有效的评估套件。

著录项

相似文献

  • 外文文献
  • 中文文献
  • 专利
获取原文

客服邮箱:kefu@zhangqiaokeyan.com

京公网安备:11010802029741号 ICP备案号:京ICP备15016152号-6 六维联合信息科技 (北京) 有限公司©版权所有
  • 客服微信

  • 服务号