首页> 外文期刊>Machine translation >Machine translation evaluation versus quality estimation
【24h】

Machine translation evaluation versus quality estimation

机译:机器翻译评估与质量评估

获取原文
获取原文并翻译 | 示例
       

摘要

Most evaluation metrics for machine translation (MT) require reference translations for each sentence in order to produce a score reflecting certain aspects of its quality. The de facto metrics, BLEU and NIST, are known to have good correlation with human evaluation at the corpus level, but this is not the case at the segment level. As an attempt to overcome these two limitations, we address the problem of evaluating the quality of MT as a prediction task, where reference-independent features are extracted from the input sentences and their translation, and a quality score is obtained based on models produced from training data. We show that this approach yields better correlation with human evaluation as compared to commonly used metrics, even with models trained on different MT systems, language-pairs and text domains.
机译:机器翻译(MT)的大多数评估指标都要求对每个句子进行参考翻译,以产生反映其质量某些方面的分数。已知事实上的指标BLEU和NIST与语料库级别的人类评估具有良好的相关性,但是在段级别却并非如此。为了克服这两个局限性,我们解决了将MT的质量评估为预测任务的问题,其中从输入的句子及其翻译中提取与参考无关的特征,并根据从训练数据。我们显示,与常用指标相比,即使使用在不同MT系统,语言对和文本域上训练的模型,该方法也能更好地与人工评估相关。

著录项

相似文献

  • 外文文献
  • 中文文献
  • 专利
获取原文

客服邮箱:kefu@zhangqiaokeyan.com

京公网安备:11010802029741号 ICP备案号:京ICP备15016152号-6 六维联合信息科技 (北京) 有限公司©版权所有
  • 客服微信

  • 服务号