首页> 外文会议>ANLP 2011 >Evaluating Human Correction Quality for Machine Translation from Crowdsourcing
【24h】

Evaluating Human Correction Quality for Machine Translation from Crowdsourcing

机译:评估人力校正质量,从众包中翻译

获取原文

摘要

Machine translation (MT) technology is becoming more and more pervasive, yet the quality of MT output is still not ideal. Thus, human corrections are used to edit the output for further studies. However, how to judge the human correction might be tricky when the annotators are not experts. We present a novel way that uses cross-validation to automatically judge the human corrections where each MT output is corrected by more than one annotator. Cross-validation among corrections for the same machine translation, and among corrections from the same annotator are both applied. We get a correlation around 40% in sentence quality for Chinese-English and Spanish-English. We also evaluate the user quality as well. At last, we rank the quality of human corrections from good to bad, which enables us to set a quality threshold to make a trade-off between the scope and the quality of the corrections.
机译:机器翻译(MT)技术越来越普遍,但MT输出的质量仍然不理想。因此,人类校正用于编辑输出以进行进一步研究。但是,当注释者不是专家时,如何判断人类更正可能很棘手。我们提出了一种新颖的方式,它使用交叉验证自动判断人类校正,其中每个MT输出都通过多个注释器校正。相同机器翻译的校正之间的交叉验证,以及应用来自相同的注释器的校正。对于汉语 - 英语和西班牙语的句子质量,我们得到了左右40%的相关性。我们还评估了用户质量。最后,我们将人类修正质量从好到糟糕排名,这使我们能够设定质量阈值,以便在校正的范围和质量之间进行权衡。

著录项

相似文献

  • 外文文献
  • 中文文献
  • 专利
获取原文

客服邮箱:kefu@zhangqiaokeyan.com

京公网安备:11010802029741号 ICP备案号:京ICP备15016152号-6 六维联合信息科技 (北京) 有限公司©版权所有
  • 客服微信

  • 服务号