【24h】

Effort-Aware Neural Automatic Post-Editing

机译:意识型神经自动后期编辑

获取原文

摘要

For this round of the WMT 2019 APE shared task, our submission focuses on addressing the "over-correction" problem in APE. Overcorrection occurs when the APE system tends to rephrase an already correct MT output, and the resulting sentence is penalized by a reference-based evaluation against human post-edits. Our intuition is that this problem can be prevented by informing the system about the predicted quality of the MT output or, in other terms, the expected amount of needed corrections. For this purpose, following the common approach in multilingual NMT, we prepend a special token to the beginning of both the source text and the MT output indicating the required amount of post-editing. Following the best submissions to the WMT 2018 APE shared task, our backbone architecture is based on multi-source Transformer to encode both the MT output and the corresponding source text. We participated both in the English-German and English-Russian sub-tasks. In the first subtask, our best submission improved the original MT output quality up to +0.98 BLEU and -0.47 TER. In the second subtask, where the higher quality of the MT output increases the risk of over-correction, none of our submitted runs was able to improve the MT output.
机译:对于本轮WMT 2019 APE共享任务,我们的提交重点在于解决APE中的“过度校正”问题。当APE系统倾向于改写已经正确的MT输出时,会发生过度校正,并且针对人工后期编辑,基于参考的评估会惩罚所得到的句子。我们的直觉是,可以通过将MT输出的预测质量或换句话说,预期的所需校正量告知系统来防止此问题。为此,遵循多语言NMT中的通用方法,我们在源文本和MT输出的开头都添加了一个特殊标记,指示所需的后期编辑量。遵循WMT 2018 APE共享任务的最佳提交方式之后,我们的主干体系结构基于多源Transformer对MT输出和相应的源文本进行编码。我们参加了英语-德语和英语-俄罗斯子任务。在第一个子任务中,我们最好的提交将原始MT输出质量提高了+0.98 BLEU和-0.47 TER。在第二个子任务中,更高质量的MT输出会增加过度校正的风险,我们提交的运行都无法改善MT输出。

著录项

相似文献

  • 外文文献
  • 中文文献
  • 专利
获取原文

客服邮箱:kefu@zhangqiaokeyan.com

京公网安备:11010802029741号 ICP备案号:京ICP备15016152号-6 六维联合信息科技 (北京) 有限公司©版权所有
  • 客服微信

  • 服务号