首页> 外文会议>Conference on machine translation;Annual meeting of the Association for Computational Linguistics >Transformer-based Automatic Post-Editing Model with Joint Encoder and Multi-source Attention of Decoder
【24h】

Transformer-based Automatic Post-Editing Model with Joint Encoder and Multi-source Attention of Decoder

机译:联合编码器与解码器多源关注的基于变压器的自动后编辑模型

获取原文

摘要

This paper describes POSTECH's submission to the WMT 2019 shared task on Automatic Post-Editing (APE). In this paper, we propose a new multi-source APE model by extending Transformer. The main contributions of our study are that we 1) reconstruct the encoder to generate a joint representation of translation (ml) and its src context, in addition to the conventional src encoding and 2) suggest two types of multi-source attention layers to compute attention between two outputs of the encoder and the decoder state in the decoder. Furthermore, we train our model by applying various teacher-forcing ratios to alleviate exposure bias. Finally, we adopt the ensemble technique across variations of our model. Experiments on the WMT19 English-German APE data set show improvements in terms of both TER and BLEU scores over the baseline. Our primary submission achieves -0.73 in TER and +1.49 in BLEU compared to the baseline, and ranks second among all submitted systems.
机译:本文介绍POSTECH提交给WMT 2019自动后期编辑(APE)共享任务的内容。在本文中,我们通过扩展Transformer提出了一种新的多源APE模型。我们研究的主要贡献在于,我们(1)重构编码器以生成翻译(ml)及其src上下文的联合表示,除了传统的src编码之外,以及2)建议两种类型的多源关注层进行计算编码器的两个输出之间的注意和解码器中的解码器状态。此外,我们通过应用各种教师强迫率来减轻暴露偏见,从而训练我们的模型。最后,我们在模型的各个变体之间采用了集成技术。在WMT19英德APE数据集上进行的实验表明,相对于基线而言,TER和BLEU得分均有所提高。与基准相比,我们的主要提交在TER上达到-0.73,在BLEU上达到+1.49,在所有提交的系统中排名第二。

著录项

相似文献

  • 外文文献
  • 中文文献
  • 专利
获取原文

客服邮箱:kefu@zhangqiaokeyan.com

京公网安备:11010802029741号 ICP备案号:京ICP备15016152号-6 六维联合信息科技 (北京) 有限公司©版权所有
  • 客服微信

  • 服务号