...
首页> 外文期刊>Neurocomputing >Effectively training neural machine translation models with monolingual data
【24h】

Effectively training neural machine translation models with monolingual data

机译:用单语数据有效地训练神经机器翻译模型

获取原文
获取原文并翻译 | 示例
   

获取外文期刊封面封底 >>

       

摘要

Improving neural machine translation models (NMT) with monolingual data has aroused more and more interests in this area and back-translation for monolingual data augmentation Sennrich et al. (2016) has been taken as a promising development recently. While the naive back-translation approach improves the translation performance substantially, we notice that its usage for monolingual data is not so effective because traditional NMT models make no distinction between the true parallel corpus and the back translated synthetic parallel corpus. This paper proposes a gate-enhanced NMT model which makes use of monolingual data more effectively. The central idea is to separate the data flow of monolingual data and parallel data into different channels by the elegant designed gate, which enables the model to perform different transformations according to the type of the input sequence, i.e., monolingual data and parallel data. Experiments on Chinese-English and English-German translation tasks show that our approach achieves substantial improvements over strong baselines and the gate-enhanced NMT model can utilize the source-side and target-side monolingual data at the same time. (C) 2018 Elsevier B.V. All rights reserved.
机译:用单语数据改进神经机器翻译模型(NMT)引起了这一领域的越来越多的兴趣,并且对单语数据增强进行了反向翻译。 (2016)最近被视为有希望的发展。尽管幼稚的逆向翻译方法大大提高了翻译性能,但我们注意到,由于传统的NMT模型没有区分真实的并行语料库和反向翻译的合成并行语料库,因此它在单语数据中的使用效果不佳。本文提出了一种门增强型NMT模型,该模型可以更有效地利用单语数据。中心思想是通过设计精美的门将单语数据和并行数据的数据流分成不同的通道,这使模型能够根据输入序列的类型(即单语数据和并行数据)执行不同的转换。在汉英和英德翻译任务上的实验表明,我们的方法在强大的基线上取得了实质性的改进,并且门增强的NMT模型可以同时利用源端和目标端的单语数据。 (C)2018 Elsevier B.V.保留所有权利。

著录项

相似文献

  • 外文文献
  • 中文文献
  • 专利
获取原文

客服邮箱:kefu@zhangqiaokeyan.com

京公网安备:11010802029741号 ICP备案号:京ICP备15016152号-6 六维联合信息科技 (北京) 有限公司©版权所有
  • 客服微信

  • 服务号