首页> 外文会议>Second conference on machine translation >Effective Domain Mixing for Neural Machine Translation
【24h】

Effective Domain Mixing for Neural Machine Translation

机译:神经机器翻译的有效域混合

获取原文
获取原文并翻译 | 示例

摘要

Neural Machine Translation (NMT) models are often trained on heterogeneous mixtures of domains, from news to parliamentary proceedings, each with unique distributions and language. In this work we show that training NMT systems on naively mixed data can degrade performance versus models fit to each constituent domain We demonstrate that this problem can be circumvented, and propose three models that do so by jointly learning domain discrimination and translation. We demonstrate the efficacy of these techniques by merging pairs of domains in three languages: Chinese, French, and Japanese. After training on composite data, each approach outperforms its domain-specific counterparts, with a model based on a discriminator network doing so most reliably. We obtain consistent performance improvements and an average increase of 1.1 BLEU.
机译:神经机器翻译(NMT)模型通常在从新闻到议会程序等领域的异类混合中接受训练,每个领域都有独特的分布和语言。在这项工作中,我们证明了在天真的混合数据上训练NMT系统会降低性能,而模型适合每个组成域。我们证明了可以解决此问题的方法,并提出了通过共同学习域区分和翻译来做到这一点的三个模型。我们通过以三种语言(中文,法文和日文)合并一对域名来证明这些技术的有效性。在对复合数据进行训练之后,每种方法都比基于特定领域的方法要好,而基于鉴别器网络的模型则最可靠。我们获得了持续的性能改进,平均增加了1.1 BLEU。

著录项

相似文献

  • 外文文献
  • 中文文献
  • 专利
获取原文

客服邮箱:kefu@zhangqiaokeyan.com

京公网安备:11010802029741号 ICP备案号:京ICP备15016152号-6 六维联合信息科技 (北京) 有限公司©版权所有
  • 客服微信

  • 服务号