首页> 外文会议>Annual meeting of the Association for Computational Linguistics >Multilingual Unsupervised NMT using Shared Encoder and Language-Specific Decoders
【24h】

Multilingual Unsupervised NMT using Shared Encoder and Language-Specific Decoders

机译:使用共享编码器和特定语言解码器的多语种无监督NMT

获取原文

摘要

In this paper, we propose a multilingual unsupervised NMT scheme which jointly trains multiple languages with a shared encoder and multiple decoders. Our approach is based on denoising autoencoding of each language and back-translating between English and multiple non-English languages. This results in a universal encoder which can encode any language participating in training into an interlingual representation, and language-specific decoders. Our experiments using only monolingual corpora show that multilingual unsupervised model performs better than the separately trained bilingual models achieving improvement of up to 1.48 BLEU points on WMT test sets. We also observe that even if we do not train the network for all possible translation directions, the network is still able to translate in a many-to-many fashion leveraging encoder's ability to generate interlingual representation.
机译:在本文中,我们提出了一种多语言无监督的NMT计划,该方案与共享编码器和多个解码器共同列举多种语言。我们的方法是基于每种语言的去噪和返回翻译的英语和多种非英语之间的返回翻译。这导致通用编码器,可以将参与培训的任何语言编码为间隔表示,以及特定于语言的解码器。我们的实验仅使用单语语料库表明,多语言无监督模型比在WMT测试组上实现高达1.48的BLEU点的分别培训双语模型更好地表现更好。我们还观察到即使我们没有为所有可能的翻译方向训练网络,网络仍然能够在利用编码器生成间歇表示的能力中转换多对多的时尚。

著录项

相似文献

  • 外文文献
  • 中文文献
  • 专利
获取原文

客服邮箱:kefu@zhangqiaokeyan.com

京公网安备:11010802029741号 ICP备案号:京ICP备15016152号-6 六维联合信息科技 (北京) 有限公司©版权所有
  • 客服微信

  • 服务号