首页> 外文会议>Conference on empirical methods in natural language processing >Coverage Embedding Models for Neural Machine Translation
【24h】

Coverage Embedding Models for Neural Machine Translation

机译:神经机器翻译的覆盖度嵌入模型

获取原文

摘要

In this paper, we enhance the attention-based neural machine translation (NMT) by adding explicit coverage embedding models to alleviate issues of repeating and dropping translations in NMT. For each source word, our model starts with a full coverage embedding vector to track the coverage status, and then keeps updating it with neural networks as the translation goes. Experiments on the large-scale Chinese-to-English task show that our enhanced model improves the translation quality significantly on various test sets over the strong large vocabulary NMT system.
机译:在本文中,我们通过添加显式覆盖嵌入模型来减轻基于NMT的重复和删除翻译的问题,从而增强了基于注意力的神经机器翻译(NMT)。对于每个源词,我们的模型都从完整的覆盖嵌入向量开始,以跟踪覆盖状态,然后随着翻译的进行,继续使用神经网络对其进行更新。对大规模汉英任务的实验表明,在强大的大词汇量NMT系统上,我们的增强模型可以显着提高各种测试集的翻译质量。

著录项

相似文献

  • 外文文献
  • 中文文献
  • 专利
获取原文

客服邮箱:kefu@zhangqiaokeyan.com

京公网安备:11010802029741号 ICP备案号:京ICP备15016152号-6 六维联合信息科技 (北京) 有限公司©版权所有
  • 客服微信

  • 服务号