首页> 外文会议>International Conference on Natural Language Processing and Chinese Computing >Look-Ahead Attention for Generation in Neural Machine Translation
【24h】

Look-Ahead Attention for Generation in Neural Machine Translation

机译:神经电机翻译中一代的展望

获取原文

摘要

The attention model has become a standard component in neural machine translation (NMT) and it guides translation process by selectively focusing on parts of the source sentence when predicting each target word. However, we find that the generation of a target word does not only depend on the source sentence, but also rely heavily on the previous generated target words, especially the distant words which are difficult to model by using recurrent neural networks. To solve this problem, we propose in this paper a novel look-ahead attention mechanism for generation in NMT, which aims at directly capturing the dependency relationship between target words. We further design three patterns to integrate our look-ahead attention into the conventional attention model. Experiments on NIST Chinese-to-English and WMT English-to-German translation tasks show that our proposed look-ahead attention mechanism achieves substantial improvements over state-of-the-art baselines.
机译:注意模型已成为神经机翻译(NMT)中的标准组件,并且在预测每个目标字时选择性地专注于源句的部分来引导翻译过程。然而,我们发现目标词的生成不仅取决于源句,而且依赖于先前生成的目标词,尤其是通过使用经常性神经网络难以模型的远端词。为了解决这个问题,我们提出了一种新的纽姆中发电的新推理注意机制,旨在直接捕获目标词之间的依赖关系。我们进一步设计了三种模式,以将我们的展望注意力整合到传统的注意力模型中。 NIST中文与英语和WMT英语到德语翻译任务的实验表明,我们提出的展望领先的注意机制实现了最先进的基线的大量改进。

著录项

相似文献

  • 外文文献
  • 中文文献
  • 专利
获取原文

客服邮箱:kefu@zhangqiaokeyan.com

京公网安备:11010802029741号 ICP备案号:京ICP备15016152号-6 六维联合信息科技 (北京) 有限公司©版权所有
  • 客服微信

  • 服务号