首页> 外文会议>Second conference on machine translation >Biasing Attention-Based Recurrent Neural Networks Using External Alignment Information
【24h】

Biasing Attention-Based Recurrent Neural Networks Using External Alignment Information

机译:使用外部比对信息对基于注意力的偏向神经网络进行偏见

获取原文
获取原文并翻译 | 示例
获取外文期刊封面目录资料

摘要

This work explores extending attention-based neural models to include alignment information as input. We modify the attention component to have dependence on the current source position. The attention model is then used as a lexical model together with an additional alignment model to generate translation. The attention model is trained using external alignment information, and it is applied in decoding by performing beam search over the lexical and alignment hypotheses. The alignment model is used to score these alignment candidates. We demonstrate that the attention layer is capable of using the alignment information to improve over the baseline attention model that uses no such alignments. Our experiments are performed on two tasks: WMT 2016 English→Romaman and WMT 2017 German→English.
机译:这项工作探索扩展基于注意力的神经模型,以包括对齐信息作为输入。我们修改注意力组件以使其依赖于当前源位置。然后,将注意力模型与附加对齐模型一起用作词汇模型,以生成翻译。使用外部对齐信息训练注意力模型,并通过对词汇和对齐假设进行波束搜索将其应用于解码。对齐模型用于对这些对齐候选进行评分。我们证明关注层能够使用对齐信息来改进不使用此类对齐的基线关注模型。我们的实验是在两项任务上执行的:WMT 2016 English→Romaman和WMT 2017 German→English。

著录项

  • 来源
  • 会议地点 Copenhagen(DK)
  • 作者

    Tamer Alkhouli; Hermann Ney;

  • 作者单位

    Human Language Technology and Pattern Recognition Group Computer Science Department RWTH Aachen University D-52056 Aachen, Germany;

    Human Language Technology and Pattern Recognition Group Computer Science Department RWTH Aachen University D-52056 Aachen, Germany;

  • 会议组织
  • 原文格式 PDF
  • 正文语种 eng
  • 中图分类
  • 关键词

相似文献

  • 外文文献
  • 中文文献
  • 专利
获取原文

客服邮箱:kefu@zhangqiaokeyan.com

京公网安备:11010802029741号 ICP备案号:京ICP备15016152号-6 六维联合信息科技 (北京) 有限公司©版权所有
  • 客服微信

  • 服务号