首页> 外文会议>Conference on empirical methods in natural language processing >Learning When to Concentrate or Divert Attention: Self-Adaptive Attention Temperature for Neural Machine Translation
【24h】

Learning When to Concentrate or Divert Attention: Self-Adaptive Attention Temperature for Neural Machine Translation

机译:学习何时集中注意力或转移注意力:神经机器翻译的自适应注意力温度

获取原文

摘要

Most of the Neural Machine Translation (NMT) models are based on the sequence-to-sequence (Seq2Seq) model with an encoder-decoder framework equipped with the attention mechanism. However, the conventional attention mechanism treats the decoding at each time step equally with the same matrix, which is problematic since the softness of the attention for different types of words (e.g. content words and function words) should differ. Therefore, we propose a new model with a mechanism called Self-Adaptive Control of Temperature (SACT) to control the softness of attention by means of an attention temperature. Experimental results on the Chinese-English translation and English-Vietnamese translation demonstrate that our model outperforms the baseline models, and the analysis and the case study show that our model can attend to the most relevant elements in the source-side contexts and generate the translation of high quality.
机译:大多数神经机器翻译(NMT)模型都是基于序列到序列(Seq2Seq)模型的,该模型具有配备了注意机制的编码器-解码器框架。但是,传统的注意机制用相同的矩阵在每个时间步上均等地对待解码,这是有问题的,因为对于不同类型的词(例如,内容词和功能词)的注意的软度应该不同。因此,我们提出了一种新的模型,该模型具有一种称为温度的自适应控制(SACT)的机制,可以通过注意力温度来控制注意力的柔和度。汉英翻译和英越南翻译的实验结果表明,我们的模型优于基准模型,而分析和案例研究表明,我们的模型可以处理源语言环境中最相关的元素并生成翻译高品质。

著录项

相似文献

  • 外文文献
  • 中文文献
  • 专利
获取原文

客服邮箱:kefu@zhangqiaokeyan.com

京公网安备:11010802029741号 ICP备案号:京ICP备15016152号-6 六维联合信息科技 (北京) 有限公司©版权所有
  • 客服微信

  • 服务号