...
首页> 外文期刊>Neurocomputing >Context-aware positional representation for self-attention networks
【24h】

Context-aware positional representation for self-attention networks

机译:上下文感知的自我关注网络的位置表示

获取原文
获取原文并翻译 | 示例
           

摘要

In self-attention networks (SANs), positional embeddings are used to model order dependencies between words in the input sentence and are added with word embeddings to gain an input representation, which enables the SAN-based neural model to perform (multi-head) and to stack (multi-layer) self-attentive functions in parallel to learn the representation of the input sentence. However, this input representation only involves static order dependencies based on discrete position indexes of words, that is, is independent of context information, which may be weak in modeling the input sentence. To address this issue, we proposed a novel positional representation method to model order dependencies based on n-gram context or sentence context in the input sentence, which allows SANs to learn a more effective sentence representation. To validate the effectiveness of the proposed method, it is applied to the neural machine translation model, which adopts a typical SAN-based neural model. Experimental results on two widely used translation tasks, i.e., WMT14 English-to-German and WMT17 Chinese-to-English, showed that the proposed approach can significantly improve the translation performance over the strong Transformer baseline. (c) 2021 Elsevier B.V. All rights reserved.
机译:在自我关注网络(SAN)中,位置嵌入式用于在输入句中的单词之间进行模拟顺序依赖性,并在Word Embeddings中添加以获得输入表示,这使得基于SAN的神经模型执行(多头)并并行堆叠(多层)自我细分功能,以了解输入句子的表示。然而,该输入表示仅涉及基于单词的离散位置索引的静态顺序依赖性,即,它与上下文信息无关,在模拟输入句时可能弱。为解决此问题,我们提出了一种基于输入句子中的n-gram上下文或句子上下文的模型顺序依赖性的新型位置表示方法,这允许SAN学习更有效的句子表示。为了验证所提出的方法的有效性,它适用于神经机翻译模型,采用典型的基于SAN的神经模型。两种广泛使用的翻译任务的实验结果,即WMT14英语到德语和WMT17中英文,表明所提出的方法可以显着提高强大变压器基线的翻译性能。 (c)2021 elestvier b.v.保留所有权利。

著录项

相似文献

  • 外文文献
  • 中文文献
  • 专利
获取原文

客服邮箱:kefu@zhangqiaokeyan.com

京公网安备:11010802029741号 ICP备案号:京ICP备15016152号-6 六维联合信息科技 (北京) 有限公司©版权所有
  • 客服微信

  • 服务号