首页> 外文期刊>Audio, Speech, and Language Processing, IEEE/ACM Transactions on >Linguistic Knowledge-Aware Neural Machine Translation
【24h】

Linguistic Knowledge-Aware Neural Machine Translation

机译:语言知识感知神经机器翻译

获取原文
获取原文并翻译 | 示例
           

摘要

Recently, researchers have shown an increasing interest in incorporating linguistic knowledge into neural machine translation (NMT). To this end, previous works choose either to alter the architecture of NMT encoder to incorporate syntactic information into the translation model, or to generalize the embedding layer of the encoder to encode additional linguistic features. The former approach mainly focuses on injecting the syntactic structure of the source sentence into the encoding process, leading to a complicated model that lacks the flexibility to incorporate other types of knowledge. The latter extends word embeddings by considering additional linguistic knowledge as features to enrich the word representation. It thus does not explicitly balance the contribution from word embeddings and the contribution from additional linguistic knowledge. To address these limitations, this paper proposes a knowledge-aware NMT approach that models additional linguistic features in parallel to the word feature. The core idea is that we propose modeling a series of linguistic features at the word level (knowledge block) using a recurrent neural network (RNN). And in sentence level, those word-corresponding feature blocks are further encoded using a RNN encoder. In decoding, we propose a knowledge gate and an attention gate to dynamically control the proportions of information contributing to the generation of target words from different sources. Extensive experiments show that our approach is capable of better accounting for importance of additional linguistic, and we observe significant improvements from 1.0 to 2.3 BLEU points on Chinesen$leftrightarrow$nEnglish and Englishn$rightarrow$nGerman translation tasks.
机译:最近,研究人员对将语言知识整合到神经机器翻译(NMT)中表现出了越来越高的兴趣。为此,先前的工作选择改变NMT编码器的体系结构以将句法信息合并到翻译模型中,或者一般化编码器的嵌入层以对其他语言特征进行编码。前一种方法主要着重于将源句子的句法结构注入编码过程,从而导致复杂的模型,缺乏整合其他类型知识的灵活性。后者通过将其他语言知识视为丰富单词表示的功能来扩展单词嵌入。因此,它没有明确地平衡单词嵌入的贡献和其他语言知识的贡献。为了解决这些局限性,本文提出了一种知识感知的NMT方法,该方法可以模拟与单词功能并行的其他语言功能。核心思想是我们建议使用递归神经网络(RNN)在单词级别(知识块)对一系列语言特征进行建模。并且在句子级别,使用RNN编码器进一步对那些单词对应的功能块进行编码。在解码中,我们提出了一个知识门和一个注意门,以动态控制有助于从不同来源生成目标词的信息比例。大量的实验表明,我们的方法能够更好地说明其他语言的重要性,并且在Chinesen $ leftrightarrow $ n英语和英语n $ rightarrow $ 德语翻译任务。

著录项

  • 来源
  • 作者单位

    Natural Language Processing Laboratory, School of Computer Science and Engineering, Northeastern University, Shenyang, China;

    Natural Language Processing and Portuguese-Chinese Machine Translation Laboratory, University of Macau, Macau, China;

    Natural Language Processing and Portuguese-Chinese Machine Translation Laboratory, University of Macau, Macau, China;

    Alibaba Inc, Hangzhou, China;

    Natural Language Processing Laboratory, School of Computer Science and Engineering, Northeastern University, Shenyang, China;

    Natural Language Processing Laboratory, School of Computer Science and Engineering, Northeastern University, Shenyang, China;

    School of Computer Science and Technology, Institute of Artificial Intelligence, Soochow University, Suzhou, China;

  • 收录信息
  • 原文格式 PDF
  • 正文语种 eng
  • 中图分类
  • 关键词

    Linguistics; Syntactics; Decoding; Speech processing; Logic gates; Recurrent neural networks; Encoding;

    机译:语言学;句法;解码;语音处理;逻辑门;递归神经网络;编码;

相似文献

  • 外文文献
  • 中文文献
  • 专利
获取原文

客服邮箱:kefu@zhangqiaokeyan.com

京公网安备:11010802029741号 ICP备案号:京ICP备15016152号-6 六维联合信息科技 (北京) 有限公司©版权所有
  • 客服微信

  • 服务号