首页> 外文期刊>Information Sciences: An International Journal >Recurrent neural network-based semantic variational autoencoder for Sequence-to-sequence learning
【24h】

Recurrent neural network-based semantic variational autoencoder for Sequence-to-sequence learning

机译:基于经常性的神经网络的语义变分自身级别,用于序列到序列学习

获取原文
获取原文并翻译 | 示例
           

摘要

Sequence-to-sequence (Seq2seq) models have played an important role in the recent success of various natural language processing methods, such as machine translation, text summarization, and speech recognition. However, current Seq2seq models have trouble preserving global latent information from a long sequence of words. Variational autoencoder (VAE) alleviates this problem by learning a continuous semantic space of the input sentence. However, it does not solve the problem completely. In this paper, we propose a new recurrent neural network (RNN)-based Seq2seq model, RNN semantic variational autoencoder (RNN-SVAE), to better capture the global latent information of a sequence of words. To suitably reflect the meanings of words in a sentence regardless of their position within the sentence, we utilized two approaches: (1) constructing a document information vector based on the attention information between the final state of the encoder and every prior hidden state, and (2) extracting the semantic vector based on the self-attention mechanism. Then, the mean and standard deviation of the continuous semantic space are learned by using this vector to take advantage of the variational method. By using the document information vector and the self-attention mechanism to find the semantic space of the sentence, it becomes possible to better capture the global latent feature of the sentence. Experimental results of three natural language tasks (i.e., language modeling, missing word imputation, paraphrase identification) confirm that the proposed RNN-SVAE yields higher performance than two benchmark models. (C) 2019 Elsevier Inc. All rights reserved.
机译:序列到序列(SEQ2Seq)模型在最近的各种自然语言处理方法的成功中发挥了重要作用,例如机器翻译,文本摘要和语音识别。但是,目前的SEQ2Seq模型无法从长时间的单词中保留全局潜在信息。变形AutoEncoder(VAE)通过学习输入句子的连续语义空间来减轻这个问题。但是,它没有完全解决问题。在本文中,我们提出了一种新的复发性神经网络(RNN)的SEQ2SEQ模型,RNN语义变形AutoEncoder(RNN-SVAE),以更好地捕获一系列单词的全局潜入信息。为了适当地反映句子中的单词的含义,无论其句子内的位置如何,我们都使用了两种方法:(1)基于编码器的最终状态和每个先前隐藏状态之间的注意信息构建文档信息矢量。 (2)基于自我关注机制提取语义矢量。然后,通过使用该向量来利用变分方法来学习连续语义空间的平均值和标准偏差。通过使用文档信息矢量和自我关注机制来查找句子的语义空间,可以更好地捕获句子的全局潜在特征。三种自然语言任务的实验结果(即语言建模,缺少词归档,解释识别)证实,所提出的RNN-SVAE产生比两个基准模型更高的性能。 (c)2019 Elsevier Inc.保留所有权利。

著录项

相似文献

  • 外文文献
  • 中文文献
  • 专利
获取原文

客服邮箱:kefu@zhangqiaokeyan.com

京公网安备:11010802029741号 ICP备案号:京ICP备15016152号-6 六维联合信息科技 (北京) 有限公司©版权所有
  • 客服微信

  • 服务号