首页> 外文会议>International conference on neural information processing >Wake-Sleep Variational Autoencoders for Language Modeling
【24h】

Wake-Sleep Variational Autoencoders for Language Modeling

机译:唤醒睡眠变体自动编码器,用于语言建模

获取原文

摘要

Variational Autoencoders (VAEs) are known to easily suffer from the KL-vanishing problem when combining with powerful autore-gressive models like recurrent neural networks (RNNs), which prohibits their wide application in natural language processing. In this paper, we tackle this problem by tearing the training procedure into two steps: learning effective mechanisms to encode and decode discrete tokens (wake step) and generalizing meaningful latent variables by reconstructing dreamed encodings (sleep step). The training pattern is similar to the wake-sleep algorithm: these two steps are trained alternatively until an equilibrium is achieved. We test our model in a language modeling task. The results demonstrate significant improvement over the current state-of-the-art latent variable models.
机译:众所周知,与自动递归模型(如递归神经网络(RNN))结合使用时,变异自动编码器(VAE)容易遭受KL消失的问题,这限制了它们在自然语言处理中的广泛应用。在本文中,我们通过将训练过程分为两个步骤来解决此问题:学习有效的机制来对离散标记进行编码和解码(唤醒步骤),并通过重构梦想的编码(睡眠步骤)来概括有意义的潜在变量。训练模式类似于唤醒睡眠算法:交替训练这两个步骤,直到达到平衡为止。我们在语言建模任务中测试我们的模型。结果表明,与当前最新的潜在变量模型相比,该方法有显着改进。

著录项

相似文献

  • 外文文献
  • 中文文献
  • 专利
获取原文

客服邮箱:kefu@zhangqiaokeyan.com

京公网安备:11010802029741号 ICP备案号:京ICP备15016152号-6 六维联合信息科技 (北京) 有限公司©版权所有
  • 客服微信

  • 服务号