首页> 外文期刊>JMLR: Workshop and Conference Proceedings >Improved Variational Autoencoders for Text Modeling using Dilated Convolutions
【24h】

Improved Variational Autoencoders for Text Modeling using Dilated Convolutions

机译:改进的变分自动编码器,用于使用膨胀卷积的文本建模

获取原文
           

摘要

Recent work on generative text modeling has found that variational autoencoders (VAE) with LSTM decoders perform worse than simpler LSTM language models (Bowman et al., 2015). This negative result is so far poorly understood, but has been attributed to the propensity of LSTM decoders to ignore conditioning information from the encoder. In this paper, we experiment with a new type of decoder for VAE: a dilated CNN. By changing the decoder’s dilation architecture, we control the size of context from previously generated words. In experiments, we find that there is a trade-off between contextual capacity of the decoder and effective use of encoding information. We show that when carefully managed, VAEs can outperform LSTM language models. We demonstrate perplexity gains on two datasets, representing the first positive language modeling result with VAE. Further, we conduct an in-depth investigation of the use of VAE (with our new decoding architecture) for semi-supervised and unsupervised labeling tasks, demonstrating gains over several strong baselines.
机译:有关生成文本建模的最新研究发现,带有LSTM解码器的变体自动编码器(VAE)的性能比简单的LSTM语言模型差(Bowman等,2015)。到目前为止,这种负面结果了解得很少,但归因于LSTM解码器倾向于忽略来自编码器的条件信息。在本文中,我们尝试使用一种新型的VAE解码器:膨胀的CNN。通过更改解码器的扩散架构,我们可以控制以前生成的单词的上下文大小。在实验中,我们发现解码器的上下文容量和编码信息的有效使用之间存在折衷。我们表明,经过精心管理,VAE可以胜过LSTM语言模型。我们展示了两个数据集上的困惑感,这代表了VAE的第一个积极的语言建模结果。此外,我们针对半监督和无监督的标记任务对VAE(使用我们的新解码体系结构)的使用进行了深入研究,展示了在多个强大基准上的收益。

著录项

相似文献

  • 外文文献
  • 中文文献
  • 专利
获取原文

客服邮箱:kefu@zhangqiaokeyan.com

京公网安备:11010802029741号 ICP备案号:京ICP备15016152号-6 六维联合信息科技 (北京) 有限公司©版权所有
  • 客服微信

  • 服务号