首页> 外文期刊>JMLR: Workshop and Conference Proceedings >Avoiding Latent Variable Collapse with Generative Skip Models
【24h】

Avoiding Latent Variable Collapse with Generative Skip Models

机译:使用生成跳过模型避免潜在的变量崩溃

获取原文
           

摘要

Variational autoencoders (VAEs) learn distributions of high-dimensional data. They model data with a deep latent-variable model and then fit the model by maximizing a lower bound of the log marginal likelihood. VAEs can capture complex distributions, but they can also suffer from an issue known as "latent variable collapse," especially if the likelihood model is powerful. Specifically, the lower bound involves an approximate posterior of the latent variables; this posterior "collapses" when it is set equal to the prior, i.e., when the approximate posterior is independent of the data. While VAEs learn good generative models, latent variable collapse prevents them from learning useful representations. In this paper, we propose a simple new way to avoid latent variable collapse by including skip connections in our generative model; these connections enforce strong links between the latent variables and the likelihood function. We study generative skip models both theoretically and empirically. Theoretically, we prove that skip models increase the mutual information between the observations and the inferred latent variables. Empirically, we study images (MNIST and Omniglot) and text (Yahoo). Compared to existing VAE architectures, we show that generative skip models maintain similar predictive performance but lead to less collapse and provide more meaningful representations of the data.
机译:可变自动编码器(VAE)学习高维数据的分布。他们使用深潜变量模型对数据建模,然后通过最大化对数边际可能性的下限来拟合模型。 VAE可以捕获复杂的分布,但是它们也可能遭受称为“潜在变量崩溃”的问题,尤其是在可能性模型强大的情况下。具体而言,下限涉及潜在变量的近似后验;当它等于先验时,即当近似后验与数据无关时,该后验“崩溃”。当VAE学习好的生成模型时,潜在变量崩溃会阻止他们学习有用的表示形式。在本文中,我们提出了一种简单的新方法,通过在生成模型中包含跳过连接来避免潜在变量崩溃。这些连接在潜在变量和似然函数之间建立了强有力的联系。我们在理论和经验上研究生成跳跃模型。从理论上讲,我们证明了跳跃模型增加了观测值和推断的潜在变量之间的相互信息。根据经验,我们研究图像(MNIST和Omniglot)和文本(Yahoo)。与现有的VAE架构相比,我们显示生成跳过模型保持相似的预测性能,但导致更少的崩溃,并提供了更有意义的数据表示形式。

著录项

相似文献

  • 外文文献
  • 中文文献
  • 专利
获取原文

客服邮箱:kefu@zhangqiaokeyan.com

京公网安备:11010802029741号 ICP备案号:京ICP备15016152号-6 六维联合信息科技 (北京) 有限公司©版权所有
  • 客服微信

  • 服务号