【24h】

Implicit Deep Latent Variable Models for Text Generation

机译:用于文本生成的隐式深潜变量模型

获取原文

摘要

Deep latent variable models (LVM) such as variational auto-encoder (VAE) have recently played an important role in text generation. One key factor is the exploitation of smooth latent structures to guide the generation. However, the representation power of VAEs is limited due to two reasons: (1) the Gaussian assumption is often made on the variational posteriors: and meanwhile (2) a notorious "posterior collapse" issue occurs. In this paper, we advocate sample-based representations of variational distributions for natural language, leading to implicit latent features, which can provide flexible representation power compared with Gaussian-based posteriors. We further develop an LVM to directly match the aggregated posterior to the prior. It can be viewed as a natural extension of VAEs with a regularization of maximizing mutual information, mitigating the "posterior collapse" issue. We demonstrate the effectiveness and versatility of our models in various text generation scenarios, including language modeling, unaligned style transfer, and dialog response generation. The source code to reproduce our experimental results is available on GitHub~1.
机译:诸如可变自动编码器(VAE)之类的深潜变量模型(LVM)最近在文本生成中发挥了重要作用。一个关键因素是利用光滑的潜伏结构来指导发电。但是,由于两个原因,VAE的表示能力受到限制:(1)高斯假设通常基于变分后验;同时(2)发生了臭名昭著的“后倒塌”问题。在本文中,我们提倡基于样本的自然语言变分分布表示,从而导致隐含的潜在特征,与基于高斯的后验者相比,它可以提供灵活的表示能力。我们进一步开发了一个LVM,以使匹配的后验与前验直接匹配。可以将其视为VAE的自然扩展,通过最大化互信息的规则化来缓解“后崩溃”问题。我们演示了我们的模型在各种文本生成方案中的有效性和多功能性,包括语言建模,不对齐的样式转换和对话框响应生成。可重现我们的实验结果的源代码可在GitHub〜1上找到。

著录项

相似文献

  • 外文文献
  • 中文文献
  • 专利
获取原文

客服邮箱:kefu@zhangqiaokeyan.com

京公网安备:11010802029741号 ICP备案号:京ICP备15016152号-6 六维联合信息科技 (北京) 有限公司©版权所有
  • 客服微信

  • 服务号