首页> 外文会议>International joint conference on natural language processing >Implicit Deep Latent Variable Models for Text Generation
【24h】

Implicit Deep Latent Variable Models for Text Generation

机译:文本生成隐含深度潜变量模型

获取原文
获取外文期刊封面目录资料

摘要

Deep latent variable models (LVM) such as variational auto-encoder (VAE) have recently played an important role in text generation. One key factor is the exploitation of smooth latent structures to guide the generation. However, the representation power of VAEs is limited due to two reasons: (1) the Gaussian assumption is often made on the variational posteriors: and meanwhile (2) a notorious "posterior collapse" issue occurs. In this paper, we advocate sample-based representations of variational distributions for natural language, leading to implicit latent features, which can provide flexible representation power compared with Gaussian-based posteriors. We further develop an LVM to directly match the aggregated posterior to the prior. It can be viewed as a natural extension of VAEs with a regularization of maximizing mutual information, mitigating the "posterior collapse" issue. We demonstrate the effectiveness and versatility of our models in various text generation scenarios, including language modeling, unaligned style transfer, and dialog response generation. The source code to reproduce our experimental results is available on GitHub~1.
机译:深潜变量模型(LVM)如变自动编码器(VAE)最近在玩文字产生了​​重要的作用。一个关键因素是光滑潜结构的剥削引导代。然而,VAES的表示功率被限制由于两个原因:(1)高斯假设通常由在变后验:同时(2)一个臭名昭著“后崩溃”问题时。在本文中,我们提倡自然语言变分布基于采样表示,导致隐含潜在功能,基于高斯后验相比,可以提供灵活的表现力。我们进一步发展的LVM到汇总后到前直接匹配。它可以被看作是VAES的自然延伸最大化互信息,减轻了“后崩溃”问题的正规化。我们证明我们的模型在不同的文本生成方案,包括语言建模,对齐风格转移和对话响应生成的有效性和通用性。源代码复制我们的实验结果可以在GitHub〜1。

著录项

相似文献

  • 外文文献
  • 中文文献
  • 专利
获取原文

客服邮箱:kefu@zhangqiaokeyan.com

京公网安备:11010802029741号 ICP备案号:京ICP备15016152号-6 六维联合信息科技 (北京) 有限公司©版权所有
  • 客服微信

  • 服务号