...
首页> 外文期刊>Pattern recognition letters >Y-Autoencoders: Disentangling latent representations via sequential encoding
【24h】

Y-Autoencoders: Disentangling latent representations via sequential encoding

机译:Y-AutoEncoders:通过顺序编码解开潜在的表示

获取原文
获取原文并翻译 | 示例
   

获取外文期刊封面封底 >>

       

摘要

In the last few years there have been important advancements in disentangling latent representations using generative models, with the two dominant approaches being Generative Adversarial Networks (GANs) and Variational Autoencoders (VAEs). However, standard Autoencoders (AEs) and closely related structures have remained popular because they are easy to train and adapt to different tasks. An interesting question is if we can achieve state-of-the-art latent disentanglement with AEs while retaining their good properties. We propose an answer to this question by introducing a new model called Y-Autoencoder (Y-AE). The structure and training procedure of a Y-AE enclose a representation into an implicit and an explicit part. The implicit part is similar to the output of an AE and the explicit part is strongly correlated with labels in the training set. The two parts are separated in the latent space by splitting the output of the encoder into two paths (forming a Y shape) before decoding and re-encoding. We then impose a number of losses, such as reconstruction loss, and a loss on dependence between the implicit and explicit parts. Additionally, the projection in the explicit manifold is monitored by a predictor, that is embedded in the encoder and trained end-to-end with no adversarial losses. We provide significant experimental results on various domains, such as separation of style and content, image-to-image translation, and inverse graphics. (C) 2020 Elsevier B.V. All rights reserved.
机译:在过去几年中,使用生成模型解开潜在的潜在表示的重要进步,两种主要方法是生成的对抗网络(GANS)和变形Autiachoders(VAES)。然而,标准的autoencoders(AES)和密切相关的结构仍然受欢迎,因为它们很容易训练并适应不同的任务。一个有趣的问题是,如果我们能够在保留他们的良好性质的同时实现最先进的潜在解剖学。我们通过引入一个名为Y-AutoEncoder(Y-AE)的新模型提出了对此问题的答案。 Y-AE的结构和训练过程将表示形式括成隐式和显式部件。隐式部分类似于AE的输出,并且显式部件与训练集中的标签强烈相关。在解码和重新编码之前,通过将编码器的输出分成两个路径(形成Y形),在潜空间中分离两部分。然后,我们强加了许多损失,例如重建损失,以及依赖隐式和显式部件的损失。另外,通过预测器监视显式歧管中的投影,其嵌入在编码器中并培训端到端,没有对抗性损失。我们在各个领域提供显着的实验结果,例如风格和内容的分离,图像到图像转换和逆图形。 (c)2020 Elsevier B.v.保留所有权利。

著录项

相似文献

  • 外文文献
  • 中文文献
  • 专利
获取原文

客服邮箱:kefu@zhangqiaokeyan.com

京公网安备:11010802029741号 ICP备案号:京ICP备15016152号-6 六维联合信息科技 (北京) 有限公司©版权所有
  • 客服微信

  • 服务号