首页> 外文期刊>IEEE Transactions on Pattern Analysis and Machine Intelligence >Exploring Explicit Domain Supervision for Latent Space Disentanglement in Unpaired Image-to-Image Translation
【24h】

Exploring Explicit Domain Supervision for Latent Space Disentanglement in Unpaired Image-to-Image Translation

机译:探索未配对图像到图像翻译中潜在空间解剖学的显式域监管

获取原文
获取原文并翻译 | 示例
       

摘要

Image-to-image translation tasks have been widely investigated with Generative Adversarial Networks (GANs). However, existing approaches are mostly designed in an unsupervised manner, while little attention has been paid to domain information within unpaired data. In this article, we treat domain information as explicit supervision and design an unpaired image-to-image translation framework, Domain-supervised GAN (DosGAN), which takes the first step towards the exploration of explicit domain supervision. In contrast to representing domain characteristics using different generators or domain codes, we pre-train a classification network to explicitly classify the domain of an image. After pre-training, this network is used to extract the domain-specific features of each image. Such features, together with the domain-independent features extracted by another encoder (shared across different domains), are used to generate image in target domain. Extensive experiments on multiple facial attribute translation, multiple identity translation, multiple season translation and conditional edges-to-shoes/handbags demonstrate the effectiveness of our method. In addition, we can transfer the domain-specific feature extractor obtained on the Facescrub dataset with domain supervision information to unseen domains, such as faces in the CelebA dataset. We also succeed in achieving conditional translation with any two images in CelebA, while previous models like StarGAN cannot handle this task.
机译:图像到图像转换任务已被生成的对抗网络(GANS)被广泛调查。但是,现有方法主要以无人监督的方式设计,虽然在未配对数据中的域信息已经支付了很少的关注。在本文中,我们将域名信息视为明确的监督和设计一个未配对的图像到图像翻译框架,域监管甘(DOSGAN),这将迈出探索显式域监管的第一步。与使用不同生成器或域代码的表示域特征相比,我们预先训练分类网络以明确地分类图像的域。在预训练之后,该网络用于提取每个图像的特定域特征。这些特征与由另一个编码器提取的域的独立功能(在不同域中共享)一起用于生成目标域中的图像。关于多个面部属性翻译的大量实验,多重识别转换,多季翻译和条件边缘到鞋/手提包展示了我们方法的有效性。此外,我们可以使用域监控信息传输在Facescrub数据集上获得的域特定的特征提取器,以解密域,例如Celeba数据集中的面部。我们还成功地与Celeba中的任何两个图像实现了条件翻译,而前以前的模型如Stargan则无法处理此任务。

著录项

相似文献

  • 外文文献
  • 中文文献
  • 专利
获取原文

客服邮箱:kefu@zhangqiaokeyan.com

京公网安备:11010802029741号 ICP备案号:京ICP备15016152号-6 六维联合信息科技 (北京) 有限公司©版权所有
  • 客服微信

  • 服务号