首页> 外文会议>IEEE/CVF Conference on Computer Vision and Pattern Recognition >Mix and Match Networks: Encoder-Decoder Alignment for Zero-Pair Image Translation
【24h】

Mix and Match Networks: Encoder-Decoder Alignment for Zero-Pair Image Translation

机译:混合和匹配网络:零对图像转换的编码器-解码器对准

获取原文

摘要

We address the problem of image translation between domains or modalities for which no direct paired data is available (i.e. zero-pair translation). We propose mix and match networks, based on multiple encoders and decoders aligned in such a way that other encoder-decoder pairs can be composed at test time to perform unseen image translation tasks between domains or modalities for which explicit paired samples were not seen during training. We study the impact of autoencoders, side information and losses in improving the alignment and transferability of trained pairwise translation models to unseen translations. We show our approach is scalable and can perform colorization and style transfer between unseen combinations of domains. We evaluate our system in a challenging cross-modal setting where semantic segmentation is estimated from depth images, without explicit access to any depth-semantic segmentation training pairs. Our model outperforms baselines based on pix2pix and CycleGAN models.
机译:我们解决了在没有直接配对数据的域或模态之间进行图像转换的问题(即零对转换)。我们提出了基于多个编码器和解码器的混合匹配网络,这些编码器和解码器以这样的方式对齐:可以在测试时组合其他编码器/解码器对,以执行在训练期间看不到显式配对样本的域或模态之间看不见的图像翻译任务。我们研究了自动编码器,辅助信息和损失对改善训练有成对的翻译模型与看不见的翻译的对齐方式和可传递性的影响。我们证明了我们的方法是可扩展的,并且可以在看不见的域组合之间执行着色和样式转换。我们在具有挑战性的跨模式设置中评估我们的系统,在该设置中,根据深度图像估计语义分割,而无需显式访问任何深度语义分割训练对。我们的模型优于基于pix2pix和CycleGAN模型的基线。

著录项

相似文献

  • 外文文献
  • 中文文献
  • 专利
获取原文

客服邮箱:kefu@zhangqiaokeyan.com

京公网安备:11010802029741号 ICP备案号:京ICP备15016152号-6 六维联合信息科技 (北京) 有限公司©版权所有
  • 客服微信

  • 服务号