首页> 外文期刊>Knowledge-Based Systems >IIT-GAT: Instance-level image transformation via unsupervised generative attention networks with disentangled representations
【24h】

IIT-GAT: Instance-level image transformation via unsupervised generative attention networks with disentangled representations

机译:IIT-GAT:通过无监督的生成关注网络具有解除不诚格表示的实例级图像转换

获取原文
获取原文并翻译 | 示例
           

摘要

Image-to-image translation is an important research field in computer vision, which is widely associated with Generative Adversarial Networks (GANs) and dual learning. However, the existing methods mainly translate the global image of the source domain to the target domain, which fails to implement instance-level image-to-image translation, and the translation results in the target domain cannot be controlled. In this paper, an instance-level image-to-image translation network (IIT-GAT) is proposed, which includes attention module and feature-encoder module. The attention module is used to guide our model to focus on more interesting instance to generate instance masks, which helps to separate instance and background of an image. The feature-encoder module is used to embed the images into two different spaces: domain-invariant content space and domain-specific attribute space. The content features and attribute features of different images are used as input to generator simultaneously to improve the controllability of image-to-image translation. To this end, we introduce a local self-reconstruction loss that encourages the network to learn the style feature of target instances. Generally, our method not only improves the quality of instance-level image-to-image translation, but also increases controllability on this basis. Extensive experiments are conducted on multiple datasets to validate the effectiveness of the proposed framework, and the results show our method has better performance than previous methods. (C) 2021 Elsevier B.V. All rights reserved.
机译:图像到图像转换是计算机视觉中的一个重要研究领域,与生成的对抗网络(GANS)和双学习广泛相关。但是,现有方法主要将源域的全局映像转换为目标域,这无法实现实例级图像到图像转换,并且无法控制目标域的转换。在本文中,提出了一个实例级图像到图像转换网络(IIT-GAT),包括注意模块和特征编码器模块。注意模块用于指导我们的模型,专注于更有趣的实例来生成实例掩码,这有助于分离图像的实例和背景。特征编码器模块用于将图像嵌入到两个不同的空格中:域不变内容空间和域特定的属性空间。不同图像的内容特征和属性特征用作同时发电机的输入,以提高图像到图像转换的可控性。为此,我们介绍了局部自我重建损失,鼓励网络了解目标实例的风格特征。通常,我们的方法不仅提高了实例级图像到图像转换的质量,而且还提高了这种基础的可控性。在多个数据集上进行了广泛的实验,以验证所提出的框架的有效性,结果表明我们的方法具有比以前的方法更好的性能。 (c)2021 elestvier b.v.保留所有权利。

著录项

相似文献

  • 外文文献
  • 中文文献
  • 专利
获取原文

客服邮箱:kefu@zhangqiaokeyan.com

京公网安备:11010802029741号 ICP备案号:京ICP备15016152号-6 六维联合信息科技 (北京) 有限公司©版权所有
  • 客服微信

  • 服务号