...
首页> 外文期刊>Computers & Graphics >Image generation from bounding box-represented semantic labels
【24h】

Image generation from bounding box-represented semantic labels

机译:从边界盒代表语义标签的图像生成

获取原文
获取原文并翻译 | 示例
   

获取外文期刊封面封底 >>

       

摘要

Image generation with pixel-wise semantic information is suitable for the development of adversarial learning techniques. In this study, we propose a method for synthesizing objects with class-specific textures and fine-scale details based on bounding box-represented semantic labels. To achieve this goal, we note that the traditional generative adversarial network (GAN) uses noise as an input to generate realistic images with sufficient textures and details, but it cannot be guided by specific targets and requirements. By contrast, conditional GAN (cGAN) can involve various types of guiding information but it often ignores specific textures and details, thereby leading to less realistic results and low resolution. Thus, we propose a new translator-enhancer framework by combining cGAN and GAN to achieve high quality image generation. cGAN is used as a translator to match the semantic constraints whereas GAN is employed as an enhancer to provide details and textures. We also propose a new form of semantic label map as an input, which is represented by instance-level bounding boxes rather than segmentation masks. The semantic label map represented by bounding boxes makes it easier for users to provide the inputs and it also gives greater flexibility when generating object boundaries. The results obtained from qualitative and quantitative experiments showed that our method can generate realistic images of objects with semantic labels represented by bounding boxes. Our method can be used to generate images of novel scenes to support learning tasks during training with various scenes, which are difficult to capture in the real world. (C) 2019 Elsevier Ltd. All rights reserved.
机译:图像生成具有像素明智的语义信息,适用于开发对抗性学习技术。在这项研究中,我们提出了一种基于边界盒代表的语义标记来合成具有专用纹理和微尺度细节的对象的方法。为了实现这一目标,我们注意到传统的生成对抗网络(GaN)使用噪声作为输入,以产生具有足够纹理和细节的现实图像,但不能被特定目标和要求引导。相比之下,条件GaN(Cgan)可以涉及各种类型的指导信息,但它通常忽略特定的纹理和细节,从而导致较差的结果和低分辨率。因此,我们通过组合Cgan和GaN来实现高质量的图像生成来提出新的翻译 - 增强者框架。 Cgan被用作转换器以匹配语义约束,而GAN被用作增强器,以提供细节和纹理。我们还提出了一种新的语义标签映射作为输入,它由实例级边界框而非分段掩码表示。边界框表示的语义标签映射使用户更容易提供输入,并且在生成对象边界时也会提供更大的灵活性。从定性和定量实验中获得的结果表明,我们的方法可以利用由边界框表示的语义标签产生物质图像。我们的方法可用于生成新颖的场景的图像,以支持使用各种场景的培训期间学习任务,这很难在现实世界中捕获。 (c)2019 Elsevier Ltd.保留所有权利。

著录项

  • 来源
    《Computers & Graphics》 |2019年第6期|32-40|共9页
  • 作者单位

    Tsinghua Univ BNRist Beijing 100084 Peoples R China|Tsinghua Univ Sch Software Beijing 100084 Peoples R China;

    Yi Tunnel Technol Co Ltd Beijing 100084 Peoples R China;

    Tsinghua Univ BNRist Beijing 100084 Peoples R China|Tsinghua Univ Sch Software Beijing 100084 Peoples R China;

    Tsinghua Univ BNRist Beijing 100084 Peoples R China|Tsinghua Univ Sch Software Beijing 100084 Peoples R China;

  • 收录信息 美国《科学引文索引》(SCI);美国《工程索引》(EI);
  • 原文格式 PDF
  • 正文语种 eng
  • 中图分类
  • 关键词

    Computer vision; Generative adversarial networks; Image generation;

    机译:计算机愿景;生成的对抗网络;图像生成;

相似文献

  • 外文文献
  • 中文文献
  • 专利
获取原文

客服邮箱:kefu@zhangqiaokeyan.com

京公网安备:11010802029741号 ICP备案号:京ICP备15016152号-6 六维联合信息科技 (北京) 有限公司©版权所有
  • 客服微信

  • 服务号