首页> 外文会议>IEEE International Conference on Multimedia and Expo >High-Resolution Driving Scene Synthesis Using Stacked Conditional Gans and Spectral Normalization
【24h】

High-Resolution Driving Scene Synthesis Using Stacked Conditional Gans and Spectral Normalization

机译:使用堆积条件甘斯和谱归一化的高分辨率驾驶场景综合

获取原文

摘要

Large-scale dataset plays a key role in the driving scene understanding for deep learning based-autonomous driving tasks. Due to the fact that the annotation for a large number of images is extremely labor-intensive and time-consuming, many researchers turn to using image-synthesis techniques for automatic construction of training data. However, traditional methods often have difficulties in producing high-definition driving scene images. To tackle this problem, in this paper, we propose a novel deep model - hdCGAN - for high-definition image-to-image translation. The hdCGAN is built on a conditional GAN in combination with a spectral normalization. Moreover, we improve the hdCGAN by using a stacked network architecture and the enhanced model is called stack-hdCGAN. With the guidance of multi-scale discriminators and the constraint of spectral normalization in the training procedure, the learned models can generate high-resolution and high-quality driving scene images from corresponding semantic segmentation maps. Quantitative and qualitative evaluations on the Cityscapes dataset demonstrate the effectiveness of the proposed models.
机译:大规模数据集在驾驶场景理解中对基于深度学习的自动驾驶任务的驾驶场景的关键作用。由于大量图像的注释是极其劳动密集型和耗时的事实,许多研究人员将转向使用图像合成技术来自动构建训练数据。然而,传统方法通常在制造高清驾驶场景图像方面具有困难。为了解决这个问题,在本文中,我们提出了一种新颖的深层模型 - HDCGAN - 用于高清图像到图像转换。 HDCGAN与条件GaN一起建立在配合GaN上,结合光谱标准化。此外,我们通过使用堆叠的网络架构来改进HDCGAN,增强型模型称为Stack-HDCGAN。随着多尺度鉴别器的指导和训练过程中的光谱归一化的约束,学习模型可以从相应的语义分割映射地图产生高分辨率和高质量的驾驶场景图像。 Citycapes DataSet上的定量和定性评估证明了所提出的模型的有效性。

著录项

相似文献

  • 外文文献
  • 中文文献
  • 专利
获取原文

客服邮箱:kefu@zhangqiaokeyan.com

京公网安备:11010802029741号 ICP备案号:京ICP备15016152号-6 六维联合信息科技 (北京) 有限公司©版权所有
  • 客服微信

  • 服务号