首页> 外文OA文献 >Depth-Map-Assisted Texture and Depth Map Super-Resolution
【2h】

Depth-Map-Assisted Texture and Depth Map Super-Resolution

机译:深度贴图辅助纹理和深度贴图超分辨率

摘要

With the development of video technology, high definition video and 3D video applications are becoming increasingly accessible to customers. The interactive and vivid 3D video experience of realistic scenes relies greatly on the amount and quality of the texture and depth map data. However, due to the limitations of video capturing hardware and transmission bandwidth, transmitted video has to be compressed which degrades, in general, the received video quality. This means that it is hard to meet the users’ requirements of high definition and visual experience; it also limits development of future applications. Therefore, image/video super-resolution techniques have been proposed to address this issue. Image super-resolution aims to reconstruct a high resolution image from single or multiple low resolution images captured of the same scene under different conditions. Based on the image type that needs to be super-resolved, image super-resolution includes texture and depth image super-resolutions. If classified based on the implementation methods, there are three main categories: interpolation-based, reconstruction-based and learning-based super-resolution algorithms. This thesis focuses on exploiting depth data in interpolation-based super-resolution algorithms for texture video and depth maps. Two novel texture and one depth super-resolution algorithms are proposed as the main contributions of this thesis. The first texture super-resolution algorithm is carried out in the Mixed Resolution (MR) multiview video system where at least one of the views is captured at Low Resolution (LR), while the others are captured at Full Resolution (FR). In order to reduce visual uncomfortableness and adapt MR video format for free-viewpoint television, the low resolution views are super-resolved to the target full resolution by the proposed virtual view assisted super resolution algorithm. The inter-view similarity is used to determine whether to fill the missing pixels in the super-resolved frame by virtual view pixels or by spatial interpolated pixels. The decision mechanism is steered by the texture characteristics of the neighbors of each missing pixel. Thus, the proposed method can recover the details in regions with edges while maintaining good quality at smooth areas by properly exploiting the high quality virtual view pixels and the directional correlation of pixels. The second texture super-resolution algorithm is based on the Multiview Video plus Depth (MVD) system, which consists of textures and the associated per-pixel depth data. In order to further reduce the transmitted data and the quality degradation of received video, a systematical framework to downsample the original MVD data and later on to super-resolved the LR views is proposed. At the encoder side, the rows of the two adjacent views are downsampled following an interlacing and complementary fashion, whereas, at the decoder side, the discarded pixels are recovered by fusing the virtual view pixels with the directional interpolated pixels from the complementary downsampled views. Consequently, with the assistance of virtual views, the proposed approach can effectively achieve these two goals. From previous two works, we can observe that depth data has big potential to be used in 3D video enhancement. However, due to the low spatial resolution of Time-of-Flight (ToF) depth camera generated depth images, their applications have been limited. Hence, in the last contribution of this thesis, a planar-surface-based depth map super-resolution approach is presented, which interpolates depth images by exploiting the equation of each detected planar surface. Both quantitative and qualitative experimental results demonstrate the effectiveness and robustness of the proposed approach over benchmark methods.
机译:随着视频技术的发展,高清视频和3D视频应用程序越来越为客户所用。逼真的场景的交互式生动3D视频体验在很大程度上取决于纹理和深度图数据的数量和质量。但是,由于视频捕获硬件和传输带宽的限制,必须压缩发送的视频,这通常会降低接收的视频质量。这意味着很难满足用户对高清和视觉体验的要求;它还限制了未来应用程序的开发。因此,已经提出了图像/视频超分辨率技术来解决这个问题。图像超分辨率旨在从在不同条件下从同一场景捕获的单个或多个低分辨率图像中重建高分辨率图像。基于需要超分辨率的图像类型,图像超分辨率包括纹理和深度图像超分辨率。如果根据实现方法进行分类,则主要分为三类:基于插值,基于重构和基于学习的超分辨率算法。本文的重点是在基于插值的纹理视频和深度图超分辨率算法中利用深度数据。提出了两种新颖的纹理和一种深度超分辨率算法,作为本文的主要贡献。第一纹理超分辨率算法在混合分辨率(MR)多视图视频系统中执行,其中至少一个视图以低分辨率(LR)捕获,而其他视图以全分辨率(FR)捕获。为了减少视觉不适感并使MR视频格式适合自由视点电视,通过提出的虚拟视点辅助超分辨率算法将低分辨率视图超分辨率为目标全分辨率。视图间相似性用于确定是通过虚拟视图像素还是通过空间插值像素填充超分辨帧中的丢失像素。决策机制由每个丢失像素的邻居的纹理特征控制。因此,通过适当地利用高质量的虚拟视点像素和像素的方向相关性,所提出的方法可以恢复具有边缘的区域中的细节,同时在平滑区域上保持良好的质量。第二种纹理超分辨率算法基于多视图视频加深度(MVD)系统,该系统由纹理和关联的每像素深度数据组成。为了进一步减少传输的数据和接收视频的质量下降,提出了一种系统框架,对原始MVD数据进行下采样,然后再对LR视图进行超解析。在编码器端,按照隔行扫描和互补方式对两个相邻视图的行进行降采样,而在解码器端,通过将虚拟视图像素与来自互补降采样视图的方向插值像素融合,来恢复丢弃的像素。因此,借助虚拟视图,所提出的方法可以有效地实现这两个目标。从前两篇作品中,我们可以看到深度数据在3D视频增强中具有很大的潜力。但是,由于飞行时间(ToF)深度相机生成的深度图像的空间分辨率较低,因此其应用受到了限制。因此,在本文的最后一项贡献中,提出了一种基于平面的深度图超分辨率方法,该方法通过利用每个检测到的平面的方程来对深度图像进行插值。定量和定性实验结果均证明了该方法相对于基准方法的有效性和鲁棒性。

著录项

  • 作者

    Jin Z;

  • 作者单位
  • 年度 2015
  • 总页数
  • 原文格式 PDF
  • 正文语种 en
  • 中图分类

相似文献

  • 外文文献
  • 中文文献
  • 专利

客服邮箱:kefu@zhangqiaokeyan.com

京公网安备:11010802029741号 ICP备案号:京ICP备15016152号-6 六维联合信息科技 (北京) 有限公司©版权所有
  • 客服微信

  • 服务号