首页> 外文期刊>Journal of signal processing systems for signal, image, and video technology >Depth Estimation and Video Synthesis for 2D to 3D Video Conversion
【24h】

Depth Estimation and Video Synthesis for 2D to 3D Video Conversion

机译:2D到3D视频转换的深度估计和视频合成

获取原文
获取原文并翻译 | 示例

摘要

With the recent progress of multi-view devices and the corresponding signal processing techniques, stereoscopic viewing experience has been introduced to the public with growing interest. To create depth perception in human vision, two different video sequences in binocular vision are required for viewers. Those videos can be either captured by 3D-enabled cameras or synthesized as needed. The primary contribution of this paper is to establish two transformation models for stationary scenes and non-stationary objects in a given view, respectively. The models can be used for the production of corresponding stereoscopic videos as a viewer would have seen at the original event of the scene. The transformation model to estimate the depth information for stationary scenes is based on the information of the vanishing point and vanishing lines of the given video. The transformation model for non-stationary regions is the result of combining the motion analysis of the non-stationary regions and the transformation model for stationary scenes to estimate the depth information. The performance of the models is evaluated using subjective 3D video quality evaluation and objective quality evaluation on the synthesized views. Performance comparison with the ground truth and a famous multi-view video synthesis algorithm, VSRS, which requires six views to complete synthesis, is also presented. It is shown that the proposed method can provide better perceptual 3D video quality with natural depth perception.
机译:随着多视点设备和相应的信号处理技术的最新发展,立体视听体验已经越来越引起人们的兴趣。为了在人类视觉中创建深度感知,观众需要两个不同的双目视觉视频序列。这些视频可以通过启用3D的相机捕获,也可以根据需要进行合成。本文的主要贡献是分别为给定视图中的静止场景和非静止对象建立两个转换模型。该模型可用于生成相应的立体视频,就像观众在场景的原始事件中所看到的一样。用于估计静止场景的深度信息的变换模型基于给定视频的消失点和消失线的信息。非平稳区域的变换模型是将非平稳区域的运动分析与平稳场景的变换模型相结合以估计深度信息的结果。使用主观3D视频质量评估和综合视图上的客观质量评估来评估模型的性能。还介绍了与地面实况的性能比较以及著名的多视图视频合成算法VSRS,该算法需要六个视图才能完成合成。结果表明,所提出的方法可以提供具有自然深度感知的更好的3D视频感知质量。

著录项

相似文献

  • 外文文献
  • 中文文献
  • 专利
获取原文

客服邮箱:kefu@zhangqiaokeyan.com

京公网安备:11010802029741号 ICP备案号:京ICP备15016152号-6 六维联合信息科技 (北京) 有限公司©版权所有
  • 客服微信

  • 服务号