首页> 外文期刊>ACM Transactions on Graphics >3D Ken Burns Effect from a Single Image
【24h】

3D Ken Burns Effect from a Single Image

机译:单个图像的3D Ken Burns效果

获取原文
获取原文并翻译 | 示例
           

摘要

The Ken Burns effect allows animating still images with a virtual camera scanand zoom. Adding parallax, which results in the 3D Ken Burns effect, enablessignificantly more compelling results. Creating such effects manually is timeconsumingand demands sophisticated editing skills. Existing automaticmethods, however, require multiple input images from varying viewpoints.In this paper, we introduce a framework that synthesizes the 3D Ken Burnseffect from a single image, supporting both a fully automatic mode and an interactive mode with the user controlling the camera. Our framework firstleverages a depth prediction pipeline, which estimates scene depth that issuitable for view synthesis tasks. To address the limitations of existing depthestimation methods such as geometric distortions, semantic distortions, andinaccurate depth boundaries, we develop a semantic-aware neural networkfor depth prediction, couple its estimate with a segmentation-based depthadjustment process, and employ a refinement neural network that facilitatesaccurate depth predictions at object boundaries. According to this depthestimate, our framework then maps the input image to a point cloud andsynthesizes the resulting video frames by rendering the point cloud from thecorresponding camera positions. To address disocclusions while maintaininggeometrically and temporally coherent synthesis results, we utilize contextawarecolor- and depth-inpainting to fill in the missing information in theextreme views of the camera path, thus extending the scene geometry of thepoint cloud. Experiments with a wide variety of image content show thatour method enables realistic synthesis results. Our study demonstrates thatour system allows users to achieve better results while requiring little effortcompared to existing solutions for the 3D Ken Burns effect creation.
机译:肯·伯恩斯(Ken Burns)效果允许使用虚拟相机扫描和缩放来对静止图像进行动画处理。添加视差会导致3D Ken Burns效果,从而使效果更加引人注目。手动创建此类效果非常耗时,并且需要复杂的编辑技能。然而,现有的自动方法需要从不同角度输入多个输入图像。在本文中,我们介绍了一种框架,该框架可以从单个图像合成3D Ken Burnseffect,同时支持全自动模式和用户控制相机的交互式模式。我们的框架首先利用了深度预测管道,该管道可以估计适合视图合成任务的场景深度。为了解决现有深度估计方法的局限性,例如几何失真,语义失真和不正确的深度边界,我们开发了一种语义感知的神经网络进行深度预测,将其估计与基于分段的深度调整过程相结合,并采用了精简的神经网络来促进精确物体边界处的深度预测。根据该深度估计,我们的框架然后将输入图像映射到点云,并通过从对应的摄像机位置渲染点云来合成生成的视频帧。为了在保持几何和时间上连贯的综合结果的同时解决遮挡问题,我们利用上下文感知的颜色和深度修补来填充摄像机路径极端视图中的缺失信息,从而扩展了点云的场景几何。具有多种图像内容的实验表明,我们的方法可以实现逼真的合成结果。我们的研究表明,与3D Ken Burns效果创建的现有解决方案相比,我们的系统可以使用户获得更好的结果,而所需的精力却很少。

著录项

相似文献

  • 外文文献
  • 中文文献
  • 专利
获取原文

客服邮箱:kefu@zhangqiaokeyan.com

京公网安备:11010802029741号 ICP备案号:京ICP备15016152号-6 六维联合信息科技 (北京) 有限公司©版权所有
  • 客服微信

  • 服务号