首页> 外文学位 >Toward 3D reconstruction of static and dynamic objects.
【24h】

Toward 3D reconstruction of static and dynamic objects.

机译:进行静态和动态对象的3D重建。

获取原文
获取原文并翻译 | 示例

摘要

The goal of image-based 3D reconstruction is to construct a spatial understanding of the world from a collection of images. For applications that seek to model generic real-world scenes, it is important that the reconstruction methods used are able to characterize both static scene elements (e.g. trees and buildings) as well as dynamic objects (e.g. cars and pedestrians). However, due to many inherent ambiguities in the reconstruction problem, recovering this 3D information with accuracy, robustness, and efficiency is a considerable challenge. To advance the research frontier for image-based 3D modeling, this dissertation focuses on three challenging problems in static scene and dynamic object reconstruction.;We first target the problem of static scene depthmap estimation from crowd-sourced datasets (i.e. photos collected from the Internet). While achieving high-quality depthmaps using images taken under a controlled environment is already a difficult task, heterogeneous crowd-sourced data presents a unique set of challenges for multi-view depth estimation, including varying illumination and occasional occlusions. We propose a depthmap estimation method that demonstrates high accuracy, robustness, and scalability on a large number of photos collected from the Internet.;Compared to static scene reconstruction, the problem of dynamic object reconstruction from monocular images is fundamentally ambiguous when not imposing any additional assumptions. This is because having only a single observation of an object is insufficient for valid 3D triangulation, which typically requires concurrent observations of the object from multiple viewpoints. Assuming that dynamic objects of the same class (e.g. all the pedestrians walking on a sidewalk) move in a common path in the real world, we develop a method that estimates the 3D positions of the dynamic objects from unstructured monocular images. Experiments on both synthetic and real datasets illustrate the solvability of the problem and the effectiveness of our approach.;Finally, we address the problem of dynamic object reconstruction from a set of unsynchronized videos capturing the same dynamic event. This problem is of great interest because, due to the increased availability of portable capture devices, captures using multiple unsynchronized videos are common in the real world. To resolve the challenges that arises from non-concurrent captures and unknown temporal overlap among video streams, we propose a self-expressive dictionary learning framework, where the dictionary entries are defined as the collection of temporally varying structures. Experiments demonstrate the effectiveness of this approach to the previously unsolved problem.
机译:基于图像的3D重建的目标是从图像集合中构建对世界的空间理解。对于寻求为一般现实世界场景建模的应用程序,重要的是所使用的重建方法必须能够表征静态场景元素(例如树木和建筑物)以及动态对象(例如汽车和行人)。但是,由于在重建问题中存在许多固有的模糊性,因此以准确性,鲁棒性和效率来恢复此3D信息是一项巨大的挑战。为了提高基于图像的3D建模的研究前沿,本文着重研究静态场景和动态对象重建中的三个挑战性问题。我们首先针对从众包数据集(即从Internet收集的照片)估计静态场景深度图的问题)。虽然使用在受控环境下拍摄的图像来获得高质量的深度图已经是一项艰巨的任务,但异构人群源数据为多视图深度估计提出了一系列独特的挑战,包括变化的照明和偶尔的遮挡。我们提出了一种深度图估计方法,该方法可对从互联网上收集的大量照片展示出高精度,鲁棒性和可扩展性。;与静态场景重建相比,从单眼图像重建动态对象的问题在不施加任何其他要求的情况下基本上是模棱两可的假设。这是因为仅对对象的一次观察不足以进行有效的3D三角测量,这通常需要从多个视点同时观察对象。假设相同类别的动态对象(例如,所有在人行道上行走的行人)在现实世界中的共同路径中移动,我们开发了一种从非结构化单眼图像估算动态对象3D位置的方法。在合成数据集和真实数据集上进行的实验说明了该问题的可解决性以及我们方法的有效性。最后,我们从一组捕获相同动态事件的非同步视频中解决了动态对象重建的问题。由于便携式捕获设备的可用性越来越高,使用多个不同步视频进行捕获在现实世界中非常普遍,因此这个问题引起了极大的关注。为了解决非并发捕获和视频流之间未知的时间重叠所带来的挑战,我们提出了一种自表达式字典学习框架,其中字典条目被定义为时间变化结构的集合。实验证明了这种方法对于以前未解决的问题的有效性。

著录项

  • 作者

    Zheng, Enliang.;

  • 作者单位

    The University of North Carolina at Chapel Hill.;

  • 授予单位 The University of North Carolina at Chapel Hill.;
  • 学科 Computer science.
  • 学位 Ph.D.
  • 年度 2016
  • 页码 120 p.
  • 总页数 120
  • 原文格式 PDF
  • 正文语种 eng
  • 中图分类
  • 关键词

相似文献

  • 外文文献
  • 中文文献
  • 专利
获取原文

客服邮箱:kefu@zhangqiaokeyan.com

京公网安备:11010802029741号 ICP备案号:京ICP备15016152号-6 六维联合信息科技 (北京) 有限公司©版权所有
  • 客服微信

  • 服务号