首页> 外文学位 >Visual tracking and illumination recovery via sparse representation.
【24h】

Visual tracking and illumination recovery via sparse representation.

机译:通过稀疏表示进行视觉跟踪和照明恢复。

获取原文
获取原文并翻译 | 示例

摘要

Compressive sensing, or sparse representation, has played a fundamental role in many fields of science. It shows that the signals and images can be reconstructed from far fewer measurements than what is usually considered to be necessary. In this dissertation, we present the results of a study of applying sparse representation to illumination recovery, object tracking, and simultaneous tracking and recognition.;Illumination recovery, also known as inverse lighting, is the problem of recovering an illumination distribution in a scene from the appearance of objects located in the scene. It is used for Augmented Reality, where the virtual objects match the existing image and cast convincing shadows on the real scene rendered with the recovered illumination. Shadows in a scene are caused by the occlusion of incoming light, and thus contain information about the lighting of the scene. Although shadows have been used in determining the 3D shape of the object that casts shadows onto the scene, few studies have focused on the illumination information provided by the shadows. In this dissertation, we recover the illumination of a scene from a single image with cast shadows given the geometry of the scene. The images with cast shadows can be quite complex and therefore cannot be well approximated by low-dimensional linear subspaces. However, in this study we show that the set of images produced by a Lambertian scene with cast shadows can be efficiently represented by a sparse set of images generated by directional light sources. We first model an image with cast shadows as composed of a diffusive part (without cast shadows) and a residual part that captures cast shadows. Then, we express the problem in an ℓ1-regularized least squares formulation, with nonnegativity constraints (as light has to be nonnegative at any point in space).;Visual tracking, which consistently infers the motion of a desired target in a video sequence, has been an active and fruitful research topic in computer vision for decades. It has many practical applications such as surveillance, human computer interaction, medical imaging and so on. Many challenges to design a robust tracking algorithm come from the enormous unpredictable variations in the target, such as deformations, fast motion, occlusions, background clutter, and lighting changes. To tackle the challenges posed by tracking, we propose a robust visual tracking method by casting tracking as a sparse approximation problem in a particle filter framework. In this framework, occlusion, noise and other challenging issues are addressed seamlessly through a set of trivial templates. Specifically, to find the tracking target at a new frame, each target candidate is sparsely represented in the space spanned by target templates and trivial templates. The sparsity is achieved by solving an ℓ1 -regularized least squares problem. Then the candidate with the smallest projection error is taken as the tracking target. After that, tracking is continued using a Bayesian state inference framework in which a particle filter is used for propagating sample distributions over time. Three additional components further improve the robustness of our approach: (1) a velocity incorporated motion model that helps concentrate the samples on the true target location in the next frame, (2) the nonnegativity constraints that help filter out clutter that is similar to tracked targets in reversed intensity patterns, and (3) a dynamic template update scheme that keeps track of the most representative templates throughout the tracking procedure. We test the proposed approach on many challenging sequences involving heavy occlusions, drastic illumination changes, large scale changes, non-rigid object movement, out-of-plane rotation, and large pose variations. The proposed approach shows excellent performance in comparison with four previously proposed trackers. We also extend the work to simultaneous tracking and recognition in vehicle classification in IR video sequences. We attempt to resolve the uncertainties in tracking and recognition at the same time by introducing a static template set that stores target images in various conditions such as different poses, lighting, and so on. The recognition results at each frame are propagated to produce the final result for the whole video. The tracking result is evaluated at each frame and low confidence in tracking performance initiates a new cycle of tracking and classification. We demonstrate the robustness of the proposed method on vehicle tracking and classification using outdoor IR video sequences. (Abstract shortened by UMI.)
机译:压缩感测或稀疏表示已在许多科学领域中发挥了基本作用。它表明可以从比通常认为必要的测量少得多的测量结果中重建信号和图像。在本文中,我们提出了将稀疏表示应用于照明恢复,对象跟踪以及同时跟踪和识别的研究结果。照明恢复,也称为逆照明,是从场景中恢复场景中照明分布的问题。场景中对象的外观。它用于增强现实,其中虚拟对象与现有图像匹配,并在使用恢复的照明渲染的真实场景上投射令人信服的阴影。场景中的阴影是由入射光的遮挡引起的,因此包含有关场景照明的信息。尽管已使用阴影来确定将阴影投射到场景上的对象的3D形状,但很少有研究集中在阴影提供的照明信息上。在本文中,我们从给定场景几何形状的带有阴影的单个图像中恢复了场景的照明。具有投射阴影的图像可能非常复杂,因此无法通过低维线性子空间很好地近似。但是,在这项研究中,我们表明由定向阴影光源生成的稀疏图像集可以有效地表示由带阴影的朗伯场景产生的图像集。我们首先对具有投影阴影的图像进行建模,该图像由扩散部分(无投影阴影)和捕获投影阴影的残差部分组成。然后,我们用非负约束(因为光在空间中的任何一点都必须是非负的)以ℓ 1正规化的最小二乘公式来表达问题。视觉跟踪,始终如一地推断视频中所需目标的运动几十年来,序列一直是计算机视觉中活跃而富有成果的研究主题。它具有许多实际应用,例如监视,人机交互,医学成像等。设计强大的跟踪算法的许多挑战来自目标的巨大不可预测的变化,例如变形,快速运动,遮挡,背景杂波和照明变化。为了解决跟踪带来的挑战,我们提出了一种鲁棒的视觉跟踪方法,该方法通过将跟踪转换为粒子过滤器框架中的稀疏近似问题。在此框架中,可通过一组琐碎的模板无缝解决遮挡,噪音和其他挑战性问题。具体而言,为了在新帧处找到跟踪目标,在目标模板和琐碎模板所跨越的空间中稀疏地表示每个目标候选。稀疏性是通过解决ℓ 1-正则化最小二乘问题来实现的。然后将投影误差最小的候选者作为跟踪目标。之后,使用贝叶斯状态推断框架继续跟踪,其中使用了粒子滤波器来传播样本随时间的分布。三个附加组件进一步提高了我们方法的鲁棒性:(1)结合速度的运动模型,有助于将样本集中在下一帧的真实目标位置上;(2)非负性约束,有助于过滤出类似于跟踪信号的杂波(3)动态模板更新方案,在整个跟踪过程中跟踪最具代表性的模板。我们在许多具有挑战性的序列上测试了提出的方法,这些序列涉及严重的遮挡,剧烈的照明变化,大规模的变化,非刚性的物体运动,平面外旋转以及较大的姿态变化。与四个以前提出的跟踪器相比,提出的方法显示出出色的性能。我们还将工作扩展到红外视频序列中车辆分类的同时跟踪和识别。我们试图通过引入一个静态模板集来解决跟踪和识别过程中的不确定性,该模板集可以在各种条件下(例如不同的姿势,光线等)存储目标图像。传播每一帧的识别结果,以产生整个视频的最终结果。跟踪结果在每一帧进行评估,对跟踪性能的低置信度启动了跟踪和分类的新周期。我们展示了使用室外红外视频序列对车辆进行跟踪和分类的方法的鲁棒性。 (摘要由UMI缩短。)

著录项

  • 作者

    Mei, Xue.;

  • 作者单位

    University of Maryland, College Park.;

  • 授予单位 University of Maryland, College Park.;
  • 学科 Engineering Electronics and Electrical.;Computer Science.
  • 学位 Ph.D.
  • 年度 2009
  • 页码 132 p.
  • 总页数 132
  • 原文格式 PDF
  • 正文语种 eng
  • 中图分类
  • 关键词

相似文献

  • 外文文献
  • 中文文献
  • 专利
获取原文

客服邮箱:kefu@zhangqiaokeyan.com

京公网安备:11010802029741号 ICP备案号:京ICP备15016152号-6 六维联合信息科技 (北京) 有限公司©版权所有
  • 客服微信

  • 服务号