首页> 外文学位 >Learning Transformations From Video.
【24h】

Learning Transformations From Video.

机译:从视频学习转变。

获取原文
获取原文并翻译 | 示例

摘要

Our survival depends on accurate understanding of the environment around us through sensory inputs. One way to achieve this is to build models of the surrounding environment that are able to provide explanations of the data. Statistical models such as PCA, ICA and sparse coding attempt to do so by exploiting the second- and higher-order structures of sensory data. While these models have been shown to reveal key properties of the mammalian sensory system and have been successfully applied in various engineering applications, one shared weakness of these models is that they assume each observation is independent. In reality, there is often a transformational relationship between sensory data observations. Exploiting this relationship allows us to tease apart the causes of the data and reason about the environment. In this thesis, I developed an unsupervised learning framework that attempts to find the translational relationship between data and infer the causes of the observed data.;This dissertation is divided into three chapters. First, I propose an unsupervised learning framework that is able to model the transformations between data points using a continuous transformation model. I highlight the difficulties faced by previous attempts using similar models. I overcome these hurdles by proposing a learning rule that is able to compute the learning updates for an exponential model in polynomial time. I also propose an adaptive inference algorithm that is able to avoid local minima. These improvements make learning transformation possible and efficient.;Second, I perform a detailed analysis of the proposed model. I show that the adaptive inference algorithm is able to simultaneously recover multiple transformation parameters with high accuracy when given synthetic data where the transformation is known. When learned on pairs of images containing affine transformations, the algorithm correctly recovers the transformation operators. The unsupervised learning algorithm is able to discover transformations such as translation, illumination adjustment, contrast enhancement and local deformations when learned on pairs of natural movie frames. I also show that the learned models provide a better description of the underlying transformation both qualitatively and quantitatively compare to commonly used motion models.;Third, I describe a plausible application for the continuous transformation model in video coding. In a hybrid coding scheme, I propose to replace the traditionally used exhaustive search motion model with transformation models learned on natural time-varying images. A detailed analysis of the rate distortion characteristics of different learned models is documented and I show that the learned model improves the performance of traditional motion models in various settings.
机译:我们的生存取决于通过感官投入对周围环境的准确理解。实现此目的的一种方法是建立能够提供数据说明的周围环境模型。诸如PCA,ICA和稀疏编码之类的统计模型试图通过利用感官数据的二阶和高阶结构来做到这一点。虽然这些模型已显示出揭示了哺乳动物感觉系统的关键特性,并已成功地应用于各种工程应用中,但这些模型的一个共同缺点是,它们假定每次观察都是独立的。实际上,感觉数据观察之间通常存在转换关系。利用这种关系可以使我们弄清数据的原因和环境的原因。本文建立了一个无监督的学习框架,试图寻找数据之间的转换关系,并推断出观察到的数据的起因。本文分为三章。首先,我提出了一种无监督学习框架,该框架能够使用连续转换模型对数据点之间的转换进行建模。我强调了以前使用类似模型的尝试所面临的困难。我通过提出一种学习规则来克服这些障碍,该学习规则能够在多项式时间内为指数模型计算学习更新。我还提出了一种能够避免局部极小值的自适应推理算法。这些改进使学习转换成为可能和有效。第二,我对提出的模型进行了详细的分析。我表明,当已知变换的已知合成数据时,自适应推理算法能够以高精度同时恢复多个变换参数。当在包含仿射变换的图像对上获知时,该算法正确地恢复了变换算子。在自然电影帧对上学习时,无监督学习算法能够发现诸如平移,照明调整,对比度增强和局部变形之类的变换。我还表明,与常用的运动模型相比,所学模型在定性和定量方面都可以更好地描述基础变换。第三,我描述了连续变换模型在视频编码中的合理应用。在混合编码方案中,我建议用在自然时变图像上学习的变换模型替换传统使用的穷举搜索运动模型。记录了对不同学习模型的速率失真特性的详细分析,我证明了该学习模型在各种设置下可以提高传统运动模型的性能。

著录项

  • 作者

    Wang, Jimmy Ching Ming.;

  • 作者单位

    University of California, Berkeley.;

  • 授予单位 University of California, Berkeley.;
  • 学科 Computer Science.
  • 学位 Ph.D.
  • 年度 2010
  • 页码 95 p.
  • 总页数 95
  • 原文格式 PDF
  • 正文语种 eng
  • 中图分类
  • 关键词

相似文献

  • 外文文献
  • 中文文献
  • 专利
获取原文

客服邮箱:kefu@zhangqiaokeyan.com

京公网安备:11010802029741号 ICP备案号:京ICP备15016152号-6 六维联合信息科技 (北京) 有限公司©版权所有
  • 客服微信

  • 服务号