首页> 外文学位 >Human Activity Recognition in Video: Extending Statistical Features Across Time, Space and Semantic Context.
【24h】

Human Activity Recognition in Video: Extending Statistical Features Across Time, Space and Semantic Context.

机译:视频中的人类活动识别:跨越时间,空间和语义上下文扩展统计功能。

获取原文
获取原文并翻译 | 示例

摘要

This thesis explores the problem of recognizing complex human activities involving the manipulation of objects in high resolution video. Inspired by human psychophysical performance, I develop and evaluate an activity recognition feature derived from the velocity histories of tracked keypoints. These features have a much greater spatial and temporal range than existing video features. I show that a generative mixture model using these features performs comparably to local spatio-temporal features on the KTH activity recognition dataset. I additionally introduce and explore a new activity recognition dataset of activities of daily living (URADL), containing high resolution video sequences of complex activities. I demonstrate the superior performance of my velocity history feature on this dataset, and explore ways in which it can be extended. I investigate the value of a more sophisticated latent velocity model for velocity histories. I explore the addition of contextual semantic information to the model, whether fully automatic or derived from supervision, and provide a sketch for the inclusion of this information in any feature-based generative model for activity recognition or time series data. This approach performs comparably to established methods on the KTH dataset, and significantly outperforms local spatio-temporal features on the challenging new URADL dataset. I further develop another new dataset, URADL2, and explore transferring knowledge between related video activity recognition domains. Using a straightforward feature-expansion transfer learning technique, I show improved performance on one dataset using activity models transferred from the other dataset.
机译:本文探讨了识别高分辨率视频中涉及对象操纵的复杂人类活动的问题。受人类心理生理表现的启发,我开发并评估了从跟踪关键点的速度历史中得出的活动识别功能。这些功能比现有的视频功能具有更大的空间和时间范围。我证明了使用这些特征的生成混合模型的性能与KTH活动识别数据集上的局部时空特征相当。我还将介绍和探索一个新的日常生活活动识别活动数据库(URADL),其中包含复杂活动的高分辨率视频序列。我演示了我的速度历史记录功能在此数据集上的优越性能,并探讨了扩展它的方式。我研究了速度历史中更复杂的潜伏速度模型的价值。我探索了将上下文语义信息添加到模型中的过程,无论是全自动的还是从监督中获得的,并为在活动识别或时间序列数据的任何基于特征的生成模型中包含此信息提供了一个草图。这种方法与在KTH数据集上建立的方法相比具有可比性,并且在具有挑战性的新URADL数据集上明显优于局部时空特征。我进一步开发了另一个新的数据集URADL2,并探索了相关视频活动识别域之间的知识转移。使用简单的功能扩展转移学习技术,我使用从另一数据集转移来的活动模型显示了一个数据集的改进性能。

著录项

  • 作者

    Messing, Ross.;

  • 作者单位

    University of Rochester.;

  • 授予单位 University of Rochester.;
  • 学科 Computer science.
  • 学位 Ph.D.
  • 年度 2011
  • 页码 128 p.
  • 总页数 128
  • 原文格式 PDF
  • 正文语种 eng
  • 中图分类
  • 关键词

相似文献

  • 外文文献
  • 中文文献
  • 专利
获取原文

客服邮箱:kefu@zhangqiaokeyan.com

京公网安备:11010802029741号 ICP备案号:京ICP备15016152号-6 六维联合信息科技 (北京) 有限公司©版权所有
  • 客服微信

  • 服务号