首页> 外文学位 >Extracting Moving People and Categorizing their Activities in Video.
【24h】

Extracting Moving People and Categorizing their Activities in Video.

机译:在视频中提取移动人员并对其活动进行分类。

获取原文
获取原文并翻译 | 示例

摘要

The ability to automatically detect and track human movements, recognize actions and activities, understand behavior and predict goals and intentions has captured the attention of many computer vision scientists. One of the main motivations is the great potential impact that this technology can make on many applications such as video search and indexing, smart surveillance systems, medical research, video game interfaces, automatic sport commentary, human-robot interaction, among others.;In this work, we focus on two important questions: given a video sequence, where are the moving humans in the sequence? what actions or activities are they performing?;We first discuss the problem of extracting human motion volumes from video sequences. We present a fully automatic framework to detect and extract arbitrary human motion volumes from challenging real-world videos. We have explored a purely top-down methodology that estimates body configurations at every frame to achieve the extraction. We also present a much more efficient approach that carefully combines bottom-up and top-down cues, which enables fast extraction in near real time.;We are not only interesting in finding where the humans are in a given sequence, but also in understanding what they are doing. We present statistical models for the task of simple human action recognition based in spatial and spatio-temporal local features. First, we show that by adapting latent topic models we can achieve competitive simple action categorization performance in an unsupervised setting. We also present a hierarchical model for simple actions that can be characterized as a constellation-of-bags-of-features. This model leverages the spatial structure of the human body to improve action recognition.;While these models are successful at the task of simple action recognition, their performance suffers when the actions of interest are more complex. We propose a discriminative model for complex action recognition capable of leveraging the temporal structure and composition of simpler motions into complex actions. We show that the contextual information provided by the temporal structure in our model greatly improves the complex action classification accuracy over state-of-the art models for simple action recognition.
机译:自动检测和跟踪人类运动,识别动作和活动,理解行为以及预测目标和意图的能力吸引了许多计算机视觉科学家的注意力。主要动机之一是该技术可能对许多应用程序产生巨大的潜在影响,例如视频搜索和索引编制,智能监控系统,医学研究,视频游戏界面,自动体育评论,人机交互等。在这项工作中,我们重点关注两个重要问题:给定视频序列,序列中移动的人在哪里?他们执行什么动作或活动?;我们首先讨论从视频序列中提取人体运动量的问题。我们提供了一个全自动框架,可从具有挑战性的真实视频中检测并提取任意人体运动量。我们已经探索了一种纯自顶向下的方法,该方法可以估计每个帧的身体构造以实现提取。我们还提供了一种更有效的方法,该方法将自下而上和自上而下的提示进行了精心组合,可以实现近乎实时的快速提取。我们不仅对找到人类在给定序列中的位置感兴趣,而且在理解方面也很有趣他们在做什么。我们提出了基于空间和时空局部特征的简单人类动作识别任务的统计模型。首先,我们证明了通过适应潜在主题模型,我们可以在无人监督的情况下实现具有竞争性的简单动作分类性能。我们还为简单动作提供了一个层次模型,可以将其描述为功能包星座。该模型利用了人体的空间结构来改善动作识别。虽然这些模型在简单的动作识别中很成功,但是当感兴趣的动作更加复杂时,它们的性能就会受到影响。我们提出了一种区分动作复杂模型的判别模型,该模型能够利用简单动作的时间结构和组成转化为复杂动作。我们表明,与简单动作识别的最新模型相比,我们模型中时间结构提供的上下文信息大大提高了复杂动作分类的准确性。

著录项

  • 作者

    Niebles Duque, Juan Carlos.;

  • 作者单位

    Princeton University.;

  • 授予单位 Princeton University.;
  • 学科 Computer Science.
  • 学位 Ph.D.
  • 年度 2011
  • 页码 151 p.
  • 总页数 151
  • 原文格式 PDF
  • 正文语种 eng
  • 中图分类
  • 关键词

相似文献

  • 外文文献
  • 中文文献
  • 专利
获取原文

客服邮箱:kefu@zhangqiaokeyan.com

京公网安备:11010802029741号 ICP备案号:京ICP备15016152号-6 六维联合信息科技 (北京) 有限公司©版权所有
  • 客服微信

  • 服务号