首页> 外文期刊>Circuits and Systems for Video Technology, IEEE Transactions on >Boosted Exemplar Learning for Action Recognition and Annotation
【24h】

Boosted Exemplar Learning for Action Recognition and Annotation

机译:促进行为识别和注释的样例学习

获取原文
获取原文并翻译 | 示例
           

摘要

Human action recognition and annotation is an active research topic in computer vision. How to model various actions, varying with time resolution, visual appearance, and others, is a challenging task. In this paper, we propose a boosted exemplar learning (BEL) approach to model various actions in a weakly supervised manner, i.e., only action bag-level labels are provided but action instance level ones are not. The proposed BEL method can be summarized as three steps. First, for each action category, amount of class-specific candidate exemplars are learned through an optimization formulation considering their discrimination and co-occurrence. Second, each action bag is described as a set of similarities between its instances and candidate exemplars. Instead of simply using a heuristic distance measure, the similarities are decided by the exemplar-based classifiers through the multiple instance learning, in which a positive (or negative) video or image set is deemed as a positive (or negative) action bag and those frames similar to the given exemplar in Euclidean Space as action instances. Third, we formulate the selection of the most discriminative exemplars into a boosted feature selection framework and simultaneously obtain an action bag-based detector. Experimental results on two publicly available datasets: the KTH dataset and Weizmann dataset, demonstrate the validity and effectiveness of the proposed approach for action recognition. We also apply BEL to learn representations of actions by using images collected from the Web and use this knowledge to automatically annotate action in YouTube videos. Results are very impressive, which proves that the proposed algorithm is also practical in unconstraint environments.
机译:人体动作识别和注释是计算机视觉中一个活跃的研究主题。如何为随时间分辨率,视觉外观等变化的各种动作建模是一项具有挑战性的任务。在本文中,我们提出了一种增强的榜样学习(BEL)方法,以弱监督的方式对各种动作进行建模,即仅提供动作包级别的标签,而没有提供动作实例级别的标签。所提出的BEL方法可以概括为三个步骤。首先,对于每个动作类别,通过考虑其歧视和共现的优化公式来学习特定于类别的候选示例的数量。其次,每个动作包被描述为其实例与候选示例之间的一组相似性。基于示例的分类器通过多实例学习来确定相似性,而不是简单地使用启发式距离度量,其中将正(或负)视频或图像集视为正(或负)动作包类似于欧几里得空间中给定示例的框架作为动作实例。第三,我们将具有最高判别力的示例的选择公式化为增强的特征选择框架,并同时获得基于动作包的检测器。在两个公开可用的数据集(KTH数据集和Weizmann数据集)上的实验结果证明了所提出的动作识别方法的有效性和有效性。我们还使用BEL通过使用从网络上收集的图像来学习动作的表示,并使用此知识来自动注释YouTube视频中的动作。结果非常令人印象深刻,这证明了所提出的算法在不受约束的环境中也是可行的。

著录项

相似文献

  • 外文文献
  • 中文文献
  • 专利
获取原文

客服邮箱:kefu@zhangqiaokeyan.com

京公网安备:11010802029741号 ICP备案号:京ICP备15016152号-6 六维联合信息科技 (北京) 有限公司©版权所有
  • 客服微信

  • 服务号