首页> 外文期刊>Image and Vision Computing >Action recognition using saliency learned from recorded human gaze
【24h】

Action recognition using saliency learned from recorded human gaze

机译:使用从记录的人类凝视中学到的显着性进行动作识别

获取原文
获取原文并翻译 | 示例
           

摘要

This paper addresses the problem of recognition and localization of actions in image sequences, by utilizing, in the training phase only, gaze tracking data of people watching videos depicting the actions in question. First, we learn discriminative action features at the areas of gaze fixation and train a Convolutional Network that predicts areas of fixation (i.e. salient regions) from raw image data. Second, we propose a Support Vector Machine-based recognition method for joint recognition and localization, in which the bounding box of the action in question is considered as a latent variable. In our formulation the optimization attempts to both minimize the classification cost and maximize the saliency within the bounding box. We show that the results obtained with the optimization where saliency within the bounding box is maximized outperform the results obtained when saliency within the bounding box is not maximized, i.e. when only classification cost is minimized. Furthermore, the results that we obtain outperform the state-of-the-art results on the UCF sports dataset. (C) 2016 Elsevier B.V. All rights reserved.
机译:本文仅通过在训练阶段利用观看视频中描述相关动作的人们的凝视跟踪数据,来解决图像序列中动作的识别和定位问题。首先,我们在凝视注视区域学习判别动作特征,并训练一个卷积网络,该网络根据原始图像数据预测注视区域(即显着区域)。其次,我们提出了一种基于支持向量机的联合识别和定位方法,该方法将所考虑动作的边界框视为潜在变量。在我们的公式化中,优化尝试既使分类成本最小化,又使边界框内的显着性最大化。我们表明,在边界框内的显着性最大化的情况下,通过优化获得的结果优于在边界框内的显着性未最大化(即仅将分类成本最小化)时获得的结果。此外,我们获得的结果优于UCF运动数据集上的最新结果。 (C)2016 Elsevier B.V.保留所有权利。

著录项

相似文献

  • 外文文献
  • 中文文献
  • 专利
获取原文

客服邮箱:kefu@zhangqiaokeyan.com

京公网安备:11010802029741号 ICP备案号:京ICP备15016152号-6 六维联合信息科技 (北京) 有限公司©版权所有
  • 客服微信

  • 服务号