首页> 外文会议>International Conference on Computer Vision >Space-Time Robust Video Representation for Action Recognition
【24h】

Space-Time Robust Video Representation for Action Recognition

机译:用于行动识别的时空强大视频表示

获取原文

摘要

We address the problem of action recognition in unconstrained videos. We propose a novel content driven pooling that leverages space-time context while being robust toward global space-time transformations. Being robust to such transformations is of primary importance in unconstrained videos where the action localizations can drastically shift between frames. Our pooling identifies regions of interest using video structural cues estimated by different saliency functions. To combine the different structural information, we introduce an iterative structure learning algorithm, WSVM (weighted SVM), that determines the optimal saliency layout of an action model through a sparse regularizer. A new optimization method is proposed to solve the WSVM' highly non-smooth objective function. We evaluate our approach on standard action datasets (KTH, UCF50 and HMDB). Most noticeably, the accuracy of our algorithm reaches 51.8% on the challenging HMDB dataset which outperforms the state-of-the-art of 7.3% relatively.
机译:我们解决了不受约束视频中的行动认可问题。我们提出了一种新的内容驱动汇总,可以利用时空上下文,同时朝向全球时空转换的强大。对这种转换的强劲在于在不受约束的视频中具有主要重要性,其中动作本地化可以在帧之间大大移位。我们的汇集使用不同显着函数估计的视频结构提示来识别利益区域。为了结合不同的结构信息,我们介绍了一种迭代结构学习算法,WSVM(加权SVM),它通过稀疏规范器确定动作模型的最佳显着性布局。提出了一种新的优化方法来解决WSVM高度不平滑的目标函数。我们在标准动作数据集(kth,UCF50和HMDB)上评估我们的方法。最明显的是,算法的准确性在挑战的HMDB数据集中达到51.8%,以相当优于最先进的7.3%。

著录项

相似文献

  • 外文文献
  • 中文文献
  • 专利
获取原文

客服邮箱:kefu@zhangqiaokeyan.com

京公网安备:11010802029741号 ICP备案号:京ICP备15016152号-6 六维联合信息科技 (北京) 有限公司©版权所有
  • 客服微信

  • 服务号