首页> 外文期刊>Journal of Information Recording >Trajectory-Based 3D Convolutional Descriptors for Human Action Recognition
【24h】

Trajectory-Based 3D Convolutional Descriptors for Human Action Recognition

机译:基于轨迹的3D卷积描述符用于人类动作识别

获取原文
获取原文并翻译 | 示例
       

摘要

This article presents a new method for video representation, called trajectory based 3D convolutional descriptor (TCD), which incorporates the advantages of both deep learned features and hand-crafted features. We utilize deep architectures to learn discriminative convolutional feature maps, and conduct trajectory constrained pooling to aggregate these convolutional features into effective descriptors. Firstly, valid trajectories are generated by tracking the interest points within co-motion super-pixels. Secondly, we utilize the 3D ConvNet (C3D) to capture both motion and appearance information in the form of convolutional feature maps. Finally, feature maps are transformed by using two normalization methods, namely channel normalization and spatiotemporal normalization. Trajectory constrained sampling and pooling are used to aggregate deep learned features into descriptors. The proposed (TCD) contains high discriminative capacity compared with hand-crafted features and is able to boost the recognition performance. Experimental results on benchmark datasets demonstrate that our pipeline obtains superior performance over conventional algorithms in terms of both efficiency and accuracy.
机译:本文提出了一种新的视频表示方法,称为基于轨迹的3D卷积描述符(TCD),它结合了深度学习功能和手工制作功能的优点。我们利用深层架构来学习判别式卷积特征图,并进行轨迹约束合并以将这些卷积特征聚合为有效的描述符。首先,通过跟踪超运动像素内的兴趣点来生成有效轨迹。其次,我们利用3D ConvNet(C3D)以卷积特征图的形式捕获运动和外观信息。最后,使用两种归一化方法对特征图进行变换,即通道归一化和时空归一化。轨迹约束采样和合并用于将深度学习特征聚合到描述符中。与手工制作的功能相比,建议的(TCD)包含较高的辨别能力,并且能够提高识别性能。在基准数据集上的实验结果表明,与传统算法相比,我们的管道在效率和准确性方面均具有出色的性能。

著录项

相似文献

  • 外文文献
  • 中文文献
  • 专利
获取原文

客服邮箱:kefu@zhangqiaokeyan.com

京公网安备:11010802029741号 ICP备案号:京ICP备15016152号-6 六维联合信息科技 (北京) 有限公司©版权所有
  • 客服微信

  • 服务号