首页> 外文期刊>Pattern Recognition: The Journal of the Pattern Recognition Society >Learning shape and motion representations for view invariant skeleton-based action recognition
【24h】

Learning shape and motion representations for view invariant skeleton-based action recognition

机译:用于查看不变的骨架的动作识别的学习形状和运动表示

获取原文
获取原文并翻译 | 示例
       

摘要

Skeleton-based action recognition is an increasing attentioned task that analyses spatial configuration and temporal dynamics of a human action from skeleton data, which has been widely applied in intelligent video surveillance and human-computer interaction. How to design an effective framework to learn discriminative spatial and temporal characteristics for skeleton-based action recognition is still a challenging problem. The shape and motion representations of skeleton sequences are the direct embodiment of spatial and temporal characteristics respectively, which can well address for human action description. In this work, we propose an original unified framework to learn comprehensive shape and motion representations from skeleton sequences by using Geometric Algebra. We firstly construct skeleton sequence space as a subset of Geometric Algebra to represent each skeleton sequence along both the spatial and temporal dimensions. Then rotor-based view transformation method is proposed to eliminate the effect of viewpoint variation, which remains the relative spatio-temporal relations among skeleton frames in a sequence. We also construct spatio-temporal view invariant model (STVIM) to collectively integrate spatial configuration and temporal dynamics of skeleton joints and bones. In STVIM, skeleton sequence shape and motion representations which mutually compensate are jointly learned to describe skeletonbased actions comprehensively. Furthermore, a selected multi-stream Convolutional Neural Network is employed to extract and fuse deep features from mapping images of the learned representations for skeleton-based action recognition. Experimental results on NTU RGB+D, Northwestern-UCLA and UTD-MHAD datasets consistently verify the effectiveness of our proposed method and the superior performance over state-of-the-art competitors. (C) 2020 Elsevier Ltd. All rights reserved.
机译:基于骨架的动作识别是一种越来越多的关注任务,分析了从骨架数据中的人类行动的空间配置和时间动态,这已广泛应用于智能视频监控和人机交互。如何设计一个有效的框架来学习基于骨架的动作识别的歧视性空间和时间特征仍然是一个具有挑战性的问题。骨架序列的形状和运动表示分别是空间和时间特征的直接实施例,其可以用于人类行动描述的井地址。在这项工作中,我们提出了一个原始统一的框架,通过使用几何代数来学习来自骨架序列的综合形状和运动表示。我们首先将骨架序列空间构造为几何代数的子集,以表示空间和时间尺寸的每个骨架序列。然后提出基于转子的视图变换方法以消除观点变化的效果,这仍然是序列中骨架帧之间的相对时空关系。我们还构建了时空视图不变模型(STVIM),以集体集成骨架关节和骨骼的空间配置和时间动态。在STVIM,相互补偿的骨架序列形状和运动表示共同学习以全面地描述骨架行动。此外,采用选定的多流卷积神经网络来从基于骨架的动作识别的映射图像的图像中提取和熔断深度特征。 NTU RGB + D的实验结果,Northwestern-UCLA和UTD-MHAD数据集始终如一地验证了我们所提出的方法的有效性和最先进的竞争对手的卓越性能。 (c)2020 elestvier有限公司保留所有权利。

著录项

相似文献

  • 外文文献
  • 中文文献
  • 专利
获取原文

客服邮箱:kefu@zhangqiaokeyan.com

京公网安备:11010802029741号 ICP备案号:京ICP备15016152号-6 六维联合信息科技 (北京) 有限公司©版权所有
  • 客服微信

  • 服务号