首页> 外文期刊>Neurocomputing >Adaptive multi-view graph convolutional networks for skeleton-based action recognition
【24h】

Adaptive multi-view graph convolutional networks for skeleton-based action recognition

机译:基于骨架的动作识别的自适应多视图图卷积网络

获取原文
获取原文并翻译 | 示例
       

摘要

Skeleton based human action recognition has attracted more and more attentions recently thanks to the accessibility of depth sensors and the development of pose estimation techniques. Conventional approaches such as convolutional neural networks usually model skeletons with grid-shaped representations, which cannot explicitly explore the dependency between two correlated joints. In this paper, we treat the skeleton as a single graph with joints as nodes and bones as edges. Based on the skeleton graph, the improved graph convolutional network called adaptive multi-view graph convolutional networks(AMV-GCNs) is proposed to deal with skeleton based action recognition. We firstly construct a novel skeleton graph and two kinds of graph nodes are defined to model the spatial configuration and temporal dynamics respectively. Then the generated graphs along with feature vectors on graph nodes are fed into AMV-GCNs. In AMV-GCNs, an adaptive view transformation module is designed to reduce the impact of view diversity. Proposed module can automatically determine suitable viewpoints and transform skeletons to new representations under those viewpoints for better recognition. Further, we employ multiple GCNs based streams to utilize and learn action information from different viewpoints. Finally, the classification scores from multiple streams are fused to provide the recognition result. Extensive experimental evaluations on four challenging datasets, NTU RGB+D 60, NTU RGB+D 120, Northwestern-UCLA and UTD-MHAD, demonstrate the superiority of our proposed network. (c) 2020 Published by Elsevier B.V.
机译:由于深度传感器的可访问性以及姿势估算技术的发展,最近,基于骨架的人类行动识别最近引起了越来越多的注意。诸如卷积神经网络的常规方法通常模拟具有网格形表示的骨架,其不能明确探索两个相关关节之间的依赖性。在本文中,我们将骨架视为单个图形,带有关节作为节点和骨骼作为边缘。基于骨架图,提出了称为自适应多视图卷积网络(AMV-GCNS)的改进的图形卷积网络,以处理基于骨架的动作识别。我们首先构造了一种新颖的骨架图,并定义了两种图形节点以分别模拟空间配置和时间动态。然后,所生成的图表以及图形节点上的特征向量被馈送到AMV-GCN中。在AMV-GCN中,自适应视图变换模块旨在减少视图分集的影响。提出的模块可以自动确定合适的观点和将骷髅转换为新的表示,以便更好地识别。此外,我们使用多个基于GCNS的流来利用和学习来自不同视点的动作信息。最后,融合来自多个流的分类评分以提供识别结果。对四个具有挑战性的数据集,NTU RGB + D 60,NTU RGB + D 120,Nortuern-UCLA和UTD-MHAD进行了广泛的实验评估,证明了我们所提出的网络的优越性。 (c)2020由elsevier b.v发布。

著录项

相似文献

  • 外文文献
  • 中文文献
  • 专利
获取原文

客服邮箱:kefu@zhangqiaokeyan.com

京公网安备:11010802029741号 ICP备案号:京ICP备15016152号-6 六维联合信息科技 (北京) 有限公司©版权所有
  • 客服微信

  • 服务号