...
首页> 外文期刊>Pattern Recognition: The Journal of the Pattern Recognition Society >Structure-aware human pose estimation with graph convolutional networks
【24h】

Structure-aware human pose estimation with graph convolutional networks

机译:与图形卷积网络的结构感知人类姿态估计

获取原文
获取原文并翻译 | 示例
   

获取外文期刊封面封底 >>

       

摘要

Human pose estimation is the task of localizing body key points from still images. As body key points are inter-connected, it is desirable to model the structural relationships between body key points to further improve the localization performance. In this paper, based on original graph convolutional networks, we propose a novel model, termed Pose Graph Convolutional Network (PGCN), to exploit these important relationships for pose estimation. Specifically, our model builds a directed graph between body key points according to the natural compositional model of a human body. Each node (key point) is represented by a 3-D tensor consisting of multiple feature maps, initially generated by our backbone network, to retain accurate spatial information. Furthermore, attention mechanism is presented to focus on crucial edges (structured information) between key points. PGCN is then learned to map the graph into a set of structure-aware key point representations which encode both structure of human body and appearance information of specific key points. Additionally, we propose two modules for PGCN, i.e., the Local PGCN (L-PGCN) module and Non-Local PGCN (NL-PGCN) module. The former utilizes spatial attention to capture the correlations between the local areas of adjacent key points to refine the location of key points. While the latter captures long-range relationships via non-local operation to associate the challenging key points. By equipping with these two modules, our PGCN can further improve localization performance. Experiments both on single- and multi-person estimation benchmark datasets show that our method consistently outperforms competing state-of-the-art methods. (C) 2020 Elsevier Ltd. All rights reserved.
机译:人类姿势估计是从静止图像定位身体关键点的任务。由于身体关键点相互连接,期望模拟身体关键点之间的结构关系,以进一步提高定位性能。本文基于原始图卷积网络,我们提出了一种新颖的模型,称为姿势图卷积网络(PGCN),利用这些重要的姿势估计关系。具体而言,我们的模型根据人体的自然组成模型在身体关键点之间构建了一条指导的图。每个节点(关键点)由由第一特征映射组成的3-D张量表示,该特征映射最初由我们的骨干网络产生,以保留准确的空间信息。此外,提出注意机制以聚焦关键点之间的关键边缘(结构化信息)。然后学会了PGCN来将图形映射到一组结构感知键点表示,其编码人体的两个结构和特定关键点的外观信息。另外,我们提出了用于PGCN的两个模块,即局部PGCN(L-PGCN)模块和非局部PGCN(NL-PGCN)模块。前者利用空间注意,以捕获相邻关键点的局域地区之间的相关性,以优化关键点的位置。虽然后者通过非本地操作捕获远程关系,以使具有挑战性的关键点关联。通过配备这两个模块,我们的PGCN可以进一步提高本地化性能。单人和多人估计基准数据集上的实验表明,我们的方法一直始终胜过竞争最先进的方法。 (c)2020 elestvier有限公司保留所有权利。

著录项

相似文献

  • 外文文献
  • 中文文献
  • 专利
获取原文

客服邮箱:kefu@zhangqiaokeyan.com

京公网安备:11010802029741号 ICP备案号:京ICP备15016152号-6 六维联合信息科技 (北京) 有限公司©版权所有
  • 客服微信

  • 服务号