首页> 外文会议>IEEE Winter Conference on Applications of Computer Vision >Improving 3D Human Pose Estimation Via 3D Part Affinity Fields
【24h】

Improving 3D Human Pose Estimation Via 3D Part Affinity Fields

机译:通过3D零件亲和力场改善3D人体姿势估计

获取原文

摘要

3D human pose estimation from monocular images has become a heated area in computer vision recently. For years, most deep neural network based practices have adopted either an end-to-end approach, or a two-stage approach. An end-to-end network typically estimates 3D human poses directly from 2D input images, but it suffers from the shortage of 3D human pose data. It is also obscure to know if the inaccuracy stems from limited visual under-standing or 2D-to-3D mapping. Whereas a two-stage directly lifts those 2D keypoint outputs to the 3D space, after utilizing an existing network for 2D keypoint detections. However, they tend to ignore some useful contextual hints from the 2D raw image pixels. In this paper, we introduce a two-stage architecture that can eliminate the main disadvantages of both these approaches. During the first stage we use an existing state-of-the-art detector to estimate 2D poses. To add more con-textual information to help lifting 2D poses to 3D poses, we propose 3D Part Affinity Fields (3D-PAFs). We use 3D-PAFs to infer 3D limb vectors, and combine them with 2D poses to regress the 3D coordinates. We trained and tested our proposed framework on Human3.6M, the most popular 3D human pose benchmark dataset. Our approach achieves the state-of-the-art performance, which proves that with right selections of contextual information, a simple regression model can be very powerful in estimating 3D poses.
机译:最近,单眼图像的3D人体姿势估计已成为计算机视觉中的热点领域。多年来,大多数基于深度神经网络的实践都采用端到端方法或两阶段方法。端到端网络通常直接从2D输入图像中估计3D人体姿势,但是它遭受3D人体姿势数据不足的困扰。不清楚这种误差是否是由于有限的视觉理解或2D到3D映射所致。而在利用现有网络进行2D关键点检测之后,分两阶段将这些2D关键点输出直接提升到3D空间。但是,他们倾向于忽略2D原始图像像素中的一些有用的上下文提示。在本文中,我们介绍了一个两阶段的体系结构,可以消除这两种方法的主要缺点。在第一阶段中,我们使用现有的最先进的检测器来估算2D姿态。为了添加更多上下文信息以帮助将2D姿势提升为3D姿势,我们提出了3D零件相似性字段(3D-PAF)。我们使用3D-PAF来推断3D肢体向量,并将其与2D姿势结合起来以回归3D坐标。我们在Human3.6M(最流行的3D人体姿势基准数据集)上训练并测试了我们提出的框架。我们的方法达到了最先进的性能,这证明了通过正确选择上下文信息,简单的回归模型可以非常强大地估计3D姿势。

著录项

相似文献

  • 外文文献
  • 中文文献
  • 专利
获取原文

客服邮箱:kefu@zhangqiaokeyan.com

京公网安备:11010802029741号 ICP备案号:京ICP备15016152号-6 六维联合信息科技 (北京) 有限公司©版权所有
  • 客服微信

  • 服务号