首页> 外文会议>European Conference on Computer Vision >Measuring Generalisation to Unseen Viewpoints, Articulations, Shapes and Objects for 3D Hand Pose Estimation Under Hand-Object Interaction
【24h】

Measuring Generalisation to Unseen Viewpoints, Articulations, Shapes and Objects for 3D Hand Pose Estimation Under Hand-Object Interaction

机译:在手工对象交互下测量用于看不见的观点,铰接,形状和对象的3D手姿势估计

获取原文

摘要

We study how well different types of approaches generalise in the task of 3D hand pose estimation under single hand scenarios and hand-object interaction. We show that the accuracy of state-of-the-art methods can drop, and that they fail mostly on poses absent from the training set. Unfortunately, since the space of hand poses is highly dimensional, it is inherently not feasible to cover the whole space densely, despite recent efforts in collecting large-scale training datasets. This sampling problem is even more severe when hands are interacting with objects and/or inputs are RGB rather than depth images, as RGB images also vary with lighting conditions and colors. To address these issues, we designed a public challenge (HANDS'19) to evaluate the abilities of current 3D hand pose estimators (HPEs) to interpolate and extrapolate the poses of a training set. More exactly, HANDS' 19 is designed (a) to evaluate the influence of both depth and color modalities on 3D hand pose estimation, under the presence or absence of objects; (b) to assess the generalisation abilities w.r.t. four main axes: shapes, articulations, viewpoints, and objects; (c) to explore the use of a synthetic hand models to fill the gaps of current datasets. Through the challenge, the overall accuracy has dramatically improved over the baseline, especially on extrapolation tasks, from 27 mm to 13 mm mean joint error. Our analyses highlight the impacts of: Data pre-processing, ensemble approaches, the use of a parametric 3D hand model (MANO), and different HPE methods/backbones.
机译:我们研究了在单手场景和手对象交互下3D手姿势估计的任务中不同类型的方法概括。我们表明,最先进的方法可以下降的准确性,并且它们大多在训练集中缺席的姿势失败。遗憾的是,由于手姿势的空间是高度的,因此尽管最近在收集大规模训练数据集的努力最近努力,但覆盖整个空间是固有的不可行的。当双手与对象交互和/或输入时,这种采样问题更严重,而且RGB图像也与深度图像相互作用,因为RGB图像也随着照明条件和颜色而变化。为了解决这些问题,我们设计了一个公共挑战(手19),以评估当前3D手姿势估计(HPE)的能力,以插入和推断培训集的姿势。更确切地说,手机19被设计(a),以评估深度和颜色模式对3D手姿势估计的影响,在存在或不存在的情况下; (b)评估概括能力w.r.t.四个主轴:形状,关节,观点和对象; (c)探讨使用合成手机模型来填补当前数据集的差距。通过挑战,整体准确性在基线上显着改善了基线,特别是在外推任务中,从27 mm到13 mm平均接头误差。我们的分析突出了以下影响:数据预处理,集合方法,使用参数3D手模型(MANO)以及不同的HPE方法/骨干。

相似文献

  • 外文文献
  • 中文文献
  • 专利
获取原文

客服邮箱:kefu@zhangqiaokeyan.com

京公网安备:11010802029741号 ICP备案号:京ICP备15016152号-6 六维联合信息科技 (北京) 有限公司©版权所有
  • 客服微信

  • 服务号