首页> 外文期刊>Computers & Graphics >Simple and effective deep hand shape and pose regression from a single depth image
【24h】

Simple and effective deep hand shape and pose regression from a single depth image

机译:从单个深度图像简单有效的深手形状和姿势回归

获取原文
获取原文并翻译 | 示例
获取外文期刊封面目录资料

摘要

Simultaneously estimating the 3D shape and pose of a hand in real time is a new and challenging computer graphics problem, which is important for animation and interactions with 3D objects in virtual environments with personalized hand shapes. CNN-based direct hand pose estimation methods are the state-of-the-art approaches, but they can only regress a 3D hand pose from a single depth image. In this study, we developed a simple and effective real-time CNN-based direct regression approach for simultaneously estimating the 3D hand shape and pose, as well as structure constraints for both egocentric and third person viewpoints by learning from the synthetic depth. In addition, we produced the first million-scale egocentric synthetic dataset called SynHandEgo, which contains egocentric depth images with accurate shape and pose annotations, as well as color segmentation of the hand parts. Our network is trained based on combined real and synthetic datasets with full supervision of the hand pose and structure constraints, and semi-supervision of the hand mesh. Our approach performed better than the state-of-the-art methods based on the SynHand5M synthetic dataset in terms of both the 3D shape and pose recovery. By learning simultaneously using real and synthetic data, we demonstrated the feasibility of hand mesh recovery from two real hand pose datasets, i.e., BigHand2.2M and NYU. Moreover, our method obtained more accurate estimates of the 3D hand poses based on the NYU dataset compared with the existing methods that output more than joint positions. The SynHandEgo dataset has been made publicly available to promote further research in the emerging domain of hand shape and pose recovery from egocentric viewpoints (https://bit.ly/2WMWM5u). (C) 2019 Elsevier Ltd. All rights reserved.
机译:实时同时估计手的3D形状和姿势是一个新的且具有挑战性的计算机图形问题,这对于动画以及与具有个性化手形的虚拟环境中的3D对象的交互非常重要。基于CNN的直接手势估计是最新技术,但它们只能从单个深度图像中回归3D手势。在这项研究中,我们开发了一种简单有效的基于CNN的实时直接回归方法,可通过从合成深度学习来同时估算3D手形和姿势以及以自我为中心和第三人称视角的结构约束。此外,我们制作了第一个百万级以自我为中心的以自我为中心的合成数据集,名为SynHandEgo,其中包含以自我为中心的深度图像以及准确的形状和姿势注释以及手部的颜色分割。我们的网络是根据真实和合成的组合数据集进行训练的,对手的姿势和结构约束进行完全监督,并对手网格进行半监督。就3D形状和姿态恢复而言,我们的方法比基于SynHand5M合成数据集的最新方法表现更好。通过同时使用真实数据和合成数据进行学习,我们证明了从两个真实的手部姿势数据集(BigHand2.2M和NYU)恢复手部网格的可行性。此外,与现有的输出比关节位置更多的方法相比,我们的方法基于NYU数据集获得了对3D手势的更准确的估计。 SynHandEgo数据集已公开提供,以促进从自我中心的角度对新兴的手部形状和姿势恢复领域进行进一步研究(https://bit.ly/2WMWM5u)。 (C)2019 Elsevier Ltd.保留所有权利。

著录项

  • 来源
    《Computers & Graphics》 |2019年第12期|85-91|共7页
  • 作者单位

    German Res Ctr Artificial Intelligence DFKI Kaiserslautern Germany|Univ Kaiserslautern Dept Informat D-67653 Kaiserslautern Germany|NUST SEECS Islamabad 44000 Pakistan|TUKL Tech Univ Kaiserslautern Kaiserslautern Germany;

    German Res Ctr Artificial Intelligence DFKI Kaiserslautern Germany|UPM Madinah 20012 Saudi Arabia;

    German Res Ctr Artificial Intelligence DFKI Kaiserslautern Germany;

    German Res Ctr Artificial Intelligence DFKI Kaiserslautern Germany|Univ Kaiserslautern Dept Informat D-67653 Kaiserslautern Germany|TUKL Tech Univ Kaiserslautern Kaiserslautern Germany;

  • 收录信息 美国《科学引文索引》(SCI);美国《工程索引》(EI);
  • 原文格式 PDF
  • 正文语种 eng
  • 中图分类
  • 关键词

    Convolutional neural network (CNN); Depth image; Three-dimensional (3D) hand mesh and pose;

    机译:卷积神经网络(CNN);深度图像;三维(3D)手网格和姿势;

相似文献

  • 外文文献
  • 中文文献
  • 专利
获取原文

客服邮箱:kefu@zhangqiaokeyan.com

京公网安备:11010802029741号 ICP备案号:京ICP备15016152号-6 六维联合信息科技 (北京) 有限公司©版权所有
  • 客服微信

  • 服务号