首页> 外文会议>International Conference on Computer Vision >DenseRaC: Joint 3D Pose and Shape Estimation by Dense Render-and-Compare
【24h】

DenseRaC: Joint 3D Pose and Shape Estimation by Dense Render-and-Compare

机译:DenseRaC:通过Dense Render-and-Compare进行联合3D姿势和形状估计

获取原文

摘要

We present DenseRaC, a novel end-to-end framework for jointly estimating 3D human pose and body shape from a monocular RGB image. Our two-step framework takes the body pixel-to-surface correspondence map (i.e., IUV map) as proxy representation and then performs estimation of parameterized human pose and shape. Specifically, given an estimated IUV map, we develop a deep neural network optimizing 3D body reconstruction losses and further integrating a render-and-compare scheme to minimize differences between the input and the rendered output, i.e., dense body landmarks, body part masks, and adversarial priors. To boost learning, we further construct a large-scale synthetic dataset (MOCA) utilizing web-crawled Mocap sequences, 3D scans and animations. The generated data covers diversified camera views, human actions and body shapes, and is paired with full ground truth. Our model jointly learns to represent the 3D human body from hybrid datasets, mitigating the problem of unpaired training data. Our experiments show that DenseRaC obtains superior performance against state of the art on public benchmarks of various human-related tasks.
机译:我们介绍了DenseRaC,这是一个新颖的端到端框架,用于从单眼RGB图像联合估计3D人体姿势和身体形状。我们的两步框架将人体像素到表面的对应图(即IUV图)作为代理表示,然后对参数化的人体姿势和形状进行估算。具体来说,给定估算的IUV图,我们开发了一个优化3D人体重建损失的深度神经网络,并进一步集成了渲染和比较方案,以最大程度地减少输入和渲染输出之间的差异,即密集的人体标志,人体部位遮罩,和对抗先验。为了促进学习,我们进一步利用网络爬行的Mocap序列,3D扫描和动画构建了大规模的综合数据集(MOCA)。生成的数据涵盖了多种摄像机视角,人类动作和身体形状,并与全面的事实相结合。我们的模型从混合数据集中共同学习表示3D人体,从而减轻了训练数据不成对的问题。我们的实验表明,DenseRaC在各种与人类有关的任务的公开基准上均获得了优于最新技术的出色性能。

著录项

相似文献

  • 外文文献
  • 中文文献
  • 专利
获取原文

客服邮箱:kefu@zhangqiaokeyan.com

京公网安备:11010802029741号 ICP备案号:京ICP备15016152号-6 六维联合信息科技 (北京) 有限公司©版权所有
  • 客服微信

  • 服务号