首页> 外文会议>International Conference on Computer Vision >DenseRaC: Joint 3D Pose and Shape Estimation by Dense Render-and-Compare
【24h】

DenseRaC: Joint 3D Pose and Shape Estimation by Dense Render-and-Compare

机译:Denserac:密集渲染和比较的关节3D姿势和形状估计

获取原文

摘要

We present DenseRaC, a novel end-to-end framework for jointly estimating 3D human pose and body shape from a monocular RGB image. Our two-step framework takes the body pixel-to-surface correspondence map (i.e., IUV map) as proxy representation and then performs estimation of parameterized human pose and shape. Specifically, given an estimated IUV map, we develop a deep neural network optimizing 3D body reconstruction losses and further integrating a render-and-compare scheme to minimize differences between the input and the rendered output, i.e., dense body landmarks, body part masks, and adversarial priors. To boost learning, we further construct a large-scale synthetic dataset (MOCA) utilizing web-crawled Mocap sequences, 3D scans and animations. The generated data covers diversified camera views, human actions and body shapes, and is paired with full ground truth. Our model jointly learns to represent the 3D human body from hybrid datasets, mitigating the problem of unpaired training data. Our experiments show that DenseRaC obtains superior performance against state of the art on public benchmarks of various human-related tasks.
机译:我们展示了Denserac,这是一种新的端到端框架,用于联合估计来自单眼RGB图像的3D人姿势和身体形状。我们的两步框架作为代理表示,将身体像素到表面对应图(即,IUV地图)作为代理表示,然后执行参数化人姿势和形状的估计。具体地,考虑到估计的IUV地图,我们开发了一个深度神经网络,优化3D身体重建损失以及进一步集成渲染和比较方案以最小化输入和渲染输出之间的差异,即密集的身体地标,身体部件面罩,和对抗性前锋。为了提升学习,我们进一步构建了利用Web爬网Mocap序列,3D扫描和动画的大规模合成数据集(MOCA)。生成的数据涵盖了多样化的相机视图,人类动作和身体形状,并与完整的原始事实配对。我们的模型共同学会从混合数据集中代表3D人体,减轻未配对训练数据的问题。我们的实验表明,Denserac在各种与人际关系任务的公共基准上获得了卓越的绩效。

著录项

相似文献

  • 外文文献
  • 中文文献
  • 专利
获取原文

客服邮箱:kefu@zhangqiaokeyan.com

京公网安备:11010802029741号 ICP备案号:京ICP备15016152号-6 六维联合信息科技 (北京) 有限公司©版权所有
  • 客服微信

  • 服务号