首页> 外文会议>AAAI Conference on Artificial Intelligence >Synthetic Depth Transfer for Monocular 3D Object Pose Estimation in the Wild
【24h】

Synthetic Depth Transfer for Monocular 3D Object Pose Estimation in the Wild

机译:用于野生的单眼3D对象姿态估计的综合深度转移

获取原文

摘要

Monocular object pose estimation is an important yet challenging computer vision problem. Depth features can provide useful information for pose estimation. However, existing methods rely on real depth images to extract depth features, leading to its difficulty on various applications. In this paper, we aim at extracting RGB and depth features from a single RGB image with the help of synthetic RGB-depth image pairs for object pose estimation. Specifically, a deep convolutional neural network is proposed with an RGB-to-Depth Embedding module and a Synthetic-Real Adaptation module. The embedding module is trained with synthetic pair data to learn a depth-oriented embedding space between RGB and depth images optimized for object pose estimation. The adaptation module is to further align distributions from synthetic to real data. Compared to existing methods, our method does not need any real depth images and can be trained easily with large-scale synthetic data. Extensive experiments and comparisons show that our method achieves best performance on a challenging public PASCAL 3D+ dataset in all the metrics, which substantiates the superiority of our method and the above modules.
机译:单眼物体姿势估计是一个重要但具有挑战性的计算机视觉问题。深度特征可以为姿势估计提供有用的信息。但是,现有方法依赖于真正的深度图像提取深度特征,导致其在各种应用中的困难。在本文中,我们的目的在于借助于对象姿势估计的合成RGB深度图像对从单个RGB图像中提取RGB和深度特征。具体地,提出了一种利用RGB到深度嵌入模块和合成实体适配模块的深卷积神经网络。嵌入模块接受了合成对数据训练,以学习RGB和深度图像之间的深度导向的嵌入空间,优化用于对象姿势估计。适配模块是进一步将来自合成数据的分布对齐。与现有方法相比,我们的方法不需要任何真正的深度图像,并且可以用大规模的合成数据轻松培训。广泛的实验和比较表明,我们的方法在所有度量标准中实现了挑战性的公共帕斯卡3D +数据集的最佳表现,这证明了我们的方法和上述模块的优势。

著录项

相似文献

  • 外文文献
  • 中文文献
  • 专利
获取原文

客服邮箱:kefu@zhangqiaokeyan.com

京公网安备:11010802029741号 ICP备案号:京ICP备15016152号-6 六维联合信息科技 (北京) 有限公司©版权所有
  • 客服微信

  • 服务号