首页> 外文期刊>Journal of visual communication & image representation >Learning latent geometric consistency for 6D object pose estimation in heavily cluttered scenes
【24h】

Learning latent geometric consistency for 6D object pose estimation in heavily cluttered scenes

机译:学习潜在几何一致性在严重杂乱场景中的6D对象姿态估计

获取原文
获取原文并翻译 | 示例
           

摘要

6D object pose (3D rotation and translation) estimation from RGB-D image is an important and challenging task in computer vision and has been widely applied in a variety of applications such as robotic manipulation, autonomous driving, augmented reality etc. Prior works extract global feature or reason about local appearance from an individual frame, which neglect the spatial geometric relevance between two frames, limiting their performance for occluded or truncated objects in heavily cluttered scenes. In this paper, we present a dual-stream network for estimating 6D pose of a set of known objects from RGB-D images. Our novelty stands in contrast to prior work that learns latent geometric consistency in pairwise dense feature representations from multiple observations of the same objects in a self-supervised manner. We show in experiments that our method outperforms state-of-the-art approaches on 6D object pose estimation in two challenging datasets, YCB-Video and LineMOD. (C) 2020 Elsevier Inc. All rights reserved.
机译:6D对象姿势(3D旋转和翻译)RGB-D图像的估计是计算机视觉中的一个重要且具有挑战性的任务,并且已广泛应用于各种应用,例如机器人操纵,自主驾驶,增强现实等。事先作品提取全球来自单个框架的局部外观的特征或原因,这些框架忽略了两个框架之间的空间几何相关性,限制了它们在严重杂乱场景中的闭塞或截断物体的性能。在本文中,我们介绍了一种用于从RGB-D图像估计一组已知对象的6D姿势的双流网络。我们的新颖性与先前的工作相反,以自我监督方式从相同对象的多次观察中学习潜在几何一致性的先前工作。我们在实验中展示了我们的方法在两个具有挑战性的数据集,YCB-Video和LineMod中的6D对象姿态估计上表现出最先进的方法。 (c)2020 Elsevier Inc.保留所有权利。

著录项

获取原文

客服邮箱:kefu@zhangqiaokeyan.com

京公网安备:11010802029741号 ICP备案号:京ICP备15016152号-6 六维联合信息科技 (北京) 有限公司©版权所有
  • 客服微信

  • 服务号