首页> 外文会议>European Conference on Computer Vision >Pseudo RGB-D for Self-improving Monocular SLAM and Depth Prediction
【24h】

Pseudo RGB-D for Self-improving Monocular SLAM and Depth Prediction

机译:伪RGB-D用于自我提高单眼血液和深度预测

获取原文

摘要

Classical monocular Simultaneous Localization And Mapping (SLAM) and the recently emerging convolutional neural networks (CNNs) for monocular depth prediction represent two largely disjoint approaches towards building a 3D map of the surrounding environment. In this paper, we demonstrate that the coupling of these two by leveraging the strengths of each mitigates the other's shortcomings. Specifically, we propose a joint narrow and wide baseline based self-improving framework, where on the one hand the CNN-predicted depth is leveraged to perform pseudo RGB-D feature-based SLAM, leading to better accuracy and robustness than the monocular RGB SLAM baseline. On the other hand, the bundle-adjusted 3D scene structures and camera poses from the more principled geometric SLAM are injected back into the depth network through novel wide baseline losses proposed for improving the depth prediction network, which then continues to contribute towards better pose and 3D structure estimation in the next iteration. We emphasize that our framework only requires unlabeled monocular videos in both training and inference stages, and yet is able to outperform state-of-the-art self-supervised monocular and stereo depth prediction networks {e.g., Monodepth2) and feature-based monocular SLAM system (i.e., ORB-SLAM). Extensive experiments on KITTI and TUM RGB-D datasets verify the superiority of our self-improving geometry-CNN framework.
机译:用于单眼深度预测的经典单目一象同时定位和映射(SLAM)和最近的卷积神经网络(CNNS)代表了构建周围环境的3D地图的两个主要脱节方法。在本文中,我们证明了通过利用每个减灾的强度来耦合其他缺点。具体而言,我们提出了一种联合窄宽的基线基础的自我改善框架,在一方面,在一方面,CNN预测深度利用以执行基于伪RGB-D功能的SLAM,导致比单眼RGB SLAM更好的精度和鲁棒性基线。另一方面,通过更具原则的几何液体的束调节的3D场景结构和相机通过提出用于改善深度预测网络的新型基线损耗来注入深度网络,然后继续为更好的姿势提供贡献3D结构估计在下一次迭代中。我们强调我们的框架只需要在训练和推理阶段进行未标记的单像素视频,但却能够以最先进的自我监督的单眼和立体声深度预测网络(例如,Monodepth2)和基于特征的单眼猛击系统(即,ORB-SLAM)。关于基蒂和Tum RGB-D数据集的广泛实验验证了我们自我改善的几何CNN框架的优越性。

著录项

相似文献

  • 外文文献
  • 中文文献
  • 专利
获取原文

客服邮箱:kefu@zhangqiaokeyan.com

京公网安备:11010802029741号 ICP备案号:京ICP备15016152号-6 六维联合信息科技 (北京) 有限公司©版权所有
  • 客服微信

  • 服务号