首页> 外文会议>IEEE International Conference on Robotics and Automation >Fusion of Stereo and Still Monocular Depth Estimates in a Self-Supervised Learning Context
【24h】

Fusion of Stereo and Still Monocular Depth Estimates in a Self-Supervised Learning Context

机译:自我监督学习上下文中的立体声和单眼深度估计的融合

获取原文

摘要

We study how autonomous robots can learn by themselves to improve their depth estimation capability. In particular, we investigate a self-supervised learning setup in which stereo vision depth estimates serve as targets for a convolutional neural network (CNN) that transforms a single still image to a dense depth map. After training, the stereo and mono estimates are fused with a novel fusion method that preserves high confidence stereo estimates, while leveraging the CNN estimates in the low-confidence regions. The main contribution of the article is that it is shown that the fused estimates lead to a higher performance than the stereo vision estimates alone. Experiments are performed on the KITTI dataset, and on board of a Parrot SLAMDunk, showing that even rather limited CNNs can help provide stereo vision equipped robots with more reliable depth maps for autonomous navigation.
机译:我们研究自主机器人如何自学以提高其深度估计能力。特别是,我们研究了一种自我监督的学习设置,其中立体视觉深度估计用作卷积神经网络(CNN)的目标,该卷积神经网络将单个静态图像转换为密集的深度图。训练后,将立体声和单声道估计值与一种新颖的融合方法进行融合,该方法可以保留高置信度立体声估计值,同时利用低置信区域中的CNN估计值。该文章的主要贡献在于,表明融合的估计比单独的立体视觉估计具有更高的性能。在KITTI数据集和Parrot SLAMDunk船上进行的实验表明,即使有限的CNN也可以为配备立体视觉的机器人提供更可靠的深度图,以进行自动导航。

著录项

相似文献

  • 外文文献
  • 中文文献
  • 专利
获取原文

客服邮箱:kefu@zhangqiaokeyan.com

京公网安备:11010802029741号 ICP备案号:京ICP备15016152号-6 六维联合信息科技 (北京) 有限公司©版权所有
  • 客服微信

  • 服务号