首页> 外文会议>European conference on computer vision >Unsupervised CNN for Single View Depth Estimation: Geometry to the Rescue
【24h】

Unsupervised CNN for Single View Depth Estimation: Geometry to the Rescue

机译:单视图深度估计的无监督CNN:救援的几何

获取原文

摘要

A significant weakness of most current deep Convolutional Neural Networks is the need to train them using vast amounts of manually labelled data. In this work we propose a unsupervised framework to learn a deep convolutional neural network for single view depth prediction, without requiring a pre-training stage or annotated ground-truth depths. We achieve this by training the network in a manner analogous to an autoencoder. At training time we consider a pair of images, source and target, with small, known camera motion between the two such as a stereo pair. We train the convolutional encoder for the task of predicting the depth map for the source image. To do so, we explicitly generate an inverse warp of the target image using the predicted depth and known inter-view displacement, to reconstruct the source image; the photometric error in the reconstruction is the reconstruction loss for the encoder. The acquisition of this training data is considerably simpler than for equivalent systems, requiring no manual annotation, nor calibration of depth sensor to camera. We show that our network trained on less than half of the KITTI dataset gives comparable performance to that of the state-of-the-art supervised methods for single view depth estimation.
机译:大多数当前深度卷积神经网络的显着弱点是需要使用大量手动标记的数据训练它们。在这项工作中,我们提出了一个无人监督的框架来学习用于单视图深度预测的深度卷积神经网络,而无需预训练阶段或注释的地面真理深度。我们通过以类似于AutoEncoder的方式培训网络来实现这一目标。在训练时,我们考虑一对图像,源头和目标,其中两者之间具有小,已知的相机运动,例如立体对。我们训练卷积编码器,以便预测源图像的深度图的任务。为此,我们使用预测的深度和已知的视图间位移明确地生成目标图像的逆扭曲,以重建源图像;重建中的光度误差是编码器的重建损耗。收购此培训数据比同等系统更简单,不需要手动注释,也不需要对相机进行深度传感器的校准。我们展示我们的网络在不到一半的基提数据集中培训了对单视图深度估计的最先进的监督方法的性能。

著录项

相似文献

  • 外文文献
  • 中文文献
  • 专利
获取原文

客服邮箱:kefu@zhangqiaokeyan.com

京公网安备:11010802029741号 ICP备案号:京ICP备15016152号-6 六维联合信息科技 (北京) 有限公司©版权所有
  • 客服微信

  • 服务号