首页> 外文会议>International Conference on Computer Vision >Unsupervised Collaborative Learning of Keyframe Detection and Visual Odometry Towards Monocular Deep SLAM
【24h】

Unsupervised Collaborative Learning of Keyframe Detection and Visual Odometry Towards Monocular Deep SLAM

机译:面向单眼深SLAM的关键帧检测和视觉里程表的无监督协作学习

获取原文

摘要

In this paper we tackle the joint learning problem of keyframe detection and visual odometry towards monocular visual SLAM systems. As an important task in visual SLAM, keyframe selection helps efficient camera relocalization and effective augmentation of visual odometry. To benefit from it, we first present a deep network design for the keyframe selection, which is able to reliably detect keyframes and localize new frames, then an end-to-end unsupervised deep framework further proposed for simultaneously learning the keyframe selection and the visual odometry tasks. As far as we know, it is the first work to jointly optimize these two complementary tasks in a single deep framework. To make the two tasks facilitate each other in the learning, a collaborative optimization loss based on both geometric and visual metrics is proposed. Extensive experiments on publicly available datasets (ie~KITTI raw dataset and its odometry split) clearly demonstrate the effectiveness of the proposed approach, and new state-of-the-art results are established on the unsupervised depth and pose estimation from monocular videos.
机译:在本文中,我们针对单目视觉SLAM系统解决了关键帧检测和视觉里程表的联合学习问题。作为视觉SLAM中的一项重要任务,关键帧选择有助于有效地重新定位相机并有效扩大视觉里程表。为了从中受益,我们首先提出一种用于关键帧选择的深度网络设计,它能够可靠地检测关键帧并定位新帧,然后进一步提出了一种端到端无监督的深度框架,用于同时学习关键帧选择和可视化。里程表任务。据我们所知,这是在单个深度框架中共同优化这两个互补任务的第一项工作。为了使这两个任务在学习中相互促进,提出了基于几何和视觉指标的协同优化损失。在公开可用的数据集(即〜KITTI原始数据集及其里程表拆分)上进行的广泛实验清楚地证明了该方法的有效性,并且在单眼视频的无监督深度和姿势估计基础上建立了最新技术成果。

著录项

相似文献

  • 外文文献
  • 中文文献
  • 专利
获取原文

客服邮箱:kefu@zhangqiaokeyan.com

京公网安备:11010802029741号 ICP备案号:京ICP备15016152号-6 六维联合信息科技 (北京) 有限公司©版权所有
  • 客服微信

  • 服务号