...
首页> 外文期刊>Chaos >Cycle-SfM: Joint self-supervised learning of depth and camera motion from monocular image sequences
【24h】

Cycle-SfM: Joint self-supervised learning of depth and camera motion from monocular image sequences

机译:Cycle-SFM:从单眼图像序列的深度和相机运动的联合自我监督学习

获取原文
获取原文并翻译 | 示例
           

摘要

Understanding 3D scene geometry is a fundamental research topic in computer vision, including various subproblems, such as depth prediction, visual odometry, optical flow, etc. With the advent of artificial intelligence methods like deep learning, many approaches have emerged to deal with such problems in an end-to-end manner. These pipelines take the 3D understanding task as a nonlinear optimization problem, with the purpose of minimizing the cost function of the whole framework. Here, we present a self-supervised framework for jointly learning the monocular depth and camera's ego-motion from unlabeled, unstructured, and monocular video sequences. We propose a forward-backward consistency constraint on view reconstruction to capture temporal relations across adjacent frames, whose purpose is to explore and make full use of the bidirectional projection information. A simple and practicable improvement on the design of cost function is proposed to enhance the estimated accuracy. Due to the fact that our improvement is a lightweight and general module, it can be integrated into any self-supervised architectures seamlessly, and more accurate results can be obtained. The evaluation on the KITTI dataset demonstrates that our approach is highly efficient and performs better than the existing works in pose estimation, while the results in depth estimation perform comparably with the existing ones. Published under license by AIP Publishing.
机译:了解3D场景几何是计算机视觉中的基本研究主题,包括各种子问题,如深度预测,视觉内径,光学流量等,随着深度学习的人工智能方法的出现,许多方法都出现了处理这些问题以端到端的方式。这些管道将3D理解任务作为非线性优化问题,目的是最大限度地减少整个框架的成本函数。在这里,我们提出了一个自我监督的框架,用于联合学习单眼深度和相机的自我运动,从未标记,非结构化和单目视频序列。我们提出了一种前后一致性约束,查看重建以捕获相邻帧的时间关系,其目的是探索和充分利用双向投影信息。提出了对成本函数设计的简单且切实可行的改进,以提高估计的准确性。由于我们的改进是一种轻量级和一般模块,它可以无缝地集成到任何自我监督的架构中,并且可以获得更准确的结果。 Kitti DataSet的评估表明,我们的方法高效,比现有的姿势估计工作更好,而深度估计的结果与现有的结果相当执行。通过AIP发布根据许可发布。

著录项

  • 来源
    《Chaos》 |2019年第12期|共11页
  • 作者

  • 作者单位
  • 收录信息
  • 原文格式 PDF
  • 正文语种 eng
  • 中图分类 自然科学总论;
  • 关键词

相似文献

  • 外文文献
  • 中文文献
  • 专利
获取原文

客服邮箱:kefu@zhangqiaokeyan.com

京公网安备:11010802029741号 ICP备案号:京ICP备15016152号-6 六维联合信息科技 (北京) 有限公司©版权所有
  • 客服微信

  • 服务号