首页> 外文会议>IEEE International Symposium on Mixed and Augmented Reality >Learning to Fuse: A Deep Learning Approach to Visual-Inertial Camera Pose Estimation
【24h】

Learning to Fuse: A Deep Learning Approach to Visual-Inertial Camera Pose Estimation

机译:学习融合:一种用于视觉惯性相机姿势估计的深度学习方法

获取原文

摘要

Camera pose estimation is the cornerstone of Augmented Reality applications. Pose tracking based on camera images exclusively has been shown to be sensitive to motion blur, occlusions, and illumination changes. Thus, a lot of work has been conducted over the last years on visual-inertial pose tracking using acceleration and angular velocity measurements from inertial sensors in order to improve the visual tracking. Most proposed systems use statistical filtering techniques to approach the sensor fusion problem, that require complex system modelling and calibrations in order to perform adequately. In this work we present a novel approach to sensor fusion using a deep learning method to learn the relation between camera poses and inertial sensor measurements. A long short-term memory model (LSTM) is trained to provide an estimate of the current pose based on previous poses and inertial measurements. This estimates then appropriately combined with the output of a visual tracking system using a linear Kalman Filter to provide a robust final pose estimate. Our experimental results confirm the applicability and tracking performance improvement gained from the proposed sensor fusion system.
机译:相机姿势估计是增强现实应用程序的基石。已经显示,仅基于摄像机图像的姿势跟踪对运动模糊,遮挡和照明变化敏感。因此,在过去的几年中,在视觉惯性姿势跟踪方面进行了大量工作,这些惯性姿势跟踪使用了来自惯性传感器的加速度和角速度测量,以改善视觉跟踪。大多数提议的系统使用统计过滤技术来解决传感器融合问题,这需要复杂的系统建模和校准才能充分发挥作用。在这项工作中,我们提出了一种使用深度学习方法来学习传感器姿态与惯性传感器测量值之间关系的新型传感器融合方法。训练了长期短期记忆模型(LSTM),可以根据先前的姿势和惯性测量结果提供当前姿势的估计值。然后,此估计值与使用线性卡尔曼滤波器的视觉跟踪系统的输出适当组合,以提供可靠的最终姿态估计值。我们的实验结果证实了从拟议的传感器融合系统获得的适用性和跟踪性能的提高。

著录项

相似文献

  • 外文文献
  • 中文文献
  • 专利
获取原文

客服邮箱:kefu@zhangqiaokeyan.com

京公网安备:11010802029741号 ICP备案号:京ICP备15016152号-6 六维联合信息科技 (北京) 有限公司©版权所有
  • 客服微信

  • 服务号