首页> 外文会议>IEEE International Conference on Computer Vision >Learning Image Representations Tied to Ego-Motion
【24h】

Learning Image Representations Tied to Ego-Motion

机译:学习与自我运动息息相关的图像表示

获取原文

摘要

Understanding how images of objects and scenes behave in response to specific ego-motions is a crucial aspect of proper visual development, yet existing visual learning methods are conspicuously disconnected from the physical source of their images. We propose to exploit proprioceptive motor signals to provide unsupervised regularization in convolutional neural networks to learn visual representations from egocentric video. Specifically, we enforce that our learned features exhibit equivariance, i.e, they respond predictably to transformations associated with distinct ego-motions. With three datasets, we show that our unsupervised feature learning approach significantly outperforms previous approaches on visual recognition and next-best-view prediction tasks. In the most challenging test, we show that features learned from video captured on an autonomous driving platform improve large-scale scene recognition in static images from a disjoint domain.
机译:了解对象和场景的图像如何响应特定的自我动作是如何适当的视觉开发的关键方面,然而现有的视觉学习方法明显地与图像的物理来源断开连接。我们建议利用卷积神经网络中的未经监督的正则化,以便学习来自EGoCentric视频的视觉表示。具体而言,我们强制执行我们的学到的特征,即,它们可预测地响应与不同的自我运动相关的转化。使用三个数据集,我们表明我们无监督的特征学习方法显着优异地优于可视识别和下一个最佳视图预测任务的先前方法。在最具挑战性的测试中,我们显示从自主驾驶平台上捕获的视频中学习的功能提高了来自不相交域的静态图像中的大规模场景识别。

著录项

相似文献

  • 外文文献
  • 中文文献
  • 专利
获取原文

客服邮箱:kefu@zhangqiaokeyan.com

京公网安备:11010802029741号 ICP备案号:京ICP备15016152号-6 六维联合信息科技 (北京) 有限公司©版权所有
  • 客服微信

  • 服务号