首页> 外文会议>European conference on computer vision >Multi-modal Unsupervised Feature Learning for RGB-D Scene Labeling
【24h】

Multi-modal Unsupervised Feature Learning for RGB-D Scene Labeling

机译:用于RGB-D场景标签的多模式无监督特征学习

获取原文

摘要

Most of the existing approaches for RGB-D indoor scene labeling employ hand-crafted features for each modality independently and combine them in a heuristic manner. There has been some attempt on directly learning features from raw RGB-D data, but the performance is not satisfactory. In this paper, we adapt the unsupervised feature learning technique for RGB-D labeling as a multi-modality learning problem. Our learning framework performs feature learning and feature encoding simultaneously which significantly boosts the performance. By stacking basic learning structure, higher-level features are derived and combined with lower-level features for better representing RGB-D data. Experimental results on the benchmark NYU depth dataset show that our method achieves competitive performance, compared with state-of-the-art.
机译:现有的大多数RGB-D室内场景标记方法都为每种模态独立采用手工制作的功能,并以启发式方式将它们组合在一起。已经尝试从原始RGB-D数据直接学习特征,但是性能并不令人满意。在本文中,我们将无监督特征学习技术用于RGB-D标记,将其作为一种多模态学习问题。我们的学习框架同时执行特征学习和特征编码,从而大大提高了性能。通过堆叠基本学习结构,可以导出更高级别的功能并将其与更低级别的功能组合在一起,以更好地表示RGB-D数据。在基准NYU深度数据集上的实验结果表明,与最新技术相比,我们的方法具有竞争优势。

著录项

相似文献

  • 外文文献
  • 中文文献
  • 专利
获取原文

客服邮箱:kefu@zhangqiaokeyan.com

京公网安备:11010802029741号 ICP备案号:京ICP备15016152号-6 六维联合信息科技 (北京) 有限公司©版权所有
  • 客服微信

  • 服务号