首页> 外文会议>IEEE International Conference on Computer Vision >MMSS: Multi-modal Sharable and Specific Feature Learning for RGB-D Object Recognition
【24h】

MMSS: Multi-modal Sharable and Specific Feature Learning for RGB-D Object Recognition

机译:MMSS:用于RGB-D对象识别的多模式可共享和特定功能学习

获取原文

摘要

Most of the feature-learning methods for RGB-D object recognition either learn features from color and depth modalities separately, or simply treat RGB-D as undifferentiated four-channel data, which cannot adequately exploit the relationship between different modalities. Motivated by the intuition that different modalities should contain not only some modal-specific patterns but also some shared common patterns, we propose a multi-modal feature learning framework for RGB-D object recognition. We first construct deep CNN layers for color and depth separately, and then connect them with our carefully designed multi-modal layers, which fuse color and depth information by enforcing a common part to be shared by features of different modalities. In this way, we obtain features reflecting shared properties as well as modal-specific properties in different modalities. The information of the multi-modal learning frameworks is back-propagated to the early CNN layers. Experimental results show that our proposed multi-modal feature learning method outperforms state-of-the-art approaches on two widely used RGB-D object benchmark datasets.
机译:大多数用于RGB-D对象识别的特征学习方法要么分别从颜色和深度模态中学习特征,要么将RGB-D视为未区分的四通道数据,无法充分利用不同模态之间的关系。出于直觉,即不同的模式不仅应包含某些特定于模式的模式,还应包含一些共享的通用模式,我们提出了一种用于RGB-D对象识别的多模式特征学习框架。我们首先分别构造用于颜色和深度的深CNN层,然后将它们与我们精心设计的多模式层连接起来,这些层通过强制由不同模式的功能共享的公共部分来融合颜色和深度信息。通过这种方式,我们获得了反映共享特性以及不同模态中特定于模态的特性的特征。多模式学习框架的信息将反向传播到早期的CNN层。实验结果表明,在两个广泛使用的RGB-D对象基准数据集上,我们提出的多模式特征学习方法优于最新方法。

著录项

相似文献

  • 外文文献
  • 中文文献
  • 专利
获取原文

客服邮箱:kefu@zhangqiaokeyan.com

京公网安备:11010802029741号 ICP备案号:京ICP备15016152号-6 六维联合信息科技 (北京) 有限公司©版权所有
  • 客服微信

  • 服务号