...
首页> 外文期刊>Multimedia, IEEE Transactions on >Large-Margin Multi-Modal Deep Learning for RGB-D Object Recognition
【24h】

Large-Margin Multi-Modal Deep Learning for RGB-D Object Recognition

机译:大距离多模态深度学习用于RGB-D对象识别

获取原文
获取原文并翻译 | 示例
   

获取外文期刊封面封底 >>

       

摘要

Most existing feature learning-based methods for RGB-D object recognition either combine RGB and depth data in an undifferentiated manner from the outset, or learn features from color and depth separately, which do not adequately exploit different characteristics of the two modalities or utilize the shared relationship between the modalities. In this paper, we propose a general CNN-based multi-modal learning framework for RGB-D object recognition. We first construct deep CNN layers for color and depth separately, which are then connected with a carefully designed multi-modal layer. This layer is designed to not only discover the most discriminative features for each modality, but is also able to harness the complementary relationship between the two modalities. The results of the multi-modal layer are back-propagated to update parameters of the CNN layers, and the multi-modal feature learning and the back-propagation are iteratively performed until convergence. Experimental results on two widely used RGB-D object datasets show that our method for general multi-modal learning achieves comparable performance to state-of-the-art methods specifically designed for RGB-D data.
机译:现有的大多数基于特征学习的RGB-D对象识别方法要么从一开始就以无差别的方式组合RGB和深度数据,要么分别从颜色和深度学习特征,但这些方法不能充分利用两种模态的不同特征或无法利用模式之间的共享关系。在本文中,我们提出了一个基于CNN的通用多模式学习框架,用于RGB-D对象识别。我们首先分别构建用于颜色和深度的深CNN层,然后将其与精心设计的多模式层连接。该层旨在不仅发现每种模态的最具区别性的功能,而且还能够利用两种模态之间的互补关系。反向传播多模式层的结果以更新CNN层的参数,并反复执行多模式特征学习和反向传播,直到收敛为止。在两个广泛使用的RGB-D对象数据集上的实验结果表明,我们的通用多模式学习方法与为RGB-D数据专门设计的最新方法具有可比的性能。

著录项

相似文献

  • 外文文献
  • 中文文献
  • 专利
获取原文

客服邮箱:kefu@zhangqiaokeyan.com

京公网安备:11010802029741号 ICP备案号:京ICP备15016152号-6 六维联合信息科技 (北京) 有限公司©版权所有
  • 客服微信

  • 服务号