...
首页> 外文期刊>Journal of electronic imaging >Two-stream Siamese network with contrastive-center losses for RGB-D action recognition
【24h】

Two-stream Siamese network with contrastive-center losses for RGB-D action recognition

机译:具有对比度中心损失的两流连体网络,用于RGB-D动作识别

获取原文
获取原文并翻译 | 示例
   

获取外文期刊封面封底 >>

       

摘要

Many fusion methods have been developed to improve the performance of action recognition with RGB and depth data, where learning conjoint representation of heterogeneous modalities by a single network has not been paid enough attention. We present an associated representation method for RGB-D action recognition using the siamese network with contrastive-center loss. First, some samples of each class and modality data are selected as the references to construct positive and negative pairs. Each positive pair consists of a training sample and its class reference, whereas the negative pair only involves different classes reference. Then these pairs are inputted to a two-stream siamese network to learn the collaborative representation of RGB and depth data. Two ranking losses, namely intramodal and cross-modal contrastive-center loss, are developed to impose similarity/dissimilarity metric on those pairs. Specifically, the intramodal contrastive-center loss measures the relationship between samples and references from RGB or depth data. The cross-modal contrastive-center loss measures the relationship of visual and depth features in a same low-dimensional space. Finally, the ranking losses and a softmax loss are jointly optimized for action recognition. The proposed method is evaluated on two large action datasets, LAP IsoGD and NTU RGB+D, and a smaller dataset, Sheffield Kinect gesture. The experimental results demonstrate that the proposed method surpasses most of the state-of-the-art methods. (C) 2019 SPIE and IS&T
机译:已经开发了许多融合方法来改善具有RGB和深度数据的动作识别的性能,其中通过单个网络来学习异构模态的联合表示尚未引起足够的重视。我们提出了一种使用暹罗网络进行对比中心损失的RGB-D动作识别的关联表示方法。首先,选择每个类别和形态数据的一些样本作为构建正负对的参考。每个正对都包含一个训练样本及其类别参考,而负对仅包含不同的类别参考。然后将这些对输入到两个流的暹罗网络中,以学习RGB和深度数据的协作表示。开发了两个等级损失,即模态内和跨模态的对比中心损失,以在这些对上施加相似性/不相似性度量。具体而言,模态内对比中心损失测量的是RGB或深度数据中样本与参考之间的关系。跨模态的对比中心损失度量了同一低维空间中视觉和深度特征的关系。最后,联合优化排名损失和softmax损失以进行动作识别。该方法在两个大型动作数据集LAP IsoGD和NTU RGB + D以及较小的数据集Sheffield Kinect手势上进行了评估。实验结果表明,所提出的方法超越了大多数最新技术。 (C)2019 SPIE和IS&T

著录项

相似文献

  • 外文文献
  • 中文文献
  • 专利
获取原文

客服邮箱:kefu@zhangqiaokeyan.com

京公网安备:11010802029741号 ICP备案号:京ICP备15016152号-6 六维联合信息科技 (北京) 有限公司©版权所有
  • 客服微信

  • 服务号