首页> 外文会议>International symposium on multispectral image processing and pattern recognition >Action recognition in depth video from RGB perspective: A knowledge transfer manner
【24h】

Action recognition in depth video from RGB perspective: A knowledge transfer manner

机译:从RGB角度看深度视频中的动作识别:一种知识转移方式

获取原文

摘要

Different video modal for human action recognition has becoming a highly promising trend in the video analysis. In this paper, we propose a method for human action recognition from RGB video to Depth video using domain adaptation, where we use learned feature from RGB videos to do action recognition for depth videos. More specifically, we make three steps for solving this problem in this paper. First, different from image, video is more complex as it has both spatial and temporal information, in order to bettor encode this information, dynamic image method is used to represent each RGB or Depth video to one image, based on this, most methods for extracting feature in image can be used in video. Secondly, as video can be represented as image, so standard CNN model can be used for training and testing for videos, beside, CNN model can be also used for feature extracting as its powerful feature expressing ability. Thirdly, as RGB videos and Depth videos are belong to two different domains, in order to make two different feature domains has more similarity, domain adaptation is firstly used for solving this problem between RGB and Depth video, based on this, the learned feature from RGB video model can be directly used for Depth video classification. We evaluate the proposed method on one complex R.GB-D action dataset (NTU RGB-D), and our method can have more than 2% accuracy improvement using domain adaptation from RGB to Depth action recognition.
机译:用于人类动作识别的不同视频模式已成为视频分析中非常有希望的趋势。在本文中,我们提出了一种使用域自适应从RGB视频到深度视频进行人类动作识别的方法,其中我们利用RGB视频中的学习特征对深度视频进行动作识别。更具体地说,我们为解决此问题采取了三个步骤。首先,与图像不同,视频由于具有空间和时间信息而更加复杂,为了更好地对该信息进行编码,动态图像方法用于将每个RGB或深度视频表示为一个图像,基于此,大多数方法图像中的提取功能可以在视频中使用。其次,由于视频可以表示为图像,因此可以使用标准的CNN模型对视频进行训练和测试,此外,由于CNN模型具有强大的特征表达能力,因此也可以用于特征提取。第三,由于RGB视频和深度视频属于两个不同的域,为了使两个不同的特征域具有更大的相似性,因此首先使用域自适应来解决RGB和深度视频之间的问题,在此基础上, RGB视频模型可以直接用于深度视频分类。我们在一个复杂的R.GB-D动作数据集(NTU RGB-D)上评估了该方法,并且使用从RGB到深度动作识别的域自适应,我们的方法可以将准确性提高2%以上。

著录项

相似文献

  • 外文文献
  • 中文文献
  • 专利
获取原文

客服邮箱:kefu@zhangqiaokeyan.com

京公网安备:11010802029741号 ICP备案号:京ICP备15016152号-6 六维联合信息科技 (北京) 有限公司©版权所有
  • 客服微信

  • 服务号