首页> 外文会议>International Joint Conference on Neural Networks >Adversarial Action Data Augmentation for Similar Gesture Action Recognition
【24h】

Adversarial Action Data Augmentation for Similar Gesture Action Recognition

机译:相似手势动作识别的对抗动作数据增强

获取原文

摘要

Human gestures are unique for recognizing and describing human actions, and video-based human action recognition techniques are effective solutions to varies real-world applications, such as surveillance, video indexing, and human-computer interaction. Most existing video human action recognition approaches either using handcraft features from the frames or deep learning models such as convolutional neural networks (CNN) and recurrent neural networks (RNN); however, they have mostly overlooked the similar gestures between different actions when processing the frames into the models. The classifiers suffer from similar features extracted from similar gestures, which are unable to classify the actions in the video streams. In this paper, we propose a novel framework with generative adversarial networks (GAN) to generate the data augmentation for similar gesture action recognition. The contribution of our work is tri-fold: 1) we proposed a novel action data augmentation framework (ADAF) to enlarge the differences between the actions with very similar gestures; 2) the framework can boost the classification performance either on similar gesture action pairs or the whole dataset; 3) experiments conducted on both KTH and UCF101 datasets show that our data augmentation framework boost the performance on both similar gestures actions as well as the whole dataset compared with baseline methods such as 2DCNN and 3DCNN.
机译:手势是识别和描述手势的唯一方法,基于视频的手势识别技术是解决各种现实应用(例如监视,视频索引和人机交互)的有效解决方案。现有的大多数视频人类动作识别方法都是使用帧中的手工特征或深度学习模型,例如卷积神经网络(CNN)和递归神经网络(RNN);但是,在将帧处理到模型中时,他们大多忽略了不同动作之间的相似手势。分类器遭受从相似手势提取的相似特征的困扰,这些相似特征无法对视频流中的动作进行分类。在本文中,我们提出了一种具有生成对抗网络(GAN)的新颖框架,以生成用于类似手势动作识别的数据增强。我们的工作有以下三方面的贡献:1)我们提出了一种新颖的动作数据增强框架(ADAF),以通过非常相似的手势来扩大动作之间的差异; 2)框架可以在相似的手势动作对或整个数据集上提高分类性能; 3)在KTH和UCF101数据集上进行的实验表明,与基线方法(例如2DCNN和3DCNN)相比,我们的数据增强框架提高了类似手势动作以及整个数据集的性能。

著录项

相似文献

  • 外文文献
  • 中文文献
  • 专利
获取原文

客服邮箱:kefu@zhangqiaokeyan.com

京公网安备:11010802029741号 ICP备案号:京ICP备15016152号-6 六维联合信息科技 (北京) 有限公司©版权所有
  • 客服微信

  • 服务号