首页> 外文期刊>Machine Vision and Applications >One-shot learning hand gesture recognition based on modified 3d convolutional neural networks
【24h】

One-shot learning hand gesture recognition based on modified 3d convolutional neural networks

机译:基于改进的3D卷积神经网络的单次学习手势识别

获取原文
获取原文并翻译 | 示例
       

摘要

Though deep neural networks have played a very important role in the field of vision-based hand gesture recognition, however, it is challenging to acquire large numbers of annotated samples to support its deep learning or training. Furthermore, in practical applications it often encounters some case with only one single sample for a new gesture class so that conventional recognition method cannot be qualified with a satisfactory classification performance. In this paper, the methodology of transfer learning is employed to build an effective network architecture of one-shot learning so as to deal with such intractable problem. Then some useful knowledge from deep training with big dataset of relative objects can be transferred and utilized to strengthen one-shot learning hand gesture recognition (OSLHGR) rather than to train a network from scratch. According to this idea a well-designed convolutional network architecture with deeper layers, C3D (Tran et al. in: ICCV, pp 4489-4497, 2015), is modified as an effective tool to extract spatiotemporal feature by deep learning. Then continuous fine-tune training is performed on a sample of new classes to complete one-shot learning. Moreover, the test of classification is carried out by Softmax classifier and geometrical classification based on Euclidean distance. Finally, a series of experiments and tests on two benchmark datasets, VIVA (Vision for Intelligent Vehicles and Applications) and SKIG (Sheffield Kinect Gesture) are conducted to demonstrate its state-of-the-art recognition accuracy of our proposed method. Meanwhile, a special dataset of gestures, BSG, is built using SoftKinetic DS325 for the test of OSLHGR, and a series of test results verify and validate its well classification performance and real-time response speed.
机译:尽管深度神经网络在基于视觉的手势识别领域中发挥了非常重要的作用,但是,要获取大量带注释的样本以支持其深度学习或训练仍是一项挑战。此外,在实际应用中,对于新的手势类别,通常只用一个样本就遇到这种情况,因此常规的识别方法不能用令人满意的分类性能来限定。本文采用转移学习的方法来构建有效的单发学习网络架构,以解决这一棘手的问题。然后,可以将大型对象相关数据集的深度训练中的一些有用知识转移并用于加强一次性学习手势识别(OSLHGR),而不是从头开始训练网络。根据这个想法,经过精心设计的具有更深层次的卷积网络架构C3D(Tran等人:ICCV,第4489-4497页,2015年)被修改为通过深度学习提取时空特征的有效工具。然后,对新课程的样本进行连续的微调训练,以完成一次学习。此外,分类测试是通过Softmax分类器和基于欧氏距离的几何分类进行的。最后,在两个基准数据集VIVA(智能车辆及其应用的视觉)和SKIG(谢菲尔德Kinect手势)上进行了一系列实验和测试,以证明我们提出的方法的最新识别精度。同时,使用SoftKinetic DS325构建了一个特殊的手势数据集BSG,用于OSLHGR的测试,一系列测试结果验证并验证了其良好的分类性能和实时响应速度。

著录项

相似文献

  • 外文文献
  • 中文文献
  • 专利
获取原文

客服邮箱:kefu@zhangqiaokeyan.com

京公网安备:11010802029741号 ICP备案号:京ICP备15016152号-6 六维联合信息科技 (北京) 有限公司©版权所有
  • 客服微信

  • 服务号