首页> 外文会议>IEEE International Conference on Image Processing >Kinect-based multimodal gesture recognition using a two-pass fusion scheme
【24h】

Kinect-based multimodal gesture recognition using a two-pass fusion scheme

机译:使用两遍融合方案的基于Kinect的多模式手势识别

获取原文

摘要

We present a new framework for multimodal gesture recognition that is based on a two-pass fusion scheme. In this, we deal with a demanding Kinect-based multimodal dataset, which was introduced in a recent gesture recognition challenge. We employ multiple modalities, i.e., visual cues, such as colour and depth images, as well as audio, and we specifically extract feature descriptors of the hands' movement, handshape, and audio spectral properties. Based on these features, we statistically train separate unimodal gesture-word models, namely hidden Markov models, explicitly accounting for the dynamics of each modality. Multimodal recognition of unknown gesture sequences is achieved by combining these models in a late, two-pass fusion scheme that exploits a set of unimodally generated n-best recognition hypotheses. The proposed scheme achieves 88.2% gesture recognition accuracy in the Kinect-based multimodal dataset, outperforming all recently published approaches on the same challenging multimodal gesture recognition task.
机译:我们提出了一种基于两遍融合方案的多模式手势识别新框架。在此,我们处理了一个苛刻的基于Kinect的多峰数据集,该数据集是在最近的手势识别挑战中引入的。我们采用了多种模式,即视觉提示,例如彩色和深度图像以及音频,并且我们专门提取了手的运动,手形和音频频谱特性的特征描述符。基于这些功能,我们统计地训练了单独的单峰手势词模型,即隐马尔可夫模型,明确考虑了每种模态的动态性。未知手势序列的多模式识别是通过在后期的两遍融合方案中组合这些模型来实现的,该方案利用了一组单模生成的n最佳识别假设。所提出的方案在基于Kinect的多模态数据集中实现了88.2%的手势识别精度,胜过了同一挑战性多模态手势识别任务上所有最近发布的方法。

著录项

相似文献

  • 外文文献
  • 中文文献
  • 专利
获取原文

客服邮箱:kefu@zhangqiaokeyan.com

京公网安备:11010802029741号 ICP备案号:京ICP备15016152号-6 六维联合信息科技 (北京) 有限公司©版权所有
  • 客服微信

  • 服务号