首页> 外文会议>European conference on computer vision >Score-Level Multi Cue Fusion for Sign Language Recognition
【24h】

Score-Level Multi Cue Fusion for Sign Language Recognition

机译:分数级多提示融合,用于行语识别

获取原文

摘要

Sign Languages are expressed through hand and upper body gestures as well as facial expressions. Therefore, Sign Language Recognition (SLR) needs to focus on all such cues. Previous work uses hand-crafted mechanisms or network aggregation to extract the different cue features, to increase SLR performance. This is slow and involves complicated architectures. We propose a more straightforward approach that focuses on training separate cue models specializing on the dominant hand, hands, face, and upper body regions. We compare the performance of 3D Convolutional Neural Network (CNN) models specializing in these regions, combine them through score-level fusion, and use the weighted alternative. Our experimental results have shown the effectiveness of mixed convolutional models. Their fusion yields up to 19% accuracy improvement over the baseline using the full upper body. Furthermore, we include a discussion for fusion settings, which can help future work on Sign Language Translation (SLT).
机译:标志语言通过手和上半身手势表达以及面部表情。因此,手语识别(SLR)需要专注于所有此类提示。以前的工作采用手工制作的机制或网络聚合来提取不同的提示功能,以提高SLR性能。这缓慢并涉及复杂的架构。我们提出了一种更直接的方法,专注于培训专门从事主导手,手,面部和上身区域的单独提示模型。我们比较专门从事这些区域的3D卷积神经网络(CNN)模型的表现,将它们通过得分级融合,并使用加权替代品。我们的实验结果表明了混合卷积模型的有效性。它们的融合在基线上使用全部上半身的准确性提高了高达19%的准确性。此外,我们包括讨论融合设置,可以帮助未来的手语翻译(SLT)。

著录项

相似文献

  • 外文文献
  • 中文文献
  • 专利
获取原文

客服邮箱:kefu@zhangqiaokeyan.com

京公网安备:11010802029741号 ICP备案号:京ICP备15016152号-6 六维联合信息科技 (北京) 有限公司©版权所有
  • 客服微信

  • 服务号