首页> 外文会议>2018 13th IEEE International Conference on Automatic Face amp; Gesture Recognition >An Immersive System with Multi-Modal Human-Computer Interaction
【24h】

An Immersive System with Multi-Modal Human-Computer Interaction

机译:具有多模式人机交互功能的沉浸式系统

获取原文
获取原文并翻译 | 示例

摘要

We introduce an immersive system prototype that integrates face, gesture and speech recognition techniques to support multi-modal human-computer interaction capability. Embedded in an indoor room setting, a multi-camera system is developed to monitor the user facial behavior, body gesture and spatial location in the room. A server that fuses different sensor inputs in a time-sensitive manner so that our system knows who is doing what at where in real-time. When correlating with speech input, the system can better understand the user intention for interaction purpose. We evaluate the performance of core recognition techniques on both benchmark and self-collected datasets and demonstrate the benefit of the system in various use cases.
机译:我们引入了一个沉浸式系统原型,该原型集成了面部,手势和语音识别技术,以支持多模式人机交互功能。嵌入在室内房间环境中的多摄像机系统被开发出来,用于监视用户在房间中的面部表情,身体姿势和空间位置。服务器以时间敏感的方式融合不同的传感器输入,以便我们的系统实时了解谁在做什么。当与语音输入相关时,系统可以更好地理解用户意图以进行交互。我们评估了基准和自收集数据集上核心识别技术的性能,并展示了该系统在各种用例中的优势。

著录项

相似文献

  • 外文文献
  • 中文文献
  • 专利
获取原文

客服邮箱:kefu@zhangqiaokeyan.com

京公网安备:11010802029741号 ICP备案号:京ICP备15016152号-6 六维联合信息科技 (北京) 有限公司©版权所有
  • 客服微信

  • 服务号