首页> 外文会议>IEEE International Conference on Emerging Technologies and Factory Automation >Audio-visual integration for human-robot interaction in multi-person scenarios
【24h】

Audio-visual integration for human-robot interaction in multi-person scenarios

机译:多人情景中人体机器人交互的视听集成

获取原文
获取外文期刊封面目录资料

摘要

This paper presents the integration of audio-visual perception components for human robot interaction in the Robot Operating System (ROS). Visual-based nodes consist of skeleton tracking and gesture recognition using a depth camera, and face recognition using an RGB camera. Auditory perception is based on sound source localization using a microphone array. We present an integration framework of these nodes using a top-down hierarchical messaging protocol. On the top of the integration, a message carries information about the number of persons and their corresponding states (who, what, where), which are updated from many low-level perception nodes. The top message is passed to a planning node to make a reaction of the robot, according to the perception about surrounding people. This paper demonstrates human-robot interaction in multi-persons scenario where robot pays its attention to the speaking or waving hand persons. Moreover, this modularization architecture enables reusing modules for other applications. To validate this approach, two sound source localization algorithms are evaluated in real-time where ground-truth localization is provided by the face recognition module.
机译:本文介绍了机器人操作系统(ROS)中人体机器人交互的视听感知组件的集成。基于视觉的节点包括使用深度摄像机的骨架跟踪和手势识别,并使用RGB相机进行人脸识别。听觉感知基于使用麦克风阵列的声源本地化。我们使用自上而下的分层消息传递协议介绍了这些节点的集成框架。在集成之上,一条消息携带有关人数及其相应状态的信息(谁,谁,其中),其从许多低级感知节点更新。根据对周围人的看法,最重要的消息被传递给规划节点以进行机器人的反应。本文展示了多人情景中的人机互动,机器人将注意力付诸讲话或挥手人员。此外,这种模块化架构使能为其他应用程序重用模块。为了验证这种方法,在面部识别模块提供地面真实定位的实时评估两个声源定位算法。

著录项

相似文献

  • 外文文献
  • 中文文献
  • 专利
获取原文

客服邮箱:kefu@zhangqiaokeyan.com

京公网安备:11010802029741号 ICP备案号:京ICP备15016152号-6 六维联合信息科技 (北京) 有限公司©版权所有
  • 客服微信

  • 服务号