首页> 外文会议>IEEE International Conference on Robotics and Automation >Predicting human intention in visual observations of hand/object interactions
【24h】

Predicting human intention in visual observations of hand/object interactions

机译:通过视觉观察手/物体相互作用来预测人的意图

获取原文

摘要

The main contribution of this paper is a probabilistic method for predicting human manipulation intention from image sequences of human-object interaction. Predicting intention amounts to inferring the imminent manipulation task when human hand is observed to have stably grasped the object. Inference is performed by means of a probabilistic graphical model that encodes object grasping tasks over the 3D state of the observed scene. The 3D state is extracted from RGB-D image sequences by a novel vision-based, markerless hand-object 3D tracking framework. To deal with the high-dimensional state-space and mixed data types (discrete and continuous) involved in grasping tasks, we introduce a generative vector quantization method using mixture models and self-organizing maps. This yields a compact model for encoding of grasping actions, able of handling uncertain and partial sensory data. Experimentation showed that the model trained on simulated data can provide a potent basis for accurate goal-inference with partial and noisy observations of actual real-world demonstrations. We also show a grasp selection process, guided by the inferred human intention, to illustrate the use of the system for goal-directed grasp imitation.
机译:本文的主要贡献是一种从人与物体交互的图像序列中预测人为操纵意图的概率方法。当观察到人的手稳定地抓住物体时,预测意图等于推断即将到来的操纵任务。推理是通过概率图形模型执行的,该模型对观察场景的3D状态上的对象抓取任务进行编码。通过新颖的基于视觉的,无标记的手部对象3D跟踪框架,可以从RGB-D图像序列中提取3D状态。为了处理抓取任务涉及的高维状态空间和混合数据类型(离散和连续),我们引入了一种使用混合模型和自组织映射的生成矢量量化方法。这产生了用于对抓握动作进行编码的紧凑模型,该模型能够处理不确定和部分感官数据。实验表明,在模拟数据上训练的模型可以为准确的目标推断提供有效的基础,并且可以对实际现实演示进行部分和嘈杂的观察。我们还展示了在推断出的人类意图指导下的抓地力选择过程,以说明该系统用于目标导向的抓地力模仿的使用。

著录项

相似文献

  • 外文文献
  • 中文文献
  • 专利
获取原文

客服邮箱:kefu@zhangqiaokeyan.com

京公网安备:11010802029741号 ICP备案号:京ICP备15016152号-6 六维联合信息科技 (北京) 有限公司©版权所有
  • 客服微信

  • 服务号