首页> 外文会议>International Conference on Neural Information Processing >Achieving Human-Robot Collaboration with Dynamic Goal Inference by Gradient Descent
【24h】

Achieving Human-Robot Collaboration with Dynamic Goal Inference by Gradient Descent

机译:通过梯度下降实现与动态目标推断的人机合作

获取原文

摘要

Collaboration with a human partner is a challenging task expected of intelligent robots. To realize this, robots need the ability to share a particular goal with a human and dynamically infer whether the goal state is changed by the human. In this paper, we propose a neural network-based computational framework with a gradient-based optimization of the goal state that enables robots to achieve this ability. The proposed framework consists of convolutional variational autoencoders (ConvVAEs) and a recurrent neural network (RNN) with a long short-term memory (LSTM) architecture that learns to map a given goal image for collaboration to visuomotor predictions. More specifically, visual and goal feature states are first extracted by the encoder of the respective ConvVAEs. Visual feature and motor predictions are then generated by the LSTM based on their current state and are conditioned according to the extracted goal feature state. During collaboration after the learning process, the goal feature state is optimized by gradient descent to minimize errors between the predicted and actual visual feature states. This enables the robot to dynamically infer situational (goal) changes of the human partner from visual observations alone. The proposed framework is evaluated by conducting experiments on a human-robot collaboration task involving object assembly. Experimental results demonstrate that a robot equipped with the proposed framework can collaborate with a human partner through dynamic goal inference even when the situation is ambiguous.
机译:与人类伴侣的合作是智能机器人的挑战性任务。为了实现这一点,机器人需要能够与人类共享特定目标,并动态推断人类是否改变了目标状态。在本文中,我们提出了一种基于神经网络的计算框架,其基于梯度的优化实现了使机器人能够实现这种能力。所提出的框架包括卷积改变自动化器(CONVVAES)和经常性神经网络(RNN),其具有长短短期存储器(LSTM)架构,该架构学会映射给定的目标图像以进行协作到Visuomotor预测。更具体地,首先由各个Convvaes的编码器提取视觉和目标特征状态。然后基于其当前状态由LSTM生成可视特征和电动机预测,并且根据提取的目标特征状态调节。在学习过程之后的协作期间,通过梯度下降来优化目标特征状态,以最小化预测和实际视觉特征状态之间的错误。这使得机器人能够单独地动态地推断人类伴侣的情境(目标)变化。通过对涉及物体组件的人机协作任务进行实验来评估所提出的框架。实验结果表明,配备有提出的框架的机器人可以通过动态目标推理与人类伴侣进行协作,即使情况含糊不清。

著录项

相似文献

  • 外文文献
  • 中文文献
  • 专利
获取原文

客服邮箱:kefu@zhangqiaokeyan.com

京公网安备:11010802029741号 ICP备案号:京ICP备15016152号-6 六维联合信息科技 (北京) 有限公司©版权所有
  • 客服微信

  • 服务号