首页> 外文会议>IEEE International Conference on Robotics and Automation >Rearrangement with Nonprehensile Manipulation Using Deep Reinforcement Learning
【24h】

Rearrangement with Nonprehensile Manipulation Using Deep Reinforcement Learning

机译:使用深度强化学习进行非精确操纵的重排

获取原文

摘要

Rearranging objects on a tabletop surface by means of nonprehensile manipulation is a task which requires skillful interaction with the physical world. Usually, this is achieved by precisely modeling physical properties of the objects, robot, and the environment for explicit planning. In contrast, as explicitly modeling the physical environment is not always feasible and involves various uncertainties, we learn a nonprehensile rearrangement strategy with deep reinforcement learning based on only visual feedback. For this, we model the task with rewards and train a deep Q-network. Our potential field-based heuristic exploration strategy reduces the amount of collisions which lead to suboptimal outcomes and we actively balance the training set to avoid bias towards poor examples. Our training process leads to quicker learning and better performance on the task as compared to uniform exploration and standard experience replay. We demonstrate empirical evidence from simulation that our method leads to a success rate of 85%, show that our system can cope with sudden changes of the environment, and compare our performance with human level performance.
机译:借助不费力的操作将对象重新布置在桌面上的任务是一项需要与物理世界进行熟练交互的任务。通常,这是通过精确建模对象,机器人和环境的物理属性以进行显式规划来实现的。相反,由于对物理环境进行显式建模并不总是可行的,并且涉及各种不确定性,因此我们仅基于视觉反馈就可以通过深度强化学习来学习非全面的重排策略。为此,我们用奖励为任务建模并训练一个深层的Q网络。我们潜在的基于现场的启发式探索策略减少了导致次优结果的碰撞次数,并且我们积极地平衡了训练集,以避免偏向不良范例。与统一的探索和标准的体验重播相比,我们的培训过程可导致更快的学习速度和更好的任务表现。我们通过仿真证明了经验证据,表明我们的方法成功率为85%,表明我们的系统可以应对环境的突然变化,并将我们的性能与人类水平的性能进行比较。

著录项

相似文献

  • 外文文献
  • 中文文献
  • 专利
获取原文

客服邮箱:kefu@zhangqiaokeyan.com

京公网安备:11010802029741号 ICP备案号:京ICP备15016152号-6 六维联合信息科技 (北京) 有限公司©版权所有
  • 客服微信

  • 服务号