首页> 外文会议>International Joint Conference on Neural Networks >Cyber-Human Approach For Learning Human Intention And Shape Robotic Behavior Based On Task Demonstration
【24h】

Cyber-Human Approach For Learning Human Intention And Shape Robotic Behavior Based On Task Demonstration

机译:基于任务示范的网络人类学习人类意图和造型机器人行为的方法

获取原文

摘要

Recent developments in artificial intelligence enabled training of autonomous robots without human supervision. Even without human supervision during training, current models have yet to be human-engineered and have neither guarantees to match human expectation nor perform within safety bounds. This paper proposes CyberSteer to leverage human-robot interaction and align goals between humans and robotic intelligent agents. Based on human demonstration of the task, CyberSteer learns an intrinsic reward function used by the human demonstrator to pursue the goal of the task. The learned intrinsic human function shapes the robotic behavior during training through deep reinforcement learning algorithms, removing the need for environment-dependent or hand-engineered reward signal. Two different hypotheses were tested, both using non-expert human operators for initial demonstration of a given task or desired behavior: one training a deep neural network to classify human-like behavior and other training a behavior cloning deep neural network to suggest actions. In this experiment, CyberSteer was tested in a high-fidelity unmanned air system simulation environment, Microsoft AirSim. The simulated aerial robot performed collision avoidance through a clustered forest environment using forward-looking depth sensing. The performance of CyberSteer is compared to behavior cloning algorithms and reinforcement learning algorithms guided by handcrafted reward functions. Results show that the human-learned intrinsic reward function can shape the behavior of robotic systems and have better task performance guiding reinforcement learning algorithms compared to standard human-handcrafted reward functions.
机译:最近的人工智能发展使得自治机器人的培训没有人为监督。即使在培训期间没有人体监督,目前的模型也尚未人为人工程,并且既没有保证匹配人类期望,也没有在安全范围内执行。本文提出了利用人机互动并对准人类和机器人智能代理的目标来利用。基于任务的人类示范,Cyber​​steer学习人类示威者使用的内在奖励功能,以追求任务的目标。学习的内在人类功能在通过深度加强学习算法训练期间塑造了机器人行为,从而消除了对环境依赖性或手工工程奖励信号的需求。测试了两种不同的假设,两者都使用非专家人类运营商进行给定的任务或期望行为的初步演示:一个培训深度神经网络,以分类人类的行为和其他培训行为克隆深神经网络的行为来建议行动。在这个实验中,Cyber​​seeer在Microsoft Airsim的高保真无人空中系统仿真环境中进行了测试。模拟的空中机器人通过使用前瞻性深度感测来通过聚集的森林环境进行碰撞避免。将Cyber​​steer的性能与手动奖励功能引导的行为克隆算法和强化学习算法进行了比较。结果表明,人类学习的内在奖励功能可以塑造机器人系统的行为,与标准人类手工奖励功能相比,具有更好的任务性能引导钢筋学习算法。

著录项

相似文献

  • 外文文献
  • 中文文献
  • 专利
获取原文

客服邮箱:kefu@zhangqiaokeyan.com

京公网安备:11010802029741号 ICP备案号:京ICP备15016152号-6 六维联合信息科技 (北京) 有限公司©版权所有
  • 客服微信

  • 服务号