首页> 外文会议> >Extend Single-agent Reinforcement Learning Approach to a Multi-robot Cooperative Task in an Unknown Dynamic Environment
【24h】

Extend Single-agent Reinforcement Learning Approach to a Multi-robot Cooperative Task in an Unknown Dynamic Environment

机译:将单主体强化学习方法扩展到未知动态环境中的多机器人协作任务

获取原文

摘要

Machine learning technology helps multi-robot systems to carry out desired tasks in an unknown dynamic environment. In this paper, we extend the single-agent Q-learning algorithm to a multi-robot box-pushing system in an unknown dynamic environment with random obstacle distribution. There are two kinds of extensions available: directly extending MDP (Markov Decision Process) based Q-learning to the multi-robot domain, and SG-based (Stochastic Game based) Q-learning. Here, we select the first kind of extension because of its simplicity. The learning space, the box dynamics, and the reward function etc. are presented in this paper. Furthermore, a simulation system is developed and its results show effectiveness, robustness and adaptivity of this learning-based multi-robot system. Our statistical analysis of the results also shows that the robots learned correct cooperative strategy even in a dynamic environment.
机译:机器学习技术可帮助多机器人系统在未知的动态环境中执行所需的任务。在本文中,我们将单智能体Q学习算法扩展到具有未知障碍物分布的未知动态环境中的多机器人盒推系统。可以使用两种扩展:将基于MDP(马尔可夫决策过程)的Q学习直接扩展到多机器人域,以及基于SG(基于随机游戏)的Q学习。在这里,由于其简单性,我们选择了第一种扩展。本文介绍了学习空间,盒子动力学和奖励函数等。此外,开发了一个仿真系统,其结果显示了这种基于学习的多机器人系统的有效性,鲁棒性和适应性。我们对结果的统计分析还表明,即使在动态环境中,机器人也能学会正确的协作策略。

著录项

相似文献

  • 外文文献
  • 中文文献
  • 专利
获取原文

客服邮箱:kefu@zhangqiaokeyan.com

京公网安备:11010802029741号 ICP备案号:京ICP备15016152号-6 六维联合信息科技 (北京) 有限公司©版权所有
  • 客服微信

  • 服务号