首页> 外文会议>IEEE International Conference on Acoustics, Speech and Signal Processing >Efficient Multi-agent Cooperative Navigation in Unknown Environments with Interlaced Deep Reinforcement Learning
【24h】

Efficient Multi-agent Cooperative Navigation in Unknown Environments with Interlaced Deep Reinforcement Learning

机译:交错式深度强化学习在未知环境中的高效多主体协作导航

获取原文

摘要

This work addresses a multi-agent cooperative navigation problem that multiple agents work together in an unknown environment in order to reach different targets without collision and minimize the maximum navigation time they spend. Typical reinforcement learning-based solutions directly model the cooperative navigation policy as a steering policy. However, when each agent does not know which target to head for, this method could prolong convergence time and reduce overall performance. To this end, we model the navigation policy as a combination of a dynamic target selection policy and a collision avoidance policy. Since these two policies are coupled, an interlaced deep reinforcement learning method is proposed to simultaneously learn them. Additionally, a reward function is directly derived from the optimization objective function instead of using a heuristic design method. Extensive experiments demonstrate that the proposed method can converge in a fast way and generate a more efficient navigation policy compared with the state-of-the-art.
机译:这项工作解决了多智能体协作导航问题,即多个智能体在未知环境中一起工作,以达到不同目标而不会发生冲突,并最大程度地减少他们花费的最大导航时间。典型的基于强化学习的解决方案直接将协作导航策略建模为指导策略。但是,当每个代理都不知道要到达哪个目标时,此方法可能会延长收敛时间并降低总体性能。为此,我们将导航策略建模为动态目标选择策略和碰撞避免策略的组合。由于这两个策略是耦合的,因此提出了一种交错的深度强化学习方法来同时学习它们。此外,直接从优化目标函数中获得奖励函数,而不是使用启发式设计方法。大量的实验表明,与最新技术相比,该方法可以快速收敛并生成更有效的导航策略。

著录项

相似文献

  • 外文文献
  • 中文文献
  • 专利
获取原文

客服邮箱:kefu@zhangqiaokeyan.com

京公网安备:11010802029741号 ICP备案号:京ICP备15016152号-6 六维联合信息科技 (北京) 有限公司©版权所有
  • 客服微信

  • 服务号