首页> 外文期刊>Applied Intelligence: The International Journal of Artificial Intelligence, Neural Networks, and Complex Problem-Solving Technologies >Heuristically accelerated reinforcement learning modularization for multi-agent multi-objective problems
【24h】

Heuristically accelerated reinforcement learning modularization for multi-agent multi-objective problems

机译:多主体多目标问题的启发式加速强化学习模块化

获取原文
获取原文并翻译 | 示例
           

摘要

This article presents two new algorithms for finding the optimal solution of a Multi-agent Multi-objective Reinforcement Learning problem. Both algorithms make use of the concepts of modularization and acceleration by a heuristic function applied in standard Reinforcement Learning algorithms to simplify and speed up the learning process of an agent that learns in a multi-agent multi-objective environment. In order to verify performance of the proposed algorithms, we considered a predator-prey environment in which the learning agent plays the role of prey that must escape the pursuing predator while reaching for food in a fixed location. The results show that combining modularization and acceleration using a heuristics function indeed produced simplification and speeding up of the learning process in a complex problem when comparing with algorithms that do not make use of acceleration or modularization techniques, such as Q-Learning and Minimax-Q.
机译:本文介绍了两种新算法,用于寻找多主体多目标强化学习问题的最佳解决方案。两种算法都通过标准增强学习算法中应用的启发式函数来利用模块化和加速的概念,以简化和加速在多主体多目标环境中学习的主体的学习过程。为了验证所提出算法的性能,我们考虑了一种捕食者-猎物环境,其中学习代理扮演猎物的角色,当在固定位置获取食物时,猎物必须逃避追捕者的捕食。结果表明,与不使用加速或模块化技术的算法(例如Q-Learning和Minimax-Q)进行比较时,使用启发式函数结合模块化和加速确实可以简化并加快复杂问题中的学习过程。 。

相似文献

  • 外文文献
  • 中文文献
  • 专利
获取原文

客服邮箱:kefu@zhangqiaokeyan.com

京公网安备:11010802029741号 ICP备案号:京ICP备15016152号-6 六维联合信息科技 (北京) 有限公司©版权所有
  • 客服微信

  • 服务号