首页> 外文期刊>International Journal of Fuzzy Systems >A Fuzzy Curiosity-Driven Mechanism for Multi-Agent Reinforcement Learning
【24h】

A Fuzzy Curiosity-Driven Mechanism for Multi-Agent Reinforcement Learning

机译:一种模糊的多功能加固学习机构

获取原文
获取原文并翻译 | 示例
       

摘要

Many works provide intrinsic rewards to deal with sparse rewards in reinforcement learning. Due to the non-stationarity of multi-agent systems, it is impracticable to apply existing methods to multi-agent reinforcement learning directly. In this paper, a fuzzy curiosity-driven mechanism is proposed for multi-agent reinforcement learning, by which agents can explore more efficiently in a scenario with sparse extrinsic reward. First, we improve the variational auto-encoder to predict the next state through the joint-state and joint-action for agents. Then several fuzzy partitions are built according to the next joint-state, which aims at assigning the prediction error to different agents. With the proposed method, each agent in the multi-agent environment receives its individual intrinsic reward. We elaborate on the proposed method in partially observable environments and fully observable environments separately. Experimental results show that multi-agent learns joint policies more efficiently by the proposed fuzzy curiosity-driven mechanism, and it can also help agents find better policies in the training process.
机译:许多作品提供了内在的奖励,以处理加强学习中的稀疏奖励。由于多种子体系统的非实用性,将现有方法直接应用于多智能体增强学习是不切实际的。本文提出了一种模糊的好奇心驱动机构,用于多蛋白增强学习,通过该机构可以在具有稀疏外在奖励的情况下更有效地探索。首先,我们改进变分式自动编码器以通过为代理的关节状态和关节作用预测下一个状态。然后根据下一个联合状态构建多个模糊分区,其旨在将预测误差分配给不同的代理。通过提出的方法,多种子体环境中的每个代理都会收到其个体内在奖励。我们在部分可观察环境中详细说明了拟议的方法和完全可观察的环境。实验结果表明,多助手通过拟议的模糊性好奇地驱动机制更有效地学习联合政策,并且还可以帮助代理商在培训过程中找到更好的政策。

著录项

相似文献

  • 外文文献
  • 中文文献
  • 专利
获取原文

客服邮箱:kefu@zhangqiaokeyan.com

京公网安备:11010802029741号 ICP备案号:京ICP备15016152号-6 六维联合信息科技 (北京) 有限公司©版权所有
  • 客服微信

  • 服务号