首页> 外文期刊>自动化学报(英文版) >Clique-based Cooperative Multiagent Reinforcement Learning Using Factor Graphs
【24h】

Clique-based Cooperative Multiagent Reinforcement Learning Using Factor Graphs

机译:基于族群的基于因子图的协同多主体强化学习

获取原文
获取原文并翻译 | 示例
           

摘要

In this paper,we propose a clique-based sparse rein-forcement learning(RL)algorithm for solving cooperative tasks. The aim is to accelerate the learning speed of the original sparse RL algorithm and to make it applicable for tasks decomposed in a more general manner.First,a transition function is estimated and used to update the Q-value function, which greatly reduces the learning time.Second,it is more reasonable to divide agents into cliques,each of which is only responsible for a specific subtask.In this way,the global Q-value function is decomposed into the sum of several simpler local Q-value functions.Such decomposition is expressed by a factor graph and exploited by the general max-plus algorithm to obtain the greedy joint action. Experimental results show that the proposed approach outperforms others with better performance.
机译:本文提出了一种基于群体的稀疏强化学习算法来解决合作任务。目的是为了加快原始稀疏RL算法的学习速度,使其更适用于以更一般的方式分解的任务。首先,估计一个过渡函数并将其用于更新Q值函数,从而大大减少了学习第二,将代理划分为集团,每个集团仅负责一个特定的子任务是更合理的。这样,全局Q值函数被分解为几个更简单的局部Q值函数的总和。分解用一个因子图表示,并用一般的最大加法算法得到贪婪的联合作用。实验结果表明,该方法在性能上优于其他方法。

著录项

  • 来源
    《自动化学报(英文版)》 |2014年第3期|248-256|共9页
  • 作者

    Zhen Zhang; Dongbin Zhao;

  • 作者单位

    the State Key Laboratory of Management and Control for Complex Systems,Institute of Automation,Chinese Academy of Sciences, Beijing 100190, China;

    Department of Electric Engineering, College of Automation Engineering, Qingdao University, Qingdao 266071, China;

    the State Key Laboratory of Management and Control for Complex Systems,Institute of Automation,Chinese Academy of Sciences, Beijing 100190, China;

  • 收录信息
  • 原文格式 PDF
  • 正文语种 eng
  • 中图分类
  • 关键词

获取原文

客服邮箱:kefu@zhangqiaokeyan.com

京公网安备:11010802029741号 ICP备案号:京ICP备15016152号-6 六维联合信息科技 (北京) 有限公司©版权所有
  • 客服微信

  • 服务号