首页> 外文会议>IEEE Symposium Series on Computational Intelligence >Off-policy reinforcement learning for distributed output synchronization of linear multi-agent systems
【24h】

Off-policy reinforcement learning for distributed output synchronization of linear multi-agent systems

机译:线性多主体系统分布式输出同步的非策略强化学习

获取原文

摘要

In this paper, off-policy reinforcement learning (RL) is used to find a model-free optimal solution to the Houtput synchronization of heterogeneous multi-agent discrete-time systems. First, the output synchronization problem is formulated as a set of local optimal tracking problems. It is shown that optimal local synchronization control protocols can be found by solving augmented game algebraic Riccati equations (GAREs). The solutions to the GAREs require the state of the leader for all agents and the knowledge of agent dynamics. To obviate this requirement, a distributed adaptive observer is designed to estimate the leader state for all agents without requiring complete knowledge of the leader dynamics. Moreover, off-policy RL algorithm is used to learn the solution to the GAREs using only measured data and without requiring the knowledge of the agent or the leader dynamics. In the proposed approach, in contrast to other model free approaches, the disturbance input does not need to be adjusted in a specific manner. A simulation example is given to show the effectiveness of the proposed method.
机译:本文采用非政策强化学习(RL)来找到针对H的无模型最优解 异构多智能体离散时间系统的输出同步。首先,将输出同步问题表述为一组局部最优跟踪问题。结果表明,最佳局部同步控制协议可以通过求解增广的游戏代数Riccati方程(GARE)来找到。 GARE的解决方案要求所有座席的领导者状态和座席动态知识。为了消除此要求,设计了分布式自适应观察器来估计所有座席的领导者状态,而无需完全了解领导者动态。此外,不使用策略的RL算法仅使用测量数据即可学习GARE的解决方案,而无需了解代理或领导者动态。在提出的方法中,与其他无模型方法相比,不需要以特定方式调整干扰输入。仿真实例表明了该方法的有效性。

著录项

相似文献

  • 外文文献
  • 中文文献
  • 专利
获取原文

客服邮箱:kefu@zhangqiaokeyan.com

京公网安备:11010802029741号 ICP备案号:京ICP备15016152号-6 六维联合信息科技 (北京) 有限公司©版权所有
  • 客服微信

  • 服务号