首页> 外文会议>IEEE Annual Conference on Decision and Control >Off-Policy Reinforcement-Learning Algorithm to Solve Minimax Games on Graphs
【24h】

Off-Policy Reinforcement-Learning Algorithm to Solve Minimax Games on Graphs

机译:缺处策略加强学习算法,解决图形上的最小游戏

获取原文

摘要

In this paper, we formulate and find distributed minimax strategies as an alternative to Nash equilibrium strategies for multi-agent systems communicating via graph topologies, i.e., communication restrictions are taken into account for the distributed design. We provide the conditions that guarantee the existence of the minimax solutions in the game. Finally, we present an off-policy Integral Reinforcement Learning (IRL) method to solve the minimax Riccati equations and determine the optimal and worst-case policies of the agents by measuring data along the system trajectories.
机译:在本文中,我们制定并找到分布式Minimax策略,作为通过图形拓扑通信的多种代理系统的纳什均衡策略的替代方案,即考虑了分布式设计的通信限制。我们提供保证在游戏中存在最低限度解决方案的条件。最后,我们介绍了一个脱离策略的整体强化学习(IRL)方法来解决Minimax Riccati方程,并通过沿系统轨迹测量数据来确定代理的最佳和最坏情况策略。

著录项

相似文献

  • 外文文献
  • 中文文献
  • 专利
获取原文

客服邮箱:kefu@zhangqiaokeyan.com

京公网安备:11010802029741号 ICP备案号:京ICP备15016152号-6 六维联合信息科技 (北京) 有限公司©版权所有
  • 客服微信

  • 服务号