首页> 外文会议>2011 International Symposium on Artificial Intelligence and Signal Processing >Multi-skill agents coalition formation under skill uncertainty
【24h】

Multi-skill agents coalition formation under skill uncertainty

机译:技能不确定性下的多技能特工联盟形成

获取原文

摘要

In a multi-agent system, there are situations in which agents are unable to do their tasks individually. Therefore forming coalitions is inevitable. In natural settings, an agent decides to form coalition based on beliefs it has regarding the capabilities of other agents. In the previous works, it was assumed that a single type can reflects all capabilities an agent has. We introduce multi-skill agents which have a value per skill. This helps us to solve more problems and to reason about the results, more exactly. We use Bayesian Reinforcement Learning (BRL) as the learning mechanism. Through the repeated use of BRL, agents can form more rewarding coalitions. We extend existing algorithms of the repeated coalition formation under type uncertainty to the skill uncertainty and exploit them in experimental studies that type uncertainty couldn't do or reason about. Average long term discounted expected reward that agents accumulate in the learning process, is the criteria we test our methods based on. We test the algorithms on a sample soccer sub-team formation problem. To have a notion of the best performance, we solve the problem in the absence of uncertainty. Results show that the VPI method does approximately 85% of the best performance.
机译:在多代理系统中,有些情况下代理无法单独执行其任务。因此,形成联盟是不可避免的。在自然环境中,代理人基于对其他代理人的能力的信念而决定组建联盟。在以前的工作中,假设单个类型可以反映代理具有的所有功能。我们介绍具有多种技能的代理人,每个代理人都有其价值。这有助于我们解决更多问题并更准确地对结果进行推理。我们使用贝叶斯强化学习(BRL)作为学习机制。通过重复使用BRL,代理商可以组成更多有回报的联盟。我们将类型不确定性下的重复联盟形成的现有算法扩展到技能不确定性,并在类型不确定性无法或无法解释的实验研究中加以利用。代理在学习过程中积累的平均长期折现期望奖励是我们测试方法的标准。我们在样本足球子团队组建问题上测试了算法。为了获得最佳性能的概念,我们在没有不确定性的情况下解决了该问题。结果表明,VPI方法可实现约85%的最佳性能。

著录项

相似文献

  • 外文文献
  • 中文文献
  • 专利
获取原文

客服邮箱:kefu@zhangqiaokeyan.com

京公网安备:11010802029741号 ICP备案号:京ICP备15016152号-6 六维联合信息科技 (北京) 有限公司©版权所有
  • 客服微信

  • 服务号