首页> 外文会议>International Workshop on Web Intelligence >Updating Strategies of Policies for Coordinating Agent Swarm in Dynamic Environments
【24h】

Updating Strategies of Policies for Coordinating Agent Swarm in Dynamic Environments

机译:更新动态环境中协调代理群的策略策略

获取原文

摘要

This paper proposes strategies for updating action policies in dynamic environments, and discusses the influence of learning parameters in algorithms based on swarm behavior. It is shown that inappropriate choices for learning parameters may cause delays in the learning process, or lead the convergence to an unacceptable solution. Such problems are aggravated in dynamic environments, since the fit of algorithm parameter values that use rewards is not enough to guarantee a satisfactory convergence. In this context, strategy-updating policies are proposed to modify reward values, thereby improving coordination between agents operating within dynamic environments. A framework has been developed which iteratively demonstrates the influence of parameters and updating strategies. Experimental results are reported which show that it is possible to accelerate convergence to a consistent global policy, improving the results achieved by classical approaches using algorithms based on swarm behavior.
机译:本文提出了更新动态环境中的行动策略的策略,并探讨了基于群体行为的算法中学习参数的影响。 It is shown that inappropriate choices for learning parameters may cause delays in the learning process, or lead the convergence to an unacceptable solution.这种问题在动态环境中加剧,因为使用奖励的算法参数值的适合度不足以保证令人满意的会聚。在此上下文中,提出了策略更新策略来修改奖励值,从而提高了动态环境中操作的代理之间的协调。已经开发了一个框架,其迭代地展示了参数和更新策略的影响。报道了实验结果表明,可以加速趋同的全球政策,从基于群体行为使用算法来提高经典方法所实现的结果。

著录项

相似文献

  • 外文文献
  • 中文文献
  • 专利
获取原文

客服邮箱:kefu@zhangqiaokeyan.com

京公网安备:11010802029741号 ICP备案号:京ICP备15016152号-6 六维联合信息科技 (北京) 有限公司©版权所有
  • 客服微信

  • 服务号