首页> 外文会议>Intelligent Robots and Systems, 2001. Proceedings. 2001 IEEE/RSJ International Conference on >KaBaGe-RL: Kanerva-based generalisation and reinforcement learning for possession football
【24h】

KaBaGe-RL: Kanerva-based generalisation and reinforcement learning for possession football

机译:KaBaGe-RL:基于Kanerva的足球足球泛化和强化学习

获取原文

摘要

The complexity of most modem systems prohibits a hand-coded approach to decision making. In addition, many problems have continuous or large discrete state spaces; some have large or continuous action spaces. The problem of learning in large spaces is tackled through generalisation techniques, which allow compact representation of learned information and transfer of knowledge between similar states and actions. In this paper Kanerva. coding and reinforcement learning are combined to produce the KaBaGe-RL decision-making module. The purpose of KaBaGe-RL is twofold. Firstly, Kanerva coding is used as a generalisation method to produce a feature vector from the raw sensory input. Secondly, the reinforcement learning uses this feature vector in order to learn an optimal policy. The efficiency of KaBaGe-RL is tested using the "3 versus 2 possession football" challenge, a subproblem of the RoboCup domain. The results demonstrate that the learning approach outperforms a number of benchmark policies including a hand-coded one.
机译:大多数调制解调器系统的复杂性都禁止采用手动编码的方法进行决策。另外,许多问题具有连续或较大的离散状态空间。有些具有较大或连续的动作空间。大空间中的学习问题通过泛化技术解决,泛化技术可以紧凑地表示所学信息,并在相似状态和动作之间进行知识转移。在本文中,Kanerva。编码和强化学习相结合以产生KaBaGe-RL决策模块。 KaBaGe-RL的目的是双重的。首先,使用Kanerva编码作为一种概括方法,从原始的感觉输入中生成特征向量。其次,强化学习使用此特征向量来学习最佳策略。 KaBaGe-RL的效率是使用“ 3对2拥有足球”挑战赛(RoboCup域的一个子问题)进行测试的。结果表明,学习方法的性能优于许多基准策略,包括手工编码的策略。

著录项

相似文献

  • 外文文献
  • 中文文献
  • 专利
获取原文

客服邮箱:kefu@zhangqiaokeyan.com

京公网安备:11010802029741号 ICP备案号:京ICP备15016152号-6 六维联合信息科技 (北京) 有限公司©版权所有
  • 客服微信

  • 服务号