首页> 外文会议>European symposium on artificial neural networks >Extending immediate reinforcement learning on neural networks to multiple actions
【24h】

Extending immediate reinforcement learning on neural networks to multiple actions

机译:在神经网络上延伸即时加强学习到多个动作

获取原文

摘要

Between supervise and unsupervised learning, connectionism proposes a qualitative learning or reinforcement learning which is of interest for applications needing qualitative control. This learning technique is not new and the following advantages of a neural implementation of reinforcemtn learning are identified: a small memory requirement and a more effective exploration of the situations-actions space. However, the restriction of the applicability of this algorithm to problems with a limited number of actions (usually two) remains. We propose to solve this problem by correctly specifying the output coding of the action of the output cell layer of the neural network and interpreting the output values as a certainty value for doing a specified action. Experiments performed in the real world with the miniature robot Khepera confirm the possibility of extending the applicability of reinforcement learning to cases where multiple actions are possible for each situation.
机译:在监督和无监督的学习之间,联系方式提出了一种定性学习或加强学习,这对需要定性控制的应用感兴趣。该学习技术不是新的,并且确定了加强型学习的神经实现的以下优点:小的内存要求和对情况的更有效探索 - 行动空间。然而,限制该算法对具有有限数量的动作(通常是两个)的问题的适用性。我们建议通过正确指定神经网络的输出小区层的动作的输出编码来解决这个问题,并将输出值解释为执行指定动作的确定性值。具有微型机器人Khepera在现实世界中进行的实验证实了将加强学习的适用性扩展到多种行动对于每种情况的情况下延长。

著录项

相似文献

  • 外文文献
  • 中文文献
  • 专利
获取原文

客服邮箱:kefu@zhangqiaokeyan.com

京公网安备:11010802029741号 ICP备案号:京ICP备15016152号-6 六维联合信息科技 (北京) 有限公司©版权所有
  • 客服微信

  • 服务号