首页> 外文会议>Smart Grid Conference >A deep structure for option discovery in reinforcement learning
【24h】

A deep structure for option discovery in reinforcement learning

机译:强化学习中选项发现的深层结构

获取原文

摘要

Hierarchical learning as another way to scale up reinforcement learning and enable its applications to very hard learning problems. Hierarchical learning is a divide-and-conquer technique a complex learning problem is decomposed into small pieces so that they can be easily solved. The option framework is one way to using hierarchical learning in reinforcement learning. In this paper we have used the free-energy based function approximation (FE-RBM) to determine the option initiation set. Our proposed method calculates the output for each of the input (including state and subgoal) according to negative free energy of an RBM. Learning is done by stochastic gradient descent and mean-squared error. The experimental results showed that this method has efficient functionality to create options. Moreover, it has a reasonable generalization ability for unvisited states.
机译:分层学习是扩大强化学习并使其适用于非常困难的学习问题的另一种方法。分层学习是一种分而治之的技术,它将复杂的学习问题分解成小块,以便可以轻松解决。选项框架是在强化学习中使用分层学习的一种方法。在本文中,我们使用了基于自由能的函数逼近(FE-RBM)来确定期权启动集。我们提出的方法根据RBM的负自由能为每个输入(包括状态和子目标)计算输出。学习是通过随机梯度下降和均方误差完成的。实验结果表明,该方法具有创建选项的高效功能。而且,它对于未访问状态具有合理的泛化能力。

著录项

相似文献

  • 外文文献
  • 中文文献
  • 专利
获取原文

客服邮箱:kefu@zhangqiaokeyan.com

京公网安备:11010802029741号 ICP备案号:京ICP备15016152号-6 六维联合信息科技 (北京) 有限公司©版权所有
  • 客服微信

  • 服务号