首页> 外文期刊>Machine Learning >An online prediction algorithm for reinforcement learning with linear function approximation using cross entropy method
【24h】

An online prediction algorithm for reinforcement learning with linear function approximation using cross entropy method

机译:交叉熵法线性函数逼近的强化学习在线预测算法

获取原文
获取原文并翻译 | 示例
获取外文期刊封面目录资料

摘要

In this paper, we provide two new stable online algorithms for the problem of prediction in reinforcement learning, i.e., estimating the value function of a model-free Markov reward process using the linear function approximation architecture and with memory and computation costs scaling quadratically in the size of the feature set. The algorithms employ the multi-timescale stochastic approximation variant of the very popular cross entropy optimization method which is a model based search method to find the global optimum of a real-valued function. A proof of convergence of the algorithms using the ODE method is provided. We supplement our theoretical results with experimental comparisons. The algorithms achieve good performance fairly consistently on many RL benchmark problems with regards to computational efficiency, accuracy and stability.
机译:在本文中,我们针对强化学习中的预测问题提供了两种新的稳定的在线算法,即使用线性函数逼近架构估算无模型马尔可夫奖励过程的值函数,并在内存和计算成本中按比例缩放功能集的大小。该算法采用了非常流行的交叉熵优化方法的多时间尺度随机近似变量,该方法是一种基于模型的搜索方法,用于找到实值函数的全局最优值。提供了使用ODE方法进行算法收敛的证明。我们通过实验比较来补充理论结果。该算法在许多RL基准问题上,在计算效率,准确性和稳定性方面都相当稳定地取得了良好的性能。

著录项

相似文献

  • 外文文献
  • 中文文献
  • 专利
获取原文

客服邮箱:kefu@zhangqiaokeyan.com

京公网安备:11010802029741号 ICP备案号:京ICP备15016152号-6 六维联合信息科技 (北京) 有限公司©版权所有
  • 客服微信

  • 服务号