首页> 外文期刊>South African Computer Journal >Upper Bounds on the Performance of Discretisation in Reinforcement Learning
【24h】

Upper Bounds on the Performance of Discretisation in Reinforcement Learning

机译:强化学习中离散化性能的上限

获取原文
       

摘要

Reinforcement learning is a machine learning framework whereby an agent learns to perform a task by maximising its total reward received for selecting actions in each state. The policy mapping states to actions that the agent learns is either represented explicitly, or implicitly through a value function. It is common in reinforcement learning to discretise a continuous state space using tile coding or binary features. We prove an upper bound on the performance of discretisation for direct policy representation or value function approximation.
机译:强化学习是一种机器学习框架,通过该框架,代理可以通过最大化其在每种状态下选择动作所获得的总奖励来学习执行任务。策略映射声明到代理学习到的操作的过程是显式表示的,或者是通过值函数隐式表示的。在强化学习中,通常使用图块编码或二进制特征离散化连续状态空间。我们证明了直接策略表示或价值函数逼近的离散化性能的上限。

著录项

相似文献

  • 外文文献
  • 中文文献
  • 专利
获取原文

客服邮箱:kefu@zhangqiaokeyan.com

京公网安备:11010802029741号 ICP备案号:京ICP备15016152号-6 六维联合信息科技 (北京) 有限公司©版权所有
  • 客服微信

  • 服务号