首页> 外文会议>International Conference on Algorithmic Learning Theory >Asymptotic Learnability of Reinforcement Problems with Arbitrary Dependence
【24h】

Asymptotic Learnability of Reinforcement Problems with Arbitrary Dependence

机译:任意依赖性加强问题的渐近学报

获取原文

摘要

We address the problem of reinforcement learning in which observations may exhibit an arbitrary form of stochastic dependence on past observations and actions, i.e. environments more general than (PO) MDPs. The task for an agent is to attain the best possible asymptotic reward where the true generating environment is unknown but belongs to a known countable family of environments. We find some sufficient conditions on the class of environments under which an agent exists which attains the best asymptotic reward for any environment in the class. We analyze how tight these conditions are and how they relate to different probabilistic assumptions known in reinforcement learning and related fields, such as Markov Decision Processes and mixing conditions.
机译:我们解决了强化学习的问题,其中观察可能表现出对过去观察和行动的任意形式的随机依赖性,即环境比(PO)MDP更通用。代理的任务是达到真正生成环境未知的最佳可能的渐近奖励,但属于已知的可数环境。我们在存在的环境中找到了一些足够的条件,该环境的存在性存在于课堂上任何环境的最佳渐近奖励。我们分析了这些条件如何以及如何与强化学习和相关领域中已知的不同概率假设相关,例如马尔可夫决策过程和混合条件。

著录项

相似文献

  • 外文文献
  • 中文文献
  • 专利
获取原文

客服邮箱:kefu@zhangqiaokeyan.com

京公网安备:11010802029741号 ICP备案号:京ICP备15016152号-6 六维联合信息科技 (北京) 有限公司©版权所有
  • 客服微信

  • 服务号