首页> 外文会议>International Conference on Algorithmic Learning Theory >Bayesian Reinforcement Learning with Exploration
【24h】

Bayesian Reinforcement Learning with Exploration

机译:贝叶斯加固与勘探学习

获取原文

摘要

We consider a general reinforcement learning problem and show that carefully combining the Bayesian optimal policy and an exploring policy leads to minimax sample-complexity bounds in a very general class of (history-based) environments. We also prove lower bounds and show that the new algorithm displays adaptive behaviour when the environment is easier than worst-case.
机译:我们考虑了一般的加强学习问题,并谨慎地结合贝叶斯最优政策,并探索策略导致最普遍的(历史为基础的)环境中的最低限度样本复杂性界限。我们还证明了下限,并显示新算法在环境比最坏情况更容易时显示自适应行为。

著录项

相似文献

  • 外文文献
  • 中文文献
  • 专利
获取原文

客服邮箱:kefu@zhangqiaokeyan.com

京公网安备:11010802029741号 ICP备案号:京ICP备15016152号-6 六维联合信息科技 (北京) 有限公司©版权所有
  • 客服微信

  • 服务号