首页> 外文期刊>Automatica >Simulation-based optimization of Markov decision processes: An empirical process theory approach
【24h】

Simulation-based optimization of Markov decision processes: An empirical process theory approach

机译:基于模拟的马尔可夫决策过程优化:一种经验过程理论方法

获取原文
获取原文并翻译 | 示例
           

摘要

We generalize and build on the PAC Learning framework for Markov Decision Processes developed in Jain and Varaiya (2006). We consider the reward function to depend on both the state and the action. Both the state and action spaces can potentially be countably infinite. We obtain an estimate for the value function of a Markov decision process, which assigns to each policy its expected discounted reward. This expected reward can be estimated as the empirical average of the reward over many independent simulation runs. We derive bounds on the number of runs needed for the convergence of the empirical average to the expected reward uniformly for a class of policies, in terms of the V-C or pseudo dimension of the policy class. We then propose a framework to obtain an ∈-optimal policy from simulation. We provide sample complexity of such an approach.
机译:我们归纳并建立在Jain和Varaiya(2006)提出的马尔可夫决策过程的PAC学习框架中。我们认为奖励功能取决于状态和行为。状态空间和动作空间都可能是无限大的。我们获得了马尔可夫决策过程的价值函数的估计值,该估计值将给每个保单分配其预期的折现报酬。可以将预期的奖励估算为许多独立模拟运行中奖励的经验平均值。我们根据政策类别的V-C或伪维数,得出一类政策将经验平均值统一收敛到预期报酬所需的运行次数的界限。然后,我们提出了一个从仿真中获得ε最优策略的框架。我们提供了这种方法的示例复杂性。

著录项

相似文献

  • 外文文献
  • 中文文献
  • 专利
获取原文

客服邮箱:kefu@zhangqiaokeyan.com

京公网安备:11010802029741号 ICP备案号:京ICP备15016152号-6 六维联合信息科技 (北京) 有限公司©版权所有
  • 客服微信

  • 服务号