...
首页> 外文期刊>SIAM Journal on Control and Optimization >Least squares temporal difference methods: An analysis under general conditions
【24h】

Least squares temporal difference methods: An analysis under general conditions

机译:最小二乘时差法:一般条件下的分析

获取原文
获取原文并翻译 | 示例
           

摘要

We consider approximate policy evaluation for finite state and action Markov decision processes (MDP) with the least squares temporal difference (LSTD) algorithm, LSTD(λ), in an exploration-enhanced learning context, where policy costs are computed from observations of a Markov chain different from the one corresponding to the policy under evaluation. We establish for the discounted cost criterion that LSTD(λ) converges almost surely under mild, minimal conditions. We also analyze other properties of the iterates involved in the algorithm, including convergence in mean and boundedness. Our analysis draws on theories of both finite space Markov chains and weak Feller Markov chains on a topological space. Our results can be applied to other temporal difference algorithms and MDP models. As examples, we give a convergence analysis of a TD(λ) algorithm and extensions to MDP with compact state and action spaces, as well as a convergence proof of a new LSTD algorithm with state-dependent λ-parameters.
机译:我们考虑在探索性学习环境中使用最小二乘时差(LSTD)算法LSTD(λ)对有限状态和动作Markov决策过程(MDP)进行近似策略评估,其中策略成本是根据对Markov的观察来计算的与评估中的政策对应的链不同。我们为折现成本标准建立了LSTD(λ)在温和,最小的条件下几乎可以肯定地收敛。我们还分析了算法中涉及的迭代的其他属性,包括均值和有界性的收敛。我们的分析借鉴了拓扑空间上的有限空间马氏链和弱Feller马氏链的理论。我们的结果可以应用于其他时差算法和MDP模型。作为示例,我们对TD(λ)算法进行了收敛性分析,并扩展了具有紧凑状态空间和动作空间的MDP,并给出了一种新的具有状态相关性λ参数的LSTD算法的收敛性证明。

著录项

相似文献

  • 外文文献
  • 中文文献
  • 专利
获取原文

客服邮箱:kefu@zhangqiaokeyan.com

京公网安备:11010802029741号 ICP备案号:京ICP备15016152号-6 六维联合信息科技 (北京) 有限公司©版权所有
  • 客服微信

  • 服务号