首页> 外文期刊>SIAM Journal on Control and Optimization >ACTION TIME SHARING POLICIES FOR ERGODIC CONTROL OFMARKOV CHAINS
【24h】

ACTION TIME SHARING POLICIES FOR ERGODIC CONTROL OFMARKOV CHAINS

机译:马尔科夫链条的生物控制的分时行动政策

获取原文
获取原文并翻译 | 示例
           

摘要

Ergodic control for discrete time controlled Markov chains with a locally compact state space and a compact action space is considered under suitable stability, irreducibility, and Feller continuity conditions. A flexible family of controls, called action time sharing (ATS) policies,associated with a given continuous stationary Markov control, is introduced. It is shown that the long-term average cost for such a control policy, for a broad range of one-stage cost functions, is the same as that for the associated stationary Markov policy. In addition, ATS policies are well suited for a range of estimation, information collection, and adaptive control goals. To illustrate the possibilities we present two examples. The first demonstrates a construction of an ATS policy thatleads to consistent estimators for unknown model parameters while producing the desired long-term average cost value. The second example considers a setting where the target stationary Markov control q is not known but there are sampling schemes available that allow for consistent estimation of q. We construct an ATS policy which uses dynamic estimators for q for control decisions and show that the associated cost coincides with that for the unknown Markov control q.
机译:在适当的稳定性,不可约性和Feller连续性条件下,考虑了具有局部紧凑状态空间和紧凑动作空间的离散时间控制马尔可夫链的遍历控制。引入了一个灵活的控件系列,称为动作时间共享(ATS)策略,它与给定的连续固定马尔可夫控件关联。结果表明,这种控制策略的长期平均成本,针对广泛的一阶段成本函数,与相关的固定马尔可夫策略相同。此外,ATS策略非常适合一系列估计,信息收集和自适应控制目标。为了说明可能性,我们提供两个示例。第一个演示了ATS策略的构建,该策略可导致对未知模型参数进行一致的估计,同时产生所需的长期平均成本值。第二个示例考虑了一个设置,在该设置中目标静止Markov控制q未知,但是有可用的采样方案可以对q进行一致的估计。我们构造了一个ATS策略,该策略对q进行动态决策,以得出控制决策,并表明相关成本与未知Markov控制q的成本相符。

著录项

相似文献

  • 外文文献
  • 中文文献
  • 专利
获取原文

客服邮箱:kefu@zhangqiaokeyan.com

京公网安备:11010802029741号 ICP备案号:京ICP备15016152号-6 六维联合信息科技 (北京) 有限公司©版权所有
  • 客服微信

  • 服务号