...
首页> 外文期刊>Information Sciences: An International Journal >Multi-step heuristic dynamic programming for optimal control of nonlinear discrete-time systems
【24h】

Multi-step heuristic dynamic programming for optimal control of nonlinear discrete-time systems

机译:非线性离散时间系统最优控制的多步启发式动态规划

获取原文
获取原文并翻译 | 示例
           

摘要

Policy iteration and value iteration are two main iterative adaptive dynamic programming frameworks for solving optimal control problems. Policy iteration converges fast while requiring an initial stabilizing control policy, which is a strict constraint in practice. Value iteration avoids the requirement of initial admissible control policy while converging much slowly. This paper tries to utilize the advantages of policy iteration and value iteration, and avoids their drawbacks at the same time. Therefore, a multi-step heuristic dynamic programming (MsHDP) method is developed for solving the optimal control problem of nonlinear discrete-time systems. MsHDP speeds up value iteration and avoids the requirement of initial admissible control policy in policy iteration at the same time. The convergence theory of MsHDP is established by proving that it converges to the solution of the Bellman equation. For implementation purpose, the actor-critic neural network (NN) structure is developed. The critic NN is employed to estimate the value function and its NN weight vector is computed with a least-square scheme. The actor NN is used to estimate the control policy and a gradient descent method is proposed for updating its NN weight vector. According to the comparative simulation studies on two examples, the effectiveness and advantages of MsHDP are verified. (C) 2017 Elsevier Inc. All rights reserved.
机译:政策迭代和价值迭代是两个主要迭代自适应动态编程框架,用于解决最佳控制问题。政策迭代在需要初始稳定控制策略的同时收敛,这是一个严格的实践约束。价值迭代避免了初始允许控制策略的要求,同时汇总缓慢。本文试图利用政策迭代和价值迭代的优势,并避免同时缺点。因此,开发了一种用于解决非线性离散时间系统的最优控制问题的多步启发式动态编程(MSHDP)方法。 MSHDP加快了价值迭代,并避免了同时在政策迭代中的初始允许控制策略的要求。通过证明它会聚到Bellman方程的解决方案,建立了MSHDP的收敛理论。为了实现目的,开发了演员 - 评论家神经网络(NN)结构。用于估计批评者NN来估计值函数,并且其NN重量向量用最小二乘方案计算。 Actor Nn用于估计控制策略和梯度下降方法,用于更新其NN权重向量。根据两个实例的比较模拟研究,验证了MSHDP的有效性和优点。 (c)2017年Elsevier Inc.保留所有权利。

著录项

相似文献

  • 外文文献
  • 中文文献
  • 专利
获取原文

客服邮箱:kefu@zhangqiaokeyan.com

京公网安备:11010802029741号 ICP备案号:京ICP备15016152号-6 六维联合信息科技 (北京) 有限公司©版权所有
  • 客服微信

  • 服务号