首页> 外文期刊>International journal of systems science >Nearly data-based optimal control for linear discrete model-free systems with delays via reinforcement learning
【24h】

Nearly data-based optimal control for linear discrete model-free systems with delays via reinforcement learning

机译:通过强化学习对带有延迟的线性离散无模型系统的基于数据的最优控制

获取原文
获取原文并翻译 | 示例
           

摘要

In this paper, a nearly data-based optimal control scheme is proposed for linear discrete model-free systems with delays. The nearly optimal control can be obtained using only measured input/output data from systems, by reinforcement learning technology, which combines Q-learning with value iterative algorithm. First, we construct a state estimator by using the measured input/output data. Second, the quadratic functional is used to approximate the value function at each point in the state space, and the data-based control is designed by Q-learning method using the obtained state estimator. Then, the paper states the method, that is, how to solve the optimal inner kernel matrix (P) over bar in the least-square sense, by value iteration algorithm. Finally, the numerical examples are given to illustrate the effectiveness of our approach.
机译:本文针对具有延迟的线性离散无模型系统,提出了一种基于数据的最优控制方案。通过将Q学习与值迭代算法结合在一起的强化学习技术,仅使用系统中测得的输入/输出数据即可获得几乎最佳的控制。首先,我们使用测得的输入/输出数据构造一个状态估计器。其次,使用二次函数对状态空间中每个点的值函数进行逼近,并使用获得的状态估计器通过Q学习方法设计基于数据的控制。然后,阐述了该方法,即如何通过值迭代算法在最小二乘意义上求解最优的内部核矩阵(P)。最后,通过数值算例说明了该方法的有效性。

著录项

相似文献

  • 外文文献
  • 中文文献
  • 专利
获取原文

客服邮箱:kefu@zhangqiaokeyan.com

京公网安备:11010802029741号 ICP备案号:京ICP备15016152号-6 六维联合信息科技 (北京) 有限公司©版权所有
  • 客服微信

  • 服务号