首页> 外文会议>Chinese Automation Congress >Off-Policy Reinforcement Learning for Optimal Preview Tracking Control of Linear Discrete-Time systems with unknown dynamics
【24h】

Off-Policy Reinforcement Learning for Optimal Preview Tracking Control of Linear Discrete-Time systems with unknown dynamics

机译:具有未知动力学的线性离散时间系统的最佳预览跟踪控制的禁止策略加强学习

获取原文

摘要

In this paper, an off-policy reinforcement learning (RL) algorithm is presented to solve the optimal preview tracking control of discrete time systems with unknown dynamics. Firstly, an augmented state-space system that includes the available preview knowledge as a part of the state vector is constructed to cast the preview tracking control problem as a standard linear quadratic regulator (LQR) one. Secondly, the reinforcement learning technique is utilized to solve the algebraic Riccati equation (ARE) using online measurable data without requiring the a priori knowledge of the system matrices. Compared with the existing off-policy RL algorithm, the proposed scheme solves a preview tracking control problem. A numerical simulation example is given to verify the effectiveness of the proposed control scheme.
机译:本文提出了一种脱策强化学习(RL)算法以解决具有未知动态的离散时间系统的最佳预览跟踪控制。首先,构造包括作为状态矢量的一部分的可用预览知识的增强状态空间系统,以将预览跟踪控制问题作为标准的线性二次调节器(LQR)。其次,利用在线可测量数据解决代数Riccati等式(AS)的加强学习技术,而不需要先验的系统矩阵知识。与现有的截止策略RL算法相比,该方案解决了预览跟踪控制问题。给出了数值模拟示例来验证所提出的控制方案的有效性。

著录项

相似文献

  • 外文文献
  • 中文文献
  • 专利
获取原文

客服邮箱:kefu@zhangqiaokeyan.com

京公网安备:11010802029741号 ICP备案号:京ICP备15016152号-6 六维联合信息科技 (北京) 有限公司©版权所有
  • 客服微信

  • 服务号