首页> 外文会议>Chinese Automation Congress >Off-Policy Reinforcement Learning for Optimal Preview Tracking Control of Linear Discrete-Time systems with unknown dynamics
【24h】

Off-Policy Reinforcement Learning for Optimal Preview Tracking Control of Linear Discrete-Time systems with unknown dynamics

机译:非策略强化学习,用于动态未知的线性离散时间系统的最优预知跟踪控制

获取原文

摘要

In this paper., an off-policy reinforcement learning (RL) algorithm is presented to solve the optimal preview tracking control of discrete time systems with unknown dynamics. Firstly., an augmented state-space system that includes the available preview knowledge as a part of the state vector is constructed to cast the preview tracking control problem as a standard linear quadratic regulator (LQR) one. Secondly., the reinforcement learning technique is utilized to solve the algebraic Riccati equation (ARE) using online measurable data without requiring the a priori knowledge of the system matrices. Compared with the existing off-policy RL algorithm., the proposed scheme solves a preview tracking control problem. A numerical simulation example is given to verify the effectiveness of the proposed control scheme.
机译:本文提出了一种非策略强化学习算法,以解决未知动态离散时间系统的最优预知跟踪控制问题。首先,构造了包括状态信息的一部分在内的可用预览知识的增强状态空间系统,以将预览跟踪控制问题转换为标准线性二次调节器(LQR)。其次,强化学习技术被用于使用在线可测量数据来求解代数Riccati方程(ARE),而无需先验系统矩阵知识。与现有的非策略RL算法相比,该方案解决了预览跟踪控制问题。数值算例验证了所提控制方案的有效性。

著录项

相似文献

  • 外文文献
  • 中文文献
  • 专利
获取原文

客服邮箱:kefu@zhangqiaokeyan.com

京公网安备:11010802029741号 ICP备案号:京ICP备15016152号-6 六维联合信息科技 (北京) 有限公司©版权所有
  • 客服微信

  • 服务号