首页> 外文会议>IEEE International Conference on Cybernetics and Intelligent Systems;IEEE International Conference on Robotics, Automation and Mechatronics >H optimal control of unknown linear discrete-time systems: An off-policy reinforcement learning approach
【24h】

H optimal control of unknown linear discrete-time systems: An off-policy reinforcement learning approach

机译:未知线性离散时间系统的H 最优控制:一种非政策强化学习方法

获取原文

摘要

This paper proposes a model-free H control design for linear discrete-time systems using reinforcement learning (RL). A novel off-policy RL algorithm is used to solve the game algebraic Riccati equation (GARE) online using the measured data along the system trajectories. The proposed RL algorithm has the following advantages compared to existing model-free RL methods for solving H control problem: 1) It is data efficient and fast since a stream of experiences which is obtained from executing a fixed behavioral policy is reused to update many value functions correspond to different leaning policies sequentially. 2) The disturbance input does not need to be adjusted in a specific manner. 3) There is no bias as a result of adding a probing noise to the control input to maintain persistence of excitation conditions. A simulation example is used to verify the effectiveness of the proposed control scheme.
机译:本文提出了一种使用强化学习(RL)的线性离散时间系统的无模型H控制设计。一种新颖的非策略RL算法用于使用沿着系统轨迹的测量数据在线求解游戏代数Riccati方程(GARE)。与现有的解决H控制问题的无模型RL方法相比,所提出的RL算法具有以下优点:1)由于执行固定行为策略所获得的经验流可重复使用以更新许多价值,因此它具有数据高效且快速的特点。功能依次对应于不同的学习策略。 2)不需要以特定方式调整干扰输入。 3)由于向控制输入端添加了探测噪声以保持激励条件的持久性,因此没有偏差。仿真实例用于验证所提出的控制方案的有效性。

著录项

相似文献

  • 外文文献
  • 中文文献
  • 专利
获取原文

客服邮箱:kefu@zhangqiaokeyan.com

京公网安备:11010802029741号 ICP备案号:京ICP备15016152号-6 六维联合信息科技 (北京) 有限公司©版权所有
  • 客服微信

  • 服务号