首页> 外文会议>IEEE Annual Conference on Decision and Control >Model-based reinforcement learning for infinite-horizon approximate optimal tracking
【24h】

Model-based reinforcement learning for infinite-horizon approximate optimal tracking

机译:基于模型的强化学习,用于无限水平近似最优跟踪

获取原文

摘要

This paper provides an approximate online adaptive solution to the infinite-horizon optimal tracking problem for control-affine continuous-time nonlinear systems with unknown drift dynamics where model-based reinforcement learning is used to relax the persistence of excitation condition. Model-based reinforcement learning is implemented using a concurrent learning-based system identifier to simulate experience by evaluating the Bellman error over unexplored areas of the state space. Tracking of the desired trajectory and convergence of the developed policy to a neighborhood of the optimal policy is established via Lyapunov-based stability analysis.
机译:本文针对具有未知漂移动力学的仿射连续时间非线性系统的无限水平最优跟踪问题,提供了一种近似的在线自适应解决方案,其中基于模型的强化学习用于放松激励条件的持续性。使用基于并发的基于学习的系统标识符来实现基于模型的强化学习,以通过评估状态空间未探索区域上的Bellman错误来模拟体验。通过基于Lyapunov的稳定性分析,可以跟踪期望的轨迹并将已开发策略收敛到最优策略的邻域。

著录项

相似文献

  • 外文文献
  • 中文文献
  • 专利
获取原文

客服邮箱:kefu@zhangqiaokeyan.com

京公网安备:11010802029741号 ICP备案号:京ICP备15016152号-6 六维联合信息科技 (北京) 有限公司©版权所有
  • 客服微信

  • 服务号