首页> 外文会议>Annual American Control Conference >Cooperative Model-Based Reinforcement Learning for Approximate Optimal Tracking
【24h】

Cooperative Model-Based Reinforcement Learning for Approximate Optimal Tracking

机译:基于合作模型的加强学习,用于近似最佳跟踪

获取原文

摘要

This paper provides an approximate online adaptive solution to the infinite-horizon optimal tracking problem for a set of agents with homogeneous dynamics and common tracking objectives. Model-based reinforcement learning is implemented by simultaneously evaluating the Bellman error (BE) at the state of each agent and on nearby off-trajectory points, as needed, throughout the state space. Each agent will calculate and share their respective on and off-trajectory BE information with a centralized estimator, which computes updates for the approximate solution to the infinite-horizon optimal tracking problem and shares the estimate with the agents. In doing so, the computational burden associated with BE extrapolation is shared between the agents and a centralized updating resource. Edge computing is leveraged to share the computational load between the agents and a centralized resource. Uniformly ultimately bounded tracking of each agent's state to the desired state and convergence of the control policy to the neighborhood of the optimal policy is proven via a Lyapunov-like stability analysis.
机译:本文提供了一组具有均匀动态和常见跟踪目标的一组代理的无限地平线最佳跟踪问题的近似的在线自适应解决方案。基于模型的增强学习是通过同时评估每个代理的状态的Bellman误差(BES),并根据需要在整个状态空间内的附近的离轨点。每个代理将计算并共享各自的开和脱轨,以集中估计器提供信息,该信息将对无限范围最佳跟踪问题的近似解的更新计算,并与代理共享估计。在这样做时,在代理和集中更新资源之间共享与外推相关的计算负担。利用边缘计算以在代理和集中资源之间共享计算负载。通过Lyapunov样稳定性分析证明,通过Lyapunov的稳定性分析证明了对每个代理的状态和控制策略的所需状态和控制策略的趋同的均匀偏移跟踪。

著录项

相似文献

  • 外文文献
  • 中文文献
  • 专利
获取原文

客服邮箱:kefu@zhangqiaokeyan.com

京公网安备:11010802029741号 ICP备案号:京ICP备15016152号-6 六维联合信息科技 (北京) 有限公司©版权所有
  • 客服微信

  • 服务号