首页> 外文期刊>Journal of Process Control >Choice of approximator and design of penalty function for an approximate dynamic programming based control approach
【24h】

Choice of approximator and design of penalty function for an approximate dynamic programming based control approach

机译:基于近似动态规划的控制方法的近似器选择和罚函数设计

获取原文
获取原文并翻译 | 示例
           

摘要

This paper investigates the choice of function approximator for an approximate dynamic programming (ADP) based control strategy. The ADP strategy allows the user to derive an improved control policy given a simulation model and some starting control policy (or alternatively, closed-loop identification data), while circumventing the 'curse-of-dimensionality' of the traditional dynamic programming approach. In ADP, one fits a function approximator to state vs. 'cost-to-go' data and solves the Bellman equation with the approximator in an iterative manner. A proper choice and design of function approximator is critical for convergence of the iteration and the quality of final learned control policy, because an approximation error can grow quickly in the loop of optimization and function approximation. Typical classes of approximators used in related approaches are parameterized global approximators (e.g. artificial neural networks) and nonparametric local averagers (e.g. k-nearest neighbor). In this paper, we assert on the basis of some case studies and a theoretical result that a certain type of local averagers should be preferred over global approximators as the former ensures monotonic convergence of the iteration. However, a converged cost-to-go function does not necessarily lead to a stable control policy on-line due to the problem of over-extrapolation. To cope with this difficulty, we propose that a penalty term be included in the objective function in each minimization to discourage the optimizer from finding a solution in the regions of state space where the local data density is inadequately low. A nonparametric density estimator, which can be naturally combined with a local averager, is employed for this purpose. (c) 2005 Elsevier Ltd. All rights reserved.
机译:本文研究了基于近似动态规划(ADP)的控制策略的函数逼近器的选择。 ADP策略允许用户在给定模拟模型和一些启动控制策略(或闭环识别数据)的情况下,得出改进的控制策略,同时规避传统动态编程方法的“维数诅咒”。在ADP中,可以拟合一个函数逼近器来陈述与“待销成本”数据之间的关系,并以逼近器的方式用逼近器求解Bellman方程。函数逼近器的正确选择和设计对于迭代的收敛和最终学习控制策略的质量至关重要,因为在优化和函数逼近的循环中,逼近误差可能会迅速增长。相关方法中使用的典型逼近器类别是参数化的全局逼近器(例如人工神经网络)和非参数局部平均器(例如k最近邻)。在本文中,我们根据一些案例研究和理论结果断言,与全局逼近器相比,应首选某种类型的局部平均器,因为前者可确保迭代的单调收敛。但是,由于过度外推的问题,会聚成本函数功能并不一定会导致稳定的在线控制策略。为了解决此难题,我们建议在每次最小化中将惩罚项包含在目标函数中,以阻止优化器在状态数据区域中局部数据密度不足的状态中找到解决方案。为此,可以使用可以与本地平均器自然组合的非参数密度估计器。 (c)2005 Elsevier Ltd.保留所有权利。

著录项

相似文献

  • 外文文献
  • 中文文献
  • 专利
获取原文

客服邮箱:kefu@zhangqiaokeyan.com

京公网安备:11010802029741号 ICP备案号:京ICP备15016152号-6 六维联合信息科技 (北京) 有限公司©版权所有
  • 客服微信

  • 服务号