首页> 外文期刊>Neurocomputing >Optimal control of nonlinear discrete time-varying systems using a new neural network approximation structure
【24h】

Optimal control of nonlinear discrete time-varying systems using a new neural network approximation structure

机译:使用新的神经网络逼近结构的非线性离散时变系统的最优控制

获取原文
获取原文并翻译 | 示例
       

摘要

In this paper motivated by recently discovered neurocognitive models of mechanisms in the brain, a new reinforcement learning (RL) method is presented based on a novel critic neural network (NN) structure to solve the optimal tracking problem of a nonlinear discrete time-varying system in an online manner. A multiple-model approach combined with an adaptive self-organizing map (ASOM) neural network is used to detect changes in the dynamics of the system. The number of sub-models is determined adaptively and grows once a mismatch between the stored sub-models and the new data is detected. By using the ASOM neural network, a novel value function approximation (VFA) scheme is presented. Each sub-model contributes into the value function based on a responsibility signal obtained by the ASOM. The responsibility signal determines how much each sub-model contributes to the general value function. Novel policy iteration and the value iteration algorithms are presented to find the optimal control for the partially-unknown nonlinear discrete time-varying systems in an online manner. Simulation results demonstrate the effectiveness of the proposed control scheme.
机译:本文以最近发现的大脑机制的神经认知模型为动力,提出了一种基于新型批评者神经网络(NN)结构的强化学习(RL)方法,以解决非线性离散时变系统的最优跟踪问题。以在线方式。结合自适应自组织图(ASOM)神经网络的多模型方法用于检测系统动力学的变化。子模型的数量是自适应确定的,并且一旦检测到存储的子模型与新数据之间的不匹配,该子模型的数量就会增加。通过使用ASOM神经网络,提出了一种新颖的价值函数逼近(VFA)方案。每个子模型都基于由ASOM获得的责任信号对价值函数做出贡献。责任信号确定每个子模型对通用价值函数的贡献。提出了新颖的策略迭代和价值迭代算法,以在线方式找到部分未知的非线性离散时变系统的最优控制。仿真结果证明了所提出控制方案的有效性。

著录项

  • 来源
    《Neurocomputing》 |2015年第25期|157-165|共9页
  • 作者单位

    UTA Research Institute, University of Texas at Arlington, Fort Worth, TX 76118, USA, 7300 Jack Newell Blvd S, Fort Worth, TX, USA, 76118;

    UTA Research Institute, University of Texas at Arlington, Fort Worth, TX, USA,State Key Laboratory of Synthetical Automation for Process Industries, Northeastern University, Shenyang, China;

    University of Texas at Arlington, Arlington, TX, USA;

  • 收录信息 美国《科学引文索引》(SCI);美国《工程索引》(EI);
  • 原文格式 PDF
  • 正文语种 eng
  • 中图分类
  • 关键词

    Multiple-model; Optimal control; Adaptive self- organizing map; Reinforcement learning; Value function approximation;

    机译:多种模式最佳控制;自适应自组织图;强化学习;值函数近似;
  • 入库时间 2022-08-18 02:06:58

相似文献

  • 外文文献
  • 中文文献
  • 专利
获取原文

客服邮箱:kefu@zhangqiaokeyan.com

京公网安备:11010802029741号 ICP备案号:京ICP备15016152号-6 六维联合信息科技 (北京) 有限公司©版权所有
  • 客服微信

  • 服务号