首页> 外文OA文献 >Stochastic optimal generation command dispatch based on improved hierarchical reinforcement learning approach
【2h】

Stochastic optimal generation command dispatch based on improved hierarchical reinforcement learning approach

机译:基于改进层次强化学习方法的随机最优发电指挥调度

摘要

This study presents an improved hierarchical reinforcement learning (HRL) approach to deal with the curse of dimensionality in the dynamic optimisation of generation command dispatch (GCD) for automatic generation control (AGC) under control performance standards. The AGC committed units are firstly classified into several different groups according to their time delay of frequency control, and the core problem of GCD is decomposed into a set of subtasks for search of the optimal regulation participation factors with the solution algorithm. The time-varying coordination factor is introduced in the control layer to improve the learning efficiency of HRL, and the generating error, hydro capacity margin and AGC regulating costs are formulated into Markov decision process reward function. The application of the improved hierarchical Q-learning (HQL) algorithm in the China southern power grid model shows that the proposed method can reduce the convergence time in the pre-learning process, decrease the AGC regulating cost and improve the control performance of AGC systems compared with the conventional HQL, genetic algorithm and a engineering method.
机译:这项研究提出了一种改进的分层强化学习(HRL)方法,以处理在控制性能标准下自动生成控制(AGC)的生成命令调度(GCD)动态优化中的维度诅咒。首先将AGC委托单位按照其频率控制的时间延迟分为几个不同的组,然后将GCD的核心问题分解为一组子任务,以求解算法来寻找最佳调节参与因子。在控制层引入时变协调因子以提高HRL的学习效率,并将发电误差,水电容量裕度和AGC调节成本公式化为马尔可夫决策过程奖励函数。改进的分层Q学习算法在南方电网模型中的应用表明,该方法可以减少预学习过程中的收敛时间,降低AGC的调节成本,提高AGC系统的控制性能。与传统的HQL,遗传算法和工程方法相比。

著录项

  • 作者

    Yu T; Wang YM; Ye WJ; Zhou B; Chan KW;

  • 作者单位
  • 年度 2011
  • 总页数
  • 原文格式 PDF
  • 正文语种 eng
  • 中图分类
  • 入库时间 2022-08-20 20:56:12

相似文献

  • 外文文献
  • 中文文献
  • 专利

客服邮箱:kefu@zhangqiaokeyan.com

京公网安备:11010802029741号 ICP备案号:京ICP备15016152号-6 六维联合信息科技 (北京) 有限公司©版权所有
  • 客服微信

  • 服务号