首页> 外文期刊>Power Systems, IEEE Transactions on >Comparing Policy Gradient and Value Function Based Reinforcement Learning Methods in Simulated Electrical Power Trade
【24h】

Comparing Policy Gradient and Value Function Based Reinforcement Learning Methods in Simulated Electrical Power Trade

机译:模拟电力贸易中基于策略梯度和价值函数的强化学习方法比较

获取原文
获取原文并翻译 | 示例
       

摘要

In electrical power engineering, reinforcement learning algorithms can be used to model the strategies of electricity market participants. However, traditional value function based reinforcement learning algorithms suffer from convergence issues when used with value function approximators. Function approximation is required in this domain to capture the characteristics of the complex and continuous multivariate problem space. The contribution of this paper is the comparison of policy gradient reinforcement learning methods, using artificial neural networks for policy function approximation, with traditional value function based methods in simulations of electricity trade. The methods are compared using an AC optimal power flow based power exchange auction market model and a reference electric power system model.
机译:在电力工程中,强化学习算法可用于对电力市场参与者的策略进行建模。但是,传统的基于价值函数的强化学习算法在与价值函数逼近器配合使用时会遇到收敛问题。在此域中需要函数逼近来捕获复杂且连续的多元问题空间的特征。本文的贡献是将策略梯度强化学习方法与基于传统价值函数方法的电力贸易模拟方法进行了比较,该方法使用人工神经网络进行策略函数逼近。使用基于交流最优潮流的电力交易拍卖市场模型和参考电力系统模型对方法进行比较。

著录项

相似文献

  • 外文文献
  • 中文文献
  • 专利
获取原文

客服邮箱:kefu@zhangqiaokeyan.com

京公网安备:11010802029741号 ICP备案号:京ICP备15016152号-6 六维联合信息科技 (北京) 有限公司©版权所有
  • 客服微信

  • 服务号