...
首页> 外文期刊>IEICE Transactions on Fundamentals of Electronics, Communications and Computer Sciences >Reinforcement Learning for Continuous Stochastic Actions: An Approximation of Probability Density Function by Orthogonal Wave Function Expansion
【24h】

Reinforcement Learning for Continuous Stochastic Actions: An Approximation of Probability Density Function by Orthogonal Wave Function Expansion

机译:连续随机动作的强化学习:通过正交波函数展开的概率密度函数逼近

获取原文
获取原文并翻译 | 示例
           

摘要

A function approximation based on an orthonormal wave function expansion in a complex space is derived. Although a probability density function (PDF) cannot always be expanded in an orthogonal series in a real space because a PDF is a positive real function, the function approximation can approximate an arbitrary PDF with high accuracy. It is applied to an actor-critic method of reinforcement learning to derive an optimal policy expressed by an arbitrary PDF in a continuous-action continuous-state environment. A chaos control problem and a PDF approximation problem are solved using the actor-critic method with the function approximation, and it is shown that the function approximation can approximate a PDF well and that the actor-critic method with the function approximation exhibits high performance.
机译:推导基于复空间中正交波函数展开的函数逼近。尽管由于PDF是正实函数,所以概率密度函数(PDF)不能总是在实际空间中以正交序列扩展,但是函数逼近可以高精度地近似任意PDF。它将其应用于强化学习的行为者批评方法,以得出在连续作用连续状态环境中由任意PDF表示的最优策略。使用具有函数逼近的actor-critic方法可以解决混沌控制问题和PDF逼近问题,结果表明,函数逼近可以很好地逼近PDF,具有函数逼近的actor-critic方法具有较高的性能。

著录项

相似文献

  • 外文文献
  • 中文文献
  • 专利
获取原文

客服邮箱:kefu@zhangqiaokeyan.com

京公网安备:11010802029741号 ICP备案号:京ICP备15016152号-6 六维联合信息科技 (北京) 有限公司©版权所有
  • 客服微信

  • 服务号