首页> 外文会议>20th European conference on artificial intelligence >Heuristically Accelerated Reinforcement Learning: Theoretical and Experimental Results
【24h】

Heuristically Accelerated Reinforcement Learning: Theoretical and Experimental Results

机译:启发式加速强化学习:理论和实验结果

获取原文
获取原文并翻译 | 示例

摘要

Since finding control policies using Reinforcement Learning (RL) can be very time consuming, in recent years several authors have investigated how to speed up RL algorithms by making improved action selections based on heuristics. In this work we present new theoretical results - convergence and a superior limit for value estimation errors - for the class that encompasses all heuristics-based algorithms, called Heuristically Accelerated Reinforcement Learning. We also expand this new class by proposing three new algorithms, the Heuristically Accelerated Q(A), SARSA(A) and TD(A), the first algorithms that uses both heuristics and eligibility traces. Empirical evaluations were conducted in traditional control problems and results show that using heuristics significantly enhances the performance of the learning process.
机译:由于使用强化学习(RL)查找控制策略可能非常耗时,因此近年来,一些作者已经研究了如何通过基于启发式进行改进的动作选择来加快RL算法的速度。在这项工作中,我们为涵盖所有基于启发式算法(称为启发式加速强化学习)的课程提供了新的理论结果-收敛性和价值估计误差的上限值。我们还提出了三种新算法,即启发式加速Q(A),SARSA(A)和TD(A),这是第一个同时使用启发式和合格跟踪的算法,从而扩展了这一新类。对传统控制问题进行了实证评估,结果表明,使用启发式方法可以显着提高学习过程的性能。

著录项

相似文献

  • 外文文献
  • 中文文献
  • 专利
获取原文

客服邮箱:kefu@zhangqiaokeyan.com

京公网安备:11010802029741号 ICP备案号:京ICP备15016152号-6 六维联合信息科技 (北京) 有限公司©版权所有
  • 客服微信

  • 服务号