...
首页> 外文期刊>Systems, Man, and Cybernetics, Part C: Applications and Reviews, IEEE Transactions on >Experience Replay for Real-Time Reinforcement Learning Control
【24h】

Experience Replay for Real-Time Reinforcement Learning Control

机译:体验回放,用于实时强化学习控制

获取原文
获取原文并翻译 | 示例
           

摘要

Reinforcement-learning (RL) algorithms can automatically learn optimal control strategies for nonlinear, possibly stochastic systems. A promising approach for RL control is experience replay (ER), which learns quickly from a limited amount of data, by repeatedly presenting these data to an underlying RL algorithm. Despite its benefits, ER RL has been studied only sporadically in the literature, and its applications have largely been confined to simulated systems. Therefore, in this paper, we evaluate ER RL on real-time control experiments that involve a pendulum swing-up problem and the vision-based control of a goalkeeper robot. These real-time experiments are complemented by simulation studies and comparisons with traditional RL. As a preliminary, we develop a general ER framework that can be combined with essentially any incremental RL technique, and instantiate this framework for the approximate Q-learning and SARSA algorithms. The successful real-time learning results that are presented here are highly encouraging for the applicability of ER RL in practice.
机译:强化学习(RL)算法可以为非线性的,可能是随机的系统自动学习最佳控制策略。 RL控制的一种有前途的方法是体验重播(ER),它可以通过将这些数据重复呈现给底层RL算法来从有限的数据中快速学习。尽管具有ER RL的优点,但文献中仅对ER RL进行了零星研究,其应用很大程度上局限于模拟系统。因此,在本文中,我们在涉及摆摆问题和守门员机器人基于视觉的控制的实时控制实验中评估了ER RL。这些实时实验辅以模拟研究以及与传统RL的比较。首先,我们开发了一个通用的ER框架,该框架可以与基本上任何增量RL技术结合使用,并为近似Q学习和SARSA算法实例化此框架。此处介绍的成功的实时学习结果对于ER RL在实践中的适用性非常令人鼓舞。

著录项

相似文献

  • 外文文献
  • 中文文献
  • 专利
获取原文

客服邮箱:kefu@zhangqiaokeyan.com

京公网安备:11010802029741号 ICP备案号:京ICP备15016152号-6 六维联合信息科技 (北京) 有限公司©版权所有
  • 客服微信

  • 服务号