首页> 外文会议>IEEE Symposium Series on Computational Intelligence >Efficient Practice for Deep Reinforcement Learning
【24h】

Efficient Practice for Deep Reinforcement Learning

机译:高效练习深度加强学习

获取原文

摘要

Deep reinforcement learning has demonstrated its ability to solve diverse hard problems, which were not able to be solved previously. However, one of the main drawbacks is the required long training time for a task. Furthermore, trial-and-error in reinforcement learning is often inappropriate in most real-world tasks. There were some discussions on transfer learning, specifically sim-to-real transfer, but the design of new environments or defining an environment with similar or relevant goals require tedious human efforts. This research aims to leverage practice approaches, which do not require a new environmental design for transfer learning, to shorten the time for training a deep reinforcement learning agent and to achieve an optimal policy for the maximum expected rewards. We verify the efficacy of existing practice approach with Deep Q networks (DQN) in Atari games and also introduce a novel practice approach involving iterations of short periods of practice and reinforcement learning, to further improve the performance of an agent.
机译:深度加固学习已经证明了解决多种难题的能力,这无法以前无法解决。然而,其中一个主要缺点是任务所需的长期训练时间。此外,在大多数现实世界的任务中,强化学习中的试验和错误往往是不合适的。有一些关于转移学习的讨论,特别是SIM-实际转移,但新环境的设计或定义具有相似或相关目标的环境需要繁琐的人类努力。本研究旨在利用练习方法,这不需要新的环境设计进行转移学习,缩短培训深度加强学习代理的时间,并为最高预期奖励实现最佳政策。我们验证了Atari Games中的深度Q网络(DQN)现有实践方法的功效,还介绍了涉及短期实践和加固学习的迭代的新颖实践方法,以进一步提高代理人的表现。

著录项

相似文献

  • 外文文献
  • 中文文献
  • 专利
获取原文

客服邮箱:kefu@zhangqiaokeyan.com

京公网安备:11010802029741号 ICP备案号:京ICP备15016152号-6 六维联合信息科技 (北京) 有限公司©版权所有
  • 客服微信

  • 服务号