首页> 外文会议>IEEE International Conference on Multimedia and Expo >State Representation Learning For Effective Deep Reinforcement Learning
【24h】

State Representation Learning For Effective Deep Reinforcement Learning

机译:状态表示学习可进行有效的深度强化学习

获取原文

摘要

Recent years have witnessed the great success of deep reinforcement learning (DRL) on a variety of vision games. Although DNN has demonstrated strong power in representation learning, such capacity is under-explored in most DRL works whose focus is usually on optimization solvers. In fact, we discover that the state feature learning is the main obstacle for further improvement of DRL algorithms. To address this issue, we propose a new state representation learning scheme with our Adjacent State Consistency Loss (ASC Loss). The loss is defined based on the hypothesis that there are fewer changes between adjacent states than that of far apart ones, since scenes in videos generally evolve smoothly. In this paper, we exploit ASC loss as an assistant of RL loss in the training phase to boost the state feature learning. We conduct evaluation on Atari games and MuJoCo continuous control tasks, which demonstrates that our method is superior to OpenAI baselines.
机译:近年来,目睹了深度视觉增强学习(DRL)在各种视觉游戏上的巨大成功。尽管DNN在表示学习中表现出强大的功能,但是在大多数DRL工作中,这种能力的开发不足,而DRL工作通常集中在优化求解器上。实际上,我们发现状态特征学习是DRL算法进一步改进的主要障碍。为了解决这个问题,我们提出了一种新的州代表学习方案,该方案具有相邻州一致性损失(ASC Loss)。损失是基于以下假设来定义的:相邻状态之间的变化少于相距较远的状态,因为视频中的场景通常会平滑发展。在本文中,我们在训练阶段利用ASC损失作为RL损失的助手来增强状态特征学习。我们对Atari游戏和MuJoCo连续控制任务进行了评估,这表明我们的方法优于OpenAI基线。

著录项

相似文献

  • 外文文献
  • 中文文献
  • 专利
获取原文

客服邮箱:kefu@zhangqiaokeyan.com

京公网安备:11010802029741号 ICP备案号:京ICP备15016152号-6 六维联合信息科技 (北京) 有限公司©版权所有
  • 客服微信

  • 服务号