首页> 外文会议>6th Iberian Conference on Information Systems and Technologies >Hierarchical Reinforcement Learning: Learning sub-goals and state-abstraction
【24h】

Hierarchical Reinforcement Learning: Learning sub-goals and state-abstraction

机译:分层强化学习:学习子目标和状态抽象

获取原文

摘要

In this paper we present a method that allows an agent to discover and create temporal abstractions autonomously. Our method is based on the concept that to reach the goal, the agent must pass through relevant states that we will interpret as subgoals. To detect useful subgoals, our method creates intersections between several paths leading to a goal. Our research focused on domains largely used in the study of temporal abstractions. We used several versions of the room-to-room navigation problem. We determined that, in the problems tested, an agent can learn more rapidly by automatically discovering subgoals and creating abstractions.
机译:在本文中,我们提出了一种允许代理自动发现和创建时间抽象的方法。我们的方法基于以下概念:要达到目标,代理必须通过我们将解释为子目标的相关状态。为了检测有用的子目标,我们的方法会在通往目标的多条路径之间创建交点。我们的研究集中在时态抽象研究中广泛使用的领域。我们使用了多个版本的“房间到房间”导航问题。我们确定,在测试的问题中,代理可以通过自动发现子目标并创建抽象来更快地学习。

著录项

相似文献

  • 外文文献
  • 中文文献
  • 专利
获取原文

客服邮箱:kefu@zhangqiaokeyan.com

京公网安备:11010802029741号 ICP备案号:京ICP备15016152号-6 六维联合信息科技 (北京) 有限公司©版权所有
  • 客服微信

  • 服务号