首页> 外文期刊>JMLR: Workshop and Conference Proceedings >Learning by Playing Solving Sparse Reward Tasks from Scratch
【24h】

Learning by Playing Solving Sparse Reward Tasks from Scratch

机译:通过从头开始解决稀疏奖励任务来学习

获取原文
           

摘要

We propose Scheduled Auxiliary Control (SAC-X), a new learning paradigm in the context of Reinforcement Learning (RL). SAC-X enables learning of complex behaviors - from scratch - in the presence of multiple sparse reward signals. To this end, the agent is equipped with a set of general auxiliary tasks, that it attempts to learn simultaneously via off-policy RL. The key idea behind our method is that active (learned) scheduling and execution of auxiliary policies allows the agent to efficiently explore its environment - enabling it to excel at sparse reward RL. Our experiments in several challenging robotic manipulation settings demonstrate the power of our approach.
机译:我们提出了计划辅助控制(SAC-X),这是强化学习(RL)的一种新的学习范例。 SAC-X可以在存在多个稀疏奖励信号的情况下从头开始学习复杂的行为。为此,代理程序配备了一组常规辅助任务,它试图通过非策略RL同时学习。我们方法背后的关键思想是主动(学习)的调度和辅助策略的执行使代理能够有效地探索其环境-使其能够胜任稀疏奖励RL。我们在几种具有挑战性的机器人操纵设置中进行的实验证明了我们方法的强大功能。

著录项

相似文献

  • 外文文献
  • 中文文献
  • 专利
获取原文

客服邮箱:kefu@zhangqiaokeyan.com

京公网安备:11010802029741号 ICP备案号:京ICP备15016152号-6 六维联合信息科技 (北京) 有限公司©版权所有
  • 客服微信

  • 服务号